1 // RUN: %clang_cc1 -triple mips-unknown-linux -o - -emit-llvm %s | FileCheck %s -check-prefix=ALL -check-prefix=O32 2 // RUN: %clang_cc1 -triple mipsel-unknown-linux -o - -emit-llvm %s | FileCheck %s -check-prefix=ALL -check-prefix=O32 3 // RUN: %clang_cc1 -triple mips64-unknown-linux -o - -emit-llvm -target-abi n32 %s | FileCheck %s -check-prefix=ALL -check-prefix=N32 -check-prefix=NEW 4 // RUN: %clang_cc1 -triple mips64-unknown-linux -o - -emit-llvm -target-abi n32 %s | FileCheck %s -check-prefix=ALL -check-prefix=N32 -check-prefix=NEW 5 // RUN: %clang_cc1 -triple mips64-unknown-linux -o - -emit-llvm %s | FileCheck %s -check-prefix=ALL -check-prefix=N64 -check-prefix=NEW 6 // RUN: %clang_cc1 -triple mips64el-unknown-linux -o - -emit-llvm %s | FileCheck %s -check-prefix=ALL -check-prefix=N64 -check-prefix=NEW 7 8 #include <stdarg.h> 9 10 typedef int v4i32 __attribute__ ((__vector_size__ (16))); 11 12 int test_i32(char *fmt, ...) { 13 va_list va; 14 15 va_start(va, fmt); 16 int v = va_arg(va, int); 17 va_end(va); 18 19 return v; 20 } 21 22 // ALL-LABEL: define i32 @test_i32(i8*{{.*}} %fmt, ...) 23 // 24 // O32: %va = alloca i8*, align [[PTRALIGN:4]] 25 // N32: %va = alloca i8*, align [[PTRALIGN:4]] 26 // N64: %va = alloca i8*, align [[PTRALIGN:8]] 27 // ALL: [[V:%.*]] = alloca i32, align 4 28 // NEW: [[PROMOTION_TEMP:%.*]] = alloca i32, align 4 29 // 30 // ALL: [[VA:%.+]] = bitcast i8** %va to i8* 31 // ALL: call void @llvm.va_start(i8* [[VA]]) 32 // ALL: [[AP_CUR:%.+]] = load i8*, i8** %va, align [[PTRALIGN]] 33 // O32: [[AP_NEXT:%.+]] = getelementptr inbounds i8, i8* [[AP_CUR]], [[INTPTR_T:i32]] [[CHUNKSIZE:4]] 34 // NEW: [[AP_NEXT:%.+]] = getelementptr inbounds i8, i8* [[AP_CUR]], [[INTPTR_T:i32|i64]] [[CHUNKSIZE:8]] 35 // 36 // ALL: store i8* [[AP_NEXT]], i8** %va, align [[PTRALIGN]] 37 // 38 // O32: [[AP_CAST:%.+]] = bitcast i8* [[AP_CUR]] to [[CHUNK_T:i32]]* 39 // O32: [[ARG:%.+]] = load i32, i32* [[AP_CAST]], align [[CHUNKALIGN:4]] 40 // 41 // N32: [[AP_CAST:%.+]] = bitcast i8* [[AP_CUR]] to [[CHUNK_T:i64]]* 42 // N32: [[TMP:%.+]] = load i64, i64* [[AP_CAST]], align [[CHUNKALIGN:8]] 43 // N64: [[AP_CAST:%.+]] = bitcast i8* [[AP_CUR]] to [[CHUNK_T:i64]]* 44 // N64: [[TMP:%.+]] = load i64, i64* [[AP_CAST]], align [[CHUNKALIGN:8]] 45 // NEW: [[TMP2:%.+]] = trunc i64 [[TMP]] to i32 46 // NEW: store i32 [[TMP2]], i32* [[PROMOTION_TEMP]], align 4 47 // NEW: [[ARG:%.+]] = load i32, i32* [[PROMOTION_TEMP]], align 4 48 // ALL: store i32 [[ARG]], i32* [[V]], align 4 49 // 50 // ALL: [[VA1:%.+]] = bitcast i8** %va to i8* 51 // ALL: call void @llvm.va_end(i8* [[VA1]]) 52 // ALL: } 53 54 long long test_i64(char *fmt, ...) { 55 va_list va; 56 57 va_start(va, fmt); 58 long long v = va_arg(va, long long); 59 va_end(va); 60 61 return v; 62 } 63 64 // ALL-LABEL: define i64 @test_i64(i8*{{.*}} %fmt, ...) 65 // 66 // ALL: %va = alloca i8*, align [[PTRALIGN]] 67 // ALL: [[VA:%.+]] = bitcast i8** %va to i8* 68 // ALL: call void @llvm.va_start(i8* [[VA]]) 69 // ALL: [[AP_CUR:%.+]] = load i8*, i8** %va, align [[PTRALIGN]] 70 // 71 // i64 is 8-byte aligned, while this is within O32's stack alignment there's no 72 // guarantee that the offset is still 8-byte aligned after earlier reads. 73 // O32: [[TMP1:%.+]] = ptrtoint i8* [[AP_CUR]] to i32 74 // O32: [[TMP2:%.+]] = add i32 [[TMP1]], 7 75 // O32: [[TMP3:%.+]] = and i32 [[TMP2]], -8 76 // O32: [[AP_CUR:%.+]] = inttoptr i32 [[TMP3]] to i8* 77 // 78 // ALL: [[AP_NEXT:%.+]] = getelementptr inbounds i8, i8* [[AP_CUR]], [[INTPTR_T]] 8 79 // ALL: store i8* [[AP_NEXT]], i8** %va, align [[PTRALIGN]] 80 // 81 // ALL: [[AP_CAST:%.*]] = bitcast i8* [[AP_CUR]] to i64* 82 // ALL: [[ARG:%.+]] = load i64, i64* [[AP_CAST]], align 8 83 // 84 // ALL: [[VA1:%.+]] = bitcast i8** %va to i8* 85 // ALL: call void @llvm.va_end(i8* [[VA1]]) 86 // ALL: } 87 88 char *test_ptr(char *fmt, ...) { 89 va_list va; 90 91 va_start(va, fmt); 92 char *v = va_arg(va, char *); 93 va_end(va); 94 95 return v; 96 } 97 98 // ALL-LABEL: define i8* @test_ptr(i8*{{.*}} %fmt, ...) 99 // 100 // ALL: %va = alloca i8*, align [[PTRALIGN]] 101 // ALL: [[V:%.*]] = alloca i8*, align [[PTRALIGN]] 102 // N32: [[AP_CAST:%.+]] = alloca i8*, align 4 103 // ALL: [[VA:%.+]] = bitcast i8** %va to i8* 104 // ALL: call void @llvm.va_start(i8* [[VA]]) 105 // ALL: [[AP_CUR:%.+]] = load i8*, i8** %va, align [[PTRALIGN]] 106 // ALL: [[AP_NEXT:%.+]] = getelementptr inbounds i8, i8* [[AP_CUR]], [[INTPTR_T]] [[CHUNKSIZE]] 107 // ALL: store i8* [[AP_NEXT]], i8** %va, align [[PTRALIGN]] 108 // 109 // When the chunk size matches the pointer size, this is easy. 110 // O32: [[AP_CAST:%.+]] = bitcast i8* [[AP_CUR]] to i8** 111 // N64: [[AP_CAST:%.+]] = bitcast i8* [[AP_CUR]] to i8** 112 // Otherwise we need a promotion temporary. 113 // N32: [[TMP1:%.+]] = bitcast i8* [[AP_CUR]] to i64* 114 // N32: [[TMP2:%.+]] = load i64, i64* [[TMP1]], align 8 115 // N32: [[TMP3:%.+]] = trunc i64 [[TMP2]] to i32 116 // N32: [[PTR:%.+]] = inttoptr i32 [[TMP3]] to i8* 117 // N32: store i8* [[PTR]], i8** [[AP_CAST]], align 4 118 // 119 // ALL: [[ARG:%.+]] = load i8*, i8** [[AP_CAST]], align [[PTRALIGN]] 120 // ALL: store i8* [[ARG]], i8** [[V]], align [[PTRALIGN]] 121 // 122 // ALL: [[VA1:%.+]] = bitcast i8** %va to i8* 123 // ALL: call void @llvm.va_end(i8* [[VA1]]) 124 // ALL: } 125 126 int test_v4i32(char *fmt, ...) { 127 va_list va; 128 129 va_start(va, fmt); 130 v4i32 v = va_arg(va, v4i32); 131 va_end(va); 132 133 return v[0]; 134 } 135 136 // ALL-LABEL: define i32 @test_v4i32(i8*{{.*}} %fmt, ...) 137 // 138 // ALL: %va = alloca i8*, align [[PTRALIGN]] 139 // ALL: [[V]] = alloca <4 x i32>, align 16 140 // ALL: [[VA1:%.+]] = bitcast i8** %va to i8* 141 // ALL: call void @llvm.va_start(i8* [[VA1]]) 142 // ALL: [[AP_CUR:%.+]] = load i8*, i8** %va, align [[PTRALIGN]] 143 // 144 // Vectors are 16-byte aligned, however the O32 ABI has a maximum alignment of 145 // 8-bytes since the base of the stack is 8-byte aligned. 146 // O32: [[TMP1:%.+]] = ptrtoint i8* [[AP_CUR]] to i32 147 // O32: [[TMP2:%.+]] = add i32 [[TMP1]], 7 148 // O32: [[TMP3:%.+]] = and i32 [[TMP2]], -8 149 // O32: [[AP_CUR:%.+]] = inttoptr i32 [[TMP3]] to i8* 150 // 151 // NEW: [[TMP1:%.+]] = ptrtoint i8* [[AP_CUR]] to [[INTPTR_T]] 152 // NEW: [[TMP2:%.+]] = add [[INTPTR_T]] [[TMP1]], 15 153 // NEW: [[TMP3:%.+]] = and [[INTPTR_T]] [[TMP2]], -16 154 // NEW: [[AP_CUR:%.+]] = inttoptr [[INTPTR_T]] [[TMP3]] to i8* 155 // 156 // ALL: [[AP_NEXT:%.+]] = getelementptr inbounds i8, i8* [[AP_CUR]], [[INTPTR_T]] 16 157 // ALL: store i8* [[AP_NEXT]], i8** %va, align [[PTRALIGN]] 158 // 159 // ALL: [[AP_CAST:%.+]] = bitcast i8* [[AP_CUR]] to <4 x i32>* 160 // O32: [[ARG:%.+]] = load <4 x i32>, <4 x i32>* [[AP_CAST]], align 8 161 // N64: [[ARG:%.+]] = load <4 x i32>, <4 x i32>* [[AP_CAST]], align 16 162 // ALL: store <4 x i32> [[ARG]], <4 x i32>* [[V]], align 16 163 // 164 // ALL: [[VA1:%.+]] = bitcast i8** %va to i8* 165 // ALL: call void @llvm.va_end(i8* [[VA1]]) 166 // ALL: [[VECEXT:%.+]] = extractelement <4 x i32> {{.*}}, i32 0 167 // ALL: ret i32 [[VECEXT]] 168 // ALL: } 169