1 ; RUN: opt < %s -instcombine -S | FileCheck %s 2 target datalayout = "E-p:64:64:64-p1:32:32:32-a0:0:8-f32:32:32-f64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:32:64-v64:64:64-v128:128:128" 3 4 ; Instcombine should be able to prove vector alignment in the 5 ; presence of a few mild address computation tricks. 6 7 ; CHECK-LABEL: @test0( 8 ; CHECK: align 16 9 10 define void @test0(i8* %b, i64 %n, i64 %u, i64 %y) nounwind { 11 entry: 12 %c = ptrtoint i8* %b to i64 13 %d = and i64 %c, -16 14 %e = inttoptr i64 %d to double* 15 %v = mul i64 %u, 2 16 %z = and i64 %y, -2 17 %t1421 = icmp eq i64 %n, 0 18 br i1 %t1421, label %return, label %bb 19 20 bb: 21 %i = phi i64 [ %indvar.next, %bb ], [ 20, %entry ] 22 %j = mul i64 %i, %v 23 %h = add i64 %j, %z 24 %t8 = getelementptr double* %e, i64 %h 25 %p = bitcast double* %t8 to <2 x double>* 26 store <2 x double><double 0.0, double 0.0>, <2 x double>* %p, align 8 27 %indvar.next = add i64 %i, 1 28 %exitcond = icmp eq i64 %indvar.next, %n 29 br i1 %exitcond, label %return, label %bb 30 31 return: 32 ret void 33 } 34 35 ; When we see a unaligned load from an insufficiently aligned global or 36 ; alloca, increase the alignment of the load, turning it into an aligned load. 37 38 ; CHECK-LABEL: @test1( 39 ; CHECK: tmp = load 40 ; CHECK: GLOBAL{{.*}}align 16 41 42 @GLOBAL = internal global [4 x i32] zeroinitializer 43 44 define <16 x i8> @test1(<2 x i64> %x) { 45 entry: 46 %tmp = load <16 x i8>* bitcast ([4 x i32]* @GLOBAL to <16 x i8>*), align 1 47 ret <16 x i8> %tmp 48 } 49 50 @GLOBAL_as1 = internal addrspace(1) global [4 x i32] zeroinitializer 51 52 define <16 x i8> @test1_as1(<2 x i64> %x) { 53 ; CHECK-LABEL: @test1_as1( 54 ; CHECK: tmp = load 55 ; CHECK: GLOBAL_as1{{.*}}align 16 56 %tmp = load <16 x i8> addrspace(1)* bitcast ([4 x i32] addrspace(1)* @GLOBAL_as1 to <16 x i8> addrspace(1)*), align 1 57 ret <16 x i8> %tmp 58 } 59 60 @GLOBAL_as1_gep = internal addrspace(1) global [8 x i32] zeroinitializer 61 62 define <16 x i8> @test1_as1_gep(<2 x i64> %x) { 63 ; CHECK-LABEL: @test1_as1_gep( 64 ; CHECK: tmp = load 65 ; CHECK: GLOBAL_as1_gep{{.*}}align 16 66 %tmp = load <16 x i8> addrspace(1)* bitcast (i32 addrspace(1)* getelementptr ([8 x i32] addrspace(1)* @GLOBAL_as1_gep, i16 0, i16 4) to <16 x i8> addrspace(1)*), align 1 67 ret <16 x i8> %tmp 68 } 69 70 71 ; When a load or store lacks an explicit alignment, add one. 72 73 ; CHECK-LABEL: @test2( 74 ; CHECK: load double* %p, align 8 75 ; CHECK: store double %n, double* %p, align 8 76 77 define double @test2(double* %p, double %n) nounwind { 78 %t = load double* %p 79 store double %n, double* %p 80 ret double %t 81 } 82 83 declare void @llvm.memset.p0i8.i64(i8* nocapture, i8, i64, i32, i1) nounwind 84 85 declare void @use(i8*) 86 87 %struct.s = type { i32, i32, i32, i32 } 88 89 define void @test3(%struct.s* sret %a4) { 90 ; Check that the alignment is bumped up the alignment of the sret type. 91 ; CHECK-LABEL: @test3( 92 %a4.cast = bitcast %struct.s* %a4 to i8* 93 call void @llvm.memset.p0i8.i64(i8* %a4.cast, i8 0, i64 16, i32 1, i1 false) 94 ; CHECK: call void @llvm.memset.p0i8.i64(i8* %a4.cast, i8 0, i64 16, i32 4, i1 false) 95 call void @use(i8* %a4.cast) 96 ret void 97 } 98