Lines Matching full:tmp1
6 %tmp1 = load <8 x i8>, <8 x i8>* %A
7 %tmp2 = call <8 x i8> @llvm.arm.neon.vabs.v8i8(<8 x i8> %tmp1)
14 %tmp1 = load <4 x i16>, <4 x i16>* %A
15 %tmp2 = call <4 x i16> @llvm.arm.neon.vabs.v4i16(<4 x i16> %tmp1)
22 %tmp1 = load <2 x i32>, <2 x i32>* %A
23 %tmp2 = call <2 x i32> @llvm.arm.neon.vabs.v2i32(<2 x i32> %tmp1)
30 %tmp1 = load <2 x float>, <2 x float>* %A
31 %tmp2 = call <2 x float> @llvm.fabs.v2f32(<2 x float> %tmp1)
38 %tmp1 = load <16 x i8>, <16 x i8>* %A
39 %tmp2 = call <16 x i8> @llvm.arm.neon.vabs.v16i8(<16 x i8> %tmp1)
46 %tmp1 = load <8 x i16>, <8 x i16>* %A
47 %tmp2 = call <8 x i16> @llvm.arm.neon.vabs.v8i16(<8 x i16> %tmp1)
54 %tmp1 = load <4 x i32>, <4 x i32>* %A
55 %tmp2 = call <4 x i32> @llvm.arm.neon.vabs.v4i32(<4 x i32> %tmp1)
62 %tmp1 = load <4 x float>, <4 x float>* %A
63 %tmp2 = call <4 x float> @llvm.fabs.v4f32(<4 x float> %tmp1)
80 %tmp1 = load <8 x i8>, <8 x i8>* %A
81 %tmp2 = call <8 x i8> @llvm.arm.neon.vqabs.v8i8(<8 x i8> %tmp1)
88 %tmp1 = load <4 x i16>, <4 x i16>* %A
89 %tmp2 = call <4 x i16> @llvm.arm.neon.vqabs.v4i16(<4 x i16> %tmp1)
96 %tmp1 = load <2 x i32>, <2 x i32>* %A
97 %tmp2 = call <2 x i32> @llvm.arm.neon.vqabs.v2i32(<2 x i32> %tmp1)
104 %tmp1 = load <16 x i8>, <16 x i8>* %A
105 %tmp2 = call <16 x i8> @llvm.arm.neon.vqabs.v16i8(<16 x i8> %tmp1)
112 %tmp1 = load <8 x i16>, <8 x i16>* %A
113 %tmp2 = call <8 x i16> @llvm.arm.neon.vqabs.v8i16(<8 x i16> %tmp1)
120 %tmp1 = load <4 x i32>, <4 x i32>* %A
121 %tmp2 = call <4 x i32> @llvm.arm.neon.vqabs.v4i32(<4 x i32> %tmp1)