Home | History | Annotate | Download | only in ARM

Lines Matching full:tmp1

6 	%tmp1 = load <4 x i16>, <4 x i16>* %A
8 %tmp3 = call <4 x i16> @llvm.arm.neon.vpadals.v4i16.v8i8(<4 x i16> %tmp1, <8 x i8> %tmp2)
15 %tmp1 = load <2 x i32>, <2 x i32>* %A
17 %tmp3 = call <2 x i32> @llvm.arm.neon.vpadals.v2i32.v4i16(<2 x i32> %tmp1, <4 x i16> %tmp2)
24 %tmp1 = load <1 x i64>, <1 x i64>* %A
26 %tmp3 = call <1 x i64> @llvm.arm.neon.vpadals.v1i64.v2i32(<1 x i64> %tmp1, <2 x i32> %tmp2)
33 %tmp1 = load <4 x i16>, <4 x i16>* %A
35 %tmp3 = call <4 x i16> @llvm.arm.neon.vpadalu.v4i16.v8i8(<4 x i16> %tmp1, <8 x i8> %tmp2)
42 %tmp1 = load <2 x i32>, <2 x i32>* %A
44 %tmp3 = call <2 x i32> @llvm.arm.neon.vpadalu.v2i32.v4i16(<2 x i32> %tmp1, <4 x i16> %tmp2)
51 %tmp1 = load <1 x i64>, <1 x i64>* %A
53 %tmp3 = call <1 x i64> @llvm.arm.neon.vpadalu.v1i64.v2i32(<1 x i64> %tmp1, <2 x i32> %tmp2)
60 %tmp1 = load <8 x i16>, <8 x i16>* %A
62 %tmp3 = call <8 x i16> @llvm.arm.neon.vpadals.v8i16.v16i8(<8 x i16> %tmp1, <16 x i8> %tmp2)
69 %tmp1 = load <4 x i32>, <4 x i32>* %A
71 %tmp3 = call <4 x i32> @llvm.arm.neon.vpadals.v4i32.v8i16(<4 x i32> %tmp1, <8 x i16> %tmp2)
78 %tmp1 = load <2 x i64>, <2 x i64>* %A
80 %tmp3 = call <2 x i64> @llvm.arm.neon.vpadals.v2i64.v4i32(<2 x i64> %tmp1, <4 x i32> %tmp2)
87 %tmp1 = load <8 x i16>, <8 x i16>* %A
89 %tmp3 = call <8 x i16> @llvm.arm.neon.vpadalu.v8i16.v16i8(<8 x i16> %tmp1, <16 x i8> %tmp2)
96 %tmp1 = load <4 x i32>, <4 x i32>* %A
98 %tmp3 = call <4 x i32> @llvm.arm.neon.vpadalu.v4i32.v8i16(<4 x i32> %tmp1, <8 x i16> %tmp2)
105 %tmp1 = load <2 x i64>, <2 x i64>* %A
107 %tmp3 = call <2 x i64> @llvm.arm.neon.vpadalu.v2i64.v4i32(<2 x i64> %tmp1, <4 x i32> %tmp2)