HomeSort by relevance Sort by last modified time
    Searched full:v2i64 (Results 1 - 25 of 410) sorted by null

1 2 3 4 5 6 7 8 91011>>

  /prebuilts/gcc/darwin-x86/mips/mips64el-linux-android-4.9/lib/gcc/mips64el-linux-android/4.9.x/include/
msa.h 43 typedef long long v2i64 __attribute__((vector_size(16), aligned(16))); typedef
56 extern v2i64 __builtin_msa_sll_d(v2i64, v2i64);
60 extern v2i64 __builtin_msa_slli_d(v2i64, unsigned char);
64 extern v2i64 __builtin_msa_sra_d(v2i64, v2i64);
68 extern v2i64 __builtin_msa_srai_d(v2i64, unsigned char)
    [all...]
  /prebuilts/gcc/linux-x86/mips/mips64el-linux-android-4.9/lib/gcc/mips64el-linux-android/4.9.x/include/
msa.h 43 typedef long long v2i64 __attribute__((vector_size(16), aligned(16))); typedef
56 extern v2i64 __builtin_msa_sll_d(v2i64, v2i64);
60 extern v2i64 __builtin_msa_slli_d(v2i64, unsigned char);
64 extern v2i64 __builtin_msa_sra_d(v2i64, v2i64);
68 extern v2i64 __builtin_msa_srai_d(v2i64, unsigned char)
    [all...]
  /external/swiftshader/third_party/LLVM/test/CodeGen/X86/
vec_ctbits.ll 3 declare <2 x i64> @llvm.cttz.v2i64(<2 x i64>)
4 declare <2 x i64> @llvm.ctlz.v2i64(<2 x i64>)
5 declare <2 x i64> @llvm.ctpop.v2i64(<2 x i64>)
8 %c = call <2 x i64> @llvm.cttz.v2i64(<2 x i64> %a)
12 %c = call <2 x i64> @llvm.ctlz.v2i64(<2 x i64> %a)
16 %c = call <2 x i64> @llvm.ctpop.v2i64(<2 x i64> %a)
widen_conv-1.ll 5 ; truncate v2i64 to v2i32
legalizedag_vec.ll 6 ; v2i64 is a legal type but with mmx disabled, i64 is an illegal type. When
  /external/llvm/test/Analysis/CostModel/X86/
scalarize.ll 14 declare %i8 @llvm.bswap.v2i64(%i8)
17 declare %i8 @llvm.cttz.v2i64(%i8)
27 ; CHECK32: cost of 1 {{.*}}bswap.v2i64
28 ; CHECK64: cost of 1 {{.*}}bswap.v2i64
29 %r3 = call %i8 @llvm.bswap.v2i64(%i8 undef)
34 ; CHECK32: cost of 10 {{.*}}cttz.v2i64
35 ; CHECK64: cost of 6 {{.*}}cttz.v2i64
36 %r5 = call %i8 @llvm.cttz.v2i64(%i8 undef)
  /external/libvpx/libvpx/vpx_dsp/mips/
vpx_convolve_copy_msa.c 25 out0 = __msa_copy_u_d((v2i64)src0, 0);
26 out1 = __msa_copy_u_d((v2i64)src1, 0);
27 out2 = __msa_copy_u_d((v2i64)src2, 0);
28 out3 = __msa_copy_u_d((v2i64)src3, 0);
29 out4 = __msa_copy_u_d((v2i64)src4, 0);
30 out5 = __msa_copy_u_d((v2i64)src5, 0);
31 out6 = __msa_copy_u_d((v2i64)src6, 0);
32 out7 = __msa_copy_u_d((v2i64)src7, 0);
42 out0 = __msa_copy_u_d((v2i64)src0, 0);
43 out1 = __msa_copy_u_d((v2i64)src1, 0)
    [all...]
deblock_msa.c 41 out9 = (v16u8)__msa_ilvl_d((v2i64)out8, (v2i64)out8); \
42 out11 = (v16u8)__msa_ilvl_d((v2i64)out10, (v2i64)out10); \
43 out13 = (v16u8)__msa_ilvl_d((v2i64)out12, (v2i64)out12); \
44 out15 = (v16u8)__msa_ilvl_d((v2i64)out14, (v2i64)out14); \
45 out1 = (v16u8)__msa_ilvl_d((v2i64)out0, (v2i64)out0);
    [all...]
loopfilter_4_msa.c 32 p1_d = __msa_copy_u_d((v2i64)p1_out, 0);
33 p0_d = __msa_copy_u_d((v2i64)p0_out, 0);
34 q0_d = __msa_copy_u_d((v2i64)q0_out, 0);
35 q1_d = __msa_copy_u_d((v2i64)q1_out, 0);
54 thresh0 = (v16u8)__msa_ilvr_d((v2i64)thresh1, (v2i64)thresh0);
58 b_limit0 = (v16u8)__msa_ilvr_d((v2i64)b_limit1, (v2i64)b_limit0);
62 limit0 = (v16u8)__msa_ilvr_d((v2i64)limit1, (v2i64)limit0)
    [all...]
loopfilter_8_msa.c 37 flat = (v16u8)__msa_ilvr_d((v2i64)zero, (v2i64)flat);
40 p1_d = __msa_copy_u_d((v2i64)p1_out, 0);
41 p0_d = __msa_copy_u_d((v2i64)p0_out, 0);
42 q0_d = __msa_copy_u_d((v2i64)q0_out, 0);
43 q1_d = __msa_copy_u_d((v2i64)q1_out, 0);
64 p2_d = __msa_copy_u_d((v2i64)p2_out, 0);
65 p1_d = __msa_copy_u_d((v2i64)p1_out, 0);
66 p0_d = __msa_copy_u_d((v2i64)p0_out, 0);
67 q0_d = __msa_copy_u_d((v2i64)q0_out, 0)
    [all...]
loopfilter_16_msa.c 455 flat = (v16u8)__msa_ilvr_d((v2i64)zero, (v2i64)flat);
458 p1_d = __msa_copy_u_d((v2i64)p1_out, 0);
459 p0_d = __msa_copy_u_d((v2i64)p0_out, 0);
460 q0_d = __msa_copy_u_d((v2i64)q0_out, 0);
461 q1_d = __msa_copy_u_d((v2i64)q1_out, 0);
491 p2_d = __msa_copy_u_d((v2i64)p2_out, 0);
492 p1_d = __msa_copy_u_d((v2i64)p1_out, 0);
493 p0_d = __msa_copy_u_d((v2i64)p0_out, 0);
494 q0_d = __msa_copy_u_d((v2i64)q0_out, 0)
    [all...]
  /external/swiftshader/third_party/LLVM/lib/Target/CellSPU/
SPU64InstrInfo.td 21 // 4. v2i64 setcc results are v4i32, which can be converted to a FSM mask (TODO)
24 // 5. The code sequences for r64 and v2i64 are probably overly conservative,
67 // v2i64 seteq (equality): the setcc result is v4i32
71 def v2i64: CodeFrag<(i32 (COPY_TO_REGCLASS CEQv2i64compare.Fragment, R32C))>;
83 def : Pat<(seteq (v2i64 VECREG:$rA), (v2i64 VECREG:$rB)), I64EQv2i64.Fragment>;
120 def v2i64: CodeFrag<CLGTv2i64compare.Fragment>;
132 //def : Pat<(setugt (v2i64 VECREG:$rA), (v2i64 VECREG:$rB)),
154 def v2i64: CodeFrag<CLGEv2i64compare.Fragment>
    [all...]
  /external/llvm/test/CodeGen/PowerPC/
vec_popcnt.ll 2 ; In addition, check the conversions to/from the v2i64 VMX register that was also added in P8.
9 declare <2 x i64> @llvm.ctpop.v2i64(<2 x i64>) nounwind readnone
39 %vcnt = tail call <2 x i64> @llvm.ctpop.v2i64(<2 x i64> %x)
48 %vcnt = tail call <2 x i64> @llvm.ctpop.v2i64(<2 x i64> %tmp)
58 %vcnt = tail call <2 x i64> @llvm.ctpop.v2i64(<2 x i64> %tmp)
67 %vcnt = tail call <2 x i64> @llvm.ctpop.v2i64(<2 x i64> %tmp)
vec_clz.ll 8 declare <2 x i64> @llvm.ctlz.v2i64(<2 x i64>) nounwind readnone
35 %vcnt = tail call <2 x i64> @llvm.ctlz.v2i64(<2 x i64> %x)
  /external/llvm/test/CodeGen/SystemZ/
vec-move-06.ll 5 ; Test the basic v2i64 usage.
vec-shift-07.ll 71 ; Test a v2i1->v2i64 extension.
82 ; Test a v2i8->v2i64 extension.
92 ; Test a v2i16->v2i64 extension.
102 ; Test a v2i32->v2i64 extension.
112 ; Test an alternative v2i8->v2i64 extension.
122 ; Test an alternative v2i16->v2i64 extension.
132 ; Test an alternative v2i32->v2i64 extension.
142 ; Test an extraction-based v2i8->v2i64 extension.
156 ; Test an extraction-based v2i16->v2i64 extension.
170 ; Test an extraction-based v2i32->v2i64 extension
    [all...]
  /external/libvpx/libvpx/vp8/common/mips/msa/
loopfilter_filters_msa.c 223 thresh0 = (v16u8)__msa_ilvr_d((v2i64)thresh1, (v2i64)thresh0);
227 b_limit0 = (v16u8)__msa_ilvr_d((v2i64)b_limit1, (v2i64)b_limit0);
231 limit0 = (v16u8)__msa_ilvr_d((v2i64)limit1, (v2i64)limit0);
263 thresh0 = (v16u8)__msa_ilvr_d((v2i64)thresh1, (v2i64)thresh0);
267 b_limit0 = (v16u8)__msa_ilvr_d((v2i64)b_limit1, (v2i64)b_limit0)
    [all...]
  /external/llvm/test/CodeGen/AArch64/
arm64-vcvt_n.ll 35 %vcvt_n1 = tail call <2 x double> @llvm.aarch64.neon.vcvtfxu2fp.v2f64.v2i64(<2 x i64> %a, i32 12)
40 %vcvt_n1 = tail call <2 x double> @llvm.aarch64.neon.vcvtfxs2fp.v2f64.v2i64(<2 x i64> %a, i32 9)
48 declare <2 x double> @llvm.aarch64.neon.vcvtfxu2fp.v2f64.v2i64(<2 x i64>, i32) nounwind readnone
49 declare <2 x double> @llvm.aarch64.neon.vcvtfxs2fp.v2f64.v2i64(<2 x i64>, i32) nounwind readnone
arm64-detect-vec-redux.ll 20 %vpaddq_v2.i = tail call <2 x i64> @llvm.aarch64.neon.addp.v2i64(<2 x i64> undef, <2 x i64> %1) #2
21 %vqdmlal2.i = tail call <2 x i64> @llvm.aarch64.neon.sqdmull.v2i64(<2 x i32> undef, <2 x i32> undef) #2
22 %vqdmlal_v3.i = tail call <2 x i64> @llvm.aarch64.neon.sqadd.v2i64(<2 x i64> %vpaddq_v2.i, <2 x i64> %vqdmlal2.i) #2
39 declare <2 x i64> @llvm.aarch64.neon.sqdmull.v2i64(<2 x i32>, <2 x i32>) #1
42 declare <2 x i64> @llvm.aarch64.neon.sqadd.v2i64(<2 x i64>, <2 x i64>) #1
45 declare <2 x i64> @llvm.aarch64.neon.addp.v2i64(<2 x i64>, <2 x i64>) #1
  /external/llvm/test/CodeGen/X86/
legalizedag_vec.ll 6 ; v2i64 is a legal type but with mmx disabled, i64 is an illegal type. When
  /external/clang/test/CodeGen/
arm-neon-misc.c 17 // CHECK: call <2 x i64> @llvm.arm.neon.vld1.v2i64.p0i8
27 // CHECK: call <2 x i64> @llvm.arm.neon.vld1.v2i64.p0i8
  /external/llvm/test/CodeGen/ARM/
2010-06-29-PartialRedefFastAlloc.ll 19 %0 = call <2 x i64> @llvm.arm.neon.vld1.v2i64.p0i8(i8* %arg, i32 1)
25 declare <2 x i64> @llvm.arm.neon.vld1.v2i64.p0i8(i8*, i32) nounwind readonly
  /external/swiftshader/third_party/LLVM/test/CodeGen/ARM/
2010-06-29-PartialRedefFastAlloc.ll 19 %0 = call <2 x i64> @llvm.arm.neon.vld1.v2i64(i8* %arg, i32 1)
25 declare <2 x i64> @llvm.arm.neon.vld1.v2i64(i8*, i32) nounwind readonly
  /external/libyuv/files/source/
row_msa.cc 48 out_y = (v16u8)__msa_insert_d((v2i64)zero_m, 0, (int64)y_m); \
190 reg8_m = (v8u16)__msa_pckev_d((v2i64)reg4_m, (v2i64)reg0_m); \
191 reg9_m = (v8u16)__msa_pckev_d((v2i64)reg5_m, (v2i64)reg1_m); \
192 reg8_m += (v8u16)__msa_pckod_d((v2i64)reg4_m, (v2i64)reg0_m); \
193 reg9_m += (v8u16)__msa_pckod_d((v2i64)reg5_m, (v2i64)reg1_m); \
194 reg0_m = (v8u16)__msa_pckev_d((v2i64)reg6_m, (v2i64)reg2_m);
    [all...]
  /external/llvm/lib/Target/AArch64/
AArch64CallingConvention.td 27 CCIfType<[v2f64, v4f32], CCBitConvertToType<v2i64>>,
33 CCIfBigEndian<CCIfType<[v2i64, v2f64, v4i32, v4f32, v8i16, v8f16, v16i8],
76 CCIfType<[f128, v2i64, v4i32, v8i16, v16i8, v4f32, v2f64, v8f16],
84 CCIfType<[f128, v2i64, v4i32, v8i16, v16i8, v4f32, v2f64, v8f16],
90 CCIfType<[v2f64, v4f32], CCBitConvertToType<v2i64>>,
98 CCIfBigEndian<CCIfType<[v2i64, v2f64, v4i32, v4f32, v8i16, v8f16, v16i8],
114 CCIfType<[f128, v2i64, v4i32, v8i16, v16i8, v4f32, v2f64, v8f16],
125 CCIfType<[v2f64, v4f32, f128], CCBitConvertToType<v2i64>>,
165 CCIfType<[v2i64, v4i32, v8i16, v16i8, v4f32, v2f64, v8f16],
174 CCIfType<[v2i64, v4i32, v8i16, v16i8, v4f32, v2f64, v8f16]
    [all...]

Completed in 585 milliseconds

1 2 3 4 5 6 7 8 91011>>