HomeSort by relevance Sort by last modified time
    Searched full:cmp0 (Results 1 - 25 of 48) sorted by null

1 2

  /external/llvm/test/CodeGen/ARM/
movcc-double.ll 13 %cmp0 = icmp ult i32 %a0, %a1
15 %and = and i1 %cmp0, %cmp1
28 %cmp0 = icmp ult i32 %a0, %a1
30 %and = or i1 %cmp0, %cmp1
43 %cmp0 = icmp ult i32 %a0, %a1
45 %or = or i1 %cmp0, %cmp1
vuzp.ll 301 define <8 x i8> @vuzp_trunc(<8 x i8> %in0, <8 x i8> %in1, <8 x i32> %cmp0, <8 x i32> %cmp1) {
310 %c = icmp ult <8 x i32> %cmp0, %cmp1
319 <4 x i32> %cmp0, <4 x i32> %cmp1, <4 x i8> *%cmp2_ptr) {
326 %c0 = icmp ult <4 x i32> %cmp0, %cmp1
335 <4 x i32> %cmp0, <4 x i32> %cmp1, <4 x i8> *%cmp2_ptr) {
341 %c0 = icmp ult <4 x i32> %cmp0, %cmp1
348 <4 x i32> %cmp0, <4 x i32> %cmp1, <4 x i8> *%cmp2_ptr) {
354 %c0 = icmp ult <4 x i32> %cmp0, %cmp1
363 <5 x i32> %cmp0, <5 x i32> %cmp1, <5 x i8> *%cmp2_ptr) {
368 %c0 = icmp ult <5 x i32> %cmp0, %cmp
    [all...]
  /external/llvm/test/CodeGen/X86/
cmov-double.ll 14 %cmp0 = icmp ult i32 %a0, %a1
16 %and = and i1 %cmp0, %cmp1
30 %cmp0 = icmp ult i32 %a0, %a1
32 %and = or i1 %cmp0, %cmp1
45 %cmp0 = icmp ult i32 %a0, %a1
47 %or = or i1 %cmp0, %cmp1
  /external/llvm/test/Transforms/JumpThreading/
phi-known.ll 8 %cmp0 = icmp eq i8* %p, null
9 br i1 %cmp0, label %exit, label %loop
25 %cmp0 = icmp eq i8* %p, null
26 br i1 %cmp0, label %exit, label %loop
53 %cmp0 = icmp eq i8* %p, null
54 br i1 %cmp0, label %exit, label %loop
  /external/llvm/test/Transforms/Util/
flattencfg.ll 10 %cmp0 = icmp eq i32 %in_a, -1
12 %cond0 = and i1 %cmp0, %cmp1
  /external/llvm/test/CodeGen/Mips/
dagcombine_crash.ll 18 %cmp0 = icmp ult i32 %arg, 23
20 %and = and i1 %cmp0, %cmp1
  /external/llvm/test/CodeGen/SPARC/
spillsize.ll 15 %cmp0 = icmp ult i64 %val0, 385672958347594845
16 %cm80 = zext i1 %cmp0 to i64
  /external/libvpx/libvpx/vp9/encoder/x86/
vp9_highbd_block_error_intrin_sse2.c 21 __m128i max, min, cmp0, cmp1, cmp2, cmp3; local
35 cmp0 = _mm_xor_si128(_mm_cmpgt_epi32(mm_coeff, max),
43 test = _mm_movemask_epi8(_mm_or_si128(_mm_or_si128(cmp0, cmp1),
  /external/llvm/test/CodeGen/Hexagon/
ifcvt-edge-weight.ll 22 %cmp0 = call i1 @pred()
23 br i1 %cmp0, label %if.else2, label %if.then0, !prof !1
  /external/vulkan-validation-layers/libs/glm/detail/
intrinsic_common.inl 164 __m128 Cmp0 = _mm_cmplt_ps(x, zero);
167 __m128 And0 = _mm_and_ps(Cmp0, Neg);
177 __m128 cmp0 = _mm_cmplt_ps(x, rnd0);
178 __m128 and0 = _mm_and_ps(cmp0, glm::detail::_ps_1);
213 __m128 cmp0 = _mm_cmpgt_ps(x, rnd0);
214 __m128 and0 = _mm_and_ps(cmp0, glm::detail::_ps_1);
  /external/libvpx/libvpx/vpx_dsp/x86/
fwd_txfm_sse2.h 40 __m128i cmp0 = _mm_or_si128(_mm_cmpeq_epi16(*preg0, max_overflow), local
44 cmp0 = _mm_or_si128(cmp0, cmp1);
45 return _mm_movemask_epi8(cmp0);
54 __m128i cmp0 = _mm_or_si128(_mm_cmpeq_epi16(*preg0, max_overflow), local
62 cmp0 = _mm_or_si128(_mm_or_si128(cmp0, cmp1), _mm_or_si128(cmp2, cmp3));
63 return _mm_movemask_epi8(cmp0);
  /external/llvm/test/CodeGen/AMDGPU/
drop-mem-operand-move-smrd.ll 20 %cmp0 = icmp eq i32 %c, 0
21 br i1 %cmp0, label %for.body, label %exit
llvm.AMDGPU.div_fmas.ll 112 ; SI-DAG: v_cmp_eq_i32_e32 [[CMP0:vcc]], 0, v{{[0-9]+}}
114 ; SI: s_and_b64 vcc, [[CMP0]], [[CMP1]]
128 %cmp0 = icmp eq i32 %tid, 0
130 %and = and i1 %cmp0, %cmp1
165 %cmp0 = icmp eq i32 %tid, 0
166 br i1 %cmp0, label %bb, label %exit
fdiv.f64.ll 19 ; SI-DAG: v_cmp_eq_i32_e64 [[CMP0:s\[[0-9]+:[0-9]+\]]], {{v[0-9]+}}, {{v[0-9]+}}
20 ; SI-DAG: s_xor_b64 vcc, [[CMP0]], vcc
or.ll 173 %cmp0 = icmp eq i32 %a, %b
175 %or = or i1 %cmp0, %cmp1
trunc-cmp-constant.ll 136 ; XSI: v_cmp_eq_i32_e64 [[CMP0:s\[[0-9]+:[0-9]+\]]], [[TMP]], 0{{$}}
137 ; XSI-NEXT: v_cndmask_b32_e64 [[RESULT:v[0-9]+]], 0, 1, [[CMP0]]
  /external/llvm/test/Transforms/SLPVectorizer/X86/
insert-element-build-vector.ll 23 %cmp0 = icmp ne i32 %c0, 0
27 %s0 = select i1 %cmp0, float %a0, float %b0
57 %cmp0 = icmp ne i32 %c0, 0
61 %s0 = select i1 %cmp0, float %a0, float %b0
99 %cmp0 = icmp ne i32 %c0, 0
103 %s0 = select i1 %cmp0, float %a0, float %b0
134 %cmp0 = icmp ne i32 %c0, 0
138 %s0 = select i1 %cmp0, float %a0, float %b0
167 %cmp0 = icmp ne i32 %c0, 0
171 %s0 = select i1 %cmp0, float %a0, float %b
    [all...]
  /external/libavc/common/x86/
ih264_resi_trans_quant_sse42.c 114 __m128i sum0, sum1, sum2, cmp0, cmp1; local
289 cmp0 = _mm_cmpeq_epi16(temp0, zero_8x16b);
292 mask0 = _mm_movemask_epi8(cmp0);
301 cmp0 = _mm_and_si128(temp_1, cmp0);
302 sum0 = _mm_hadd_epi16(cmp0, zero_8x16b);
388 __m128i cmp0, cmp1, sum0, sum1, sum2; local
574 cmp0 = _mm_cmpeq_epi16(temp0, zero_8x16b);
577 mask0 = _mm_movemask_epi8(cmp0);
586 cmp0 = _mm_and_si128(temp_1, cmp0)
673 __m128i cmp0, cmp1, sum0, sum1, sum2; local
901 __m128i cmp, cmp0, cmp1; local
    [all...]
  /external/vulkan-validation-layers/libs/glm/gtx/
simd_vec4.inl 322 __m128 Cmp0 = _mm_cmplt_ps(x.Data, glm::detail::zero);
325 __m128 And0 = _mm_and_ps(Sub0, Cmp0);
464 __m128 cmp0 = _mm_cmpngt_ps(x.Data, edge.Data);
465 return _mm_max_ps(_mm_min_ps(cmp0, _mm_setzero_ps()), detail::one);
474 __m128 cmp0 = _mm_cmpngt_ps(x.Data, _mm_set1_ps(edge));
475 return _mm_max_ps(_mm_min_ps(cmp0, _mm_setzero_ps()), detail::one);
  /external/opencv3/modules/cudev/include/opencv2/cudev/block/
reduce.hpp 112 class Cmp0, class Cmp1, class Cmp2, class Cmp3, class Cmp4, class Cmp5, class Cmp6, class Cmp7, class Cmp8, class Cmp9>
118 const tuple<Cmp0, Cmp1, Cmp2, Cmp3, Cmp4, Cmp5, Cmp6, Cmp7, Cmp8, Cmp9>& cmp)
125 const tuple<Cmp0, Cmp1, Cmp2, Cmp3, Cmp4, Cmp5, Cmp6, Cmp7, Cmp8, Cmp9>&
  /external/llvm/test/Analysis/ScalarEvolution/
avoid-assume-hang.ll 9 %cmp0 = icmp ne i64 %i.ext, %a
10 call void @llvm.assume(i1 %cmp0)
  /external/llvm/test/Transforms/LoopUnswitch/
2015-06-17-Metadata.ll 9 %cmp0 = icmp sgt i32 %b, 0
10 br i1 %cmp0, label %for.body, label %for.cond.cleanup
  /external/opencv3/modules/core/include/opencv2/core/cuda/
reduce.hpp 103 class Cmp0, class Cmp1, class Cmp2, class Cmp3, class Cmp4, class Cmp5, class Cmp6, class Cmp7, class Cmp8, class Cmp9>
109 const thrust::tuple<Cmp0, Cmp1, Cmp2, Cmp3, Cmp4, Cmp5, Cmp6, Cmp7, Cmp8, Cmp9>& cmp)
116 const thrust::tuple<Cmp0, Cmp1, Cmp2, Cmp3, Cmp4, Cmp5, Cmp6, Cmp7, Cmp8, Cmp9>&
  /external/opencv3/modules/cudev/include/opencv2/cudev/warp/
reduce.hpp 108 class Cmp0, class Cmp1, class Cmp2, class Cmp3, class Cmp4, class Cmp5, class Cmp6, class Cmp7, class Cmp8, class Cmp9>
114 const tuple<Cmp0, Cmp1, Cmp2, Cmp3, Cmp4, Cmp5, Cmp6, Cmp7, Cmp8, Cmp9>& cmp)
121 const tuple<Cmp0, Cmp1, Cmp2, Cmp3, Cmp4, Cmp5, Cmp6, Cmp7, Cmp8, Cmp9>&
  /external/clang/test/CodeGenCXX/
microsoft-abi-member-pointers.cpp 250 // CHECK: %[[cmp0:.*]] = icmp ne i32 %[[mp0]], 0
253 // CHECK: %[[and0:.*]] = or i1 %[[cmp0]], %[[cmp1]]
272 // CHECK: %[[cmp0:.*]] = icmp ne i8* %[[mp0]], null
273 // CHECK: ret i1 %[[cmp0]]
417 // CHECK: %[[cmp0:.*]] = icmp eq i8* %[[lhs0]], %{{.*}}
431 // CHECK: %{{.*}} = and i1 %[[bits_or_null]], %[[cmp0]]
444 // CHECK: %[[cmp0:.*]] = icmp ne i8* %[[lhs0]], %{{.*}}
458 // CHECK: %{{.*}} = or i1 %[[bits_or_null]], %[[cmp0]]

Completed in 708 milliseconds

1 2