/external/llvm/test/CodeGen/ARM/ |
movcc-double.ll | 13 %cmp0 = icmp ult i32 %a0, %a1 15 %and = and i1 %cmp0, %cmp1 28 %cmp0 = icmp ult i32 %a0, %a1 30 %and = or i1 %cmp0, %cmp1 43 %cmp0 = icmp ult i32 %a0, %a1 45 %or = or i1 %cmp0, %cmp1
|
vuzp.ll | 301 define <8 x i8> @vuzp_trunc(<8 x i8> %in0, <8 x i8> %in1, <8 x i32> %cmp0, <8 x i32> %cmp1) { 310 %c = icmp ult <8 x i32> %cmp0, %cmp1 319 <4 x i32> %cmp0, <4 x i32> %cmp1, <4 x i8> *%cmp2_ptr) { 326 %c0 = icmp ult <4 x i32> %cmp0, %cmp1 335 <4 x i32> %cmp0, <4 x i32> %cmp1, <4 x i8> *%cmp2_ptr) { 341 %c0 = icmp ult <4 x i32> %cmp0, %cmp1 348 <4 x i32> %cmp0, <4 x i32> %cmp1, <4 x i8> *%cmp2_ptr) { 354 %c0 = icmp ult <4 x i32> %cmp0, %cmp1 363 <5 x i32> %cmp0, <5 x i32> %cmp1, <5 x i8> *%cmp2_ptr) { 368 %c0 = icmp ult <5 x i32> %cmp0, %cmp [all...] |
/external/llvm/test/CodeGen/X86/ |
cmov-double.ll | 14 %cmp0 = icmp ult i32 %a0, %a1 16 %and = and i1 %cmp0, %cmp1 30 %cmp0 = icmp ult i32 %a0, %a1 32 %and = or i1 %cmp0, %cmp1 45 %cmp0 = icmp ult i32 %a0, %a1 47 %or = or i1 %cmp0, %cmp1
|
/external/llvm/test/Transforms/JumpThreading/ |
phi-known.ll | 8 %cmp0 = icmp eq i8* %p, null 9 br i1 %cmp0, label %exit, label %loop 25 %cmp0 = icmp eq i8* %p, null 26 br i1 %cmp0, label %exit, label %loop 53 %cmp0 = icmp eq i8* %p, null 54 br i1 %cmp0, label %exit, label %loop
|
/external/llvm/test/Transforms/Util/ |
flattencfg.ll | 10 %cmp0 = icmp eq i32 %in_a, -1 12 %cond0 = and i1 %cmp0, %cmp1
|
/external/llvm/test/CodeGen/Mips/ |
dagcombine_crash.ll | 18 %cmp0 = icmp ult i32 %arg, 23 20 %and = and i1 %cmp0, %cmp1
|
/external/llvm/test/CodeGen/SPARC/ |
spillsize.ll | 15 %cmp0 = icmp ult i64 %val0, 385672958347594845 16 %cm80 = zext i1 %cmp0 to i64
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
vp9_highbd_block_error_intrin_sse2.c | 21 __m128i max, min, cmp0, cmp1, cmp2, cmp3; local 35 cmp0 = _mm_xor_si128(_mm_cmpgt_epi32(mm_coeff, max), 43 test = _mm_movemask_epi8(_mm_or_si128(_mm_or_si128(cmp0, cmp1),
|
/external/llvm/test/CodeGen/Hexagon/ |
ifcvt-edge-weight.ll | 22 %cmp0 = call i1 @pred() 23 br i1 %cmp0, label %if.else2, label %if.then0, !prof !1
|
/external/vulkan-validation-layers/libs/glm/detail/ |
intrinsic_common.inl | 164 __m128 Cmp0 = _mm_cmplt_ps(x, zero); 167 __m128 And0 = _mm_and_ps(Cmp0, Neg); 177 __m128 cmp0 = _mm_cmplt_ps(x, rnd0); 178 __m128 and0 = _mm_and_ps(cmp0, glm::detail::_ps_1); 213 __m128 cmp0 = _mm_cmpgt_ps(x, rnd0); 214 __m128 and0 = _mm_and_ps(cmp0, glm::detail::_ps_1);
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
fwd_txfm_sse2.h | 40 __m128i cmp0 = _mm_or_si128(_mm_cmpeq_epi16(*preg0, max_overflow), local 44 cmp0 = _mm_or_si128(cmp0, cmp1); 45 return _mm_movemask_epi8(cmp0); 54 __m128i cmp0 = _mm_or_si128(_mm_cmpeq_epi16(*preg0, max_overflow), local 62 cmp0 = _mm_or_si128(_mm_or_si128(cmp0, cmp1), _mm_or_si128(cmp2, cmp3)); 63 return _mm_movemask_epi8(cmp0);
|
/external/llvm/test/CodeGen/AMDGPU/ |
drop-mem-operand-move-smrd.ll | 20 %cmp0 = icmp eq i32 %c, 0 21 br i1 %cmp0, label %for.body, label %exit
|
llvm.AMDGPU.div_fmas.ll | 112 ; SI-DAG: v_cmp_eq_i32_e32 [[CMP0:vcc]], 0, v{{[0-9]+}} 114 ; SI: s_and_b64 vcc, [[CMP0]], [[CMP1]] 128 %cmp0 = icmp eq i32 %tid, 0 130 %and = and i1 %cmp0, %cmp1 165 %cmp0 = icmp eq i32 %tid, 0 166 br i1 %cmp0, label %bb, label %exit
|
fdiv.f64.ll | 19 ; SI-DAG: v_cmp_eq_i32_e64 [[CMP0:s\[[0-9]+:[0-9]+\]]], {{v[0-9]+}}, {{v[0-9]+}} 20 ; SI-DAG: s_xor_b64 vcc, [[CMP0]], vcc
|
or.ll | 173 %cmp0 = icmp eq i32 %a, %b 175 %or = or i1 %cmp0, %cmp1
|
trunc-cmp-constant.ll | 136 ; XSI: v_cmp_eq_i32_e64 [[CMP0:s\[[0-9]+:[0-9]+\]]], [[TMP]], 0{{$}} 137 ; XSI-NEXT: v_cndmask_b32_e64 [[RESULT:v[0-9]+]], 0, 1, [[CMP0]]
|
/external/llvm/test/Transforms/SLPVectorizer/X86/ |
insert-element-build-vector.ll | 23 %cmp0 = icmp ne i32 %c0, 0 27 %s0 = select i1 %cmp0, float %a0, float %b0 57 %cmp0 = icmp ne i32 %c0, 0 61 %s0 = select i1 %cmp0, float %a0, float %b0 99 %cmp0 = icmp ne i32 %c0, 0 103 %s0 = select i1 %cmp0, float %a0, float %b0 134 %cmp0 = icmp ne i32 %c0, 0 138 %s0 = select i1 %cmp0, float %a0, float %b0 167 %cmp0 = icmp ne i32 %c0, 0 171 %s0 = select i1 %cmp0, float %a0, float %b [all...] |
/external/libavc/common/x86/ |
ih264_resi_trans_quant_sse42.c | 114 __m128i sum0, sum1, sum2, cmp0, cmp1; local 289 cmp0 = _mm_cmpeq_epi16(temp0, zero_8x16b); 292 mask0 = _mm_movemask_epi8(cmp0); 301 cmp0 = _mm_and_si128(temp_1, cmp0); 302 sum0 = _mm_hadd_epi16(cmp0, zero_8x16b); 388 __m128i cmp0, cmp1, sum0, sum1, sum2; local 574 cmp0 = _mm_cmpeq_epi16(temp0, zero_8x16b); 577 mask0 = _mm_movemask_epi8(cmp0); 586 cmp0 = _mm_and_si128(temp_1, cmp0) 673 __m128i cmp0, cmp1, sum0, sum1, sum2; local 901 __m128i cmp, cmp0, cmp1; local [all...] |
/external/vulkan-validation-layers/libs/glm/gtx/ |
simd_vec4.inl | 322 __m128 Cmp0 = _mm_cmplt_ps(x.Data, glm::detail::zero); 325 __m128 And0 = _mm_and_ps(Sub0, Cmp0); 464 __m128 cmp0 = _mm_cmpngt_ps(x.Data, edge.Data); 465 return _mm_max_ps(_mm_min_ps(cmp0, _mm_setzero_ps()), detail::one); 474 __m128 cmp0 = _mm_cmpngt_ps(x.Data, _mm_set1_ps(edge)); 475 return _mm_max_ps(_mm_min_ps(cmp0, _mm_setzero_ps()), detail::one);
|
/external/opencv3/modules/cudev/include/opencv2/cudev/block/ |
reduce.hpp | 112 class Cmp0, class Cmp1, class Cmp2, class Cmp3, class Cmp4, class Cmp5, class Cmp6, class Cmp7, class Cmp8, class Cmp9> 118 const tuple<Cmp0, Cmp1, Cmp2, Cmp3, Cmp4, Cmp5, Cmp6, Cmp7, Cmp8, Cmp9>& cmp) 125 const tuple<Cmp0, Cmp1, Cmp2, Cmp3, Cmp4, Cmp5, Cmp6, Cmp7, Cmp8, Cmp9>&
|
/external/llvm/test/Analysis/ScalarEvolution/ |
avoid-assume-hang.ll | 9 %cmp0 = icmp ne i64 %i.ext, %a 10 call void @llvm.assume(i1 %cmp0)
|
/external/llvm/test/Transforms/LoopUnswitch/ |
2015-06-17-Metadata.ll | 9 %cmp0 = icmp sgt i32 %b, 0 10 br i1 %cmp0, label %for.body, label %for.cond.cleanup
|
/external/opencv3/modules/core/include/opencv2/core/cuda/ |
reduce.hpp | 103 class Cmp0, class Cmp1, class Cmp2, class Cmp3, class Cmp4, class Cmp5, class Cmp6, class Cmp7, class Cmp8, class Cmp9> 109 const thrust::tuple<Cmp0, Cmp1, Cmp2, Cmp3, Cmp4, Cmp5, Cmp6, Cmp7, Cmp8, Cmp9>& cmp) 116 const thrust::tuple<Cmp0, Cmp1, Cmp2, Cmp3, Cmp4, Cmp5, Cmp6, Cmp7, Cmp8, Cmp9>&
|
/external/opencv3/modules/cudev/include/opencv2/cudev/warp/ |
reduce.hpp | 108 class Cmp0, class Cmp1, class Cmp2, class Cmp3, class Cmp4, class Cmp5, class Cmp6, class Cmp7, class Cmp8, class Cmp9> 114 const tuple<Cmp0, Cmp1, Cmp2, Cmp3, Cmp4, Cmp5, Cmp6, Cmp7, Cmp8, Cmp9>& cmp) 121 const tuple<Cmp0, Cmp1, Cmp2, Cmp3, Cmp4, Cmp5, Cmp6, Cmp7, Cmp8, Cmp9>&
|
/external/clang/test/CodeGenCXX/ |
microsoft-abi-member-pointers.cpp | 250 // CHECK: %[[cmp0:.*]] = icmp ne i32 %[[mp0]], 0 253 // CHECK: %[[and0:.*]] = or i1 %[[cmp0]], %[[cmp1]] 272 // CHECK: %[[cmp0:.*]] = icmp ne i8* %[[mp0]], null 273 // CHECK: ret i1 %[[cmp0]] 417 // CHECK: %[[cmp0:.*]] = icmp eq i8* %[[lhs0]], %{{.*}} 431 // CHECK: %{{.*}} = and i1 %[[bits_or_null]], %[[cmp0]] 444 // CHECK: %[[cmp0:.*]] = icmp ne i8* %[[lhs0]], %{{.*}} 458 // CHECK: %{{.*}} = or i1 %[[bits_or_null]], %[[cmp0]]
|