HomeSort by relevance Sort by last modified time
    Searched full:src2 (Results 151 - 175 of 559) sorted by null

1 2 3 4 5 67 8 91011>>

  /external/skia/tests/
CPlusPlusEleven.cpp 25 Moveable src2, dst2; dst2 = Move(src2); local
  /external/llvm/lib/Target/X86/
X86InstrSSE.td 247 def rr : SI<opc, MRMSrcReg, (outs RC:$dst), (ins RC:$src1, RC:$src2),
249 !strconcat(OpcodeStr, "\t{$src2, $dst|$dst, $src2}"),
250 !strconcat(OpcodeStr, "\t{$src2, $src1, $dst|$dst, $src1, $src2}")),
251 [(set RC:$dst, (OpNode RC:$src1, RC:$src2))], itins.rr>,
254 def rm : SI<opc, MRMSrcMem, (outs RC:$dst), (ins RC:$src1, x86memop:$src2),
256 !strconcat(OpcodeStr, "\t{$src2, $dst|$dst, $src2}"),
257 !strconcat(OpcodeStr, "\t{$src2, $src1, $dst|$dst, $src1, $src2}"))
    [all...]
X86InstrAVX512.td 416 (ins VR512:$src1, From.RC:$src2, u8imm:$src3),
418 "\t{$src3, $src2, $src1, $dst|"
419 "$dst, $src1, $src2, $src3}",
421 (From.VT From.RC:$src2),
427 (ins VR512:$src1, From.MemOp:$src2, u8imm:$src3),
429 "\t{$src3, $src2, $src1, $dst|"
430 "$dst, $src1, $src2, $src3}",
448 (AltTo.VT VR512:$src1), (AltFrom.VT From.RC:$src2), (iPTR imm)),
450 VR512:$src1, From.RC:$src2,
489 (ins VR128X:$src1, VR128X:$src2, u8imm:$src3)
    [all...]
X86InstrCompiler.td 554 MRMDestMem, (outs), (ins i8mem:$dst, GR8:$src2),
556 "{$src2, $dst|$dst, $src2}"),
560 MRMDestMem, (outs), (ins i16mem:$dst, GR16:$src2),
562 "{$src2, $dst|$dst, $src2}"),
566 MRMDestMem, (outs), (ins i32mem:$dst, GR32:$src2),
568 "{$src2, $dst|$dst, $src2}"),
572 MRMDestMem, (outs), (ins i64mem:$dst, GR64:$src2),
    [all...]
X86InstrMMX.td 99 (ins VR64:$src1, VR64:$src2),
100 !strconcat(OpcodeStr, "\t{$src2, $dst|$dst, $src2}"),
101 [(set VR64:$dst, (IntId VR64:$src1, VR64:$src2))], itins.rr>,
106 (ins VR64:$src1, i64mem:$src2),
107 !strconcat(OpcodeStr, "\t{$src2, $dst|$dst, $src2}"),
109 (bitconvert (load_mmx addr:$src2))))],
117 (ins VR64:$src1, VR64:$src2),
118 !strconcat(OpcodeStr, "\t{$src2, $dst|$dst, $src2}")
    [all...]
  /external/llvm/test/CodeGen/SystemZ/
int-cmp-38.ll 16 %src2 = load i32 , i32 *@g
17 %cond = icmp slt i32 %src1, %src2
34 %src2 = load i32 , i32 *@g
35 %cond = icmp ult i32 %src1, %src2
52 %src2 = load i32 , i32 *@g
53 %cond = icmp eq i32 %src1, %src2
70 %src2 = load i32 , i32 *@g
71 %cond = icmp ne i32 %src1, %src2
89 %src2 = load i32 , i32 *@h, align 2
90 %cond = icmp slt i32 %src1, %src2
    [all...]
memcmp-02.ll 5 declare i64 @memcmp(i8 *%src1, i8 *%src2, i64 %size)
8 define i64 @f1(i8 *%src1, i8 *%src2) {
12 %res = call i64 @memcmp(i8 *%src1, i8 *%src2, i64 0)
17 define i64 @f2(i8 *%src1, i8 *%src2) {
25 %res = call i64 @memcmp(i8 *%src1, i8 *%src2, i64 2)
30 define void @f3(i8 *%src1, i8 *%src2, i64 *%dest) {
35 %res = call i64 @memcmp(i8 *%src1, i8 *%src2, i64 3)
48 define void @f4(i8 *%src1, i8 *%src2, i64 *%dest) {
54 %res = call i64 @memcmp(i8 *%src1, i8 *%src2, i64 4)
67 define void @f5(i8 *%src1, i8 *%src2, i64 *%dest)
    [all...]
strcmp-01.ll 5 declare signext i32 @strcmp(i8 *%src1, i8 *%src2)
8 define i32 @f1(i8 *%src1, i8 *%src2) {
19 %res = call i32 @strcmp(i8 *%src1, i8 *%src2)
24 define void @f2(i8 *%src1, i8 *%src2, i32 *%dest) {
33 %res = call i32 @strcmp(i8 *%src1, i8 *%src2)
47 define i32 @f3(i8 *%src1, i8 *%src2, i32 *%dest) {
60 %res = call i32 @strcmp(i8 *%src1, i8 *%src2)
strcmp-02.ll 5 declare i64 @strcmp(i8 *%src1, i8 *%src2)
8 define i64 @f1(i8 *%src1, i8 *%src2) {
20 %res = call i64 @strcmp(i8 *%src1, i8 *%src2)
25 define void @f2(i8 *%src1, i8 *%src2, i64 *%dest) {
34 %res = call i64 @strcmp(i8 *%src1, i8 *%src2)
48 define i64 @f3(i8 *%src1, i8 *%src2, i64 *%dest) {
62 %res = call i64 @strcmp(i8 *%src1, i8 *%src2)
memcmp-01.ll 5 declare signext i32 @memcmp(i8 *%src1, i8 *%src2, i64 %size)
8 define i32 @f1(i8 *%src1, i8 *%src2) {
12 %res = call i32 @memcmp(i8 *%src1, i8 *%src2, i64 0)
17 define i32 @f2(i8 *%src1, i8 *%src2) {
24 %res = call i32 @memcmp(i8 *%src1, i8 *%src2, i64 2)
29 define void @f3(i8 *%src1, i8 *%src2, i32 *%dest) {
34 %res = call i32 @memcmp(i8 *%src1, i8 *%src2, i64 3)
47 define void @f4(i8 *%src1, i8 *%src2, i32 *%dest) {
53 %res = call i32 @memcmp(i8 *%src1, i8 *%src2, i64 4)
66 define void @f5(i8 *%src1, i8 *%src2, i32 *%dest)
    [all...]
int-cmp-36.ll 17 %src2 = sext i16 %val to i32
18 %cond = icmp slt i32 %src1, %src2
35 %src2 = sext i16 %val to i32
36 %cond = icmp ult i32 %src1, %src2
54 %src2 = sext i16 %val to i32
55 %cond = icmp eq i32 %src1, %src2
73 %src2 = sext i16 %val to i32
74 %cond = icmp ne i32 %src1, %src2
93 %src2 = sext i16 %val to i32
94 %cond = icmp slt i32 %src1, %src2
    [all...]
int-cmp-37.ll 17 %src2 = zext i16 %val to i32
18 %cond = icmp ult i32 %src1, %src2
35 %src2 = zext i16 %val to i32
36 %cond = icmp slt i32 %src1, %src2
54 %src2 = zext i16 %val to i32
55 %cond = icmp eq i32 %src1, %src2
73 %src2 = zext i16 %val to i32
74 %cond = icmp ne i32 %src1, %src2
93 %src2 = zext i16 %val to i32
94 %cond = icmp ult i32 %src1, %src2
    [all...]
int-cmp-39.ll 17 %src2 = sext i16 %val to i64
18 %cond = icmp slt i64 %src1, %src2
35 %src2 = sext i16 %val to i64
36 %cond = icmp ult i64 %src1, %src2
54 %src2 = sext i16 %val to i64
55 %cond = icmp eq i64 %src1, %src2
73 %src2 = sext i16 %val to i64
74 %cond = icmp ne i64 %src1, %src2
93 %src2 = sext i16 %val to i64
94 %cond = icmp slt i64 %src1, %src2
    [all...]
int-cmp-40.ll 17 %src2 = zext i16 %val to i64
18 %cond = icmp ult i64 %src1, %src2
35 %src2 = zext i16 %val to i64
36 %cond = icmp slt i64 %src1, %src2
54 %src2 = zext i16 %val to i64
55 %cond = icmp eq i64 %src1, %src2
73 %src2 = zext i16 %val to i64
74 %cond = icmp ne i64 %src1, %src2
93 %src2 = zext i16 %val to i64
94 %cond = icmp ult i64 %src1, %src2
    [all...]
int-cmp-41.ll 17 %src2 = sext i32 %val to i64
18 %cond = icmp slt i64 %src1, %src2
35 %src2 = sext i32 %val to i64
36 %cond = icmp ult i64 %src1, %src2
54 %src2 = sext i32 %val to i64
55 %cond = icmp eq i64 %src1, %src2
73 %src2 = sext i32 %val to i64
74 %cond = icmp ne i64 %src1, %src2
93 %src2 = sext i32 %val to i64
94 %cond = icmp slt i64 %src1, %src2
    [all...]
int-cmp-42.ll 17 %src2 = zext i32 %val to i64
18 %cond = icmp ult i64 %src1, %src2
35 %src2 = zext i32 %val to i64
36 %cond = icmp slt i64 %src1, %src2
54 %src2 = zext i32 %val to i64
55 %cond = icmp eq i64 %src1, %src2
73 %src2 = zext i32 %val to i64
74 %cond = icmp ne i64 %src1, %src2
93 %src2 = zext i32 %val to i64
94 %cond = icmp ult i64 %src1, %src2
    [all...]
int-cmp-43.ll 16 %src2 = load i64 , i64 *@g
17 %cond = icmp slt i64 %src1, %src2
34 %src2 = load i64 , i64 *@g
35 %cond = icmp ult i64 %src1, %src2
52 %src2 = load i64 , i64 *@g
53 %cond = icmp eq i64 %src1, %src2
70 %src2 = load i64 , i64 *@g
71 %cond = icmp ne i64 %src1, %src2
89 %src2 = load i64 , i64 *@h, align 4
90 %cond = icmp slt i64 %src1, %src2
    [all...]
  /external/bison/lib/
bitset_stats.c 460 bitset_stats_and (bitset dst, bitset src1, bitset src2)
462 BITSET_CHECK3_ (dst, src1, src2);
463 BITSET_AND_ (dst->s.bset, src1->s.bset, src2->s.bset);
468 bitset_stats_and_cmp (bitset dst, bitset src1, bitset src2)
470 BITSET_CHECK3_ (dst, src1, src2);
471 return BITSET_AND_CMP_ (dst->s.bset, src1->s.bset, src2->s.bset);
476 bitset_stats_andn (bitset dst, bitset src1, bitset src2)
478 BITSET_CHECK3_ (dst, src1, src2);
479 BITSET_ANDN_ (dst->s.bset, src1->s.bset, src2->s.bset);
484 bitset_stats_andn_cmp (bitset dst, bitset src1, bitset src2)
    [all...]
bitset.c 409 bitset_op4_cmp (bitset dst, bitset src1, bitset src2, bitset src3,
428 bitset_or (tmp, src1, src2);
433 bitset_and (tmp, src1, src2);
438 bitset_andn (tmp, src1, src2);
448 /* DST = (SRC1 & SRC2) | SRC3. */
450 bitset_and_or_ (bitset dst, bitset src1, bitset src2, bitset src3)
452 bitset_and_or_cmp_ (dst, src1, src2, src3);
456 /* DST = (SRC1 & SRC2) | SRC3. Return non-zero if
457 DST != (SRC1 & SRC2) | SRC3. */
459 bitset_and_or_cmp_ (bitset dst, bitset src1, bitset src2, bitset src3
    [all...]
  /art/test/077-method-override/src2/
Base.java 26 /* src2: removed */
41 /* src2: public */
56 /* src2: public */
71 /* src2: non-static */
  /external/llvm/test/CodeGen/X86/
vsplit-and.ll 3 define void @t0(<2 x i64>* %dst, <2 x i64> %src1, <2 x i64> %src2) nounwind readonly {
8 %cmp2 = icmp ne <2 x i64> %src2, zeroinitializer
15 define void @t2(<3 x i64>* %dst, <3 x i64> %src1, <3 x i64> %src2) nounwind readonly {
20 %cmp2 = icmp ne <3 x i64> %src2, zeroinitializer
  /frameworks/av/media/libeffects/lvm/lib/Common/src/
MixSoft_2St_D32C31_SAT.c 32 const LVM_INT32 *src2,
46 src2, dst, n);
57 src2, dst, n);
61 Core_MixHard_2St_D32C31_SAT( pInstance, src1, src2, dst, n);
  /external/opencv/cv/src/
_cvmatrix.h 63 #define icvAddMatrix_32f( src1, src2, dst, w, h ) \
64 icvAddVector_32f( (src1), (src2), (dst), (w)*(h))
66 #define icvSubMatrix_32f( src1, src2, dst, w, h ) \
67 icvSubVector_32f( (src1), (src2), (dst), (w)*(h))
91 CV_INLINE double icvDotProduct_32f( const float* src1, const float* src2, int len )
94 for( int i = 0; i < len; i++ ) s += src1[i]*src2[i];
102 CV_INLINE double icvDotProduct_64f( const double* src1, const double* src2, int len )
105 for( int i = 0; i < len; i++ ) s += src1[i]*src2[i];
113 CV_INLINE void icvMulVectors_32f( const float* src1, const float* src2,
118 dst[i] = src1[i] * src2[i]
    [all...]
  /external/libavc/common/arm/
ih264_inter_pred_luma_bilinear_a9q.s 137 vld1.8 {q2}, [r1], r4 @// Load row0 ;src2
140 vld1.8 {q3}, [r1], r4 @// Load row1 ;src2
146 vld1.8 {q6}, [r1], r4 @// Load row2 ;src2
148 vld1.8 {q7}, [r1], r4 @// Load row3 ;src2
163 vld1.8 {q2}, [r1], r4 @// Load row4 ;src2
165 vld1.8 {q3}, [r1], r4 @// Load row5 ;src2
175 vld1.8 {q6}, [r1], r4 @// Load row6 ;src2
177 vld1.8 {q7}, [r1], r4 @// Load row7 ;src2
189 vld1.8 {q2}, [r1], r4 @// Load row8 ;src2
201 vld1.8 {q3}, [r1], r4 @// Load row9 ;src2
    [all...]
  /art/test/077-method-override/src/
Base.java 26 /* src2: removed */
41 /* src2: public */
56 /* src2: public */
71 /* src2: non-static */
79 /* src2: static */

Completed in 264 milliseconds

1 2 3 4 5 67 8 91011>>