HomeSort by relevance Sort by last modified time
    Searched full:x86 (Results 226 - 250 of 5961) sorted by null

1 2 3 4 5 6 7 8 91011>>

  /dalvik/vm/mterp/x86/
OP_AND_LONG.S 2 %include "x86/binopWide.S" {"instr1":"andl (rFP,%ecx,4),rIBASE", "instr2":"andl 4(rFP,%ecx,4),%eax"}
OP_AND_LONG_2ADDR.S 2 %include "x86/binopWide2addr.S" {"instr1":"andl %eax,(rFP,rINST,4)","instr2":"andl %ecx,4(rFP,rINST,4)"}
OP_IGET_BYTE.S 3 %include "x86/OP_IGET.S" { "load":"movsbl", "sqnum":"2" }
OP_IGET_CHAR.S 3 %include "x86/OP_IGET.S" { "load":"movzwl", "sqnum":"3" }
OP_IGET_SHORT.S 3 %include "x86/OP_IGET.S" { "load":"movswl", "sqnum":"4" }
OP_LONG_TO_INT.S 3 %include "x86/OP_MOVE.S"
OP_OR_LONG.S 2 %include "x86/binopWide.S" {"instr1":"orl (rFP,%ecx,4),rIBASE", "instr2":"orl 4(rFP,%ecx,4),%eax"}
OP_OR_LONG_2ADDR.S 2 %include "x86/binopWide2addr.S" {"instr1":"orl %eax,(rFP,rINST,4)","instr2":"orl %ecx,4(rFP,rINST,4)"}
OP_SUB_LONG.S 2 %include "x86/binopWide.S" {"instr1":"subl (rFP,%ecx,4),rIBASE", "instr2":"sbbl 4(rFP,%ecx,4),%eax"}
OP_SUB_LONG_2ADDR.S 2 %include "x86/binopWide2addr.S" {"instr1":"subl %eax,(rFP,rINST,4)","instr2":"sbbl %ecx,4(rFP,rINST,4)"}
OP_XOR_LONG.S 2 %include "x86/binopWide.S" {"instr1":"xorl (rFP,%ecx,4),rIBASE", "instr2":"xorl 4(rFP,%ecx,4),%eax"}
OP_XOR_LONG_2ADDR.S 2 %include "x86/binopWide2addr.S" {"instr1":"xorl %eax,(rFP,rINST,4)","instr2":"xorl %ecx,4(rFP,rINST,4)"}
  /external/clang/test/CodeGen/
rtm-builtins.c 9 // CHECK: i32 @llvm.x86.xbegin()
15 // CHECK: void @llvm.x86.xend()
21 // CHECK: void @llvm.x86.xabort(i8 2)
26 // CHECK: i32 @llvm.x86.xtest()
xop-builtins.c 9 // CHECK: @llvm.x86.xop.vpmacssww
14 // CHECK: @llvm.x86.xop.vpmacsww
19 // CHECK: @llvm.x86.xop.vpmacsswd
24 // CHECK: @llvm.x86.xop.vpmacswd
29 // CHECK: @llvm.x86.xop.vpmacssdd
34 // CHECK: @llvm.x86.xop.vpmacsdd
39 // CHECK: @llvm.x86.xop.vpmacssdql
44 // CHECK: @llvm.x86.xop.vpmacsdql
49 // CHECK: @llvm.x86.xop.vpmacssdqh
54 // CHECK: @llvm.x86.xop.vpmacsdq
    [all...]
  /external/linux-tools-perf/util/include/asm/
alternative-asm.h 4 /* Just disable it so we can build arch/x86/lib/memcpy_64.S for perf bench: */
cpufeature.h 5 /* cpufeature.h ... dummy header file for including arch/x86/lib/memcpy_64.S */
dwarf2.h 5 /* dwarf2.h ... dummy header file for including arch/x86/lib/memcpy_64.S */
  /external/llvm/test/CodeGen/X86/
2004-02-14-InefficientStackPointer.ll 1 ; RUN: llc < %s -march=x86 | grep -i ESP | not grep sub
2007-02-25-FastCCStack.ll 1 ; RUN: llc < %s -march=x86 -mcpu=pentium3
barrier.ll 1 ; RUN: llc < %s -march=x86 -mattr=-sse2 | grep lock
i128-immediate.ll 1 ; RUN: llc < %s -march=x86-64 | grep movq | count 2
maskmovdqu.ll 1 ; RUN: llc < %s -march=x86 -mattr=+sse2 | grep -i EDI
2 ; RUN: llc < %s -march=x86-64 -mattr=+sse2 | grep -i RDI
7 tail call void @llvm.x86.sse2.maskmov.dqu( <16 x i8> %a, <16 x i8> %b, i8* %c )
11 declare void @llvm.x86.sse2.maskmov.dqu(<16 x i8>, <16 x i8>, i8*) nounwind
ret-i64-0.ll 1 ; RUN: llc < %s -march=x86 | grep xor | count 2
utf8.ll 1 ; RUN: llc < %s -march=x86 | FileCheck %s
  /external/llvm/test/MC/ARM/AlignedBundling/
lit.local.cfg 4 if not 'X86' in targets:

Completed in 880 milliseconds

1 2 3 4 5 6 7 8 91011>>