/dalvik/vm/mterp/x86/ |
OP_AND_LONG.S | 2 %include "x86/binopWide.S" {"instr1":"andl (rFP,%ecx,4),rIBASE", "instr2":"andl 4(rFP,%ecx,4),%eax"}
|
OP_AND_LONG_2ADDR.S | 2 %include "x86/binopWide2addr.S" {"instr1":"andl %eax,(rFP,rINST,4)","instr2":"andl %ecx,4(rFP,rINST,4)"}
|
OP_IGET_BYTE.S | 3 %include "x86/OP_IGET.S" { "load":"movsbl", "sqnum":"2" }
|
OP_IGET_CHAR.S | 3 %include "x86/OP_IGET.S" { "load":"movzwl", "sqnum":"3" }
|
OP_IGET_SHORT.S | 3 %include "x86/OP_IGET.S" { "load":"movswl", "sqnum":"4" }
|
OP_LONG_TO_INT.S | 3 %include "x86/OP_MOVE.S"
|
OP_OR_LONG.S | 2 %include "x86/binopWide.S" {"instr1":"orl (rFP,%ecx,4),rIBASE", "instr2":"orl 4(rFP,%ecx,4),%eax"}
|
OP_OR_LONG_2ADDR.S | 2 %include "x86/binopWide2addr.S" {"instr1":"orl %eax,(rFP,rINST,4)","instr2":"orl %ecx,4(rFP,rINST,4)"}
|
OP_SUB_LONG.S | 2 %include "x86/binopWide.S" {"instr1":"subl (rFP,%ecx,4),rIBASE", "instr2":"sbbl 4(rFP,%ecx,4),%eax"}
|
OP_SUB_LONG_2ADDR.S | 2 %include "x86/binopWide2addr.S" {"instr1":"subl %eax,(rFP,rINST,4)","instr2":"sbbl %ecx,4(rFP,rINST,4)"}
|
OP_XOR_LONG.S | 2 %include "x86/binopWide.S" {"instr1":"xorl (rFP,%ecx,4),rIBASE", "instr2":"xorl 4(rFP,%ecx,4),%eax"}
|
OP_XOR_LONG_2ADDR.S | 2 %include "x86/binopWide2addr.S" {"instr1":"xorl %eax,(rFP,rINST,4)","instr2":"xorl %ecx,4(rFP,rINST,4)"}
|
/external/clang/test/CodeGen/ |
rtm-builtins.c | 9 // CHECK: i32 @llvm.x86.xbegin() 15 // CHECK: void @llvm.x86.xend() 21 // CHECK: void @llvm.x86.xabort(i8 2) 26 // CHECK: i32 @llvm.x86.xtest()
|
xop-builtins.c | 9 // CHECK: @llvm.x86.xop.vpmacssww 14 // CHECK: @llvm.x86.xop.vpmacsww 19 // CHECK: @llvm.x86.xop.vpmacsswd 24 // CHECK: @llvm.x86.xop.vpmacswd 29 // CHECK: @llvm.x86.xop.vpmacssdd 34 // CHECK: @llvm.x86.xop.vpmacsdd 39 // CHECK: @llvm.x86.xop.vpmacssdql 44 // CHECK: @llvm.x86.xop.vpmacsdql 49 // CHECK: @llvm.x86.xop.vpmacssdqh 54 // CHECK: @llvm.x86.xop.vpmacsdq [all...] |
/external/linux-tools-perf/util/include/asm/ |
alternative-asm.h | 4 /* Just disable it so we can build arch/x86/lib/memcpy_64.S for perf bench: */
|
cpufeature.h | 5 /* cpufeature.h ... dummy header file for including arch/x86/lib/memcpy_64.S */
|
dwarf2.h | 5 /* dwarf2.h ... dummy header file for including arch/x86/lib/memcpy_64.S */
|
/external/llvm/test/CodeGen/X86/ |
2004-02-14-InefficientStackPointer.ll | 1 ; RUN: llc < %s -march=x86 | grep -i ESP | not grep sub
|
2007-02-25-FastCCStack.ll | 1 ; RUN: llc < %s -march=x86 -mcpu=pentium3
|
barrier.ll | 1 ; RUN: llc < %s -march=x86 -mattr=-sse2 | grep lock
|
i128-immediate.ll | 1 ; RUN: llc < %s -march=x86-64 | grep movq | count 2
|
maskmovdqu.ll | 1 ; RUN: llc < %s -march=x86 -mattr=+sse2 | grep -i EDI 2 ; RUN: llc < %s -march=x86-64 -mattr=+sse2 | grep -i RDI 7 tail call void @llvm.x86.sse2.maskmov.dqu( <16 x i8> %a, <16 x i8> %b, i8* %c ) 11 declare void @llvm.x86.sse2.maskmov.dqu(<16 x i8>, <16 x i8>, i8*) nounwind
|
ret-i64-0.ll | 1 ; RUN: llc < %s -march=x86 | grep xor | count 2
|
utf8.ll | 1 ; RUN: llc < %s -march=x86 | FileCheck %s
|
/external/llvm/test/MC/ARM/AlignedBundling/ |
lit.local.cfg | 4 if not 'X86' in targets:
|