/external/boringssl/win-x86_64/crypto/bn/ |
x86_64-mont.asm | 610 movdqu xmm2,XMMWORD[r14*1+rsp] 614 pxor xmm2,xmm1 616 pand xmm2,xmm0 618 pxor xmm2,xmm1 620 movdqu XMMWORD[r14*1+rdi],xmm2
|
modexp512-x86_64.asm | 1503 movdqu xmm2,XMMWORD PTR[32+rsi] 1511 movdqa XMMWORD PTR[576+rsp],xmm2 1517 movdqu xmm2,XMMWORD PTR[32+rdx] 1598 movdqa XMMWORD PTR[96+rsp],xmm2 1710 movdqu xmm2,XMMWORD PTR[32+rdx] 1718 movdqa XMMWORD PTR[544+rsp],xmm2
|
/external/valgrind/none/tests/amd64/ |
redundantRexW.c | 111 "\tmovupd 32(%%r14), %%xmm2\n" \ 129 "\tmovupd %%xmm2, 32(%%r14)\n" \ 145 "xmm0", "xmm1", "xmm2", "xmm3", "xmm4", "xmm5", "xmm6", "xmm7", \ 462 /* movsd mem, reg f2 48 0f 10 11 rex.W movsd (%rcx),%xmm2 */ 476 after_test( "rex.W movsd (%rcx),%xmm2", regs, mem );
|
/external/llvm/lib/CodeGen/ |
MachineCopyPropagation.cpp | 209 // %xmm9<def> = copy %xmm2 211 // %xmm2<def> = copy %xmm0 213 // %xmm2<def> = copy %xmm9
|
/external/boringssl/src/crypto/ |
cpu-x86-asm.pl | 250 &pxor ("xmm2","xmm2");
|
/external/google-breakpad/src/common/ |
dwarf_cfi_to_module.cc | 58 "$xmm0", "$xmm1", "$xmm2", "$xmm3", "$xmm4", "$xmm5", "$xmm6", "$xmm7", 73 "$xmm0","$xmm1","$xmm2", "$xmm3", "$xmm4", "$xmm5", "$xmm6", "$xmm7",
|
/external/llvm/test/CodeGen/X86/ |
cmovcmov.ll | 100 ; CMOV-NEXT: movaps %xmm2, %xmm3 129 ; CMOV-NEXT: movaps %xmm2, %xmm3
|
anyregcc.ll | 399 ;SSE-NEXT: movaps %xmm2 434 call void asm sideeffect "", "~{rax},~{rbx},~{rcx},~{rdx},~{rsi},~{rdi},~{r8},~{r9},~{r10},~{r11},~{r12},~{r13},~{r14},~{r15},~{rbp},~{xmm0},~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15}"() 448 %a2 = call <2 x double> asm sideeffect "", "={xmm2}"() nounwind 463 call void asm sideeffect "", "{xmm0},{xmm1},{xmm2},{xmm3},{xmm4},{xmm5},{xmm6},{xmm7},{xmm8},{xmm9},{xmm10},{xmm11},{xmm12},{xmm13},{xmm14},{xmm15}"(<2 x double> %a0, <2 x double> %a1, <2 x double> %a2, <2 x double> %a3, <2 x double> %a4, <2 x double> %a5, <2 x double> %a6, <2 x double> %a7, <2 x double> %a8, <2 x double> %a9, <2 x double> %a10, <2 x double> %a11, <2 x double> %a12, <2 x double> %a13, <2 x double> %a14, <2 x double> %a15)
|
avx512bwvl-intrinsics.ll | [all...] |
sse3.ll | 44 ; X64-NEXT: movdqa {{.*#+}} xmm2 = [0,65535,65535,0,65535,65535,65535,65535] 45 ; X64-NEXT: pand %xmm2, %xmm0 47 ; X64-NEXT: pandn %xmm1, %xmm2 48 ; X64-NEXT: por %xmm2, %xmm0
|
vaargs.ll | 14 ; CHECK-NEXT: vmovaps %xmm2, 80(%rsp)
|
vector-shuffle-256-v8.ll | 272 ; AVX1-NEXT: vunpckhps {{.*#+}} xmm2 = xmm0[2],xmm1[2],xmm0[3],xmm1[3] 274 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0 292 ; AVX1-NEXT: vshufps {{.*#+}} xmm2 = xmm0[0,0],xmm1[0,0] 293 ; AVX1-NEXT: vshufps {{.*#+}} xmm2 = xmm2[0,2],xmm1[1,1] 314 ; AVX1-NEXT: vpermilps {{.*#+}} xmm2 = xmm0[0,1,1,3] 333 ; AVX1-NEXT: vmovshdup {{.*#+}} xmm2 = xmm0[1,1,3,3] 334 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0 775 ; AVX1-NEXT: vextractf128 $1, %ymm1, %xmm2 776 ; AVX1-NEXT: vshufps {{.*#+}} xmm1 = xmm1[1,3],xmm2[1,3 [all...] |
/external/boringssl/win-x86_64/crypto/sha/ |
sha256-x86_64.asm | 1816 movdqu xmm2,XMMWORD[32+rsi] 1829 paddd xmm6,xmm2 1989 movdqa xmm4,xmm2 2160 paddd xmm2,xmm7 2198 paddd xmm2,xmm4 2227 paddd xmm2,xmm7 2231 pshufd xmm7,xmm2,80 2268 paddd xmm2,xmm7 2272 paddd xmm6,xmm2 2280 movdqa xmm7,xmm2 [all...] |
/art/compiler/dex/quick/x86/ |
target_x86.cc | [all...] |
/external/boringssl/win-x86/crypto/sha/ |
sha256-586.asm | [all...] |
/art/compiler/optimizing/ |
code_generator_x86_64.h | 35 { XMM0, XMM1, XMM2, XMM3, XMM4, XMM5, XMM6, XMM7 };
|
/art/runtime/arch/x86/ |
context_x86.cc | 110 "movsd 16(%%ebx), %%xmm2\n\t"
|
quick_entrypoints_x86.S | 99 movsd %xmm2, 16(%esp) 139 movsd %xmm2, 16(%esp) 151 movsd 20(%esp), %xmm2 174 movsd 16(%esp), %xmm2 317 movsd 16(%esp), %xmm2 431 LOOP_OVER_SHORTY_LOADING_XMMS xmm2, esi, edi, al, .Lxmm_setup_finished 528 LOOP_OVER_SHORTY_LOADING_XMMS xmm2, esi, edi, al, .Lxmm_setup_finished2 [all...] |
/external/google-breakpad/src/third_party/libdisasm/ |
ia32_reg.c | 115 { REG_SIMD_SIZE, reg_simd, 0, "xmm2" },
|
/external/llvm/test/MC/Disassembler/X86/ |
intel-syntax.txt | 108 # CHECK: vshufpd xmm0, xmm1, xmm2, 1
|
/external/llvm/test/MC/X86/ |
x86_64-avx-encoding.s | 403 // CHECK: vcmpordps -4(%rbx,%rcx,8), %xmm6, %xmm2 405 vcmpordps -4(%rbx,%rcx,8), %xmm6, %xmm2 467 // CHECK: vcmpordpd -4(%rbx,%rcx,8), %xmm6, %xmm2 469 vcmpordpd -4(%rbx,%rcx,8), %xmm6, %xmm2 531 // CHECK: vcmpordss -4(%rbx,%rcx,8), %xmm6, %xmm2 533 vcmpordss -4(%rbx,%rcx,8), %xmm6, %xmm2 595 // CHECK: vcmpordsd -4(%rbx,%rcx,8), %xmm6, %xmm2 597 vcmpordsd -4(%rbx,%rcx,8), %xmm6, %xmm2 723 // CHECK: vcmpgtps -4(%rbx,%rcx,8), %xmm6, %xmm2 725 vcmpgtps -4(%rbx,%rcx,8), %xmm6, %xmm2 [all...] |
/art/runtime/arch/x86_64/ |
quick_entrypoints_x86_64.S | 176 movq %xmm2, 32(%rsp) 218 movq %xmm2, 32(%rsp) 238 movq 32(%rsp), %xmm2 477 LOOP_OVER_SHORTY_LOADING_XMMS xmm2, .Lxmm_setup_finished 571 LOOP_OVER_SHORTY_LOADING_XMMS xmm2, .Lxmm_setup_finished2 657 movq 16(%rsi), %xmm2 [all...] |
/external/boringssl/mac-x86_64/crypto/bn/ |
modexp512-x86_64.S | 1494 movdqu 32(%rsi),%xmm2 1502 movdqa %xmm2,576(%rsp) 1508 movdqu 32(%rdx),%xmm2 1589 movdqa %xmm2,96(%rsp) 1701 movdqu 32(%rdx),%xmm2 1709 movdqa %xmm2,544(%rsp)
|
/external/zlib/src/contrib/amd64/ |
amd64-match.S | 294 movdqu (%prev, %rdx), %xmm2 295 pcmpeqb %xmm2, %xmm1
|
/art/compiler/utils/x86/ |
assembler_x86_test.cc | 68 new x86::XmmRegister(x86::XMM2),
|