HomeSort by relevance Sort by last modified time
    Searched full:xmm15 (Results 1 - 25 of 90) sorted by null

1 2 3 4

  /external/chromium_org/third_party/openssl/openssl/crypto/aes/asm/
bsaes-x86_64.S 14 pxor %xmm8,%xmm15
49 movdqa %xmm15,%xmm10
50 psrlq $1,%xmm15
52 pxor %xmm0,%xmm15
54 pand %xmm7,%xmm15
57 pxor %xmm15,%xmm0
58 psllq $1,%xmm15
60 pxor %xmm10,%xmm15
78 movdqa %xmm15,%xmm10
79 psrlq $2,%xmm15
    [all...]
aesni-x86_64.S 893 movdqa .Lbswap_mask(%rip),%xmm15
1138 movups (%r9),%xmm15
1144 xorps %xmm0,%xmm15
1159 pcmpgtd %xmm15,%xmm14
1162 movdqa %xmm15,%xmm10
1163 paddq %xmm15,%xmm15
1165 pcmpgtd %xmm15,%xmm14
1166 pxor %xmm9,%xmm15
1169 movdqa %xmm15,%xmm1
    [all...]
aesni-sha1-x86_64.S 79 movups 32(%r15),%xmm15
132 movups 64(%r15),%xmm15
187 movups 96(%r15),%xmm15
243 movups 128(%r15),%xmm15
297 movups 160(%r15),%xmm15
329 movups 192(%r15),%xmm15
334 movups 224(%r15),%xmm15
370 movups 32(%r15),%xmm15
415 movups 64(%r15),%xmm15
477 movups 96(%r15),%xmm15
    [all...]
  /external/openssl/crypto/aes/asm/
bsaes-x86_64.S 14 pxor %xmm8,%xmm15
49 movdqa %xmm15,%xmm10
50 psrlq $1,%xmm15
52 pxor %xmm0,%xmm15
54 pand %xmm7,%xmm15
57 pxor %xmm15,%xmm0
58 psllq $1,%xmm15
60 pxor %xmm10,%xmm15
78 movdqa %xmm15,%xmm10
79 psrlq $2,%xmm15
    [all...]
aesni-x86_64.S 893 movdqa .Lbswap_mask(%rip),%xmm15
1138 movups (%r9),%xmm15
1144 xorps %xmm0,%xmm15
1159 pcmpgtd %xmm15,%xmm14
1162 movdqa %xmm15,%xmm10
1163 paddq %xmm15,%xmm15
1165 pcmpgtd %xmm15,%xmm14
1166 pxor %xmm9,%xmm15
1169 movdqa %xmm15,%xmm1
    [all...]
aesni-sha1-x86_64.S 79 movups 32(%r15),%xmm15
132 movups 64(%r15),%xmm15
187 movups 96(%r15),%xmm15
243 movups 128(%r15),%xmm15
297 movups 160(%r15),%xmm15
329 movups 192(%r15),%xmm15
334 movups 224(%r15),%xmm15
370 movups 32(%r15),%xmm15
415 movups 64(%r15),%xmm15
477 movups 96(%r15),%xmm15
    [all...]
  /external/llvm/test/CodeGen/X86/
coldcc64.ll 20 ; CHECK: movaps %xmm15
22 call void asm sideeffect "", "~{xmm15},~{xmm0},~{rbp},~{r15},~{r14},~{r13},~{r12},~{r11},~{r10},~{r9},~{r8},~{rdi},~{rsi},~{rdx},~{rcx},~{rbx}"()
preserve_allcc64.ll 21 ;SSE: movaps %xmm15
68 call void asm sideeffect "", "~{rax},~{rbx},~{rcx},~{rdx},~{rsi},~{rdi},~{r8},~{r9},~{r10},~{r11},~{r12},~{r13},~{r14},~{r15},~{rbp},~{xmm0},~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15}"()
100 %a23 = call <2 x double> asm sideeffect "", "={xmm15}"() nounwind
102 call void asm sideeffect "", "{rax},{rcx},{rdx},{r8},{r9},{r10},{r11},{xmm2},{xmm3},{xmm4},{xmm5},{xmm6},{xmm7},{xmm8},{xmm9},{xmm10},{xmm11},{xmm12},{xmm13},{xmm14},{xmm15}"(i64 %a0, i64 %a1, i64 %a2, i64 %a3, i64 %a4, i64 %a5, i64 %a6, <2 x double> %a10, <2 x double> %a11, <2 x double> %a12, <2 x double> %a13, <2 x double> %a14, <2 x double> %a15, <2 x double> %a16, <2 x double> %a17, <2 x double> %a18, <2 x double> %a19, <2 x double> %a20, <2 x double> %a21, <2 x double> %a22, <2 x double> %a23)
preserve_mostcc64.ll 37 call void asm sideeffect "", "~{rax},~{rbx},~{rcx},~{rdx},~{rsi},~{rdi},~{r8},~{r9},~{r10},~{r11},~{r12},~{r13},~{r14},~{r15},~{rbp},~{xmm0},~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15}"()
60 ;SSE: movaps %xmm15
82 %a23 = call <2 x double> asm sideeffect "", "={xmm15}"() nounwind
84 call void asm sideeffect "", "{rax},{rcx},{rdx},{r8},{r9},{r10},{r11},{xmm2},{xmm3},{xmm4},{xmm5},{xmm6},{xmm7},{xmm8},{xmm9},{xmm10},{xmm11},{xmm12},{xmm13},{xmm14},{xmm15}"(i64 %a0, i64 %a1, i64 %a2, i64 %a3, i64 %a4, i64 %a5, i64 %a6, <2 x double> %a10, <2 x double> %a11, <2 x double> %a12, <2 x double> %a13, <2 x double> %a14, <2 x double> %a15, <2 x double> %a16, <2 x double> %a17, <2 x double> %a18, <2 x double> %a19, <2 x double> %a20, <2 x double> %a21, <2 x double> %a22, <2 x double> %a23)
  /external/chromium_org/third_party/libvpx/source/libvpx/vp8/encoder/x86/
ssim_opt.asm 15 paddusw xmm15, xmm3 ; sum_s
80 pxor xmm15,xmm15 ;sum_s
112 SUM_ACROSS_W xmm15
119 movd [rdi], xmm15;
170 pxor xmm15,xmm15 ;sum_s
193 SUM_ACROSS_W xmm15
200 movd [rdi], xmm15;
  /external/chromium_org/third_party/libvpx/source/libvpx/vp9/encoder/x86/
vp9_ssim_opt.asm 15 paddusw xmm15, xmm3 ; sum_s
80 pxor xmm15,xmm15 ;sum_s
112 SUM_ACROSS_W xmm15
119 movd [rdi], xmm15;
170 pxor xmm15,xmm15 ;sum_s
193 SUM_ACROSS_W xmm15
200 movd [rdi], xmm15;
  /external/libvpx/libvpx/vp8/encoder/x86/
ssim_opt.asm 15 paddusw xmm15, xmm3 ; sum_s
80 pxor xmm15,xmm15 ;sum_s
112 SUM_ACROSS_W xmm15
119 movd [rdi], xmm15;
170 pxor xmm15,xmm15 ;sum_s
193 SUM_ACROSS_W xmm15
200 movd [rdi], xmm15;
  /external/libvpx/libvpx/vp9/encoder/x86/
vp9_ssim_opt.asm 15 paddusw xmm15, xmm3 ; sum_s
80 pxor xmm15,xmm15 ;sum_s
112 SUM_ACROSS_W xmm15
119 movd [rdi], xmm15;
170 pxor xmm15,xmm15 ;sum_s
193 SUM_ACROSS_W xmm15
200 movd [rdi], xmm15;
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/encoder/x86/
ssim_opt.asm 15 paddusw xmm15, xmm3 ; sum_s
80 pxor xmm15,xmm15 ;sum_s
112 SUM_ACROSS_W xmm15
119 movd [rdi], xmm15;
170 pxor xmm15,xmm15 ;sum_s
193 SUM_ACROSS_W xmm15
200 movd [rdi], xmm15;
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/encoder/x86/
vp9_ssim_opt.asm 15 paddusw xmm15, xmm3 ; sum_s
80 pxor xmm15,xmm15 ;sum_s
112 SUM_ACROSS_W xmm15
119 movd [rdi], xmm15;
170 pxor xmm15,xmm15 ;sum_s
193 SUM_ACROSS_W xmm15
200 movd [rdi], xmm15;
  /external/llvm/test/TableGen/
cast.td 61 def XMM15: Register<"xmm15">;
66 XMM12, XMM13, XMM14, XMM15]>;
MultiPat.td 65 def XMM15: Register<"xmm15">;
70 XMM12, XMM13, XMM14, XMM15]>;
Slice.td 55 def XMM15: Register<"xmm15">;
60 XMM12, XMM13, XMM14, XMM15]>;
TargetInstrSpec.td 62 def XMM15: Register<"xmm15">;
67 XMM12, XMM13, XMM14, XMM15]>;
  /art/runtime/arch/x86_64/
jni_entrypoints_x86_64.S 45 movq %xmm15, 88(%rsp)
62 movq 88(%rsp), %xmm15
registers_x86_64.cc 38 if (rhs >= XMM0 && rhs <= XMM15) {
registers_x86_64.h 67 XMM15 = 15,
quick_method_frame_info_x86_64.h 39 (1 << art::x86_64::XMM14) | (1 << art::x86_64::XMM15);
  /external/valgrind/main/memcheck/tests/amd64/
fxsave-amd64.c 64 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm15");
81 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm15");
115 asm __volatile__("movaps %xmm1, %xmm15");
  /art/compiler/jni/quick/x86_64/
calling_convention_x86_64.cc 136 callee_save_regs_.push_back(X86_64ManagedRegister::FromXmmRegister(XMM15));
145 return 1 << XMM12 | 1 << XMM13 | 1 << XMM14 | 1 << XMM15;

Completed in 1308 milliseconds

1 2 3 4