/dalvik/vm/compiler/codegen/x86/libenc/ |
encoder.h | 524 ENCODER_DECLARE_EXPORT char * sse_mov(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl); 528 ENCODER_DECLARE_EXPORT char * sse_add(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl); 531 ENCODER_DECLARE_EXPORT char * sse_sub(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl); 534 ENCODER_DECLARE_EXPORT char * sse_mul(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl); 537 ENCODER_DECLARE_EXPORT char * sse_div(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl); 540 ENCODER_DECLARE_EXPORT char * sse_xor(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1); 542 ENCODER_DECLARE_EXPORT char * sse_compare(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl); 549 ENCODER_DECLARE_EXPORT char * sse_cvt_fp2dq(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl); 550 ENCODER_DECLARE_EXPORT char * sse_cvt_dq2fp(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl); 552 ENCODER_DECLARE_EXPORT char * sse_d2s(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1); [all...] |
encoder.inl | 386 ENCODER_DECLARE_EXPORT char * sse_mov(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl) { 389 add_xmm(args, xmm1, dbl); 401 ENCODER_DECLARE_EXPORT char * sse_add(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl) { 404 add_xmm(args, xmm1, dbl); 415 ENCODER_DECLARE_EXPORT char * sse_sub(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl) { 418 add_xmm(args, xmm1, dbl); 429 ENCODER_DECLARE_EXPORT char * sse_mul(char * stream, const XMM_Opnd& xmm0, const XMM_Opnd& xmm1, bool dbl) { 432 add_xmm(args, xmm1, dbl); 443 ENCODER_DECLARE_EXPORT char * sse_div(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl) { 446 add_xmm(args, xmm1, dbl) [all...] |
/external/llvm/lib/Target/X86/ |
X86CompilationCallback_Win64.asm | 36 movaps [rsp+16+32], xmm1 50 movaps xmm1, [rsp+16+32]
|
/external/llvm/test/MC/X86/ |
x86_64-encoding.s | 118 // CHECK: pshufb CPI1_0(%rip), %xmm1 121 pshufb CPI1_0(%rip), %xmm1
|
x86-64.s | 509 cvtsd2si %xmm1, %rax 510 // CHECK: cvtsd2si %xmm1, %rax 512 cvtsd2si %xmm1, %eax 513 // CHECK: cvtsd2si %xmm1, %eax 521 cvttpd2dq %xmm1, %xmm0 // CHECK: cvttpd2dq %xmm1, %xmm0 524 cvttps2dq %xmm1, %xmm0 // CHECK: cvttps2dq %xmm1, %xmm0 991 movdqu %xmm0, %xmm1 // CHECK: movdqu %xmm0, %xmm1 # encoding: [0xf3,0x0f,0x6f,0xc8 [all...] |
/external/openssl/crypto/modes/asm/ |
ghash-x86_64.S | 680 movdqa %xmm0,%xmm1 689 pxor %xmm1,%xmm3 694 pxor %xmm3,%xmm1 707 pxor %xmm4,%xmm1 715 pxor %xmm1,%xmm4 730 movdqa %xmm0,%xmm1 739 pxor %xmm1,%xmm3 744 pxor %xmm3,%xmm1 757 pxor %xmm4,%xmm1 765 pxor %xmm1,%xmm [all...] |
/external/libvpx/libvpx/vp8/common/x86/ |
sad_sse2.asm | 45 movq xmm1, QWORD PTR [rdi] 53 punpcklbw xmm1, xmm3 55 psadbw xmm0, xmm1 357 movdqu xmm1, XMMWORD PTR [rsi + 16] 371 movdqa XMMWORD PTR [rdi + 16], xmm1 393 movdqu xmm1, XMMWORD PTR [rsi + 16] 397 movdqa XMMWORD PTR [rdi + 16], xmm1
|
/external/libvpx/libvpx/vp8/encoder/x86/ |
ssim_opt.asm | 17 movdqa xmm1, xmm3 18 pmaddwd xmm1, xmm1 19 paddd xmm13, xmm1 ; sum_sq_s 41 movdqa xmm1, %1 43 punpckhwd xmm1,xmm0 44 paddd %1, xmm1
|
/external/openssl/crypto/rc4/asm/ |
rc4-x86_64.S | 182 psllq $8,%xmm1 189 pxor %xmm1,%xmm2 196 pxor %xmm1,%xmm1 203 pinsrw $0,(%rdi,%rbx,4),%xmm1 219 pinsrw $1,(%rdi,%rbx,4),%xmm1 235 pinsrw $2,(%rdi,%rbx,4),%xmm1 251 pinsrw $3,(%rdi,%rbx,4),%xmm1 267 pinsrw $4,(%rdi,%rbx,4),%xmm1 283 pinsrw $5,(%rdi,%rbx,4),%xmm1 [all...] |
rc4-md5-x86_64.pl | 250 psllq \$8,%xmm1 252 pxor %xmm1,%xmm2 290 psllq \$8,%xmm1 292 pxor %xmm1,%xmm3 329 psllq \$8,%xmm1 331 pxor %xmm1,%xmm4 370 psllq \$8,%xmm1 372 pxor %xmm1,%xmm5
|
/external/eigen/Eigen/src/Core/arch/SSE/ |
MathFunctions.h | 202 Packet4f xmm1, xmm2 = _mm_setzero_ps(), xmm3, sign_bit, y; local 241 xmm1 = pmul(y, p4f_minus_cephes_DP1); 244 x = padd(x, xmm1); 301 Packet4f xmm1, xmm2 = _mm_setzero_ps(), xmm3, y; local 330 xmm1 = pmul(y, p4f_minus_cephes_DP1); 333 x = padd(x, xmm1);
|
/external/libffi/src/x86/ |
unix64.S | 170 movd %xmm1, %r11 195 movdqa 64(%r10), %xmm1 299 %rax/%xmm0, %xmm0/%xmm1. We collapse two by always loading 300 both rdx and xmm1 with the second word. For the remaining, 305 movq -16(%rsp), %xmm1 319 movdqa %xmm1, 64(%rsp)
|
darwin64.S | 166 movd %xmm1, %r11 191 movdqa 64(%r10), %xmm1 288 %rax/%xmm0, %xmm0/%xmm1. We collapse two by always loading 289 both rdx and xmm1 with the second word. For the remaining, 294 movq -16(%rsp), %xmm1 308 movdqa %xmm1, 64(%rsp)
|
/external/openssl/crypto/aes/asm/ |
aesni-x86_64.pl | 196 $rndkey0="%xmm0"; $rndkey1="%xmm1"; [all...] |
aesni-sha1-x86_64.S | 51 movdqu 16(%r10),%xmm1 60 paddd %xmm9,%xmm1 64 movdqa %xmm1,16(%rsp) 65 psubd %xmm9,%xmm1 73 movdqa %xmm1,%xmm4 156 pxor %xmm1,%xmm5 359 pxor %xmm1,%xmm0 416 pxor %xmm5,%xmm1 421 pxor %xmm2,%xmm1 428 pxor %xmm10,%xmm1 [all...] |
/external/openssl/crypto/ |
x86_64cpuid.pl | 218 pxor %xmm1,%xmm1 251 pxor %xmm1,%xmm1
|
x86_64cpuid.S | 195 pxor %xmm1,%xmm1
|
/external/zlib/src/contrib/amd64/ |
amd64-match.S | 293 movdqu (%windowbestlen, %rdx), %xmm1 295 pcmpeqb %xmm2, %xmm1 307 pmovmskb %xmm1, %rax
|
/external/v8/test/cctest/ |
test-assembler-ia32.cc | 272 __ movdbl(xmm1, Operand(esp, 3 * kPointerSize)); 273 __ addsd(xmm0, xmm1); 274 __ mulsd(xmm0, xmm1); 275 __ subsd(xmm0, xmm1); 276 __ divsd(xmm0, xmm1);
|
/external/openssl/crypto/sha/asm/ |
sha1-x86_64.S | 1317 movdqu 16(%r9),%xmm1 1326 paddd %xmm9,%xmm1 1330 movdqa %xmm1,16(%rsp) 1331 psubd %xmm9,%xmm1 1337 movdqa %xmm1,%xmm4 1411 pxor %xmm1,%xmm5 1588 pxor %xmm1,%xmm0 1635 pxor %xmm5,%xmm1 1640 pxor %xmm2,%xmm1 1647 pxor %xmm10,%xmm1 [all...] |
/external/openssl/crypto/bn/asm/ |
x86_64-gf2m.pl | 41 ($R,$Tx)=("%xmm0","%xmm1"); 174 movq $b1,%xmm1 185 movdqa %xmm1,%xmm5 186 pclmulqdq \$0,%xmm1,%xmm0 # a1·b1
|
modexp512-x86_64.pl | 1134 movdqu (+16*1)(%rsi), %xmm1 1142 movdqa %xmm1, (+$tmp16_offset+16*3)(%rsp) 1148 movdqu (+16*1)(%rdx), %xmm1 1214 movdqa %xmm1, (+$exp_offset+16*1)(%rsp) 1299 movdqu (+16*1)(%rdx), %xmm1 1307 movdqa %xmm1, (+$tmp16_offset+16*1)(%rsp) [all...] |
/external/v8/src/x64/ |
code-stubs-x64.cc | 559 // Load the operands from rdx and rax into xmm0 and xmm1, as doubles. 1526 __ movsd(Operand(rsp, 0), xmm1); local 1628 __ movsd(FieldOperand(rax, HeapNumber::kValueOffset), xmm1); local 1644 __ movsd(Operand(rsp, 0), xmm1); local 1672 __ movsd(FieldOperand(rax, HeapNumber::kValueOffset), xmm1); local [all...] |
/external/v8/src/ia32/ |
codegen-ia32.cc | 217 __ movdqa(xmm1, Operand(src, 0x10)); 221 __ movdqa(Operand(dst, 0x10), xmm1); local 263 __ movdqu(xmm1, Operand(src, 0x10)); 267 __ movdqa(Operand(dst, 0x10), xmm1); local 426 XMMRegister the_hole_nan = xmm1;
|
/external/libyuv/files/source/ |
convert_from.cc | 256 movdqa xmm1, xmm0 local 258 punpckhbw xmm1, xmm2 local 260 movdqa [edi + 16], xmm1 294 movdqa xmm1, xmm2 local 296 punpcklbw xmm1, xmm0 // UYVY local 298 movdqa [edi], xmm1 325 "movdqa %%xmm0,%%xmm1 \n" 327 "punpckhbw %%xmm2,%%xmm1 \n" 329 "movdqa %%xmm1,0x10(%3) \n" 341 , "xmm0", "xmm1", "xmm2", "xmm3 [all...] |