HomeSort by relevance Sort by last modified time
    Searched refs:xmm1 (Results 76 - 100 of 128) sorted by null

1 2 34 5 6

  /dalvik/vm/compiler/codegen/x86/libenc/
encoder.h 524 ENCODER_DECLARE_EXPORT char * sse_mov(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl);
528 ENCODER_DECLARE_EXPORT char * sse_add(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl);
531 ENCODER_DECLARE_EXPORT char * sse_sub(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl);
534 ENCODER_DECLARE_EXPORT char * sse_mul(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl);
537 ENCODER_DECLARE_EXPORT char * sse_div(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl);
540 ENCODER_DECLARE_EXPORT char * sse_xor(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1);
542 ENCODER_DECLARE_EXPORT char * sse_compare(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl);
549 ENCODER_DECLARE_EXPORT char * sse_cvt_fp2dq(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl);
550 ENCODER_DECLARE_EXPORT char * sse_cvt_dq2fp(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl);
552 ENCODER_DECLARE_EXPORT char * sse_d2s(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1);
    [all...]
encoder.inl 386 ENCODER_DECLARE_EXPORT char * sse_mov(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl) {
389 add_xmm(args, xmm1, dbl);
401 ENCODER_DECLARE_EXPORT char * sse_add(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl) {
404 add_xmm(args, xmm1, dbl);
415 ENCODER_DECLARE_EXPORT char * sse_sub(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl) {
418 add_xmm(args, xmm1, dbl);
429 ENCODER_DECLARE_EXPORT char * sse_mul(char * stream, const XMM_Opnd& xmm0, const XMM_Opnd& xmm1, bool dbl) {
432 add_xmm(args, xmm1, dbl);
443 ENCODER_DECLARE_EXPORT char * sse_div(char * stream, const XMM_Opnd & xmm0, const XMM_Opnd & xmm1, bool dbl) {
446 add_xmm(args, xmm1, dbl)
    [all...]
  /external/llvm/lib/Target/X86/
X86CompilationCallback_Win64.asm 36 movaps [rsp+16+32], xmm1
50 movaps xmm1, [rsp+16+32]
  /external/llvm/test/MC/X86/
x86_64-encoding.s 118 // CHECK: pshufb CPI1_0(%rip), %xmm1
121 pshufb CPI1_0(%rip), %xmm1
x86-64.s 509 cvtsd2si %xmm1, %rax
510 // CHECK: cvtsd2si %xmm1, %rax
512 cvtsd2si %xmm1, %eax
513 // CHECK: cvtsd2si %xmm1, %eax
521 cvttpd2dq %xmm1, %xmm0 // CHECK: cvttpd2dq %xmm1, %xmm0
524 cvttps2dq %xmm1, %xmm0 // CHECK: cvttps2dq %xmm1, %xmm0
991 movdqu %xmm0, %xmm1 // CHECK: movdqu %xmm0, %xmm1 # encoding: [0xf3,0x0f,0x6f,0xc8
    [all...]
  /external/openssl/crypto/modes/asm/
ghash-x86_64.S 680 movdqa %xmm0,%xmm1
689 pxor %xmm1,%xmm3
694 pxor %xmm3,%xmm1
707 pxor %xmm4,%xmm1
715 pxor %xmm1,%xmm4
730 movdqa %xmm0,%xmm1
739 pxor %xmm1,%xmm3
744 pxor %xmm3,%xmm1
757 pxor %xmm4,%xmm1
765 pxor %xmm1,%xmm
    [all...]
  /external/libvpx/libvpx/vp8/common/x86/
sad_sse2.asm 45 movq xmm1, QWORD PTR [rdi]
53 punpcklbw xmm1, xmm3
55 psadbw xmm0, xmm1
357 movdqu xmm1, XMMWORD PTR [rsi + 16]
371 movdqa XMMWORD PTR [rdi + 16], xmm1
393 movdqu xmm1, XMMWORD PTR [rsi + 16]
397 movdqa XMMWORD PTR [rdi + 16], xmm1
  /external/libvpx/libvpx/vp8/encoder/x86/
ssim_opt.asm 17 movdqa xmm1, xmm3
18 pmaddwd xmm1, xmm1
19 paddd xmm13, xmm1 ; sum_sq_s
41 movdqa xmm1, %1
43 punpckhwd xmm1,xmm0
44 paddd %1, xmm1
  /external/openssl/crypto/rc4/asm/
rc4-x86_64.S 182 psllq $8,%xmm1
189 pxor %xmm1,%xmm2
196 pxor %xmm1,%xmm1
203 pinsrw $0,(%rdi,%rbx,4),%xmm1
219 pinsrw $1,(%rdi,%rbx,4),%xmm1
235 pinsrw $2,(%rdi,%rbx,4),%xmm1
251 pinsrw $3,(%rdi,%rbx,4),%xmm1
267 pinsrw $4,(%rdi,%rbx,4),%xmm1
283 pinsrw $5,(%rdi,%rbx,4),%xmm1
    [all...]
rc4-md5-x86_64.pl 250 psllq \$8,%xmm1
252 pxor %xmm1,%xmm2
290 psllq \$8,%xmm1
292 pxor %xmm1,%xmm3
329 psllq \$8,%xmm1
331 pxor %xmm1,%xmm4
370 psllq \$8,%xmm1
372 pxor %xmm1,%xmm5
  /external/eigen/Eigen/src/Core/arch/SSE/
MathFunctions.h 202 Packet4f xmm1, xmm2 = _mm_setzero_ps(), xmm3, sign_bit, y; local
241 xmm1 = pmul(y, p4f_minus_cephes_DP1);
244 x = padd(x, xmm1);
301 Packet4f xmm1, xmm2 = _mm_setzero_ps(), xmm3, y; local
330 xmm1 = pmul(y, p4f_minus_cephes_DP1);
333 x = padd(x, xmm1);
  /external/libffi/src/x86/
unix64.S 170 movd %xmm1, %r11
195 movdqa 64(%r10), %xmm1
299 %rax/%xmm0, %xmm0/%xmm1. We collapse two by always loading
300 both rdx and xmm1 with the second word. For the remaining,
305 movq -16(%rsp), %xmm1
319 movdqa %xmm1, 64(%rsp)
darwin64.S 166 movd %xmm1, %r11
191 movdqa 64(%r10), %xmm1
288 %rax/%xmm0, %xmm0/%xmm1. We collapse two by always loading
289 both rdx and xmm1 with the second word. For the remaining,
294 movq -16(%rsp), %xmm1
308 movdqa %xmm1, 64(%rsp)
  /external/openssl/crypto/aes/asm/
aesni-x86_64.pl 196 $rndkey0="%xmm0"; $rndkey1="%xmm1";
    [all...]
aesni-sha1-x86_64.S 51 movdqu 16(%r10),%xmm1
60 paddd %xmm9,%xmm1
64 movdqa %xmm1,16(%rsp)
65 psubd %xmm9,%xmm1
73 movdqa %xmm1,%xmm4
156 pxor %xmm1,%xmm5
359 pxor %xmm1,%xmm0
416 pxor %xmm5,%xmm1
421 pxor %xmm2,%xmm1
428 pxor %xmm10,%xmm1
    [all...]
  /external/openssl/crypto/
x86_64cpuid.pl 218 pxor %xmm1,%xmm1
251 pxor %xmm1,%xmm1
x86_64cpuid.S 195 pxor %xmm1,%xmm1
  /external/zlib/src/contrib/amd64/
amd64-match.S 293 movdqu (%windowbestlen, %rdx), %xmm1
295 pcmpeqb %xmm2, %xmm1
307 pmovmskb %xmm1, %rax
  /external/v8/test/cctest/
test-assembler-ia32.cc 272 __ movdbl(xmm1, Operand(esp, 3 * kPointerSize));
273 __ addsd(xmm0, xmm1);
274 __ mulsd(xmm0, xmm1);
275 __ subsd(xmm0, xmm1);
276 __ divsd(xmm0, xmm1);
  /external/openssl/crypto/sha/asm/
sha1-x86_64.S 1317 movdqu 16(%r9),%xmm1
1326 paddd %xmm9,%xmm1
1330 movdqa %xmm1,16(%rsp)
1331 psubd %xmm9,%xmm1
1337 movdqa %xmm1,%xmm4
1411 pxor %xmm1,%xmm5
1588 pxor %xmm1,%xmm0
1635 pxor %xmm5,%xmm1
1640 pxor %xmm2,%xmm1
1647 pxor %xmm10,%xmm1
    [all...]
  /external/openssl/crypto/bn/asm/
x86_64-gf2m.pl 41 ($R,$Tx)=("%xmm0","%xmm1");
174 movq $b1,%xmm1
185 movdqa %xmm1,%xmm5
186 pclmulqdq \$0,%xmm1,%xmm0 # a1·b1
modexp512-x86_64.pl 1134 movdqu (+16*1)(%rsi), %xmm1
1142 movdqa %xmm1, (+$tmp16_offset+16*3)(%rsp)
1148 movdqu (+16*1)(%rdx), %xmm1
1214 movdqa %xmm1, (+$exp_offset+16*1)(%rsp)
1299 movdqu (+16*1)(%rdx), %xmm1
1307 movdqa %xmm1, (+$tmp16_offset+16*1)(%rsp)
    [all...]
  /external/v8/src/x64/
code-stubs-x64.cc 559 // Load the operands from rdx and rax into xmm0 and xmm1, as doubles.
1526 __ movsd(Operand(rsp, 0), xmm1); local
1628 __ movsd(FieldOperand(rax, HeapNumber::kValueOffset), xmm1); local
1644 __ movsd(Operand(rsp, 0), xmm1); local
1672 __ movsd(FieldOperand(rax, HeapNumber::kValueOffset), xmm1); local
    [all...]
  /external/v8/src/ia32/
codegen-ia32.cc 217 __ movdqa(xmm1, Operand(src, 0x10));
221 __ movdqa(Operand(dst, 0x10), xmm1); local
263 __ movdqu(xmm1, Operand(src, 0x10));
267 __ movdqa(Operand(dst, 0x10), xmm1); local
426 XMMRegister the_hole_nan = xmm1;
  /external/libyuv/files/source/
convert_from.cc 256 movdqa xmm1, xmm0 local
258 punpckhbw xmm1, xmm2 local
260 movdqa [edi + 16], xmm1
294 movdqa xmm1, xmm2 local
296 punpcklbw xmm1, xmm0 // UYVY local
298 movdqa [edi], xmm1
325 "movdqa %%xmm0,%%xmm1 \n"
327 "punpckhbw %%xmm2,%%xmm1 \n"
329 "movdqa %%xmm1,0x10(%3) \n"
341 , "xmm0", "xmm1", "xmm2", "xmm3
    [all...]

Completed in 1615 milliseconds

1 2 34 5 6