/external/compiler-rt/lib/i386/ |
floatdisf.S | 21 movd 4(%esp), %xmm0 22 movd 8(%esp), %xmm1
|
floatdixf.S | 21 movd 4(%esp), %xmm0 22 movd 8(%esp), %xmm1
|
ashldi3.S | 21 movd 12(%esp), %xmm2 // Load count 23 movd 4(%esp), %xmm0 24 movd 8(%esp), %xmm1 30 movd %xmm0, %eax 32 movd %xmm0, %edx
|
lshrdi3.S | 21 movd 12(%esp), %xmm2 // Load count 23 movd 4(%esp), %xmm0 24 movd 8(%esp), %xmm1 30 movd %xmm0, %eax 32 movd %xmm0, %edx
|
ashrdi3.S | 14 movd 12(%esp), %xmm2 // Load count 17 movd 4(%esp), %xmm0 18 movd 8(%esp), %xmm1 40 1: movd %xmm0, %eax 42 movd %xmm0, %edx
|
floatundisf.S | 33 movd 8(%esp), %xmm1 34 movd 4(%esp), %xmm0 73 movd 8(%esp), %xmm1 74 movd 4(%esp), %xmm0 85 movd %edx, %xmm3 95 movd %xmm0, 4(%esp)
|
/external/llvm/test/MC/X86/ |
x86_64-encoding.s | 85 // CHECK: movd %r8, %mm1 87 movd %r8, %mm1 label 89 // CHECK: movd %r8d, %mm1 91 movd %r8d, %mm1 label 93 // CHECK: movd %rdx, %mm1 95 movd %rdx, %mm1 label 97 // CHECK: movd %edx, %mm1 99 movd %edx, %mm1 label 101 // CHECK: movd %mm1, %r8 103 movd %mm1, %r label 107 movd %mm1, %r8d label 111 movd %mm1, %rdx label 115 movd %mm1, %edx label [all...] |
/external/compiler-rt/lib/x86_64/ |
floatundidf.S | 34 movd %edi, %xmm0 // low 32 bits of a 38 movd %rdi, %xmm1
|
/external/openssl/crypto/bn/asm/ |
bn-586.pl | 42 &movd("mm0",&wparam(3)); # mm0 = w 47 &movd("mm3",&DWP(0,$r,"",0)); # mm3 = r[0] 49 &movd("mm2",&DWP(0,$a,"",0)); # mm2 = a[0] 51 &movd("mm4",&DWP(4,$a,"",0)); # mm4 = a[1] 53 &movd("mm6",&DWP(8,$a,"",0)); # mm6 = a[2] 55 &movd("mm7",&DWP(12,$a,"",0)); # mm7 = a[3] 58 &movd("mm3",&DWP(4,$r,"",0)); # mm3 = r[1] 60 &movd("mm5",&DWP(8,$r,"",0)); # mm5 = r[2] 62 &movd("mm4",&DWP(12,$r,"",0)); # mm4 = r[3] 64 &movd(&DWP(0,$r,"",0),"mm1") [all...] |
x86-mont.pl | 122 &movd ($mask,"eax"); # mask 32 lower bits 131 &movd ($mul0,&DWP(0,$bp)); # bp[0] 132 &movd ($mul1,&DWP(0,$ap)); # ap[0] 133 &movd ($car1,&DWP(0,$np)); # np[0] 137 &movq ($acc0,$mul1); # I wish movd worked for 145 &movd ($acc1,&DWP(4,$np)); # np[1] 146 &movd ($acc0,&DWP(4,$ap)); # ap[1] 160 &movd ($acc1,&DWP(4,$np,$j,4)); # np[j+1] 162 &movd ($acc0,&DWP(4,$ap,$j,4)); # ap[j+1] 164 &movd (&DWP($frame-4,"esp",$j,4),$car1); # tp[j-1] [all...] |
x86-gf2m.S | 13 movd %eax,%mm2 14 movd %ebx,%mm3 43 movd (%esp,%esi,4),%mm0 47 movd (%esp,%edi,4),%mm2 53 movd (%esp,%esi,4),%mm1 59 movd (%esp,%edi,4),%mm2 65 movd (%esp,%esi,4),%mm1 71 movd (%esp,%edi,4),%mm2 77 movd (%esp,%esi,4),%mm1 83 movd (%esp,%edi,4),%mm [all...] |
x86-gf2m.pl | 64 &movd ($A,$a); 65 &movd ($B,$b); 94 &movd ($R,&DWP(0,"esp",@i[0],4)); 99 &movd (@T[1],&DWP(0,"esp",@i[1],4)); 108 &movd (@T[1],&DWP(0,"esp",@i[1],4)); 113 &movd (@T[0],&DWP(0,"esp",@i[0],4));
|
/external/libvpx/libvpx/vp8/common/x86/ |
sad_sse3.asm | 403 movd [rcx], xmm0 409 movd [rcx+4], xmm0 415 movd [rcx+8], xmm0 441 movd [rcx], xmm0 447 movd [rcx+4], xmm0 453 movd [rcx+8], xmm0 482 movd [rcx+8], mm7 507 movd [rcx+8], mm7 522 movd mm0, DWORD PTR [src_ptr] 523 movd mm1, DWORD PTR [ref_ptr [all...] |
iwalsh_mmx.asm | 27 movd mm7, rax 98 movd eax, mm1 99 movd ecx, mm0 108 movd eax, mm1 109 movd ecx, mm0 117 movd eax, mm6 118 movd ecx, mm5 127 movd eax, mm6 128 movd ecx, mm5
|
iwalsh_sse2.asm | 56 movd xmm0, eax 78 movd eax, xmm5 79 movd ecx, xmm4 88 movd eax, xmm5 89 movd ecx, xmm4 99 movd eax, xmm5 100 movd ecx, xmm4 109 movd eax, xmm5 110 movd ecx, xmm4
|
idctllm_mmx.asm | 186 movd mm4, [rsi] 190 movd [rdx], mm0 192 movd mm4, [rsi+rax] 196 movd [rdx+rdi], mm1 198 movd mm4, [rsi+2*rax] 202 movd [rdx+rdi*2], mm2 207 movd mm4, [rsi+2*rax] 211 movd [rdx+rdi*2], mm5 235 movd mm5, arg(0) ;input_dc 250 movd mm1, [rax [all...] |
dequantize_mmx.asm | 216 movd mm4, [rdx] 220 movd [rdx], mm0 222 movd mm4, [rdx+rdi] 226 movd [rdx+rdi], mm1 228 movd mm4, [rdx+2*rdi] 232 movd [rdx+rdi*2], mm2 236 movd mm4, [rdx+2*rdi] 240 movd [rdx+rdi*2], mm5
|
sad_mmx.asm | 273 movd mm0, DWORD PTR [rsi] 274 movd mm1, DWORD PTR [rdi] 276 movd mm2, DWORD PTR [rsi+rax] 277 movd mm3, DWORD PTR [rdi+rdx] 299 movd mm4, DWORD PTR [rsi] 300 movd mm5, DWORD PTR [rdi] 302 movd mm6, DWORD PTR [rsi+rax] 303 movd mm7, DWORD PTR [rdi+rdx]
|
sad_sse2.asm | 224 movd mm0, DWORD PTR [rsi] 225 movd mm1, DWORD PTR [rdi] 227 movd mm2, DWORD PTR [rsi+rax] 228 movd mm3, DWORD PTR [rdi+rdx] 237 movd mm4, DWORD PTR [rsi] 239 movd mm5, DWORD PTR [rdi] 240 movd mm6, DWORD PTR [rsi+rax] 242 movd mm7, DWORD PTR [rdi+rdx]
|
/external/openssl/crypto/modes/asm/ |
ghash-x86.S | 216 movd %mm0,%ebp 226 movd %mm0,%ebx 238 movd %mm0,%ebp 250 movd %mm0,%ebx 262 movd %mm0,%ebp 274 movd %mm0,%ebx 286 movd %mm0,%ebp 298 movd %mm0,%ebx 310 movd %mm0,%ebp 322 movd %mm0,%eb [all...] |
/external/libvpx/libvpx/vpx_ports/ |
x86_abi_support.asm | 43 movd %1,%2 45 movd %1,%2 47 movd %1,%2 49 movd %1,%2 51 movd %1,%2 53 movd %1,%2 55 movd %1,%2 57 movd %1,%2 59 movd %1,%2 61 movd %1,% [all...] |
/external/libvpx/libvpx/vpx_scale/win32/ |
scaleopt.c | 102 movd mm1, eax // mm1 = 01 02 02 03 xx xx xx xx 107 movd mm0, ebx // mm0 = 00 01 01 02 125 movd DWORD Ptr [edi-4], mm0 147 movd mm1, eax // mm1 = 01 02 02 02 xx xx xx xx 152 movd mm0, ebx // mm0 = 00 01 01 02 165 movd DWORD Ptr [edi+1], mm0 224 movd DWORD PTR [edi], mm0 // write output 00 xx xx xx 233 movd DWORD PTR [edi+5], mm2 // write ouput 05 xx xx xx 245 movd DWORD PTR [edi+1], mm0 // write output 01 02 03 04 257 movd DWORD PTR [edi-4], mm2 // writeoutput 06 07 08 0 [all...] |
/external/libvpx/libvpx/vp8/encoder/x86/ |
subtract_mmx.asm | 34 movd mm0, [rsi] 35 movd mm1, [rax] 42 movd mm0, [rsi+rdx] 43 movd mm1, [rax+rcx] 50 movd mm0, [rsi+rdx*2] 51 movd mm1, [rax+rcx*2] 62 movd mm0, [rsi+rdx] 63 movd mm1, [rax+rcx]
|
ssim_opt.asm | 119 movd [rdi], xmm15; 121 movd [rdi], xmm14; 123 movd [rdi], xmm13; 125 movd [rdi], xmm12; 127 movd [rdi], xmm11; 200 movd [rdi], xmm15; 202 movd [rdi], xmm14; 204 movd [rdi], xmm13; 206 movd [rdi], xmm12; 208 movd [rdi], xmm11 [all...] |
/external/zlib/src/contrib/inflate86/ |
inffast.S | 884 movd %ebp, hold_mm 894 movd lmask(%esp), lmask_mm 896 movd dmask(%esp), dmask_mm 918 movd bitslong_r, tmp_mm 919 movd (in_r), %mm7 927 movd lmask_mm, %eax 933 movd %ecx, used_mm 954 movd %eax, used_mm 955 movd hold_mm, %ecx 966 movd bitslong_r, tmp_m [all...] |