/dalvik/vm/mterp/x86-atom/ |
OP_INT_TO_FLOAT.S | 34 cvtsi2ss (rFP,%eax,4), %xmm0 # %xmm0<- vB 35 movss %xmm0, (rFP, rINST, 4) # vA<- %xmm0
|
OP_IPUT_WIDE_QUICK.S | 36 movq (rFP, rINST, 4), %xmm0 # %xmm0<- fp[A] 37 movq %xmm0, (%edx, %ecx) # object field<- %xmm0; fp[A]
|
OP_MOVE_WIDE_16.S | 34 movq (rFP, %edx, 4), %xmm0 # %xmm0<- vB 35 movq %xmm0, (rFP, %ecx, 4) # vA<- vB; %xmm0
|
OP_SHL_LONG.S | 35 movss (rFP, %eax, 4), %xmm0 # %xmm0<- vCC 36 pand %xmm2, %xmm0 # %xmm0<- masked shift bits 38 psllq %xmm0, %xmm1 # %xmm1<- shifted vBB
|
OP_SHL_LONG_2ADDR.S | 35 movss (rFP, %edx, 4), %xmm0 # %xmm0<- vB 38 pand %xmm2, %xmm0 # %xmm0<- masked shift bits 39 psllq %xmm0, %xmm1 # %xmm1<- shifted vA
|
OP_USHR_LONG_2ADDR.S | 36 movss (rFP, %edx, 4), %xmm0 # %xmm0<- vB 38 pand %xmm2, %xmm0 # %xmm0<- masked shift bits 39 psrlq %xmm0, %xmm1 # %xmm1<- shifted vA
|
OP_SHR_LONG.S | 35 movss (rFP, %eax, 4), %xmm0 # %xmm0<- vCC 37 pand %xmm2, %xmm0 # %xmm0<- masked for the shift bits 38 psrlq %xmm0, %xmm1 # %xmm1<- shifted vBB 46 psubq %xmm0, %xmm3 # %xmm3<- 64 - shift amount
|
OP_SHR_LONG_2ADDR.S | 35 movss (rFP, %edx, 4), %xmm0 # %xmm0<- vB 38 pand %xmm2, %xmm0 # %xmm0<- masked for the shift bits 39 psrlq %xmm0, %xmm1 # %xmm1<- shifted vBB 47 psubq %xmm0, %xmm3 # %xmm3<- 64 - shift amount
|
OP_AGET_WIDE.S | 41 movq offArrayObject_contents(%ecx, %edx, 8), %xmm0 # %xmm0<- vBB[vCC] 42 movq %xmm0, (rFP, rINST, 4) # vAA<- %xmm0; value
|
OP_APUT_WIDE.S | 41 movq (rFP, rINST, 4), %xmm0 # %xmm0<- vAA 42 movq %xmm0, offArrayObject_contents(%ecx, %edx, 8) # vBB[vCC]<- %xmm0; value
|
/external/libvpx/vp8/encoder/x86/ |
ssim_opt.asm | 30 punpckldq %1,xmm0 31 punpckhdq xmm2,xmm0 34 punpcklqdq %1,xmm0 35 punpckhqdq xmm2,xmm0 42 punpcklwd %1,xmm0 43 punpckhwd xmm1,xmm0 78 pxor xmm0, xmm0 93 punpckhbw xmm3, xmm0 ; high_s 94 punpckhbw xmm4, xmm0 ; high_ [all...] |
quantize_sse2.asm | 62 movdqa xmm0, [rdx] 69 movdqa xmm1, xmm0 73 psraw xmm0, 15 77 pxor xmm1, xmm0 81 psubw xmm1, xmm0 183 pxor xmm2, xmm0 186 psubw xmm2, xmm0 190 movdqa xmm0, [rcx] 195 pmullw xmm0, xmm2 200 movdqa [rsi], xmm0 ; store dqcoef [all...] |
/external/llvm/test/CodeGen/X86/ |
pmulld.ll | 9 ; WIN64-NEXT: movdqa (%rcx), %xmm0 10 ; WIN64-NEXT: pmulld (%rdx), %xmm0 20 ; WIN64-NEXT: movdqa (%rcx), %xmm0 21 ; WIN64-NEXT: pmulld (%rdx), %xmm0
|
vec_shuffle-37.ll | 7 ; CHECK: movaps ({{%rdi|%rcx}}), %xmm0 8 ; CHECK: movaps %xmm0, %xmm1 10 ; CHECK-NEXT: shufps $36, %xmm1, %xmm0 19 ; CHECK_O0: movsd (%eax), %xmm0 20 ; CHECK_O0: unpcklpd %xmm0, %xmm0 32 ; CHECK: movq %xmm0, ({{%rsi|%rdx}})
|
bc-extract.ll | 14 ; CHECK: pxor %xmm0, %xmm0
|
illegal-vector-args-return.ll | 2 ; RUN: llc < %s -march=x86 -mattr=+sse2 | grep {mulpd %xmm2, %xmm0} 4 ; RUN: llc < %s -march=x86 -mattr=+sse2 | grep {addps %xmm2, %xmm0}
|
rip-rel-address.ll | 12 ; PIC64: movsd _a(%rip), %xmm0 13 ; STATIC64: movsd a(%rip), %xmm0
|
/frameworks/compile/libbcc/runtime/lib/i386/ |
floatundixf.S | 26 movss 8(%esp), %xmm0 // hi 32 bits of input 28 orpd REL_ADDR(twop84), %xmm0 // 2^84 + hi (as a double) 30 addsd REL_ADDR(twop84_plus_twop52_neg), %xmm0 // hi - 2^52 (no rounding occurs) 33 movsd %xmm0, 4(%esp)
|
/external/llvm/test/MC/Disassembler/X86/ |
simple-tests.txt | 85 # CHECK: vcvtps2pd %xmm0, %ymm0 91 # CHECK: vcvtss2sil %xmm0, %eax 94 # CHECK: vcvtsd2si %xmm0, %eax 97 # CHECK: vcvtsd2si %xmm0, %rax 100 # CHECK: vmaskmovpd %xmm0, %xmm1, (%rax) 103 # CHECK: vmovapd %xmm0, %xmm2 177 # CHECK: movups %xmm1, %xmm0 180 # CHECK: movups %xmm0, %xmm1 183 # CHECK: movaps %xmm1, %xmm0 186 # CHECK: movaps %xmm0, %xmm [all...] |
x86-32.txt | 120 # CHECK: vcvtps2pd %xmm0, %ymm0 126 # CHECK: vcvtss2sil %xmm0, %eax 129 # CHECK: vcvtsd2si %xmm0, %eax 132 # CHECK: vcvtsd2si %xmm0, %eax 135 # CHECK: vmaskmovpd %xmm0, %xmm1, (%eax) 138 # CHECK: vmovapd %xmm0, %xmm2 208 # CHECK: movups %xmm1, %xmm0 211 # CHECK: movups %xmm0, %xmm1 214 # CHECK: movaps %xmm1, %xmm0 217 # CHECK: movaps %xmm0, %xmm [all...] |
/bionic/libc/arch-x86/string/ |
ssse3-memcmp3-new.S | 143 movdqu (%edx), %xmm0 146 pcmpeqb %xmm0, %xmm3 227 movdqa (%esi), %xmm0 229 pcmpeqb (%edi), %xmm0 234 pand %xmm0, %xmm2 237 movdqa %xmm0, %xmm1 238 movdqa 32(%esi), %xmm0 241 pcmpeqb 32(%edi), %xmm0 247 pand %xmm0, %xmm2 257 movdqa %xmm0, %xmm [all...] |
/external/libvpx/vp8/common/x86/ |
idctllm_sse2.asm | 59 movq xmm0, [rax] 65 punpcklbw xmm0, xmm7 74 paddw xmm0, xmm4 80 packuswb xmm0, xmm7 86 movq [rax], xmm0 124 movdqa xmm0, [rax] 136 pmullw xmm0, [rdx] 142 movdqa xmm4, xmm0 143 punpckldq xmm0, xmm1 146 pshufd xmm0, xmm0, 11011000 [all...] |
loopfilter_sse2.asm | 54 movdqa xmm0, xmm5 ; q0 58 psubusb xmm3, xmm0 ; q1-=q0 128 movdqa xmm3, xmm0 ; q0 182 movdqa xmm0, [rdx+32] ; q0 192 pxor xmm0, [GLOBAL(t80)] ; offset to convert to signed values 194 movdqa xmm3, xmm0 ; q0 195 psubsb xmm0, xmm6 ; q0 - p0 197 paddsb xmm2, xmm0 ; 1 * (q0 - p0) + hvm(p1 - q1) 199 paddsb xmm2, xmm0 ; 2 * (q0 - p0) + hvm(p1 - q1) 201 paddsb xmm2, xmm0 ; 3 * (q0 - p0) + hvm(p1 - q1 [all...] |
/external/libyuv/files/source/ |
planar_functions.cc | 76 movdqa xmm0, [eax] 79 movdqa xmm2, xmm0 81 pand xmm0, xmm7 // even bytes 83 packuswb xmm0, xmm1 84 movdqa [edx], xmm0 107 "movdqa (%0),%%xmm0\n" 110 "movdqa %%xmm0,%%xmm2\n" 112 "pand %%xmm7,%%xmm0\n" 114 "packuswb %%xmm1,%%xmm0\n" 115 "movdqa %%xmm0,(%1)\n 517 movdqa xmm0, [eax] local 527 psrlw xmm0, 8 \/\/ YUYV -> UVUV local 529 packuswb xmm0, xmm1 local 531 pand xmm0, xmm7 \/\/ U local 532 packuswb xmm0, xmm0 local 669 movdqa xmm0, [eax] local 672 pand xmm0, xmm7 \/\/ even bytes are Y local 674 packuswb xmm0, xmm1 local 698 movdqa xmm0, [eax] local 703 pavgb xmm0, xmm2 local 705 psrlw xmm0, 8 \/\/ YUYV -> UVUV local 707 packuswb xmm0, xmm1 local 709 pand xmm0, xmm7 \/\/ U local 710 packuswb xmm0, xmm0 local 736 movdqa xmm0, [eax] local 739 psrlw xmm0, 8 \/\/ odd bytes are Y local 741 packuswb xmm0, xmm1 local 765 movdqa xmm0, [eax] local 770 pavgb xmm0, xmm2 local 772 pand xmm0, xmm7 \/\/ UYVY -> UVUV local 774 packuswb xmm0, xmm1 local 776 pand xmm0, xmm7 \/\/ U local 777 packuswb xmm0, xmm0 local 1214 movq xmm0, qword ptr [eax] local 1216 punpcklbw xmm0, xmm0 local 1218 punpcklwd xmm0, xmm0 local 1220 por xmm0, xmm7 local 1242 movdqa xmm0, [eax] local 1244 pshufb xmm0, xmm7 local 1264 movdqa xmm0, [eax] local 1266 pshufb xmm0, xmm7 local [all...] |
rotate.cc | 85 movq xmm0, qword ptr [eax] local 89 punpcklbw xmm0, xmm1 local 91 movdqa xmm1, xmm0 111 punpcklwd xmm0, xmm2 local 113 movdqa xmm2, xmm0 125 punpckldq xmm0, xmm4 local 126 movq qword ptr [edx], xmm0 127 movdqa xmm4, xmm0 184 movdqa xmm0, [eax] local 187 movdqa xmm7, xmm0 // use xmm7 as temp register 188 punpcklbw xmm0, xmm1 local 218 punpcklwd xmm0, xmm2 local 238 punpckldq xmm0, xmm4 local 248 movdqa xmm0, xmm2 \/\/ use xmm0 as the temp register. local 252 punpckhdq xmm0, xmm6 local 257 movdqa xmm0, xmm1 \/\/ use xmm0 as the temp register. local 261 punpckhdq xmm0, xmm5 local 266 movdqa xmm0, xmm3 \/\/ use xmm0 as the temp register. local 270 punpckhdq xmm0, xmm7 local 871 movdqa xmm0, [eax] local 873 pshufb xmm0, xmm7 local 1082 movdqa xmm0, [eax] local 1084 pshufb xmm0, xmm7 local [all...] |