/dalvik/vm/mterp/x86-atom/ |
OP_USHR_LONG_2ADDR.S | 36 movss (rFP, %edx, 4), %xmm0 # %xmm0<- vB 38 pand %xmm2, %xmm0 # %xmm0<- masked shift bits 39 psrlq %xmm0, %xmm1 # %xmm1<- shifted vA
|
OP_SHR_LONG.S | 35 movss (rFP, %eax, 4), %xmm0 # %xmm0<- vCC 37 pand %xmm2, %xmm0 # %xmm0<- masked for the shift bits 38 psrlq %xmm0, %xmm1 # %xmm1<- shifted vBB 46 psubq %xmm0, %xmm3 # %xmm3<- 64 - shift amount
|
OP_SHR_LONG_2ADDR.S | 35 movss (rFP, %edx, 4), %xmm0 # %xmm0<- vB 38 pand %xmm2, %xmm0 # %xmm0<- masked for the shift bits 39 psrlq %xmm0, %xmm1 # %xmm1<- shifted vBB 47 psubq %xmm0, %xmm3 # %xmm3<- 64 - shift amount
|
OP_AGET_WIDE.S | 41 movq offArrayObject_contents(%ecx, %edx, 8), %xmm0 # %xmm0<- vBB[vCC] 42 movq %xmm0, (rFP, rINST, 4) # vAA<- %xmm0; value
|
OP_APUT_WIDE.S | 41 movq (rFP, rINST, 4), %xmm0 # %xmm0<- vAA 42 movq %xmm0, offArrayObject_contents(%ecx, %edx, 8) # vBB[vCC]<- %xmm0; value
|
unopWide.S | 21 * performs, for example, "%xmm0 = op %xmm1". 33 %default {"preinstr":"","result":"%xmm0"} 39 movq (rFP, rINST, 4), %xmm0 # %xmm0<- vB
|
OP_MOVE_RESULT_WIDE.S | 35 movq offGlue_retval(%eax), %xmm0 # %xmm0<- glue->retval 36 movq %xmm0, (rFP, rINST, 4) # vA<- glue->retval
|
OP_MOVE_WIDE.S | 35 movq (rFP, %edx, 4), %xmm0 # %xmm0<- vB 36 movq %xmm0, (rFP, rINST, 4) # vA<- vB
|
OP_MOVE_WIDE_FROM16.S | 32 movq (rFP, %edx, 4), %xmm0 # %xmm0<- vB 33 movq %xmm0, (rFP, rINST, 4) # vA<- vB
|
OP_RETURN_WIDE.S | 32 movq (rFP, rINST, 4), %xmm0 # %xmm0<- vAA 33 movq %xmm0, offGlue_retval(%edx)# glue->retval<- vAA
|
/external/libvpx/vp8/common/x86/ |
loopfilter_sse2.asm | 54 movdqa xmm0, xmm5 ; q0 58 psubusb xmm3, xmm0 ; q1-=q0 128 movdqa xmm3, xmm0 ; q0 182 movdqa xmm0, [rdx+32] ; q0 192 pxor xmm0, [GLOBAL(t80)] ; offset to convert to signed values 194 movdqa xmm3, xmm0 ; q0 195 psubsb xmm0, xmm6 ; q0 - p0 197 paddsb xmm2, xmm0 ; 1 * (q0 - p0) + hvm(p1 - q1) 199 paddsb xmm2, xmm0 ; 2 * (q0 - p0) + hvm(p1 - q1) 201 paddsb xmm2, xmm0 ; 3 * (q0 - p0) + hvm(p1 - q1 [all...] |
/external/libyuv/files/source/ |
planar_functions.cc | 76 movdqa xmm0, [eax] 79 movdqa xmm2, xmm0 81 pand xmm0, xmm7 // even bytes 83 packuswb xmm0, xmm1 84 movdqa [edx], xmm0 107 "movdqa (%0),%%xmm0\n" 110 "movdqa %%xmm0,%%xmm2\n" 112 "pand %%xmm7,%%xmm0\n" 114 "packuswb %%xmm1,%%xmm0\n" 115 "movdqa %%xmm0,(%1)\n 517 movdqa xmm0, [eax] local 527 psrlw xmm0, 8 \/\/ YUYV -> UVUV local 529 packuswb xmm0, xmm1 local 531 pand xmm0, xmm7 \/\/ U local 532 packuswb xmm0, xmm0 local 669 movdqa xmm0, [eax] local 672 pand xmm0, xmm7 \/\/ even bytes are Y local 674 packuswb xmm0, xmm1 local 698 movdqa xmm0, [eax] local 703 pavgb xmm0, xmm2 local 705 psrlw xmm0, 8 \/\/ YUYV -> UVUV local 707 packuswb xmm0, xmm1 local 709 pand xmm0, xmm7 \/\/ U local 710 packuswb xmm0, xmm0 local 736 movdqa xmm0, [eax] local 739 psrlw xmm0, 8 \/\/ odd bytes are Y local 741 packuswb xmm0, xmm1 local 765 movdqa xmm0, [eax] local 770 pavgb xmm0, xmm2 local 772 pand xmm0, xmm7 \/\/ UYVY -> UVUV local 774 packuswb xmm0, xmm1 local 776 pand xmm0, xmm7 \/\/ U local 777 packuswb xmm0, xmm0 local 1214 movq xmm0, qword ptr [eax] local 1216 punpcklbw xmm0, xmm0 local 1218 punpcklwd xmm0, xmm0 local 1220 por xmm0, xmm7 local 1242 movdqa xmm0, [eax] local 1244 pshufb xmm0, xmm7 local 1264 movdqa xmm0, [eax] local 1266 pshufb xmm0, xmm7 local [all...] |
rotate.cc | 85 movq xmm0, qword ptr [eax] local 89 punpcklbw xmm0, xmm1 local 91 movdqa xmm1, xmm0 111 punpcklwd xmm0, xmm2 local 113 movdqa xmm2, xmm0 125 punpckldq xmm0, xmm4 local 126 movq qword ptr [edx], xmm0 127 movdqa xmm4, xmm0 184 movdqa xmm0, [eax] local 187 movdqa xmm7, xmm0 // use xmm7 as temp register 188 punpcklbw xmm0, xmm1 local 218 punpcklwd xmm0, xmm2 local 238 punpckldq xmm0, xmm4 local 248 movdqa xmm0, xmm2 \/\/ use xmm0 as the temp register. local 252 punpckhdq xmm0, xmm6 local 257 movdqa xmm0, xmm1 \/\/ use xmm0 as the temp register. local 261 punpckhdq xmm0, xmm5 local 266 movdqa xmm0, xmm3 \/\/ use xmm0 as the temp register. local 270 punpckhdq xmm0, xmm7 local 871 movdqa xmm0, [eax] local 873 pshufb xmm0, xmm7 local 1082 movdqa xmm0, [eax] local 1084 pshufb xmm0, xmm7 local [all...] |
/external/llvm/test/CodeGen/X86/ |
vec_shuffle-37.ll | 7 ; CHECK: movaps ({{%rdi|%rcx}}), %[[XMM0:xmm[0-9]+]] 8 ; CHECK: movaps %[[XMM0]], %[[XMM1:xmm[0-9]+]] 10 ; CHECK-NEXT: shufps $36, %[[XMM1]], %[[XMM0]] 19 ; CHECK_O0: movsd (%eax), %xmm0 20 ; CHECK_O0: unpcklpd %xmm0, %xmm0
|
bc-extract.ll | 14 ; CHECK: xorps %xmm0, %xmm0
|
illegal-vector-args-return.ll | 2 ; RUN: llc < %s -march=x86 -mattr=+sse2 | grep {mulpd %xmm2, %xmm0} 4 ; RUN: llc < %s -march=x86 -mattr=+sse2 | grep {addps %xmm2, %xmm0}
|
rip-rel-address.ll | 12 ; PIC64: movsd _a(%rip), %xmm0 13 ; STATIC64: movsd a(%rip), %xmm0
|
avx-vextractf128.ll | 22 ; CHECK-NOT: vextractf128 $0, %ymm0, %xmm0 23 ; CHECK-NOT: vmovaps %xmm0, (%rdi) 36 ; CHECK-NOT: vextractf128 $0, %ymm0, %xmm0 37 ; CHECK-NOT: vmovups %xmm0, (%rdi) 50 ; CHECK-NOT: vextractf128 $0, %ymm0, %xmm0 51 ; CHECK-NOT: vmovaps %xmm0, (%rdi) 64 ; CHECK-NOT: vextractf128 $0, %ymm0, %xmm0 65 ; CHECK-NOT: vmovups %xmm0, (%rdi) 78 ; CHECK-NOT: vextractf128 $0, %ymm0, %xmm0 79 ; CHECK-NOT: vmovaps %xmm0, (%rdi [all...] |
/external/libvpx/vp8/encoder/x86/ |
quantize_sse2.asm | 62 movdqa xmm0, [rdx] 69 movdqa xmm1, xmm0 73 psraw xmm0, 15 77 pxor xmm1, xmm0 81 psubw xmm1, xmm0 183 pxor xmm2, xmm0 186 psubw xmm2, xmm0 190 movdqa xmm0, [rcx] 195 pmullw xmm0, xmm2 200 movdqa [rsi], xmm0 ; store dqcoef [all...] |
/external/valgrind/main/none/tests/amd64/ |
pcmpxstrx64.c | 60 + 2/*inout:xmm0*/ // 6 48 79 "movupd 48(%0), %%xmm0" "\n\t" 82 "movupd %%xmm0, 48(%0)" "\n\t" 89 : /*trash*/"memory","cc","xmm2","xmm13","xmm0","rdx","rax","rcx","r15" 92 printf(" xmm0 "); 108 "movupd 48(%0), %%xmm0" "\n\t" 111 "movupd %%xmm0, 48(%0)" "\n\t" 118 : /*trash*/"memory","cc","xmm2","xmm13","xmm0","rdx","rax","rcx","r15" 121 printf(" xmm0 "); 137 "movupd 48(%0), %%xmm0" "\n\t [all...] |
/external/compiler-rt/lib/i386/ |
floatdisf.S | 21 movd 4(%esp), %xmm0 23 punpckldq %xmm1, %xmm0 24 movq %xmm0, 4(%esp)
|
floatdixf.S | 21 movd 4(%esp), %xmm0 23 punpckldq %xmm1, %xmm0 24 movq %xmm0, 4(%esp)
|
/frameworks/compile/libbcc/runtime/lib/i386/ |
floatdisf.S | 21 movd 4(%esp), %xmm0 23 punpckldq %xmm1, %xmm0 24 movq %xmm0, 4(%esp)
|
/external/llvm/test/MC/X86/ |
x86_64-xop-encoding.s | 11 // CHECK: vphsubwd %xmm0, %xmm1 13 vphsubwd %xmm0, %xmm1 19 // CHECK: vphsubdq %xmm0, %xmm1 21 vphsubdq %xmm0, %xmm1 51 // CHECK: vphadduwq %xmm7, %xmm0 53 vphadduwq %xmm7, %xmm0 75 // CHECK: vphaddubw %xmm5, %xmm0 77 vphaddubw %xmm5, %xmm0 115 // CHECK: vphaddbq %xmm2, %xmm0 117 vphaddbq %xmm2, %xmm0 [all...] |
/external/llvm/lib/Target/X86/ |
README-SSE.txt | 33 pshufd $1, %xmm0, %xmm1 ## xmm1 = xmm0[1,0,0,0] 34 addss %xmm0, %xmm1 35 pshufd $3, %xmm0, %xmm2 ## xmm2 = xmm0[3,0,0,0] 36 movhlps %xmm0, %xmm0 ## xmm0 = xmm0[1,1] 37 movaps %xmm0, %xmm [all...] |