HomeSort by relevance Sort by last modified time
    Searched full:xmm0 (Results 76 - 100 of 341) sorted by null

1 2 34 5 6 7 8 91011>>

  /dalvik/vm/mterp/x86-atom/
OP_USHR_LONG_2ADDR.S 36 movss (rFP, %edx, 4), %xmm0 # %xmm0<- vB
38 pand %xmm2, %xmm0 # %xmm0<- masked shift bits
39 psrlq %xmm0, %xmm1 # %xmm1<- shifted vA
OP_SHR_LONG.S 35 movss (rFP, %eax, 4), %xmm0 # %xmm0<- vCC
37 pand %xmm2, %xmm0 # %xmm0<- masked for the shift bits
38 psrlq %xmm0, %xmm1 # %xmm1<- shifted vBB
46 psubq %xmm0, %xmm3 # %xmm3<- 64 - shift amount
OP_SHR_LONG_2ADDR.S 35 movss (rFP, %edx, 4), %xmm0 # %xmm0<- vB
38 pand %xmm2, %xmm0 # %xmm0<- masked for the shift bits
39 psrlq %xmm0, %xmm1 # %xmm1<- shifted vBB
47 psubq %xmm0, %xmm3 # %xmm3<- 64 - shift amount
OP_AGET_WIDE.S 41 movq offArrayObject_contents(%ecx, %edx, 8), %xmm0 # %xmm0<- vBB[vCC]
42 movq %xmm0, (rFP, rINST, 4) # vAA<- %xmm0; value
OP_APUT_WIDE.S 41 movq (rFP, rINST, 4), %xmm0 # %xmm0<- vAA
42 movq %xmm0, offArrayObject_contents(%ecx, %edx, 8) # vBB[vCC]<- %xmm0; value
unopWide.S 21 * performs, for example, "%xmm0 = op %xmm1".
33 %default {"preinstr":"","result":"%xmm0"}
39 movq (rFP, rINST, 4), %xmm0 # %xmm0<- vB
OP_MOVE_RESULT_WIDE.S 35 movq offGlue_retval(%eax), %xmm0 # %xmm0<- glue->retval
36 movq %xmm0, (rFP, rINST, 4) # vA<- glue->retval
OP_MOVE_WIDE.S 35 movq (rFP, %edx, 4), %xmm0 # %xmm0<- vB
36 movq %xmm0, (rFP, rINST, 4) # vA<- vB
OP_MOVE_WIDE_FROM16.S 32 movq (rFP, %edx, 4), %xmm0 # %xmm0<- vB
33 movq %xmm0, (rFP, rINST, 4) # vA<- vB
OP_RETURN_WIDE.S 32 movq (rFP, rINST, 4), %xmm0 # %xmm0<- vAA
33 movq %xmm0, offGlue_retval(%edx)# glue->retval<- vAA
  /external/libvpx/vp8/common/x86/
loopfilter_sse2.asm 54 movdqa xmm0, xmm5 ; q0
58 psubusb xmm3, xmm0 ; q1-=q0
128 movdqa xmm3, xmm0 ; q0
182 movdqa xmm0, [rdx+32] ; q0
192 pxor xmm0, [GLOBAL(t80)] ; offset to convert to signed values
194 movdqa xmm3, xmm0 ; q0
195 psubsb xmm0, xmm6 ; q0 - p0
197 paddsb xmm2, xmm0 ; 1 * (q0 - p0) + hvm(p1 - q1)
199 paddsb xmm2, xmm0 ; 2 * (q0 - p0) + hvm(p1 - q1)
201 paddsb xmm2, xmm0 ; 3 * (q0 - p0) + hvm(p1 - q1
    [all...]
  /external/libyuv/files/source/
planar_functions.cc 76 movdqa xmm0, [eax]
79 movdqa xmm2, xmm0
81 pand xmm0, xmm7 // even bytes
83 packuswb xmm0, xmm1
84 movdqa [edx], xmm0
107 "movdqa (%0),%%xmm0\n"
110 "movdqa %%xmm0,%%xmm2\n"
112 "pand %%xmm7,%%xmm0\n"
114 "packuswb %%xmm1,%%xmm0\n"
115 "movdqa %%xmm0,(%1)\n
517 movdqa xmm0, [eax] local
527 psrlw xmm0, 8 \/\/ YUYV -> UVUV local
529 packuswb xmm0, xmm1 local
531 pand xmm0, xmm7 \/\/ U local
532 packuswb xmm0, xmm0 local
669 movdqa xmm0, [eax] local
672 pand xmm0, xmm7 \/\/ even bytes are Y local
674 packuswb xmm0, xmm1 local
698 movdqa xmm0, [eax] local
703 pavgb xmm0, xmm2 local
705 psrlw xmm0, 8 \/\/ YUYV -> UVUV local
707 packuswb xmm0, xmm1 local
709 pand xmm0, xmm7 \/\/ U local
710 packuswb xmm0, xmm0 local
736 movdqa xmm0, [eax] local
739 psrlw xmm0, 8 \/\/ odd bytes are Y local
741 packuswb xmm0, xmm1 local
765 movdqa xmm0, [eax] local
770 pavgb xmm0, xmm2 local
772 pand xmm0, xmm7 \/\/ UYVY -> UVUV local
774 packuswb xmm0, xmm1 local
776 pand xmm0, xmm7 \/\/ U local
777 packuswb xmm0, xmm0 local
1214 movq xmm0, qword ptr [eax] local
1216 punpcklbw xmm0, xmm0 local
1218 punpcklwd xmm0, xmm0 local
1220 por xmm0, xmm7 local
1242 movdqa xmm0, [eax] local
1244 pshufb xmm0, xmm7 local
1264 movdqa xmm0, [eax] local
1266 pshufb xmm0, xmm7 local
    [all...]
rotate.cc 85 movq xmm0, qword ptr [eax] local
89 punpcklbw xmm0, xmm1 local
91 movdqa xmm1, xmm0
111 punpcklwd xmm0, xmm2 local
113 movdqa xmm2, xmm0
125 punpckldq xmm0, xmm4 local
126 movq qword ptr [edx], xmm0
127 movdqa xmm4, xmm0
184 movdqa xmm0, [eax] local
187 movdqa xmm7, xmm0 // use xmm7 as temp register
188 punpcklbw xmm0, xmm1 local
218 punpcklwd xmm0, xmm2 local
238 punpckldq xmm0, xmm4 local
248 movdqa xmm0, xmm2 \/\/ use xmm0 as the temp register. local
252 punpckhdq xmm0, xmm6 local
257 movdqa xmm0, xmm1 \/\/ use xmm0 as the temp register. local
261 punpckhdq xmm0, xmm5 local
266 movdqa xmm0, xmm3 \/\/ use xmm0 as the temp register. local
270 punpckhdq xmm0, xmm7 local
871 movdqa xmm0, [eax] local
873 pshufb xmm0, xmm7 local
1082 movdqa xmm0, [eax] local
1084 pshufb xmm0, xmm7 local
    [all...]
  /external/llvm/test/CodeGen/X86/
vec_shuffle-37.ll 7 ; CHECK: movaps ({{%rdi|%rcx}}), %[[XMM0:xmm[0-9]+]]
8 ; CHECK: movaps %[[XMM0]], %[[XMM1:xmm[0-9]+]]
10 ; CHECK-NEXT: shufps $36, %[[XMM1]], %[[XMM0]]
19 ; CHECK_O0: movsd (%eax), %xmm0
20 ; CHECK_O0: unpcklpd %xmm0, %xmm0
bc-extract.ll 14 ; CHECK: xorps %xmm0, %xmm0
illegal-vector-args-return.ll 2 ; RUN: llc < %s -march=x86 -mattr=+sse2 | grep {mulpd %xmm2, %xmm0}
4 ; RUN: llc < %s -march=x86 -mattr=+sse2 | grep {addps %xmm2, %xmm0}
rip-rel-address.ll 12 ; PIC64: movsd _a(%rip), %xmm0
13 ; STATIC64: movsd a(%rip), %xmm0
avx-vextractf128.ll 22 ; CHECK-NOT: vextractf128 $0, %ymm0, %xmm0
23 ; CHECK-NOT: vmovaps %xmm0, (%rdi)
36 ; CHECK-NOT: vextractf128 $0, %ymm0, %xmm0
37 ; CHECK-NOT: vmovups %xmm0, (%rdi)
50 ; CHECK-NOT: vextractf128 $0, %ymm0, %xmm0
51 ; CHECK-NOT: vmovaps %xmm0, (%rdi)
64 ; CHECK-NOT: vextractf128 $0, %ymm0, %xmm0
65 ; CHECK-NOT: vmovups %xmm0, (%rdi)
78 ; CHECK-NOT: vextractf128 $0, %ymm0, %xmm0
79 ; CHECK-NOT: vmovaps %xmm0, (%rdi
    [all...]
  /external/libvpx/vp8/encoder/x86/
quantize_sse2.asm 62 movdqa xmm0, [rdx]
69 movdqa xmm1, xmm0
73 psraw xmm0, 15
77 pxor xmm1, xmm0
81 psubw xmm1, xmm0
183 pxor xmm2, xmm0
186 psubw xmm2, xmm0
190 movdqa xmm0, [rcx]
195 pmullw xmm0, xmm2
200 movdqa [rsi], xmm0 ; store dqcoef
    [all...]
  /external/valgrind/main/none/tests/amd64/
pcmpxstrx64.c 60 + 2/*inout:xmm0*/ // 6 48
79 "movupd 48(%0), %%xmm0" "\n\t"
82 "movupd %%xmm0, 48(%0)" "\n\t"
89 : /*trash*/"memory","cc","xmm2","xmm13","xmm0","rdx","rax","rcx","r15"
92 printf(" xmm0 ");
108 "movupd 48(%0), %%xmm0" "\n\t"
111 "movupd %%xmm0, 48(%0)" "\n\t"
118 : /*trash*/"memory","cc","xmm2","xmm13","xmm0","rdx","rax","rcx","r15"
121 printf(" xmm0 ");
137 "movupd 48(%0), %%xmm0" "\n\t
    [all...]
  /external/compiler-rt/lib/i386/
floatdisf.S 21 movd 4(%esp), %xmm0
23 punpckldq %xmm1, %xmm0
24 movq %xmm0, 4(%esp)
floatdixf.S 21 movd 4(%esp), %xmm0
23 punpckldq %xmm1, %xmm0
24 movq %xmm0, 4(%esp)
  /frameworks/compile/libbcc/runtime/lib/i386/
floatdisf.S 21 movd 4(%esp), %xmm0
23 punpckldq %xmm1, %xmm0
24 movq %xmm0, 4(%esp)
  /external/llvm/test/MC/X86/
x86_64-xop-encoding.s 11 // CHECK: vphsubwd %xmm0, %xmm1
13 vphsubwd %xmm0, %xmm1
19 // CHECK: vphsubdq %xmm0, %xmm1
21 vphsubdq %xmm0, %xmm1
51 // CHECK: vphadduwq %xmm7, %xmm0
53 vphadduwq %xmm7, %xmm0
75 // CHECK: vphaddubw %xmm5, %xmm0
77 vphaddubw %xmm5, %xmm0
115 // CHECK: vphaddbq %xmm2, %xmm0
117 vphaddbq %xmm2, %xmm0
    [all...]
  /external/llvm/lib/Target/X86/
README-SSE.txt 33 pshufd $1, %xmm0, %xmm1 ## xmm1 = xmm0[1,0,0,0]
34 addss %xmm0, %xmm1
35 pshufd $3, %xmm0, %xmm2 ## xmm2 = xmm0[3,0,0,0]
36 movhlps %xmm0, %xmm0 ## xmm0 = xmm0[1,1]
37 movaps %xmm0, %xmm
    [all...]

Completed in 1005 milliseconds

1 2 34 5 6 7 8 91011>>