HomeSort by relevance Sort by last modified time
    Searched full:xmm0 (Results 51 - 75 of 297) sorted by null

1 23 4 5 6 7 8 91011>>

  /dalvik/vm/mterp/x86-atom/
OP_INT_TO_FLOAT.S 34 cvtsi2ss (rFP,%eax,4), %xmm0 # %xmm0<- vB
35 movss %xmm0, (rFP, rINST, 4) # vA<- %xmm0
OP_IPUT_WIDE_QUICK.S 36 movq (rFP, rINST, 4), %xmm0 # %xmm0<- fp[A]
37 movq %xmm0, (%edx, %ecx) # object field<- %xmm0; fp[A]
OP_MOVE_WIDE_16.S 34 movq (rFP, %edx, 4), %xmm0 # %xmm0<- vB
35 movq %xmm0, (rFP, %ecx, 4) # vA<- vB; %xmm0
OP_SHL_LONG.S 35 movss (rFP, %eax, 4), %xmm0 # %xmm0<- vCC
36 pand %xmm2, %xmm0 # %xmm0<- masked shift bits
38 psllq %xmm0, %xmm1 # %xmm1<- shifted vBB
OP_SHL_LONG_2ADDR.S 35 movss (rFP, %edx, 4), %xmm0 # %xmm0<- vB
38 pand %xmm2, %xmm0 # %xmm0<- masked shift bits
39 psllq %xmm0, %xmm1 # %xmm1<- shifted vA
OP_USHR_LONG_2ADDR.S 36 movss (rFP, %edx, 4), %xmm0 # %xmm0<- vB
38 pand %xmm2, %xmm0 # %xmm0<- masked shift bits
39 psrlq %xmm0, %xmm1 # %xmm1<- shifted vA
OP_SHR_LONG.S 35 movss (rFP, %eax, 4), %xmm0 # %xmm0<- vCC
37 pand %xmm2, %xmm0 # %xmm0<- masked for the shift bits
38 psrlq %xmm0, %xmm1 # %xmm1<- shifted vBB
46 psubq %xmm0, %xmm3 # %xmm3<- 64 - shift amount
OP_SHR_LONG_2ADDR.S 35 movss (rFP, %edx, 4), %xmm0 # %xmm0<- vB
38 pand %xmm2, %xmm0 # %xmm0<- masked for the shift bits
39 psrlq %xmm0, %xmm1 # %xmm1<- shifted vBB
47 psubq %xmm0, %xmm3 # %xmm3<- 64 - shift amount
OP_AGET_WIDE.S 41 movq offArrayObject_contents(%ecx, %edx, 8), %xmm0 # %xmm0<- vBB[vCC]
42 movq %xmm0, (rFP, rINST, 4) # vAA<- %xmm0; value
OP_APUT_WIDE.S 41 movq (rFP, rINST, 4), %xmm0 # %xmm0<- vAA
42 movq %xmm0, offArrayObject_contents(%ecx, %edx, 8) # vBB[vCC]<- %xmm0; value
  /external/libvpx/vp8/encoder/x86/
ssim_opt.asm 30 punpckldq %1,xmm0
31 punpckhdq xmm2,xmm0
34 punpcklqdq %1,xmm0
35 punpckhqdq xmm2,xmm0
42 punpcklwd %1,xmm0
43 punpckhwd xmm1,xmm0
78 pxor xmm0, xmm0
93 punpckhbw xmm3, xmm0 ; high_s
94 punpckhbw xmm4, xmm0 ; high_
    [all...]
quantize_sse2.asm 62 movdqa xmm0, [rdx]
69 movdqa xmm1, xmm0
73 psraw xmm0, 15
77 pxor xmm1, xmm0
81 psubw xmm1, xmm0
183 pxor xmm2, xmm0
186 psubw xmm2, xmm0
190 movdqa xmm0, [rcx]
195 pmullw xmm0, xmm2
200 movdqa [rsi], xmm0 ; store dqcoef
    [all...]
  /external/llvm/test/CodeGen/X86/
pmulld.ll 9 ; WIN64-NEXT: movdqa (%rcx), %xmm0
10 ; WIN64-NEXT: pmulld (%rdx), %xmm0
20 ; WIN64-NEXT: movdqa (%rcx), %xmm0
21 ; WIN64-NEXT: pmulld (%rdx), %xmm0
vec_shuffle-37.ll 7 ; CHECK: movaps ({{%rdi|%rcx}}), %xmm0
8 ; CHECK: movaps %xmm0, %xmm1
10 ; CHECK-NEXT: shufps $36, %xmm1, %xmm0
19 ; CHECK_O0: movsd (%eax), %xmm0
20 ; CHECK_O0: unpcklpd %xmm0, %xmm0
32 ; CHECK: movq %xmm0, ({{%rsi|%rdx}})
bc-extract.ll 14 ; CHECK: pxor %xmm0, %xmm0
illegal-vector-args-return.ll 2 ; RUN: llc < %s -march=x86 -mattr=+sse2 | grep {mulpd %xmm2, %xmm0}
4 ; RUN: llc < %s -march=x86 -mattr=+sse2 | grep {addps %xmm2, %xmm0}
rip-rel-address.ll 12 ; PIC64: movsd _a(%rip), %xmm0
13 ; STATIC64: movsd a(%rip), %xmm0
  /frameworks/compile/libbcc/runtime/lib/i386/
floatundixf.S 26 movss 8(%esp), %xmm0 // hi 32 bits of input
28 orpd REL_ADDR(twop84), %xmm0 // 2^84 + hi (as a double)
30 addsd REL_ADDR(twop84_plus_twop52_neg), %xmm0 // hi - 2^52 (no rounding occurs)
33 movsd %xmm0, 4(%esp)
  /external/llvm/test/MC/Disassembler/X86/
simple-tests.txt 85 # CHECK: vcvtps2pd %xmm0, %ymm0
91 # CHECK: vcvtss2sil %xmm0, %eax
94 # CHECK: vcvtsd2si %xmm0, %eax
97 # CHECK: vcvtsd2si %xmm0, %rax
100 # CHECK: vmaskmovpd %xmm0, %xmm1, (%rax)
103 # CHECK: vmovapd %xmm0, %xmm2
177 # CHECK: movups %xmm1, %xmm0
180 # CHECK: movups %xmm0, %xmm1
183 # CHECK: movaps %xmm1, %xmm0
186 # CHECK: movaps %xmm0, %xmm
    [all...]
x86-32.txt 120 # CHECK: vcvtps2pd %xmm0, %ymm0
126 # CHECK: vcvtss2sil %xmm0, %eax
129 # CHECK: vcvtsd2si %xmm0, %eax
132 # CHECK: vcvtsd2si %xmm0, %eax
135 # CHECK: vmaskmovpd %xmm0, %xmm1, (%eax)
138 # CHECK: vmovapd %xmm0, %xmm2
208 # CHECK: movups %xmm1, %xmm0
211 # CHECK: movups %xmm0, %xmm1
214 # CHECK: movaps %xmm1, %xmm0
217 # CHECK: movaps %xmm0, %xmm
    [all...]
  /bionic/libc/arch-x86/string/
ssse3-memcmp3-new.S 143 movdqu (%edx), %xmm0
146 pcmpeqb %xmm0, %xmm3
227 movdqa (%esi), %xmm0
229 pcmpeqb (%edi), %xmm0
234 pand %xmm0, %xmm2
237 movdqa %xmm0, %xmm1
238 movdqa 32(%esi), %xmm0
241 pcmpeqb 32(%edi), %xmm0
247 pand %xmm0, %xmm2
257 movdqa %xmm0, %xmm
    [all...]
  /external/libvpx/vp8/common/x86/
idctllm_sse2.asm 59 movq xmm0, [rax]
65 punpcklbw xmm0, xmm7
74 paddw xmm0, xmm4
80 packuswb xmm0, xmm7
86 movq [rax], xmm0
124 movdqa xmm0, [rax]
136 pmullw xmm0, [rdx]
142 movdqa xmm4, xmm0
143 punpckldq xmm0, xmm1
146 pshufd xmm0, xmm0, 11011000
    [all...]
loopfilter_sse2.asm 54 movdqa xmm0, xmm5 ; q0
58 psubusb xmm3, xmm0 ; q1-=q0
128 movdqa xmm3, xmm0 ; q0
182 movdqa xmm0, [rdx+32] ; q0
192 pxor xmm0, [GLOBAL(t80)] ; offset to convert to signed values
194 movdqa xmm3, xmm0 ; q0
195 psubsb xmm0, xmm6 ; q0 - p0
197 paddsb xmm2, xmm0 ; 1 * (q0 - p0) + hvm(p1 - q1)
199 paddsb xmm2, xmm0 ; 2 * (q0 - p0) + hvm(p1 - q1)
201 paddsb xmm2, xmm0 ; 3 * (q0 - p0) + hvm(p1 - q1
    [all...]
  /external/libyuv/files/source/
planar_functions.cc 76 movdqa xmm0, [eax]
79 movdqa xmm2, xmm0
81 pand xmm0, xmm7 // even bytes
83 packuswb xmm0, xmm1
84 movdqa [edx], xmm0
107 "movdqa (%0),%%xmm0\n"
110 "movdqa %%xmm0,%%xmm2\n"
112 "pand %%xmm7,%%xmm0\n"
114 "packuswb %%xmm1,%%xmm0\n"
115 "movdqa %%xmm0,(%1)\n
517 movdqa xmm0, [eax] local
527 psrlw xmm0, 8 \/\/ YUYV -> UVUV local
529 packuswb xmm0, xmm1 local
531 pand xmm0, xmm7 \/\/ U local
532 packuswb xmm0, xmm0 local
669 movdqa xmm0, [eax] local
672 pand xmm0, xmm7 \/\/ even bytes are Y local
674 packuswb xmm0, xmm1 local
698 movdqa xmm0, [eax] local
703 pavgb xmm0, xmm2 local
705 psrlw xmm0, 8 \/\/ YUYV -> UVUV local
707 packuswb xmm0, xmm1 local
709 pand xmm0, xmm7 \/\/ U local
710 packuswb xmm0, xmm0 local
736 movdqa xmm0, [eax] local
739 psrlw xmm0, 8 \/\/ odd bytes are Y local
741 packuswb xmm0, xmm1 local
765 movdqa xmm0, [eax] local
770 pavgb xmm0, xmm2 local
772 pand xmm0, xmm7 \/\/ UYVY -> UVUV local
774 packuswb xmm0, xmm1 local
776 pand xmm0, xmm7 \/\/ U local
777 packuswb xmm0, xmm0 local
1214 movq xmm0, qword ptr [eax] local
1216 punpcklbw xmm0, xmm0 local
1218 punpcklwd xmm0, xmm0 local
1220 por xmm0, xmm7 local
1242 movdqa xmm0, [eax] local
1244 pshufb xmm0, xmm7 local
1264 movdqa xmm0, [eax] local
1266 pshufb xmm0, xmm7 local
    [all...]
rotate.cc 85 movq xmm0, qword ptr [eax] local
89 punpcklbw xmm0, xmm1 local
91 movdqa xmm1, xmm0
111 punpcklwd xmm0, xmm2 local
113 movdqa xmm2, xmm0
125 punpckldq xmm0, xmm4 local
126 movq qword ptr [edx], xmm0
127 movdqa xmm4, xmm0
184 movdqa xmm0, [eax] local
187 movdqa xmm7, xmm0 // use xmm7 as temp register
188 punpcklbw xmm0, xmm1 local
218 punpcklwd xmm0, xmm2 local
238 punpckldq xmm0, xmm4 local
248 movdqa xmm0, xmm2 \/\/ use xmm0 as the temp register. local
252 punpckhdq xmm0, xmm6 local
257 movdqa xmm0, xmm1 \/\/ use xmm0 as the temp register. local
261 punpckhdq xmm0, xmm5 local
266 movdqa xmm0, xmm3 \/\/ use xmm0 as the temp register. local
270 punpckhdq xmm0, xmm7 local
871 movdqa xmm0, [eax] local
873 pshufb xmm0, xmm7 local
1082 movdqa xmm0, [eax] local
1084 pshufb xmm0, xmm7 local
    [all...]

Completed in 315 milliseconds

1 23 4 5 6 7 8 91011>>