HomeSort by relevance Sort by last modified time
    Searched full:xmm0 (Results 1 - 25 of 341) sorted by null

1 2 3 4 5 6 7 8 91011>>

  /external/valgrind/main/none/tests/amd64/
pcmpxstrx64.stdout.exp 3 istri $0x4A: xmm0 55555555555555555555555555555555 rcx 5555555555550006 flags 00000881
4 istri $0x0A: xmm0 55555555555555555555555555555555 rcx 5555555555550000 flags 00000881
5 istrm $0x4A: xmm0 000000000000000000ffffffffffffff rcx 5555555555555555 flags 00000881
6 istrm $0x0A: xmm0 0000000000000000000000000000007f rcx 5555555555555555 flags 00000881
7 estri $0x4A: xmm0 55555555555555555555555555555555 rcx 555555555555000f flags 000008c1
8 estri $0x0A: xmm0 55555555555555555555555555555555 rcx 5555555555550000 flags 000008c1
9 estrm $0x4A: xmm0 ffffffffffffffffffffffffffffffff rcx 5555555555555555 flags 000008c1
10 estrm $0x0A: xmm0 0000000000000000000000000000ffff rcx 5555555555555555 flags 000008c1
13 istri $0x4A: xmm0 55555555555555555555555555555555 rcx 555555555555000f flags 000000c1
14 istri $0x0A: xmm0 55555555555555555555555555555555 rcx 5555555555550007 flags 000000c
    [all...]
  /external/llvm/test/MC/X86/
x86_64-fma4-encoding.s 4 // CHECK: vfmaddss (%rcx), %xmm1, %xmm0, %xmm0
6 vfmaddss (%rcx), %xmm1, %xmm0, %xmm0
8 // CHECK: vfmaddss %xmm1, (%rcx), %xmm0, %xmm0
10 vfmaddss %xmm1, (%rcx),%xmm0, %xmm0
12 // CHECK: vfmaddss %xmm2, %xmm1, %xmm0, %xmm0
    [all...]
  /external/chromium/base/
cpu_unittest.cc 27 __asm xorps xmm0, xmm0; local
32 __asm psrldq xmm0, 0; local
37 __asm addsubpd xmm0, xmm0; local
42 __asm psignb xmm0, xmm0; local
47 __asm pmuldq xmm0, xmm0; local
62 __asm__ __volatile__("xorps %%xmm0, %%xmm0\n" : : : "xmm0")
    [all...]
  /external/llvm/test/CodeGen/X86/
v2f32.ll 14 ; X64-NEXT: pshufd $1, %xmm0, %xmm1
15 ; X64-NEXT: addss %xmm0, %xmm1
20 ; W64-NEXT: movdqa (%rcx), %xmm0
21 ; W64-NEXT: pshufd $1, %xmm0, %xmm1
22 ; W64-NEXT: addss %xmm0, %xmm1
27 ; X32-NEXT: pshufd $1, %xmm0, %xmm1
28 ; X32-NEXT: addss %xmm0, %xmm1
40 ; X64-NEXT: addps %xmm1, %xmm0
44 ; W64-NEXT: movaps (%rcx), %xmm0
45 ; W64-NEXT: addps (%rdx), %xmm0
    [all...]
vec_set-9.ll 2 ; RUN: llc < %s -march=x86-64 | grep {movlhps.*%xmm0, %xmm0}
sse-minmax.ll 19 ; CHECK-NEXT: maxsd %xmm1, %xmm0
22 ; UNSAFE-NEXT: maxsd %xmm1, %xmm0
25 ; FINITE-NEXT: maxsd %xmm1, %xmm0
34 ; CHECK-NEXT: minsd %xmm1, %xmm0
37 ; UNSAFE-NEXT: minsd %xmm1, %xmm0
40 ; FINITE-NEXT: minsd %xmm1, %xmm0
49 ; CHECK-NEXT: minsd %xmm0, %xmm1
50 ; CHECK-NEXT: movap{{[sd]}} %xmm1, %xmm0
53 ; UNSAFE-NEXT: minsd %xmm0, %xmm1
54 ; UNSAFE-NEXT: movap{{[sd]}} %xmm1, %xmm0
    [all...]
break-sse-dep.ll 7 ; CHECK: movss ([[A0:%rdi|%rcx]]), %xmm0
8 ; CHECK: cvtss2sd %xmm0, %xmm0
18 ; CHECK: cvtsd2ss ([[A0]]), %xmm0
27 ; CHECK: movss ([[A0]]), %xmm0
28 ; CHECK: sqrtss %xmm0, %xmm0
37 ; CHECK: movsd ([[A0]]), %xmm0
38 ; CHECK: sqrtsd %xmm0, %xmm0
    [all...]
peep-vector-extract-concat.ll 2 ; CHECK: pshufd $3, %xmm0, %xmm0
6 ; WIN64: movss 12(%rcx), %xmm0
  /external/libvpx/vp8/encoder/x86/
subtract_sse2.asm 95 movdqa xmm0, XMMWORD PTR [rsi] ; src
98 movdqa xmm2, xmm0
99 psubb xmm0, xmm1
105 movdqa xmm2, xmm0
107 punpcklbw xmm0, xmm1 ; put sign back to subtraction
110 movdqa XMMWORD PTR [rdi], xmm0
169 movq xmm0, MMWORD PTR [rsi] ; src
172 punpcklqdq xmm0, xmm2
174 movdqa xmm2, xmm0
175 psubb xmm0, xmm1 ; subtraction with sign misse
    [all...]
sad_ssse3.asm 16 movdqa xmm0, XMMWORD PTR [rsi]
21 psadbw xmm5, xmm0
22 psadbw xmm6, xmm0
23 psadbw xmm7, xmm0
25 movdqa xmm0, XMMWORD PTR [rsi]
30 psadbw xmm1, xmm0
31 psadbw xmm2, xmm0
32 psadbw xmm3, xmm0
38 movdqa xmm0, XMMWORD PTR [rsi+rax]
46 psadbw xmm1, xmm0
    [all...]
fwalsh_sse2.asm 31 movq xmm0, MMWORD PTR [rsi] ; load input
37 punpcklwd xmm0, xmm1
40 movdqa xmm1, xmm0
41 punpckldq xmm0, xmm2 ; ip[1] ip[0]
44 movdqa xmm2, xmm0
45 paddw xmm0, xmm1
48 psllw xmm0, 2 ; d1 a1
51 movdqa xmm1, xmm0
52 punpcklqdq xmm0, xmm2 ; b1 a1
56 movq xmm6, xmm0
    [all...]
sad_sse4.asm 16 movdqa xmm0, XMMWORD PTR [rsi]
24 mpsadbw xmm1, xmm0, 0x0
25 mpsadbw xmm2, xmm0, 0x5
27 psrldq xmm0, 8
30 mpsadbw xmm3, xmm0, 0x0
31 mpsadbw xmm4, xmm0, 0x5
37 movdqa xmm0, XMMWORD PTR [rsi]
45 mpsadbw xmm5, xmm0, 0x0
46 mpsadbw xmm2, xmm0, 0x5
48 psrldq xmm0,
    [all...]
temporal_filter_apply_sse2.asm 64 movd xmm0, arg(5) ; filter_weight
65 pshuflw xmm0, xmm0, 0
66 punpcklwd xmm0, xmm0
67 movdqa [rsp + filter_weight], xmm0
78 movq xmm0, [rsi] ; first row
80 punpcklbw xmm0, xmm7 ; src[ 0- 7]
87 movdqa xmm0, [rsi] ; src (frame1)
89 movdqa xmm1, xmm0
    [all...]
  /bionic/libc/arch-x86/string/
sse2-memset5-atom.S 283 /* Fill xmm0 with the pattern. */
285 pxor %xmm0, %xmm0
287 movd %eax, %xmm0
288 pshufd $0, %xmm0, %xmm0
294 movdqu %xmm0, (%edx)
300 movd %xmm0, %eax
350 movdqa %xmm0, (%edx)
351 movdqa %xmm0, 0x10(%edx
    [all...]
  /external/valgrind/main/memcheck/tests/amd64/
bug279698.c 11 ("movdqu (%0), %%xmm0 \n"
12 "packuswb %%xmm0, %%xmm0 \n"
13 "movdqu %%xmm0, 16(%0) \n"
15 :"memory","xmm0"
xor-undef-amd64.c 66 "movups 0(%0), %%xmm0\n\t"
68 "xorps %%xmm8, %%xmm0\n\t"
69 "movups %%xmm0, 32(%0)\n\t"
73 : : "r"(junk) : "rax", "xmm8", "xmm0", "cc", "memory"
78 "movups 0(%0), %%xmm0\n\t"
80 "xorps %%xmm0, %%xmm0\n\t"
81 "movups %%xmm0, 32(%0)\n\t"
85 : : "r"(junk) : "rax", "xmm8", "xmm0", "cc", "memory"
93 "movups 0(%0), %%xmm0\n\t
    [all...]
  /system/core/libcutils/arch-x86/
sse2-memset32-atom.S 230 pxor %xmm0, %xmm0
232 movd %eax, %xmm0
233 pshufd $0, %xmm0, %xmm0
239 movdqu %xmm0, (%edx)
245 movd %xmm0, %eax
295 movdqa %xmm0, (%edx)
296 movdqa %xmm0, 0x10(%edx)
297 movdqa %xmm0, 0x20(%edx
    [all...]
sse2-memset16-atom.S 314 pxor %xmm0, %xmm0
316 movd %eax, %xmm0
317 pshufd $0, %xmm0, %xmm0
323 movdqu %xmm0, (%edx)
329 movd %xmm0, %eax
381 movdqa %xmm0, (%edx)
382 movdqa %xmm0, 0x10(%edx)
383 movdqa %xmm0, 0x20(%edx
    [all...]
  /dalvik/vm/mterp/x86-atom/
binopF.S 20 * specify an instruction that performs "%xmm0 = %xmm0 op %xmm1"
35 movss (rFP, %ecx, 4), %xmm0 # %xmm0<-vBB
37 $instr # %xmm0<- vBB op vCC
38 movss %xmm0, (rFP, rINST, 4) # vAA<- %xmm0; result
  /external/compiler-rt/lib/x86_64/
floatundisf.S 23 cvtsi2ssq %rdi, %xmm0
29 cvtsi2ssq %rdi, %xmm0
30 mulss REL_ADDR(two), %xmm0
  /external/valgrind/main/memcheck/tests/x86/
xor-undef-x86.c 68 "movups 0(%0), %%xmm0\n\t"
70 "xorps %%xmm7, %%xmm0\n\t"
71 "movups %%xmm0, 32(%0)\n\t"
77 : : "r"(junk) : "esi", "xmm7", "xmm0", "cc", "memory"
82 "movups 0(%0), %%xmm0\n\t"
84 "xorps %%xmm0, %%xmm0\n\t"
85 "movups %%xmm0, 32(%0)\n\t"
91 : : "r"(junk) : "esi", "xmm7", "xmm0", "cc", "memory"
99 "movups 0(%0), %%xmm0\n\t
    [all...]
  /frameworks/compile/libbcc/runtime/lib/x86_64/
floatundisf.S 23 cvtsi2ssq %rdi, %xmm0
29 cvtsi2ssq %rdi, %xmm0
30 mulss REL_ADDR(two), %xmm0
  /external/libvpx/vp8/common/x86/
recon_sse2.asm 27 pxor xmm0, xmm0
30 punpcklbw xmm1, xmm0
32 packuswb xmm1, xmm0 ; pack and unpack to saturate
37 punpcklbw xmm2, xmm0
39 packuswb xmm2, xmm0 ; pack and unpack to saturate
44 punpcklbw xmm3, xmm0
46 packuswb xmm3, xmm0 ; pack and unpack to saturate
51 punpcklbw xmm4, xmm0
53 packuswb xmm4, xmm0 ; pack and unpack to saturat
    [all...]
  /external/openssl/crypto/aes/asm/
vpaes-x86.pl 169 ## AES-encrypt %xmm0.
172 ## %xmm0 = input
176 ## Output in %xmm0
185 &pandn ("xmm1","xmm0");
188 &pand ("xmm0","xmm6");
189 &pshufb ("xmm2","xmm0");
190 &movdqa ("xmm0",&QWP($k_ipt+16,$const));
191 &pshufb ("xmm0","xmm1");
193 &pxor ("xmm0","xmm2");
204 &movdqa ("xmm0",&QWP($k_sb1+16,$const));# 0 : sb1
    [all...]
  /frameworks/compile/libbcc/runtime/lib/i386/
floatdidf.S 23 movss 4(%esp), %xmm0 // low 32 bits of a
29 orpd %xmm2, %xmm0 // 0x1p52 + a_lo (no rounding occurs)
30 addsd %xmm1, %xmm0 // a_hi + a_lo (round happens here)
31 movsd %xmm0, 4(%esp)

Completed in 338 milliseconds

1 2 3 4 5 6 7 8 91011>>