HomeSort by relevance Sort by last modified time
    Searched full:xmm2 (Results 76 - 100 of 435) sorted by null

1 2 34 5 6 7 8 91011>>

  /external/boringssl/win-x86_64/crypto/aes/
aesni-x86_64.asm 12 movups xmm2,XMMWORD[rcx]
17 xorps xmm2,xmm0
27 movups XMMWORD[rdx],xmm2
28 pxor xmm2,xmm2
36 movups xmm2,XMMWORD[rcx]
41 xorps xmm2,xmm0
51 movups XMMWORD[rdx],xmm2
52 pxor xmm2,xmm2
    [all...]
  /external/llvm/test/CodeGen/X86/
vec_cast2.ll 94 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15]
95 ; CHECK-NEXT: vpshufb %xmm2, %xmm1, %xmm1
96 ; CHECK-NEXT: vpshufb %xmm2, %xmm0, %xmm0
105 ; CHECK-WIDE-NEXT: vmovshdup %xmm0, %xmm2 ## xmm2 = xmm0[1,1,3,3]
106 ; CHECK-WIDE-NEXT: vcvttss2si %xmm2, %eax
108 ; CHECK-WIDE-NEXT: vpermilpd $1, %xmm0, %xmm2 ## xmm2 = xmm0[1,0]
109 ; CHECK-WIDE-NEXT: vcvttss2si %xmm2, %eax
111 ; CHECK-WIDE-NEXT: vpermilps $231, %xmm0, %xmm2 ## xmm2 = xmm0[3,1,2,3
    [all...]
illegal-vector-args-return.ll 2 ; RUN: llc < %s -march=x86 -mattr=+sse2 -mcpu=nehalem | grep "mulpd %xmm2, %xmm0"
4 ; RUN: llc < %s -march=x86 -mattr=+sse2 -mcpu=nehalem | grep "addps %xmm2, %xmm0"
fma_patterns.ll 6 ; CHECK: vfmadd213ps %xmm2, %xmm1, %xmm0
9 ; CHECK_FMA4: vfmaddps %xmm2, %xmm1, %xmm0, %xmm0
18 ; CHECK: fmsub213ps %xmm2, %xmm1, %xmm0
21 ; CHECK_FMA4: vfmsubps %xmm2, %xmm1, %xmm0, %xmm0
30 ; CHECK: fnmadd213ps %xmm2, %xmm1, %xmm0
33 ; CHECK_FMA4: vfnmaddps %xmm2, %xmm1, %xmm0, %xmm0
42 ; CHECK: fnmsub213ps %xmm2, %xmm1, %xmm0
45 ; CHECK_FMA4: fnmsubps %xmm2, %xmm1, %xmm0, %xmm0
125 ; CHECK: vfmsub213pd %xmm2, %xmm1, %xmm0
128 ; CHECK_FMA4: vfmsubpd %xmm2, %xmm1, %xmm0, %xmm
    [all...]
avx-cvt-2.ll 14 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15]
15 ; CHECK-NEXT: vpshufb %xmm2, %xmm1, %xmm1
16 ; CHECK-NEXT: vpshufb %xmm2, %xmm0, %xmm0
31 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15]
32 ; CHECK-NEXT: vpshufb %xmm2, %xmm1, %xmm1
33 ; CHECK-NEXT: vpshufb %xmm2, %xmm0, %xmm0
48 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15]
49 ; CHECK-NEXT: vpshufb %xmm2, %xmm1, %xmm1
50 ; CHECK-NEXT: vpshufb %xmm2, %xmm0, %xmm0
66 ; CHECK-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,4,5,8,9,12,13,8,9,12,13,12,13,14,15
    [all...]
vector-shuffle-256-v16.ll 316 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,1,0,1,0,1,0,1,0,1,0,1,0,1,0,1]
317 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1
318 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
334 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [14,15,14,15,14,15,14,15,14,15,14,15,14,15,14,15]
335 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1
336 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
543 ; AVX1-NEXT: vextractf128 $1, %ymm1, %xmm2
545 ; AVX1-NEXT: vpblendw {{.*#+}} xmm2 = xmm3[0],xmm2[1],xmm3[2],xmm2[3],xmm3[4],xmm2[5],xmm3[6],xmm2[7
    [all...]
  /external/libvpx/libvpx/vp9/common/x86/
vp9_postproc_sse2.asm 48 movd xmm2, dword ptr arg(6) ;flimit
49 punpcklwd xmm2, xmm2
50 punpckldq xmm2, xmm2
51 punpcklqdq xmm2, xmm2
78 pcmpgtw xmm7, xmm2
89 pcmpgtw xmm6, xmm2
103 pcmpgtw xmm6, xmm2
    [all...]
  /external/llvm/test/MC/Disassembler/X86/
marked-up.txt 5 # CHECK: xorps <reg:%xmm1>, <reg:%xmm2>
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/x86/
vp9_postproc_sse2.asm 48 movd xmm2, dword ptr arg(6) ;flimit
49 punpcklwd xmm2, xmm2
50 punpckldq xmm2, xmm2
51 punpcklqdq xmm2, xmm2
78 pcmpgtw xmm7, xmm2
89 pcmpgtw xmm6, xmm2
103 pcmpgtw xmm6, xmm2
    [all...]
  /external/compiler-rt/lib/builtins/i386/
floatdidf.S 30 movsd REL_ADDR(twop52), %xmm2 // 0x1.0p52
31 subsd %xmm2, %xmm1 // a_hi - 0x1p52 (no rounding occurs)
32 orpd %xmm2, %xmm0 // 0x1p52 + a_lo (no rounding occurs)
ashrdi3.S 14 movd 12(%esp), %xmm2 // Load count
24 psrlq %xmm2, %xmm0 // unsigned shift input by count
33 pandn %xmm1, %xmm2 // 63 - count
35 psubq %xmm1, %xmm2 // 64 - count
36 psllq %xmm2, %xmm1 // -1 << (64 - count) = leading sign bits
  /external/libvpx/libvpx/vp8/encoder/x86/
temporal_filter_apply_sse2.asm 94 movdqa xmm2, [rdx] ; predictor (frame2)
95 movdqa xmm3, xmm2
96 punpcklbw xmm2, xmm7 ; pred[ 0- 7]
100 psubw xmm0, xmm2 ; src - pred[ 0- 7]
122 movdqa xmm2, [GLOBAL(_const_16w)]
124 psubusw xmm2, xmm0
127 pmullw xmm2, [rsp + filter_weight]
134 paddw xmm4, xmm2
150 pmullw xmm0, xmm2
154 movdqa xmm2, xmm
    [all...]
  /external/libvpx/libvpx/vp9/encoder/x86/
vp9_temporal_filter_apply_sse2.asm 99 movdqa xmm2, [rdx] ; predictor (frame2)
100 movdqa xmm3, xmm2
101 punpcklbw xmm2, xmm7 ; pred[ 0- 7]
105 psubw xmm0, xmm2 ; src - pred[ 0- 7]
127 movdqa xmm2, [GLOBAL(_const_16w)]
129 psubusw xmm2, xmm0
132 pmullw xmm2, [rsp + filter_weight]
139 paddw xmm4, xmm2
155 pmullw xmm0, xmm2
159 movdqa xmm2, xmm
    [all...]
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/encoder/x86/
temporal_filter_apply_sse2.asm 94 movdqa xmm2, [rdx] ; predictor (frame2)
95 movdqa xmm3, xmm2
96 punpcklbw xmm2, xmm7 ; pred[ 0- 7]
100 psubw xmm0, xmm2 ; src - pred[ 0- 7]
122 movdqa xmm2, [GLOBAL(_const_16w)]
124 psubusw xmm2, xmm0
127 pmullw xmm2, [rsp + filter_weight]
134 paddw xmm4, xmm2
150 pmullw xmm0, xmm2
154 movdqa xmm2, xmm
    [all...]
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/encoder/x86/
vp9_temporal_filter_apply_sse2.asm 94 movdqa xmm2, [rdx] ; predictor (frame2)
95 movdqa xmm3, xmm2
96 punpcklbw xmm2, xmm7 ; pred[ 0- 7]
100 psubw xmm0, xmm2 ; src - pred[ 0- 7]
122 movdqa xmm2, [GLOBAL(_const_16w)]
124 psubusw xmm2, xmm0
127 pmullw xmm2, [rsp + filter_weight]
134 paddw xmm4, xmm2
150 pmullw xmm0, xmm2
154 movdqa xmm2, xmm
    [all...]
  /external/libvpx/libvpx/third_party/libyuv/source/
scale_posix.cc 136 "movdqa %%xmm0,%%xmm2 \n"
140 "pand %%xmm5,%%xmm2 \n"
142 "pavgw %%xmm2,%%xmm0 \n"
170 MEMOPREG(movdqa,0x00,0,3,1,xmm2) // movdqa (%0,%3,1),%%xmm2
174 "pavgb %%xmm2,%%xmm0 \n"
176 "movdqa %%xmm0,%%xmm2 \n"
180 "pand %%xmm5,%%xmm2 \n"
182 "pavgw %%xmm2,%%xmm0 \n"
198 , "xmm0", "xmm1", "xmm2", "xmm3", "xmm5
    [all...]
  /external/boringssl/linux-x86/crypto/aes/
vpaes-x86.S 77 movdqa (%ebp),%xmm2
83 pxor %xmm5,%xmm2
88 pxor %xmm2,%xmm0
101 movdqa 80(%ebp),%xmm2
105 pxor %xmm5,%xmm2
108 pxor %xmm2,%xmm0
129 movdqa %xmm7,%xmm2
133 pxor %xmm0,%xmm2
155 movdqa -64(%ebx),%xmm2
167 pxor %xmm5,%xmm2
    [all...]
  /external/boringssl/mac-x86/crypto/aes/
vpaes-x86.S 74 movdqa (%ebp),%xmm2
80 pxor %xmm5,%xmm2
85 pxor %xmm2,%xmm0
98 movdqa 80(%ebp),%xmm2
102 pxor %xmm5,%xmm2
105 pxor %xmm2,%xmm0
126 movdqa %xmm7,%xmm2
130 pxor %xmm0,%xmm2
150 movdqa -64(%ebx),%xmm2
162 pxor %xmm5,%xmm2
    [all...]
  /external/boringssl/win-x86/crypto/aes/
vpaes-x86.asm 85 movdqa xmm2,[ebp]
91 pxor xmm2,xmm5
96 pxor xmm0,xmm2
109 movdqa xmm2,[80+ebp]
113 pxor xmm2,xmm5
116 pxor xmm0,xmm2
137 movdqa xmm2,xmm7
141 pxor xmm2,xmm0
160 movdqa xmm2,[ebx-64]
172 pxor xmm2,xmm
    [all...]
  /bionic/libc/arch-x86/atom/string/
sse2-wcslen-atom.S 100 pxor %xmm2, %xmm2
105 pcmpeqd (%eax), %xmm2
106 pmovmskb %xmm2, %edx
130 pcmpeqd (%eax), %xmm2
131 pmovmskb %xmm2, %edx
154 pcmpeqd (%eax), %xmm2
155 pmovmskb %xmm2, %edx
178 pcmpeqd (%eax), %xmm2
179 pmovmskb %xmm2, %ed
    [all...]
  /external/valgrind/VEX/test/
fxsave.c 39 asm __volatile__("movups vecZ, %xmm2");
63 asm __volatile__("xorps %xmm2, %xmm2");
64 asm __volatile__("movaps %xmm2, %xmm3");
65 asm __volatile__("movaps %xmm2, %xmm4");
66 asm __volatile__("movaps %xmm2, %xmm5");
67 asm __volatile__("movaps %xmm2, %xmm6");
  /external/valgrind/memcheck/tests/x86/
fxsave.c 40 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm2");
64 asm __volatile__("xorps %xmm2, %xmm2");
65 asm __volatile__("movaps %xmm2, %xmm3");
66 asm __volatile__("movaps %xmm2, %xmm4");
67 asm __volatile__("movaps %xmm2, %xmm5");
68 asm __volatile__("movaps %xmm2, %xmm6");
  /bionic/libc/arch-x86/silvermont/string/
sse2-strlen-slm.S 116 pxor %xmm2, %xmm2
128 pcmpeqb 48(%eax), %xmm2
129 pmovmskb %xmm2, %ecx
149 pcmpeqb 48(%eax), %xmm2
150 pmovmskb %xmm2, %ecx
170 pcmpeqb 48(%eax), %xmm2
171 pmovmskb %xmm2, %ecx
191 pcmpeqb 48(%eax), %xmm2
192 pmovmskb %xmm2, %ec
    [all...]
  /bionic/libc/arch-x86_64/string/
sse2-strlen-slm.S 95 pxor %xmm2, %xmm2
107 pcmpeqb 48(%rax), %xmm2
108 pmovmskb %xmm2, %edx
128 pcmpeqb 48(%rax), %xmm2
129 pmovmskb %xmm2, %edx
149 pcmpeqb 48(%rax), %xmm2
150 pmovmskb %xmm2, %edx
170 pcmpeqb 48(%rax), %xmm2
171 pmovmskb %xmm2, %ed
    [all...]
  /external/libvpx/libvpx/vp8/common/x86/
sad_ssse3.asm 27 lddqu xmm2, XMMWORD PTR [rdi+1]
31 psadbw xmm2, xmm0
35 paddw xmm6, xmm2
40 lddqu xmm2, XMMWORD PTR [rdi+rdx+1]
47 psadbw xmm2, xmm0
51 paddw xmm6, xmm2
80 movdqa xmm2, xmm3
81 palignr xmm2, xmm4, (%2+1)
86 psadbw xmm2, xmm0
90 paddw xmm6, xmm2
    [all...]

Completed in 431 milliseconds

1 2 34 5 6 7 8 91011>>