/toolchain/binutils/binutils-2.25/gas/testsuite/gas/i386/ |
avx512er.s | 67 vrcp28ss %xmm4, %xmm5, %xmm6{%k7} # AVX512ER 68 vrcp28ss %xmm4, %xmm5, %xmm6{%k7}{z} # AVX512ER 69 vrcp28ss {sae}, %xmm4, %xmm5, %xmm6{%k7} # AVX512ER 70 vrcp28ss (%ecx), %xmm5, %xmm6{%k7} # AVX512ER 71 vrcp28ss -123456(%esp,%esi,8), %xmm5, %xmm6{%k7} # AVX512ER 72 vrcp28ss 508(%edx), %xmm5, %xmm6{%k7} # AVX512ER Disp8 73 vrcp28ss 512(%edx), %xmm5, %xmm6{%k7} # AVX512ER 74 vrcp28ss -512(%edx), %xmm5, %xmm6{%k7} # AVX512ER Disp8 75 vrcp28ss -516(%edx), %xmm5, %xmm6{%k7} # AVX512ER 77 vrcp28sd %xmm4, %xmm5, %xmm6{%k7} # AVX512E [all...] |
x86-64-avx-gather.s | 43 vgatherdps %xmm5,0x8(,%xmm4,1),%xmm6 44 vgatherdps %xmm5,-0x8(,%xmm4,1),%xmm6 45 vgatherdps %xmm5,(,%xmm4,1),%xmm6 46 vgatherdps %xmm5,0x298(,%xmm4,1),%xmm6 47 vgatherdps %xmm5,0x8(,%xmm4,8),%xmm6 48 vgatherdps %xmm5,-0x8(,%xmm4,8),%xmm6 49 vgatherdps %xmm5,(,%xmm4,8),%xmm6 50 vgatherdps %xmm5,0x298(,%xmm4,8),%xmm6 52 vgatherdps %xmm5,0x8(,%xmm14,1),%xmm6 53 vgatherdps %xmm5,-0x8(,%xmm14,1),%xmm [all...] |
avx512f_vl-opts-intel.d | 12 [ ]*[a-f0-9]+:[ ]*62 f1 fd 0f 28 f5[ ]*vmovapd xmm6\{k7\},xmm5 13 [ ]*[a-f0-9]+:[ ]*62 f1 fd 0f 29 ee[ ]*vmovapd\.s xmm6\{k7\},xmm5 14 [ ]*[a-f0-9]+:[ ]*62 f1 fd 8f 28 f5[ ]*vmovapd xmm6\{k7\}\{z\},xmm5 15 [ ]*[a-f0-9]+:[ ]*62 f1 fd 8f 29 ee[ ]*vmovapd\.s xmm6\{k7\}\{z\},xmm5 16 [ ]*[a-f0-9]+:[ ]*62 f1 fd 0f 28 f5[ ]*vmovapd xmm6\{k7\},xmm5 17 [ ]*[a-f0-9]+:[ ]*62 f1 fd 0f 29 ee[ ]*vmovapd\.s xmm6\{k7\},xmm5 18 [ ]*[a-f0-9]+:[ ]*62 f1 fd 8f 28 f5[ ]*vmovapd xmm6\{k7\}\{z\},xmm5 19 [ ]*[a-f0-9]+:[ ]*62 f1 fd 8f 29 ee[ ]*vmovapd\.s xmm6\{k7\}\{z\},xmm5 28 [ ]*[a-f0-9]+:[ ]*62 f1 7c 0f 28 f5[ ]*vmovaps xmm6\{k7\},xmm5 29 [ ]*[a-f0-9]+:[ ]*62 f1 7c 0f 29 ee[ ]*vmovaps\.s xmm6\{k7\},xmm5 [all...] |
avx512bw_vl.s | 6 vpabsb %xmm5, %xmm6{%k7} # AVX512{BW,VL} 7 vpabsb %xmm5, %xmm6{%k7}{z} # AVX512{BW,VL} 22 vpabsw %xmm5, %xmm6{%k7} # AVX512{BW,VL} 23 vpabsw %xmm5, %xmm6{%k7}{z} # AVX512{BW,VL} 38 vpackssdw %xmm4, %xmm5, %xmm6{%k7} # AVX512{BW,VL} 39 vpackssdw %xmm4, %xmm5, %xmm6{%k7}{z} # AVX512{BW,VL} 40 vpackssdw (%ecx), %xmm5, %xmm6{%k7} # AVX512{BW,VL} 41 vpackssdw -123456(%esp,%esi,8), %xmm5, %xmm6{%k7} # AVX512{BW,VL} 42 vpackssdw (%eax){1to4}, %xmm5, %xmm6{%k7} # AVX512{BW,VL} 43 vpackssdw 2032(%edx), %xmm5, %xmm6{%k7} # AVX512{BW,VL} Disp [all...] |
avx512f_vl.s | 6 vaddpd %xmm4, %xmm5, %xmm6{%k7} # AVX512{F,VL} 7 vaddpd %xmm4, %xmm5, %xmm6{%k7}{z} # AVX512{F,VL} 8 vaddpd (%ecx), %xmm5, %xmm6{%k7} # AVX512{F,VL} 9 vaddpd -123456(%esp,%esi,8), %xmm5, %xmm6{%k7} # AVX512{F,VL} 10 vaddpd (%eax){1to2}, %xmm5, %xmm6{%k7} # AVX512{F,VL} 11 vaddpd 2032(%edx), %xmm5, %xmm6{%k7} # AVX512{F,VL} Disp8 12 vaddpd 2048(%edx), %xmm5, %xmm6{%k7} # AVX512{F,VL} 13 vaddpd -2048(%edx), %xmm5, %xmm6{%k7} # AVX512{F,VL} Disp8 14 vaddpd -2064(%edx), %xmm5, %xmm6{%k7} # AVX512{F,VL} 15 vaddpd 1016(%edx){1to2}, %xmm5, %xmm6{%k7} # AVX512{F,VL} Disp [all...] |
katmai.s | 10 andnps %xmm6,%xmm5 17 cmpps $0x6,%xmm5,%xmm4 18 cmpps $0x7,(%esi),%xmm5 25 cmpltps %xmm5,%xmm4 26 cmpltps (%esi),%xmm5 34 cmpunordps %xmm6,%xmm5 41 cmpnltps %xmm5,%xmm4 42 cmpnltps (%esi),%xmm5 50 cmpordps %xmm6,%xmm5 58 cvtsi2ss (%esi),%xmm5 [all...] |
avx512f.s | 45 vaddsd %xmm4, %xmm5, %xmm6{%k7} # AVX512F 46 vaddsd %xmm4, %xmm5, %xmm6{%k7}{z} # AVX512F 47 vaddsd {rn-sae}, %xmm4, %xmm5, %xmm6{%k7} # AVX512F 48 vaddsd {ru-sae}, %xmm4, %xmm5, %xmm6{%k7} # AVX512F 49 vaddsd {rd-sae}, %xmm4, %xmm5, %xmm6{%k7} # AVX512F 50 vaddsd {rz-sae}, %xmm4, %xmm5, %xmm6{%k7} # AVX512F 51 vaddsd (%ecx), %xmm5, %xmm6{%k7} # AVX512F 52 vaddsd -123456(%esp,%esi,8), %xmm5, %xmm6{%k7} # AVX512F 53 vaddsd 1016(%edx), %xmm5, %xmm6{%k7} # AVX512F Disp8 54 vaddsd 1024(%edx), %xmm5, %xmm6{%k7} # AVX512 [all...] |
sse2.s | 11 andnpd %xmm6,%xmm5 18 cmppd $0x6,%xmm5,%xmm4 19 cmppd $0x7,(%esi),%xmm5 26 cmpltpd %xmm5,%xmm4 27 cmpltpd (%esi),%xmm5 35 cmpunordpd %xmm6,%xmm5 42 cmpnltpd %xmm5,%xmm4 43 cmpnltpd (%esi),%xmm5 51 cmpordpd %xmm6,%xmm5 59 cvtsi2sd (%esi),%xmm5 [all...] |
avx512vbmi_vl-intel.d | 12 [ ]*[a-f0-9]+:[ ]*62 f2 55 0f 8d f4[ ]*vpermb xmm6\{k7\},xmm5,xmm4 13 [ ]*[a-f0-9]+:[ ]*62 f2 55 8f 8d f4[ ]*vpermb xmm6\{k7\}\{z\},xmm5,xmm4 14 [ ]*[a-f0-9]+:[ ]*62 f2 55 0f 8d 31[ ]*vpermb xmm6\{k7\},xmm5,XMMWORD PTR \[ecx\] 15 [ ]*[a-f0-9]+:[ ]*62 f2 55 0f 8d b4 f4 c0 1d fe ff[ ]*vpermb xmm6\{k7\},xmm5,XMMWORD PTR \[esp\+esi\*8-0x1e240\] 16 [ ]*[a-f0-9]+:[ ]*62 f2 55 0f 8d 72 7f[ ]*vpermb xmm6\{k7\},xmm5,XMMWORD PTR \[edx\+0x7f0\] 17 [ ]*[a-f0-9]+:[ ]*62 f2 55 0f 8d b2 00 08 00 00[ ]*vpermb xmm6\{k7\},xmm5,XMMWORD PTR \[edx\+0x800\] 18 [ ]*[a-f0-9]+:[ ]*62 f2 55 0f 8d 72 80[ ]*vpermb xmm6\{k7\},xmm5,XMMWORD PTR \[edx-0x800\] 19 [ ]*[a-f0-9]+:[ ]*62 f2 55 0f 8d b2 f0 f7 ff ff[ ]*vpermb xmm6\{k7\},xmm5,XMMWORD PTR \[edx-0x810\] 28 [ ]*[a-f0-9]+:[ ]*62 f2 55 0f 75 f4[ ]*vpermi2b xmm6\{k7\},xmm5,xmm4 29 [ ]*[a-f0-9]+:[ ]*62 f2 55 8f 75 f4[ ]*vpermi2b xmm6\{k7\}\{z\},xmm5,xmm [all...] |
evex-wig.s | 7 vpmovsxbd %xmm5, %zmm6{%k7} # AVX512 8 vpmovsxbd %xmm5, %zmm6{%k7}{z} # AVX512 16 vpmovsxbq %xmm5, %zmm6{%k7} # AVX512 17 vpmovsxbq %xmm5, %zmm6{%k7}{z} # AVX512 34 vpmovsxwq %xmm5, %zmm6{%k7} # AVX512 35 vpmovsxwq %xmm5, %zmm6{%k7}{z} # AVX512 43 vpmovzxbd %xmm5, %zmm6{%k7} # AVX512 44 vpmovzxbd %xmm5, %zmm6{%k7}{z} # AVX512 52 vpmovzxbq %xmm5, %zmm6{%k7} # AVX512 53 vpmovzxbq %xmm5, %zmm6{%k7}{z} # AVX51 [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/x86/ |
sad_sse4.asm | 38 movq xmm5, MMWORD PTR [rdi] 41 punpcklqdq xmm5, xmm3 44 movdqa xmm2, xmm5 45 mpsadbw xmm5, xmm0, 0x0 54 paddw xmm5, xmm2 55 paddw xmm5, xmm3 56 paddw xmm5, xmm4 58 paddw xmm1, xmm5 61 movq xmm5, MMWORD PTR [rdi+ rdx] 64 punpcklqdq xmm5, xmm [all...] |
idctllm_sse2.asm | 34 movd xmm5, [rdx] 37 pinsrw xmm5, [rdx], 4 39 pmullw xmm4, xmm5 41 ; Zero out xmm5, for use unpacking 42 pxor xmm5, xmm5 45 movd [rax], xmm5 46 movd [rax+32], xmm5 64 punpcklbw xmm0, xmm5 65 punpcklbw xmm1, xmm5 [all...] |
iwalsh_sse2.asm | 65 movdqa xmm5, xmm4 67 punpckhqdq xmm5, xmm3 ;c1 b1 69 movdqa xmm1, xmm5 ;c1 b1 70 paddw xmm5, xmm4 ;dl+cl a1+b1 aka op[4] op[0] 73 paddw xmm5, xmm0 75 psraw xmm5, 3 78 movd eax, xmm5 80 psrldq xmm5, 4 88 movd eax, xmm5 90 psrldq xmm5, [all...] |
/external/libjpeg-turbo/simd/ |
jidctint-sse2-64.asm | 147 movdqa xmm5, XMMWORD [XMMBLOCK(0,0,rsi,SIZEOF_JCOEF)] 148 pmullw xmm5, XMMWORD [XMMBLOCK(0,0,rdx,SIZEOF_ISLOW_MULT_TYPE)] 150 psllw xmm5,PASS1_BITS 152 movdqa xmm4,xmm5 ; xmm5=in0=(00 01 02 03 04 05 06 07) 153 punpcklwd xmm5,xmm5 ; xmm5=(00 00 01 01 02 02 03 03) 156 pshufd xmm7,xmm5,0x00 ; xmm7=col0=(00 00 00 00 00 00 00 00) 157 pshufd xmm6,xmm5,0x55 ; xmm6=col1=(01 01 01 01 01 01 01 01 [all...] |
jfdctint-sse2-64.asm | 131 movdqa xmm5,xmm2 ; transpose coefficients(phase 1) 133 punpckhwd xmm5,xmm3 ; xmm5=(24 34 25 35 26 36 27 37) 144 movdqa XMMWORD [wk(1)], xmm5 ; wk(1)=(24 34 25 35 26 36 27 37) 149 movdqa xmm5,xmm1 ; transpose coefficients(phase 1) 151 punpckhwd xmm5,xmm3 ; xmm5=(64 74 65 75 66 76 67 77) 157 punpckldq xmm2,xmm5 ; xmm2=(44 54 64 74 45 55 65 75) 158 punpckhdq xmm3,xmm5 ; xmm3=(46 56 66 76 47 57 67 77) 161 movdqa xmm5, XMMWORD [wk(1)] ; xmm5=(24 34 25 35 26 36 27 37 [all...] |
jfdctint-sse2.asm | 137 movdqa xmm5,xmm2 ; transpose coefficients(phase 1) 139 punpckhwd xmm5,xmm3 ; xmm5=(24 34 25 35 26 36 27 37) 150 movdqa XMMWORD [wk(1)], xmm5 ; wk(1)=(24 34 25 35 26 36 27 37) 155 movdqa xmm5,xmm1 ; transpose coefficients(phase 1) 157 punpckhwd xmm5,xmm3 ; xmm5=(64 74 65 75 66 76 67 77) 163 punpckldq xmm2,xmm5 ; xmm2=(44 54 64 74 45 55 65 75) 164 punpckhdq xmm3,xmm5 ; xmm3=(46 56 66 76 47 57 67 77) 167 movdqa xmm5, XMMWORD [wk(1)] ; xmm5=(24 34 25 35 26 36 27 37 [all...] |
jidctint-sse2.asm | 153 movdqa xmm5, XMMWORD [XMMBLOCK(0,0,esi,SIZEOF_JCOEF)] 154 pmullw xmm5, XMMWORD [XMMBLOCK(0,0,edx,SIZEOF_ISLOW_MULT_TYPE)] 156 psllw xmm5,PASS1_BITS 158 movdqa xmm4,xmm5 ; xmm5=in0=(00 01 02 03 04 05 06 07) 159 punpcklwd xmm5,xmm5 ; xmm5=(00 00 01 01 02 02 03 03) 162 pshufd xmm7,xmm5,0x00 ; xmm7=col0=(00 00 00 00 00 00 00 00) 163 pshufd xmm6,xmm5,0x55 ; xmm6=col1=(01 01 01 01 01 01 01 01 [all...] |
jidctfst-sse2-64.asm | 143 pshufd xmm5,xmm0,0xAA ; xmm5=col2=(02 02 02 02 02 02 02 02) 168 movdqa xmm5,xmm1 172 paddw xmm5,xmm3 ; xmm5=tmp13 176 psubw xmm1,xmm5 ; xmm1=tmp12 180 psubw xmm4,xmm5 ; xmm4=tmp3 182 paddw xmm6,xmm5 ; xmm6=tmp0 194 movdqa xmm5, XMMWORD [XMMBLOCK(5,0,rsi,SIZEOF_JCOEF)] 196 pmullw xmm5, XMMWORD [XMMBLOCK(5,0,rdx,SIZEOF_IFAST_MULT_TYPE) [all...] |
jidctfst-sse2.asm | 149 pshufd xmm5,xmm0,0xAA ; xmm5=col2=(02 02 02 02 02 02 02 02) 175 movdqa xmm5,xmm1 179 paddw xmm5,xmm3 ; xmm5=tmp13 183 psubw xmm1,xmm5 ; xmm1=tmp12 187 psubw xmm4,xmm5 ; xmm4=tmp3 189 paddw xmm6,xmm5 ; xmm6=tmp0 201 movdqa xmm5, XMMWORD [XMMBLOCK(5,0,esi,SIZEOF_JCOEF)] 203 pmullw xmm5, XMMWORD [XMMBLOCK(5,0,edx,SIZEOF_IFAST_MULT_TYPE) [all...] |
jfdctfst-sse2-64.asm | 110 movdqa xmm5,xmm2 ; transpose coefficients(phase 1) 112 punpckhwd xmm5,xmm3 ; xmm5=(24 34 25 35 26 36 27 37) 123 movdqa XMMWORD [wk(1)], xmm5 ; wk(1)=(24 34 25 35 26 36 27 37) 128 movdqa xmm5,xmm1 ; transpose coefficients(phase 1) 130 punpckhwd xmm5,xmm3 ; xmm5=(64 74 65 75 66 76 67 77) 136 punpckldq xmm2,xmm5 ; xmm2=(44 54 64 74 45 55 65 75) 137 punpckhdq xmm3,xmm5 ; xmm3=(46 56 66 76 47 57 67 77) 140 movdqa xmm5, XMMWORD [wk(1)] ; xmm5=(24 34 25 35 26 36 27 37 [all...] |
/external/libvpx/libvpx/vp8/common/x86/ |
idctllm_sse2.asm | 34 movd xmm5, [rdx] 37 pinsrw xmm5, [rdx], 4 39 pmullw xmm4, xmm5 41 ; Zero out xmm5, for use unpacking 42 pxor xmm5, xmm5 45 movd [rax], xmm5 46 movd [rax+32], xmm5 64 punpcklbw xmm0, xmm5 65 punpcklbw xmm1, xmm5 [all...] |
iwalsh_sse2.asm | 65 movdqa xmm5, xmm4 67 punpckhqdq xmm5, xmm3 ;c1 b1 69 movdqa xmm1, xmm5 ;c1 b1 70 paddw xmm5, xmm4 ;dl+cl a1+b1 aka op[4] op[0] 73 paddw xmm5, xmm0 75 psraw xmm5, 3 78 movd eax, xmm5 80 psrldq xmm5, 4 88 movd eax, xmm5 90 psrldq xmm5, [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/encoder/x86/ |
quantize_ssse3.asm | 63 movdqa xmm5, xmm4 70 pabsw xmm5, xmm5 73 paddw xmm5, xmm3 77 pmulhw xmm5, [rdx + 16] 84 movdqa xmm3, xmm5 87 pxor xmm5, xmm4 89 psubw xmm5, xmm4 92 movdqa [rax + 16], xmm5 98 pmullw xmm4, xmm5 [all...] |
quantize_sse4.asm | 85 movdqa xmm5, [rcx + 16] 89 paddw xmm5, xmm7 96 psubw xmm7, xmm5 100 movdqa xmm5, [rdx + 16] 108 paddw xmm3, xmm5 112 movdqa xmm5, [rcx + 16] 116 pmulhw xmm5, xmm3 120 paddw xmm3, xmm5 131 movdqa xmm5, [rax] 148 pextrb ecx, xmm5, %1 ; quant_shift[rc [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/encoder/x86/ |
vp9_variance_impl_sse2.asm | 167 pxor xmm5, xmm5 170 punpckhwd xmm5, xmm7 171 psrad xmm5, 16 174 paddd xmm6, xmm5 437 movq xmm5, QWORD PTR [rsi] ; xmm5 = s0,s1,s2..s8 439 pavgb xmm5, xmm3 ; xmm5 = avg(xmm1,xmm3) horizontal line 1 453 pavgb xmm5, xmm1 ; xmm = vertical average of the abov [all...] |