/external/valgrind/main/none/tests/amd64/ |
redundantRexW.stdout.exp | 9 %xmm 0 ................................ 10 %xmm 1 .2.6.6.2.527faf9.8.8.8.f.1.3.1.7 11 %xmm 2 ................................ 12 %xmm 3 ................................ 13 %xmm 4 ................................ 14 %xmm 5 ................................ 15 %xmm 6 ................................ 16 %xmm 7 ................................ 17 %xmm 8 ................................ 18 %xmm 9 ............................... [all...] |
gen_insn_test.pl | 13 xmm => "reg128_t", 219 # my @xmmregs = map { "xmm$_" } (4,5,0,1,2,3,6,7); 220 my @xmmregs = map { "xmm$_" } (12,13,8,9,10,11,14,15); 354 if ($arg =~ /^([abcd]l|[abcd]x|e[abcd]x|r[abcd]x|r8|r16|r32|r64|mm|xmm|m8|m16|m32|m64|m128)\.(sb|ub|sw|uw|sd|ud|sq|uq|ps|pd)\[([^\]]+)\]$/) 463 elsif ($arg->{type} =~ /^(xmm|m128)$/) 631 if ($result->{type} eq "xmm") 646 if ($arg->{type} eq "xmm") 695 elsif ($arg->{type} eq "xmm") 752 if ($arg->{type} =~ /^(r8|r16|r32|r64|mm|xmm)$/) 809 elsif ($result->{type} eq "xmm") [all...] |
/external/openssl/crypto/perlasm/ |
x86asm.pl | 88 if ("$dst:$src" =~ /(e[a-dsd][ixp]):xmm([0-7])/) 96 if ("$dst:$src" =~ /xmm([0-7]):(e[a-dsd][ixp])/) 104 if ("$dst:$src" =~ /xmm([0-7]):xmm([0-7])/) 112 if ("$dst:$src" =~ /xmm([0-7]):xmm([0-7])/) 120 if ("$dst:$src" =~ /xmm([0-7]):xmm([0-7])/)
|
x86_64-xlate.pl | 125 } elsif ($self->{op} =~ /movq/ && $line =~ /%xmm/) { 669 if ($arg =~ /%xmm([0-9]+),\s*%r(\w+)/) { 676 } elsif ($arg =~ /%r(\w+),\s*%xmm([0-9]+)/) { 689 if (shift =~ /\$([0-9]+),\s*%xmm([0-9]+),\s*(%\w+)/) { 707 if (shift =~ /\$([0-9]+),\s*(%\w+),\s*%xmm([0-9]+)/) { 725 if (shift =~ /%xmm([0-9]+),\s*%xmm([0-9]+)/) { 737 if (shift =~ /\$([0-9]+),\s*%xmm([0-9]+),\s*%xmm([0-9]+)/) { 750 if (shift =~ /\$([x0-9a-f]+),\s*%xmm([0-9]+),\s*%xmm([0-9]+)/) [all...] |
/external/openssl/crypto/rc4/asm/ |
rc4-md5-x86_64.pl | 219 my $xmm="%xmm".($j&1); 222 $code.=" pxor $xmm,$xmm\n" if ($rc4 && $j<=1); 239 #rc4# pinsrw \$`($j>>1)&7`,($dat,$TX[0],4),$xmm\n 259 my $xmm="%xmm".($j&1); 262 $code.=" pxor $xmm,$xmm\n" if ($rc4 && $j<=1); 279 #rc4# pinsrw \$`($j>>1)&7`,($dat,$TX[0],4),$xmm\ [all...] |
rc4-x86_64.pl | 257 my $xmm="%xmm".($j&1); 265 $code.=" pxor $xmm,$xmm\n" if ($i<=1); 274 $code.=" pinsrw \$`($j>>1)&7`,($dat,$TX[0],4),$xmm\n";
|
/external/valgrind/main/VEX/priv/ |
guest_amd64_helpers.c | 1627 U128* xmm = (U128*)(addr + 160); local 1731 U128* xmm = (U128*)(addr + 160); local [all...] |
guest_x86_helpers.c | 1672 U128* xmm = (U128*)(addr + 160); local 1764 U128* xmm = (U128*)(addr + 160); local [all...] |
/external/valgrind/main/none/tests/x86/ |
gen_insn_test.pl | 12 xmm => "reg128_t", 197 my @xmmregs = map { "xmm$_" } (4,5,0,1,2,3,6,7); 331 if ($arg =~ /^([abcd]l|[abcd]x|e[abcd]x|r8|r16|r32|mm|xmm|m8|m16|m32|m64|m128)\.(sb|ub|sw|uw|sd|ud|sq|uq|ps|pd)\[([^\]]+)\]$/) 440 elsif ($arg->{type} =~ /^(xmm|m128)$/) 608 if ($result->{type} eq "xmm") 623 if ($arg->{type} eq "xmm") 664 elsif ($arg->{type} eq "xmm") 720 if ($arg->{type} =~ /^(r8|r16|r32|mm|xmm)$/) 773 elsif ($result->{type} eq "xmm") 825 if ($result->{type} eq "xmm") [all...] |
/prebuilts/gcc/linux-x86/host/i686-linux-glibc2.7-4.4.3/sysroot/usr/include/asm/ |
kvm.h | 127 __u8 xmm[16][16]; member in struct:kvm_fpu
|
/prebuilts/gcc/linux-x86/host/i686-linux-glibc2.7-4.6/sysroot/usr/include/asm/ |
kvm.h | 127 __u8 xmm[16][16]; member in struct:kvm_fpu
|
/prebuilts/gcc/linux-x86/host/x86_64-linux-glibc2.7-4.6/sysroot/usr/include/asm/ |
kvm.h | 127 __u8 xmm[16][16]; member in struct:kvm_fpu
|
/external/openssl/crypto/sha/asm/ |
sha1-586.pl | 56 # a) XMM units became faster and wider; 107 $xmm=$ymm=0; 108 for (@ARGV) { $xmm=1 if (/-DOPENSSL_IA32_SSE2/); } 110 $ymm=1 if ($xmm && 115 $ymm=1 if ($xmm && !$ymm && $ARGV[0] eq "win32n" && 119 &external_label("OPENSSL_ia32cap_P") if ($xmm); 297 if ($xmm) { 391 if ($xmm) { 395 # %xmm[0-7] are used as ring @X[] buffer containing quadruples of last 403 # no %xmm registers left to spare. For this reason part of the rin [all...] |
/external/qemu/android/config/linux-x86/asm/ |
kvm.h | 138 __u8 xmm[16][16]; member in struct:kvm_fpu
|
/external/qemu/android/config/linux-x86_64/asm/ |
kvm.h | 138 __u8 xmm[16][16]; member in struct:kvm_fpu
|
/external/openssl/crypto/aes/asm/ |
aesni-sha1-x86_64.pl | 110 my @X=map("%xmm$_",(4..7,0..3)); 111 my @Tx=map("%xmm$_",(8..10)); 116 my ($iv,$in,$rndkey0)=map("%xmm$_",(11..13)); 185 movdqu 0($inp),@X[-4&7] # load input to %xmm[0-3] 639 my @X=map("%xmm$_",(4..7,0..3)); 640 my @Tx=map("%xmm$_",(8..10)); 705 vmovdqu 0($inp),@X[-4&7] # load input to %xmm[0-3] 1232 if ($line=~/(aes[a-z]+)\s+%xmm([0-9]+),\s*%xmm([0-9]+)/) { 1246 $code =~ s/\b(aes.*%xmm[0-9]+).*$/aesni($1)/gem [all...] |
aesni-x86.pl | 80 if ("$dst:$src" =~ /xmm([0-7]):xmm([0-7])/) 85 if ("$dst:$src" =~ /xmm([0-7]):xmm([0-7])/) 205 # in XMM registers addreassable in 32-bit mode and therefore 6x is [all...] |
aesni-x86_64.pl | 194 # %xmm register layout 1016 my ($in0,$in1,$in2,$in3)=map("%xmm$_",(8..11)); [all...] |
/external/qemu/target-i386/ |
kvm.c | 381 memcpy(fpu.xmm, env->xmm_regs, sizeof env->xmm_regs); 490 memcpy(env->xmm_regs, fpu.xmm, sizeof env->xmm_regs);
|
/external/libvpx/vp8/encoder/x86/ |
variance_impl_sse2.asm | 844 pavgb xmm5, xmm1 ; xmm = vertical average of the above 956 pavgb xmm5, xmm1 ; xmm = vertical average of the above [all...] |