/external/valgrind/main/none/tests/amd64/ |
redundantRexW.stdout.exp | 9 %xmm 0 ................................ 10 %xmm 1 .2.6.6.2.527faf9.8.8.8.f.1.3.1.7 11 %xmm 2 ................................ 12 %xmm 3 ................................ 13 %xmm 4 ................................ 14 %xmm 5 ................................ 15 %xmm 6 ................................ 16 %xmm 7 ................................ 17 %xmm 8 ................................ 18 %xmm 9 ............................... [all...] |
gen_insn_test.pl | 13 xmm => "reg128_t", 219 # my @xmmregs = map { "xmm$_" } (4,5,0,1,2,3,6,7); 220 my @xmmregs = map { "xmm$_" } (12,13,8,9,10,11,14,15); 354 if ($arg =~ /^([abcd]l|[abcd]x|e[abcd]x|r[abcd]x|r8|r16|r32|r64|mm|xmm|m8|m16|m32|m64|m128)\.(sb|ub|sw|uw|sd|ud|sq|uq|ps|pd)\[([^\]]+)\]$/) 463 elsif ($arg->{type} =~ /^(xmm|m128)$/) 631 if ($result->{type} eq "xmm") 646 if ($arg->{type} eq "xmm") 695 elsif ($arg->{type} eq "xmm") 752 if ($arg->{type} =~ /^(r8|r16|r32|r64|mm|xmm)$/) 809 elsif ($result->{type} eq "xmm") [all...] |
/external/openssl/crypto/perlasm/ |
x86asm.pl | 88 if ("$dst:$src" =~ /(e[a-dsd][ixp]):xmm([0-7])/) 96 if ("$dst:$src" =~ /xmm([0-7]):(e[a-dsd][ixp])/) 104 if ("$dst:$src" =~ /xmm([0-7]):xmm([0-7])/) 112 if ("$dst:$src" =~ /xmm([0-7]):xmm([0-7])/) 120 if ("$dst:$src" =~ /xmm([0-7]):xmm([0-7])/)
|
x86_64-xlate.pl | 125 } elsif ($self->{op} =~ /movq/ && $line =~ /%xmm/) { 669 if ($arg =~ /%xmm([0-9]+),\s*%r(\w+)/) { 676 } elsif ($arg =~ /%r(\w+),\s*%xmm([0-9]+)/) { 689 if (shift =~ /\$([0-9]+),\s*%xmm([0-9]+),\s*(%\w+)/) { 707 if (shift =~ /\$([0-9]+),\s*(%\w+),\s*%xmm([0-9]+)/) { 725 if (shift =~ /%xmm([0-9]+),\s*%xmm([0-9]+)/) { 737 if (shift =~ /\$([0-9]+),\s*%xmm([0-9]+),\s*%xmm([0-9]+)/) { 750 if (shift =~ /\$([x0-9a-f]+),\s*%xmm([0-9]+),\s*%xmm([0-9]+)/) [all...] |
/dalvik/vm/compiler/codegen/x86/libenc/ |
encoder.inl | 84 inline static void add_xmm(EncoderBase::Operands & args, const XMM_Opnd & xmm, bool dbl) { 86 // XMM registers indexes in Reg_No enum are shifted by xmm0_reg, their indexes 88 // xmm.get_idx() value 89 assert(xmm.get_idx() >= xmm0_reg); 90 return args.add((RegName)( (dbl ? RegName_XMM0D : RegName_XMM0S) + xmm.get_idx() - 326 ENCODER_DECLARE_EXPORT char * movd(char * stream, const RM_Opnd & rm, const XMM_Opnd & xmm) { 329 add_xmm(args, xmm, false); 333 ENCODER_DECLARE_EXPORT char * movd(char * stream, const XMM_Opnd & xmm, const RM_Opnd & rm) { 335 add_xmm(args, xmm, false); 340 ENCODER_DECLARE_EXPORT char * movq(char * stream, const RM_Opnd & rm, const XMM_Opnd & xmm) { [all...] |
encoder.h | 195 enum Tag { SignedImm, UnsignedImm, Reg, Mem, FP, XMM }; 396 XMM_Opnd(unsigned _idx): Opnd(XMM), m_idx(_idx) {}; 401 XMM_Opnd(const XMM_Opnd &): Opnd(XMM) { assert(false); } 516 ENCODER_DECLARE_EXPORT char * movd(char * stream, const RM_Opnd & rm, const XMM_Opnd & xmm); 517 ENCODER_DECLARE_EXPORT char * movd(char * stream, const XMM_Opnd & xmm, const RM_Opnd & rm); 518 ENCODER_DECLARE_EXPORT char * movq(char * stream, const RM_Opnd & rm, const XMM_Opnd & xmm); 519 ENCODER_DECLARE_EXPORT char * movq(char * stream, const XMM_Opnd & xmm, const RM_Opnd & rm); 522 ENCODER_DECLARE_EXPORT char * sse_mov(char * stream, const XMM_Opnd & xmm, const M_Opnd & mem, bool dbl); 523 ENCODER_DECLARE_EXPORT char * sse_mov(char * stream, const M_Opnd & mem, const XMM_Opnd & xmm, bool dbl); 527 ENCODER_DECLARE_EXPORT char * sse_add(char * stream, const XMM_Opnd & xmm, const M_Opnd & mem, bool dbl) [all...] |
/external/openssl/crypto/rc4/asm/ |
rc4-md5-x86_64.pl | 220 my $xmm="%xmm".($j&1); 223 $code.=" pxor $xmm,$xmm\n" if ($rc4 && $j<=1); 240 #rc4# pinsrw \$`($j>>1)&7`,($dat,$TX[0],4),$xmm\n 260 my $xmm="%xmm".($j&1); 263 $code.=" pxor $xmm,$xmm\n" if ($rc4 && $j<=1); 280 #rc4# pinsrw \$`($j>>1)&7`,($dat,$TX[0],4),$xmm\ [all...] |
rc4-x86_64.pl | 258 my $xmm="%xmm".($j&1); 266 $code.=" pxor $xmm,$xmm\n" if ($i<=1); 275 $code.=" pinsrw \$`($j>>1)&7`,($dat,$TX[0],4),$xmm\n";
|
/external/valgrind/main/none/tests/x86/ |
gen_insn_test.pl | 12 xmm => "reg128_t", 197 my @xmmregs = map { "xmm$_" } (4,5,0,1,2,3,6,7); 331 if ($arg =~ /^([abcd]l|[abcd]x|e[abcd]x|r8|r16|r32|mm|xmm|m8|m16|m32|m64|m128)\.(sb|ub|sw|uw|sd|ud|sq|uq|ps|pd)\[([^\]]+)\]$/) 440 elsif ($arg->{type} =~ /^(xmm|m128)$/) 608 if ($result->{type} eq "xmm") 623 if ($arg->{type} eq "xmm") 664 elsif ($arg->{type} eq "xmm") 720 if ($arg->{type} =~ /^(r8|r16|r32|mm|xmm)$/) 773 elsif ($result->{type} eq "xmm") 825 if ($result->{type} eq "xmm") [all...] |
/external/valgrind/main/VEX/priv/ |
guest_amd64_helpers.c | 1652 U128* xmm = (U128*)(addr + 160); local 1756 U128* xmm = (U128*)(addr + 160); local [all...] |
guest_x86_helpers.c | 1672 U128* xmm = (U128*)(addr + 160); local 1764 U128* xmm = (U128*)(addr + 160); local [all...] |
/prebuilts/gcc/linux-x86/host/i686-linux-glibc2.7-4.4.3/sysroot/usr/include/asm/ |
kvm.h | 127 __u8 xmm[16][16]; member in struct:kvm_fpu
|
/prebuilts/gcc/linux-x86/host/i686-linux-glibc2.7-4.6/sysroot/usr/include/asm/ |
kvm.h | 127 __u8 xmm[16][16]; member in struct:kvm_fpu
|
/prebuilts/gcc/linux-x86/host/x86_64-linux-glibc2.7-4.6/sysroot/usr/include/asm/ |
kvm.h | 127 __u8 xmm[16][16]; member in struct:kvm_fpu
|
/external/libvpx/libvpx/vpx_ports/ |
x86_abi_support.asm | 313 %error Only xmm registers 6-15 must be preserved 321 movxmm [rsp + ((i - 6) * 16)], xmm %+ i 332 movxmm xmm %+ i, [rsp +((i - 6) * 16)]
|
/external/openssl/crypto/sha/asm/ |
sha1-586.pl | 56 # a) XMM units became faster and wider; 107 $xmm=$ymm=0; 108 for (@ARGV) { $xmm=1 if (/-DOPENSSL_IA32_SSE2/); } 110 $ymm=1 if ($xmm && 115 $ymm=1 if ($xmm && !$ymm && $ARGV[0] eq "win32n" && 119 &external_label("OPENSSL_ia32cap_P") if ($xmm); 297 if ($xmm) { 391 if ($xmm) { 395 # %xmm[0-7] are used as ring @X[] buffer containing quadruples of last 403 # no %xmm registers left to spare. For this reason part of the rin [all...] |
/external/qemu/android/config/linux-x86/asm/ |
kvm.h | 138 __u8 xmm[16][16]; member in struct:kvm_fpu
|
/external/qemu/android/config/linux-x86_64/asm/ |
kvm.h | 138 __u8 xmm[16][16]; member in struct:kvm_fpu
|
/external/libvpx/libvpx/vp8/common/x86/ |
loopfilter_block_sse2.asm | 380 ; using xmm[0124] 448 ; using xmm[0124] 617 ; using xmm[0124] 685 ; using xmm[0124]
|
variance_impl_sse2.asm | 735 pavgb xmm5, xmm1 ; xmm = vertical average of the above 848 pavgb xmm5, xmm1 ; xmm = vertical average of the above [all...] |
/external/openssl/crypto/aes/asm/ |
aesni-sha1-x86_64.pl | 111 my @X=map("%xmm$_",(4..7,0..3)); 112 my @Tx=map("%xmm$_",(8..10)); 117 my ($iv,$in,$rndkey0)=map("%xmm$_",(11..13)); 186 movdqu 0($inp),@X[-4&7] # load input to %xmm[0-3] 640 my @X=map("%xmm$_",(4..7,0..3)); 641 my @Tx=map("%xmm$_",(8..10)); 706 vmovdqu 0($inp),@X[-4&7] # load input to %xmm[0-3] 1233 if ($line=~/(aes[a-z]+)\s+%xmm([0-9]+),\s*%xmm([0-9]+)/) { 1247 $code =~ s/\b(aes.*%xmm[0-9]+).*$/aesni($1)/gem [all...] |
aesni-x86.pl | 80 if ("$dst:$src" =~ /xmm([0-7]):xmm([0-7])/) 85 if ("$dst:$src" =~ /xmm([0-7]):xmm([0-7])/) 205 # in XMM registers addreassable in 32-bit mode and therefore 6x is [all...] |
aesni-x86_64.pl | 195 # %xmm register layout 1017 my ($in0,$in1,$in2,$in3)=map("%xmm$_",(8..11)); [all...] |
/external/qemu/target-i386/ |
kvm.c | 381 memcpy(fpu.xmm, env->xmm_regs, sizeof env->xmm_regs); 490 memcpy(env->xmm_regs, fpu.xmm, sizeof env->xmm_regs);
|