HomeSort by relevance Sort by last modified time
    Searched refs:xmm (Results 1 - 24 of 24) sorted by null

  /external/valgrind/main/none/tests/amd64/
redundantRexW.stdout.exp 9 %xmm 0 ................................
10 %xmm 1 .2.6.6.2.527faf9.8.8.8.f.1.3.1.7
11 %xmm 2 ................................
12 %xmm 3 ................................
13 %xmm 4 ................................
14 %xmm 5 ................................
15 %xmm 6 ................................
16 %xmm 7 ................................
17 %xmm 8 ................................
18 %xmm 9 ...............................
    [all...]
gen_insn_test.pl 13 xmm => "reg128_t",
219 # my @xmmregs = map { "xmm$_" } (4,5,0,1,2,3,6,7);
220 my @xmmregs = map { "xmm$_" } (12,13,8,9,10,11,14,15);
354 if ($arg =~ /^([abcd]l|[abcd]x|e[abcd]x|r[abcd]x|r8|r16|r32|r64|mm|xmm|m8|m16|m32|m64|m128)\.(sb|ub|sw|uw|sd|ud|sq|uq|ps|pd)\[([^\]]+)\]$/)
463 elsif ($arg->{type} =~ /^(xmm|m128)$/)
631 if ($result->{type} eq "xmm")
646 if ($arg->{type} eq "xmm")
695 elsif ($arg->{type} eq "xmm")
752 if ($arg->{type} =~ /^(r8|r16|r32|r64|mm|xmm)$/)
809 elsif ($result->{type} eq "xmm")
    [all...]
  /external/openssl/crypto/perlasm/
x86asm.pl 88 if ("$dst:$src" =~ /(e[a-dsd][ixp]):xmm([0-7])/)
96 if ("$dst:$src" =~ /xmm([0-7]):(e[a-dsd][ixp])/)
104 if ("$dst:$src" =~ /xmm([0-7]):xmm([0-7])/)
112 if ("$dst:$src" =~ /xmm([0-7]):xmm([0-7])/)
120 if ("$dst:$src" =~ /xmm([0-7]):xmm([0-7])/)
x86_64-xlate.pl 125 } elsif ($self->{op} =~ /movq/ && $line =~ /%xmm/) {
669 if ($arg =~ /%xmm([0-9]+),\s*%r(\w+)/) {
676 } elsif ($arg =~ /%r(\w+),\s*%xmm([0-9]+)/) {
689 if (shift =~ /\$([0-9]+),\s*%xmm([0-9]+),\s*(%\w+)/) {
707 if (shift =~ /\$([0-9]+),\s*(%\w+),\s*%xmm([0-9]+)/) {
725 if (shift =~ /%xmm([0-9]+),\s*%xmm([0-9]+)/) {
737 if (shift =~ /\$([0-9]+),\s*%xmm([0-9]+),\s*%xmm([0-9]+)/) {
750 if (shift =~ /\$([x0-9a-f]+),\s*%xmm([0-9]+),\s*%xmm([0-9]+)/)
    [all...]
  /dalvik/vm/compiler/codegen/x86/libenc/
encoder.inl 84 inline static void add_xmm(EncoderBase::Operands & args, const XMM_Opnd & xmm, bool dbl) {
86 // XMM registers indexes in Reg_No enum are shifted by xmm0_reg, their indexes
88 // xmm.get_idx() value
89 assert(xmm.get_idx() >= xmm0_reg);
90 return args.add((RegName)( (dbl ? RegName_XMM0D : RegName_XMM0S) + xmm.get_idx() -
326 ENCODER_DECLARE_EXPORT char * movd(char * stream, const RM_Opnd & rm, const XMM_Opnd & xmm) {
329 add_xmm(args, xmm, false);
333 ENCODER_DECLARE_EXPORT char * movd(char * stream, const XMM_Opnd & xmm, const RM_Opnd & rm) {
335 add_xmm(args, xmm, false);
340 ENCODER_DECLARE_EXPORT char * movq(char * stream, const RM_Opnd & rm, const XMM_Opnd & xmm) {
    [all...]
encoder.h 195 enum Tag { SignedImm, UnsignedImm, Reg, Mem, FP, XMM };
396 XMM_Opnd(unsigned _idx): Opnd(XMM), m_idx(_idx) {};
401 XMM_Opnd(const XMM_Opnd &): Opnd(XMM) { assert(false); }
516 ENCODER_DECLARE_EXPORT char * movd(char * stream, const RM_Opnd & rm, const XMM_Opnd & xmm);
517 ENCODER_DECLARE_EXPORT char * movd(char * stream, const XMM_Opnd & xmm, const RM_Opnd & rm);
518 ENCODER_DECLARE_EXPORT char * movq(char * stream, const RM_Opnd & rm, const XMM_Opnd & xmm);
519 ENCODER_DECLARE_EXPORT char * movq(char * stream, const XMM_Opnd & xmm, const RM_Opnd & rm);
522 ENCODER_DECLARE_EXPORT char * sse_mov(char * stream, const XMM_Opnd & xmm, const M_Opnd & mem, bool dbl);
523 ENCODER_DECLARE_EXPORT char * sse_mov(char * stream, const M_Opnd & mem, const XMM_Opnd & xmm, bool dbl);
527 ENCODER_DECLARE_EXPORT char * sse_add(char * stream, const XMM_Opnd & xmm, const M_Opnd & mem, bool dbl)
    [all...]
  /external/openssl/crypto/rc4/asm/
rc4-md5-x86_64.pl 220 my $xmm="%xmm".($j&1);
223 $code.=" pxor $xmm,$xmm\n" if ($rc4 && $j<=1);
240 #rc4# pinsrw \$`($j>>1)&7`,($dat,$TX[0],4),$xmm\n
260 my $xmm="%xmm".($j&1);
263 $code.=" pxor $xmm,$xmm\n" if ($rc4 && $j<=1);
280 #rc4# pinsrw \$`($j>>1)&7`,($dat,$TX[0],4),$xmm\
    [all...]
rc4-x86_64.pl 258 my $xmm="%xmm".($j&1);
266 $code.=" pxor $xmm,$xmm\n" if ($i<=1);
275 $code.=" pinsrw \$`($j>>1)&7`,($dat,$TX[0],4),$xmm\n";
  /external/valgrind/main/none/tests/x86/
gen_insn_test.pl 12 xmm => "reg128_t",
197 my @xmmregs = map { "xmm$_" } (4,5,0,1,2,3,6,7);
331 if ($arg =~ /^([abcd]l|[abcd]x|e[abcd]x|r8|r16|r32|mm|xmm|m8|m16|m32|m64|m128)\.(sb|ub|sw|uw|sd|ud|sq|uq|ps|pd)\[([^\]]+)\]$/)
440 elsif ($arg->{type} =~ /^(xmm|m128)$/)
608 if ($result->{type} eq "xmm")
623 if ($arg->{type} eq "xmm")
664 elsif ($arg->{type} eq "xmm")
720 if ($arg->{type} =~ /^(r8|r16|r32|mm|xmm)$/)
773 elsif ($result->{type} eq "xmm")
825 if ($result->{type} eq "xmm")
    [all...]
  /external/valgrind/main/VEX/priv/
guest_amd64_helpers.c 1652 U128* xmm = (U128*)(addr + 160); local
1756 U128* xmm = (U128*)(addr + 160); local
    [all...]
guest_x86_helpers.c 1672 U128* xmm = (U128*)(addr + 160); local
1764 U128* xmm = (U128*)(addr + 160); local
    [all...]
  /prebuilts/gcc/linux-x86/host/i686-linux-glibc2.7-4.4.3/sysroot/usr/include/asm/
kvm.h 127 __u8 xmm[16][16]; member in struct:kvm_fpu
  /prebuilts/gcc/linux-x86/host/i686-linux-glibc2.7-4.6/sysroot/usr/include/asm/
kvm.h 127 __u8 xmm[16][16]; member in struct:kvm_fpu
  /prebuilts/gcc/linux-x86/host/x86_64-linux-glibc2.7-4.6/sysroot/usr/include/asm/
kvm.h 127 __u8 xmm[16][16]; member in struct:kvm_fpu
  /external/libvpx/libvpx/vpx_ports/
x86_abi_support.asm 313 %error Only xmm registers 6-15 must be preserved
321 movxmm [rsp + ((i - 6) * 16)], xmm %+ i
332 movxmm xmm %+ i, [rsp +((i - 6) * 16)]
  /external/openssl/crypto/sha/asm/
sha1-586.pl 56 # a) XMM units became faster and wider;
107 $xmm=$ymm=0;
108 for (@ARGV) { $xmm=1 if (/-DOPENSSL_IA32_SSE2/); }
110 $ymm=1 if ($xmm &&
115 $ymm=1 if ($xmm && !$ymm && $ARGV[0] eq "win32n" &&
119 &external_label("OPENSSL_ia32cap_P") if ($xmm);
297 if ($xmm) {
391 if ($xmm) {
395 # %xmm[0-7] are used as ring @X[] buffer containing quadruples of last
403 # no %xmm registers left to spare. For this reason part of the rin
    [all...]
  /external/qemu/android/config/linux-x86/asm/
kvm.h 138 __u8 xmm[16][16]; member in struct:kvm_fpu
  /external/qemu/android/config/linux-x86_64/asm/
kvm.h 138 __u8 xmm[16][16]; member in struct:kvm_fpu
  /external/libvpx/libvpx/vp8/common/x86/
loopfilter_block_sse2.asm 380 ; using xmm[0124]
448 ; using xmm[0124]
617 ; using xmm[0124]
685 ; using xmm[0124]
variance_impl_sse2.asm 735 pavgb xmm5, xmm1 ; xmm = vertical average of the above
848 pavgb xmm5, xmm1 ; xmm = vertical average of the above
    [all...]
  /external/openssl/crypto/aes/asm/
aesni-sha1-x86_64.pl 111 my @X=map("%xmm$_",(4..7,0..3));
112 my @Tx=map("%xmm$_",(8..10));
117 my ($iv,$in,$rndkey0)=map("%xmm$_",(11..13));
186 movdqu 0($inp),@X[-4&7] # load input to %xmm[0-3]
640 my @X=map("%xmm$_",(4..7,0..3));
641 my @Tx=map("%xmm$_",(8..10));
706 vmovdqu 0($inp),@X[-4&7] # load input to %xmm[0-3]
1233 if ($line=~/(aes[a-z]+)\s+%xmm([0-9]+),\s*%xmm([0-9]+)/) {
1247 $code =~ s/\b(aes.*%xmm[0-9]+).*$/aesni($1)/gem
    [all...]
aesni-x86.pl 80 if ("$dst:$src" =~ /xmm([0-7]):xmm([0-7])/)
85 if ("$dst:$src" =~ /xmm([0-7]):xmm([0-7])/)
205 # in XMM registers addreassable in 32-bit mode and therefore 6x is
    [all...]
aesni-x86_64.pl 195 # %xmm register layout
1017 my ($in0,$in1,$in2,$in3)=map("%xmm$_",(8..11));
    [all...]
  /external/qemu/target-i386/
kvm.c 381 memcpy(fpu.xmm, env->xmm_regs, sizeof env->xmm_regs);
490 memcpy(env->xmm_regs, fpu.xmm, sizeof env->xmm_regs);

Completed in 638 milliseconds