/external/chromium_org/third_party/openssl/openssl/crypto/modes/asm/ |
ghash-sparcv9.pl | 49 $Zlo="%o1"; 99 ldx [$Htblo+$nlo],$Zlo 105 and $Zlo,0xf,$remi 109 srlx $Zlo,4,$Zlo 112 xor $Tlo,$Zlo,$Zlo 114 xor $Zlo,$tmp,$Zlo 117 and $Zlo,0xf,$rem [all...] |
ghash-alpha.pl | 41 $Zlo="t9"; 59 ldq $Zlo,8($nlo) 63 and $Zlo,0x0f,$remp 71 srl $Zlo,4,$Zlo 75 xor $t0,$Zlo,$Zlo 78 xor $Tlo1,$Zlo,$Zlo 89 and $Zlo,0x0f,$rem [all...] |
ghash-s390x.pl | 56 $Zlo="%r1"; 113 lg $Zlo,8+1($Xi) # Xi 152 lg $Zlo,8+1($Xi) # Xi 157 xg $Zlo,8($inp) 159 stg $Zlo,8+1($Xi) 164 sllg $nlo,$Zlo,4 165 srlg $xi,$Zlo,8 # extract second byte 167 lgr $nhi,$Zlo 171 lg $Zlo,8($nlo,$Htbl) 175 sllg $rem0,$Zlo, [all...] |
ghash-ia64.pl | 53 # &rem_4bit[Zlo&0xf]. It works, because rem_4bit is aligned at 128 59 (p19) dep rem=Zlo,rem_4bitp,3,4 } 63 (p19) shrp Zlo=Zhi,Zlo,4 } 67 (p18) xor Zlo=Zlo,Hlo 73 (p18) dep rem=Zlo,rem_4bitp,3,4 } 77 (p18) shrp Zlo=Zhi,Zlo,4 } 81 (p18) xor Zlo=Zlo,Hl [all...] |
ghash-x86_64.pl | 59 $Zlo="%r8"; 92 mov `&LB("$Zlo")`,`&LB("$nlo")` 93 mov `&LB("$Zlo")`,`&LB("$nhi")` 96 mov 8($Htbl,$nlo),$Zlo 99 mov $Zlo,$rem 104 shr \$4,$Zlo 109 xor 8($Htbl,$nhi),$Zlo 114 mov $Zlo,$rem 116 xor $tmp,$Zlo 120 shr \$4,$Zlo [all...] |
ghash-x86.pl | 360 my ($Zlo,$Zhi) = ("mm0","mm1"); 368 &movq ($Zlo,&QWP(8,$Htbl,$nlo)); 370 &movd ($rem[0],$Zlo); 377 &psrlq ($Zlo,4); 380 &pxor ($Zlo,&QWP(8,$Htbl,$nix)); 388 &movd ($rem[1],$Zlo); 389 &pxor ($Zlo,$tmp); 396 &psrlq ($Zlo,32); # lower part of Zlo is already there 399 &movd ($Zlh,$Zlo); [all...] |
/external/openssl/crypto/modes/asm/ |
ghash-sparcv9.pl | 49 $Zlo="%o1"; 99 ldx [$Htblo+$nlo],$Zlo 105 and $Zlo,0xf,$remi 109 srlx $Zlo,4,$Zlo 112 xor $Tlo,$Zlo,$Zlo 114 xor $Zlo,$tmp,$Zlo 117 and $Zlo,0xf,$rem [all...] |
ghash-alpha.pl | 41 $Zlo="t9"; 59 ldq $Zlo,8($nlo) 63 and $Zlo,0x0f,$remp 71 srl $Zlo,4,$Zlo 75 xor $t0,$Zlo,$Zlo 78 xor $Tlo1,$Zlo,$Zlo 89 and $Zlo,0x0f,$rem [all...] |
ghash-s390x.pl | 56 $Zlo="%r1"; 113 lg $Zlo,8+1($Xi) # Xi 152 lg $Zlo,8+1($Xi) # Xi 157 xg $Zlo,8($inp) 159 stg $Zlo,8+1($Xi) 164 sllg $nlo,$Zlo,4 165 srlg $xi,$Zlo,8 # extract second byte 167 lgr $nhi,$Zlo 171 lg $Zlo,8($nlo,$Htbl) 175 sllg $rem0,$Zlo, [all...] |
ghash-ia64.pl | 53 # &rem_4bit[Zlo&0xf]. It works, because rem_4bit is aligned at 128 59 (p19) dep rem=Zlo,rem_4bitp,3,4 } 63 (p19) shrp Zlo=Zhi,Zlo,4 } 67 (p18) xor Zlo=Zlo,Hlo 73 (p18) dep rem=Zlo,rem_4bitp,3,4 } 77 (p18) shrp Zlo=Zhi,Zlo,4 } 81 (p18) xor Zlo=Zlo,Hl [all...] |
ghash-x86_64.pl | 59 $Zlo="%r8"; 92 mov `&LB("$Zlo")`,`&LB("$nlo")` 93 mov `&LB("$Zlo")`,`&LB("$nhi")` 96 mov 8($Htbl,$nlo),$Zlo 99 mov $Zlo,$rem 104 shr \$4,$Zlo 109 xor 8($Htbl,$nhi),$Zlo 114 mov $Zlo,$rem 116 xor $tmp,$Zlo 120 shr \$4,$Zlo [all...] |
ghash-x86.pl | 360 my ($Zlo,$Zhi) = ("mm0","mm1"); 368 &movq ($Zlo,&QWP(8,$Htbl,$nlo)); 370 &movd ($rem[0],$Zlo); 377 &psrlq ($Zlo,4); 380 &pxor ($Zlo,&QWP(8,$Htbl,$nix)); 388 &movd ($rem[1],$Zlo); 389 &pxor ($Zlo,$tmp); 396 &psrlq ($Zlo,32); # lower part of Zlo is already there 399 &movd ($Zlh,$Zlo); [all...] |