/external/chromium_org/third_party/boringssl/src/crypto/modes/asm/ |
ghash-x86_64.pl | 122 $Xi="%rdi"; 233 movzb 15($Xi),$Zlo 236 &loop ($Xi); 238 mov $Zlo,8($Xi) 239 mov $Zhi,($Xi) 302 &mov ($Zlo,"8($Xi)"); 303 &mov ($Zhi,"0($Xi)"); 313 &mov ("($Xi)",$Zhi); 314 &mov ("8($Xi)","%rdx"); 349 &mov ($dat,"$j($Xi)") if (--$j%4==0) [all...] |
ghash-x86.pl | 265 &mov ($inp,&wparam(0)); # load Xi 268 &mov ($Zhh,&DWP(0,$inp)); # load Xi[16] 275 &mov (&DWP(0,"esp"),$Zhh); # copy Xi[16] on stack 298 &mov ($Zll,&wparam(0)); # load Xi 305 &mov ($Zhh,&DWP(0,$Zll)); # load Xi[16] 336 &mov ($inp,&wparam(0)); # load Xi 426 &mov ($inp,&wparam(0)); # load Xi 438 &mov ($inp,&wparam(0)); # load Xi 450 &mov ($Zhh,&wparam(0)); # load Xi 464 &mov ($Zll,&DWP(12,$Zhh)); # load Xi[16 [all...] |
aesni-gcm-x86_64.pl | 68 $Z0,$Z1,$Z2,$Z3,$Xi) = map("%xmm$_",(0..8)); 125 vpxor $Z0,$Xi,$Xi # modulo-scheduled 136 vpxor 16+8(%rsp),$Xi,$Xi # modulo-scheduled [vpxor $Z3,$Xi,$Xi] 200 vpxor 0x70+8(%rsp),$Xi,$Xi # accumulate I[0] 214 vpclmulqdq \$0x10,$Hkey,$Xi,$Z [all...] |
ghash-armv4.pl | 77 $Xi="r0"; # argument block 105 str $_,[$Xi,#$i] 107 str $_,[$Xi,#$i] 110 strb $_,[$Xi,#$i+3] 112 strb $Tlh,[$Xi,#$i+2] 114 strb $Thl,[$Xi,#$i+1] 115 strb $Thh,[$Xi,#$i] 162 ldrb $nhi,[$Xi,#15] 180 ldrb $nhi,[$Xi,#14] 213 ldrbpl $Tll,[$Xi,$cnt [all...] |
/external/openssl/crypto/modes/asm/ |
ghash-x86_64.pl | 64 $Xi="%rdi"; 174 movzb 15($Xi),$Zlo 177 &loop ($Xi); 179 mov $Zlo,8($Xi) 180 mov $Zhi,($Xi) 243 &mov ($Zlo,"8($Xi)"); 244 &mov ($Zhi,"0($Xi)"); 254 &mov ("($Xi)",$Zhi); 255 &mov ("8($Xi)","%rdx"); 290 &mov ($dat,"$j($Xi)") if (--$j%4==0) [all...] |
ghash-x86.pl | 253 &mov ($inp,&wparam(0)); # load Xi 256 &mov ($Zhh,&DWP(0,$inp)); # load Xi[16] 263 &mov (&DWP(0,"esp"),$Zhh); # copy Xi[16] on stack 286 &mov ($Zll,&wparam(0)); # load Xi 293 &mov ($Zhh,&DWP(0,$Zll)); # load Xi[16] 324 &mov ($inp,&wparam(0)); # load Xi 414 &mov ($inp,&wparam(0)); # load Xi 426 &mov ($inp,&wparam(0)); # load Xi 438 &mov ($Zhh,&wparam(0)); # load Xi 452 &mov ($Zll,&DWP(12,$Zhh)); # load Xi[16 [all...] |
ghash-s390x.pl | 58 $Xi="%r2"; # argument block 67 $xi="%r10"; 96 la %r1,0($Xi) # H lies right after Xi in gcm128_context 108 aghi $Xi,-1 113 lg $Zlo,8+1($Xi) # Xi 134 la %r1,0($Xi) # H lies right after Xi in gcm128_context 147 aghi $Xi,- [all...] |
ghashv8-armx.pl | 30 $Xi="x0"; # argument block 78 vld1.64 {$t1},[$Xi] @ load Xi 90 mov $inp,$Xi 98 vld1.64 {$Xl},[$Xi] @ load [rotated] Xi 119 veor $IN,$IN,$Xl @ inp^=Xi 120 veor $t1,$t1,$t2 @ $t1 is rotated inp^Xi 123 vpmull.p64 $Xl,$H,$IN @ H.lo·Xi.lo 125 vpmull2.p64 $Xh,$H,$IN @ H.hi·Xi.h [all...] |
ghashv8-armx.S | 34 vld1.64 {q9},[r0] @ load Xi 54 vld1.64 {q0},[r0] @ load [rotated] Xi 75 veor q3,q3,q0 @ inp^=Xi 76 veor q9,q9,q10 @ q9 is rotated inp^Xi 79 .byte 0x86,0x0e,0xa8,0xf2 @ pmull q0,q12,q3 @ H.lo·Xi.lo 81 .byte 0x87,0x4e,0xa9,0xf2 @ pmull2 q2,q12,q3 @ H.hi·Xi.hi 83 .byte 0xa2,0x2e,0xaa,0xf2 @ pmull q1,q13,q9 @ (H.lo+H.hi)·(Xi.lo+Xi.hi) 111 vst1.64 {q0},[r0] @ write out Xi
|
ghash-sparcv9.pl | 64 $Xi="%i0"; # input argument block 86 ldub [$Xi+15],$xi0 87 ldub [$Xi+14],$xi1 136 ldub [$Xi+$cnt],$xi1 188 stx $Zlo,[$Xi+8] 190 stx $Zhi,[$Xi] 208 stx $Zlo,[$Xi+8] 210 stx $Zhi,[$Xi] 226 ldub [$Xi+15],$nlo 238 ldub [$Xi+14],$nl [all...] |
ghash-parisc.pl | 53 $Xi="%r26"; # argument block 125 ldb 15($Xi),$nlo 137 ldb 14($Xi),$nlo 166 ldbx $cnt($Xi),$nlo 206 std $Zll,8($Xi) 207 std $Zhh,0($Xi) 216 ldb 15($Xi),$nlo 229 ldb 14($Xi),$nlo 260 ldbx $cnt($Xi),$nlo 315 stw $Zll,12($Xi) [all...] |
ghash-armv4.pl | 77 $Xi="r0"; # argument block 105 str $_,[$Xi,#$i] 107 str $_,[$Xi,#$i] 110 strb $_,[$Xi,#$i+3] 112 strb $Tlh,[$Xi,#$i+2] 114 strb $Thl,[$Xi,#$i+1] 115 strb $Thh,[$Xi,#$i] 158 ldrb $nhi,[$Xi,#15] 176 ldrb $nhi,[$Xi,#14] 209 ldrplb $Tll,[$Xi,$cnt [all...] |
/external/chromium_org/third_party/boringssl/src/crypto/modes/ |
gcm.c | 150 static void gcm_gmult_4bit(uint64_t Xi[2], const u128 Htable[16]) { 159 nlo = ((const uint8_t *)Xi)[15]; 183 nlo = ((const uint8_t *)Xi)[cnt]; 202 Xi[0] = BSWAP8(Z.hi); 203 Xi[1] = BSWAP8(Z.lo); 205 uint8_t *p = (uint8_t *)Xi; 217 Xi[0] = Z.hi; 218 Xi[1] = Z.lo; 227 static void gcm_ghash_4bit(uint64_t Xi[2], const u128 Htable[16], const uint8_t *inp, 239 nlo = ((const uint8_t *)Xi)[15] [all...] |
internal.h | 159 } Yi, EKi, EK0, len, Xi, H; 161 /* Relative position of Xi, H and pre-computed Htable is used in some 164 void (*gmult)(uint64_t Xi[2], const u128 Htable[16]); 165 void (*ghash)(uint64_t Xi[2], const u128 Htable[16], const uint8_t *inp,
|
/external/openssl/crypto/sha/asm/ |
sha1-sparcv9.pl | 32 $Xi="%g4"; 54 my $xi=($i&1)?@X[($i/2)%8]:$Xi; 68 add $xi,$e,$e 72 " srlx @X[(($i+1)/2)%8],32,$Xi\n"; 91 sllx @X[($j+6)%8],32,$Xi ! Xupdate($i) 96 or $tmp1,$Xi,$Xi 98 xor $Xi,@X[$j%8],@X[$j%8] 99 srlx @X[$j%8],31,$Xi [all...] |
sha1-586.pl | 149 # with xi, also note that e becomes 154 &lea($tmp1,&DWP(0x5a827999,$tmp1,$e)); # tmp1+=K_00_19+xi 170 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd) 180 &mov(&swtmp($n%16),$f); # xi=f 187 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd) 197 &mov(&swtmp($n%16),$f); # xi=f 214 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd) 222 &mov(&swtmp($n%16),$f) if($n<77);# xi=f 232 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd) 242 &mov(&swtmp($n%16),$f) if($n<77);# xi= [all...] |
sha1-thumb.pl | 40 $Xi="r12"; 133 mov $Xi,sp 170 mov $t0,$Xi 175 mov $Xi,$t1 181 cmp $Xi,$t0 187 mov $Xi,$t1 193 cmp $Xi,$t0 201 mov $Xi,$t1 207 cmp $Xi,$t0 211 mov $Xi,s [all...] |
sha1-x86_64.pl | 100 @xi=("%edx","%ebp"); 113 mov `4*$i`($inp),$xi[0] 114 bswap $xi[0] 115 mov $xi[0],`4*$i`(%rsp) 119 mov `4*$j`($inp),$xi[1] 122 bswap $xi[1] 124 lea 0x5a827999($xi[0],$e),$e 126 mov $xi[1],`4*$j`(%rsp) 133 mov `4*($j%16)`(%rsp),$xi[1] 136 xor `4*(($j+2)%16)`(%rsp),$xi[1 [all...] |
sha1-armv4-large.pl | 85 $Xi="r14"; 91 ldr $t0,[$Xi,#15*4] 92 ldr $t1,[$Xi,#13*4] 93 ldr $t2,[$Xi,#7*4] 95 ldr $t3,[$Xi,#2*4] 102 str $t0,[$Xi,#-4]! 135 str $t0,[$Xi,#-4]! 191 mov $Xi,sp 202 teq $Xi,sp 221 teq $Xi,sp @ preserve carr [all...] |
sha1-sparcv9a.pl | 58 $Xi="%o7"; 152 ld [$Xfer+`4*($i%16)`],$Xi 160 add $Xi,$e,$e 171 ld [$Xfer+`4*($i%16)`],$Xi 180 add $Xi,$e,$e 204 ld [$Xfer+`4*($i%16)`],$Xi 217 add $Xi,$e,$e 222 ld [$Xfer+`4*($i%16)`],$Xi 237 add $Xi,$e,$e 242 ld [$Xfer+`4*($i%16)`],$Xi [all...] |
/external/chromium_org/third_party/boringssl/src/crypto/sha/asm/ |
sha1-586.pl | 166 # with xi, also note that e becomes 171 &lea($tmp1,&DWP(0x5a827999,$tmp1,$e)); # tmp1+=K_00_19+xi 187 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd) 197 &mov(&swtmp($n%16),$f); # xi=f 204 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd) 214 &mov(&swtmp($n%16),$f); # xi=f 231 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd) 239 &mov(&swtmp($n%16),$f) if($n<77);# xi=f 249 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd) 259 &mov(&swtmp($n%16),$f) if($n<77);# xi= [all...] |
sha1-armv4-large.pl | 71 $Xi="r14"; 77 ldr $t0,[$Xi,#15*4] 78 ldr $t1,[$Xi,#13*4] 79 ldr $t2,[$Xi,#7*4] 81 ldr $t3,[$Xi,#2*4] 88 str $t0,[$Xi,#-4]! 121 str $t0,[$Xi,#-4]! 169 mov $Xi,sp 180 teq $Xi,sp 199 teq $Xi,sp @ preserve carr [all...] |
sha1-x86_64.pl | 132 @xi=("%edx","%ebp","%r14d"); 145 mov `4*$i`($inp),$xi[0] 146 bswap $xi[0] 149 mov `4*$j`($inp),$xi[1] 151 mov $xi[0],`4*$i`(%rsp) 153 bswap $xi[1] 157 lea 0x5a827999($xi[0],$e),$e 164 xor `4*($j%16)`(%rsp),$xi[1] 166 mov $xi[0],`4*($i%16)`(%rsp) 168 xor `4*(($j+2)%16)`(%rsp),$xi[1 [all...] |
/external/openssl/crypto/modes/ |
gcm128.c | 145 static void gcm_gmult_8bit(u64 Xi[2], const u128 Htable[256]) 148 const u8 *xi = (const u8 *)Xi+15; local 149 size_t rem, n = *xi; 221 if ((u8 *)Xi==xi) break; 223 n = *(--xi); 236 Xi[0] = BSWAP8(Z.hi); 237 Xi[1] = BSWAP8(Z.lo); 239 u8 *p = (u8 *)Xi; 590 const long *xi = (const long *)Xi; local [all...] |
/external/openssl/crypto/aes/asm/ |
aesni-sha1-x86_64.pl | 110 my $Xi=4; 252 sub Xupdate_ssse3_16_31() # recall that $Xi starts wtih 4 285 &movdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU 321 &movdqa (@Tx[2],eval(16*(($Xi)/5))."($K_XX_XX)"); # K_XX_XX 329 $Xi++; push(@X,shift(@X)); # "rotate" X[] 339 &movdqa (@Tx[0],@X[-1&7]) if ($Xi==8); 350 if ($Xi%5) { 353 &movdqa (@Tx[2],eval(16*($Xi/5))."($K_XX_XX)"); 366 &movdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU 386 &movdqa (@Tx[1],@X[0]) if ($Xi<19) [all...] |