HomeSort by relevance Sort by last modified time
    Searched refs:Xi (Results 1 - 25 of 38) sorted by null

1 2

  /external/chromium_org/third_party/boringssl/src/crypto/modes/asm/
ghash-x86_64.pl 122 $Xi="%rdi";
233 movzb 15($Xi),$Zlo
236 &loop ($Xi);
238 mov $Zlo,8($Xi)
239 mov $Zhi,($Xi)
302 &mov ($Zlo,"8($Xi)");
303 &mov ($Zhi,"0($Xi)");
313 &mov ("($Xi)",$Zhi);
314 &mov ("8($Xi)","%rdx");
349 &mov ($dat,"$j($Xi)") if (--$j%4==0)
    [all...]
ghash-x86.pl 265 &mov ($inp,&wparam(0)); # load Xi
268 &mov ($Zhh,&DWP(0,$inp)); # load Xi[16]
275 &mov (&DWP(0,"esp"),$Zhh); # copy Xi[16] on stack
298 &mov ($Zll,&wparam(0)); # load Xi
305 &mov ($Zhh,&DWP(0,$Zll)); # load Xi[16]
336 &mov ($inp,&wparam(0)); # load Xi
426 &mov ($inp,&wparam(0)); # load Xi
438 &mov ($inp,&wparam(0)); # load Xi
450 &mov ($Zhh,&wparam(0)); # load Xi
464 &mov ($Zll,&DWP(12,$Zhh)); # load Xi[16
    [all...]
aesni-gcm-x86_64.pl 68 $Z0,$Z1,$Z2,$Z3,$Xi) = map("%xmm$_",(0..8));
125 vpxor $Z0,$Xi,$Xi # modulo-scheduled
136 vpxor 16+8(%rsp),$Xi,$Xi # modulo-scheduled [vpxor $Z3,$Xi,$Xi]
200 vpxor 0x70+8(%rsp),$Xi,$Xi # accumulate I[0]
214 vpclmulqdq \$0x10,$Hkey,$Xi,$Z
    [all...]
ghash-armv4.pl 77 $Xi="r0"; # argument block
105 str $_,[$Xi,#$i]
107 str $_,[$Xi,#$i]
110 strb $_,[$Xi,#$i+3]
112 strb $Tlh,[$Xi,#$i+2]
114 strb $Thl,[$Xi,#$i+1]
115 strb $Thh,[$Xi,#$i]
162 ldrb $nhi,[$Xi,#15]
180 ldrb $nhi,[$Xi,#14]
213 ldrbpl $Tll,[$Xi,$cnt
    [all...]
  /external/openssl/crypto/modes/asm/
ghash-x86_64.pl 64 $Xi="%rdi";
174 movzb 15($Xi),$Zlo
177 &loop ($Xi);
179 mov $Zlo,8($Xi)
180 mov $Zhi,($Xi)
243 &mov ($Zlo,"8($Xi)");
244 &mov ($Zhi,"0($Xi)");
254 &mov ("($Xi)",$Zhi);
255 &mov ("8($Xi)","%rdx");
290 &mov ($dat,"$j($Xi)") if (--$j%4==0)
    [all...]
ghash-x86.pl 253 &mov ($inp,&wparam(0)); # load Xi
256 &mov ($Zhh,&DWP(0,$inp)); # load Xi[16]
263 &mov (&DWP(0,"esp"),$Zhh); # copy Xi[16] on stack
286 &mov ($Zll,&wparam(0)); # load Xi
293 &mov ($Zhh,&DWP(0,$Zll)); # load Xi[16]
324 &mov ($inp,&wparam(0)); # load Xi
414 &mov ($inp,&wparam(0)); # load Xi
426 &mov ($inp,&wparam(0)); # load Xi
438 &mov ($Zhh,&wparam(0)); # load Xi
452 &mov ($Zll,&DWP(12,$Zhh)); # load Xi[16
    [all...]
ghash-s390x.pl 58 $Xi="%r2"; # argument block
67 $xi="%r10";
96 la %r1,0($Xi) # H lies right after Xi in gcm128_context
108 aghi $Xi,-1
113 lg $Zlo,8+1($Xi) # Xi
134 la %r1,0($Xi) # H lies right after Xi in gcm128_context
147 aghi $Xi,-
    [all...]
ghashv8-armx.pl 30 $Xi="x0"; # argument block
78 vld1.64 {$t1},[$Xi] @ load Xi
90 mov $inp,$Xi
98 vld1.64 {$Xl},[$Xi] @ load [rotated] Xi
119 veor $IN,$IN,$Xl @ inp^=Xi
120 veor $t1,$t1,$t2 @ $t1 is rotated inp^Xi
123 vpmull.p64 $Xl,$H,$IN @ H.lo·Xi.lo
125 vpmull2.p64 $Xh,$H,$IN @ H.hi·Xi.h
    [all...]
ghashv8-armx.S 34 vld1.64 {q9},[r0] @ load Xi
54 vld1.64 {q0},[r0] @ load [rotated] Xi
75 veor q3,q3,q0 @ inp^=Xi
76 veor q9,q9,q10 @ q9 is rotated inp^Xi
79 .byte 0x86,0x0e,0xa8,0xf2 @ pmull q0,q12,q3 @ H.lo·Xi.lo
81 .byte 0x87,0x4e,0xa9,0xf2 @ pmull2 q2,q12,q3 @ H.hi·Xi.hi
83 .byte 0xa2,0x2e,0xaa,0xf2 @ pmull q1,q13,q9 @ (H.lo+H.hi)·(Xi.lo+Xi.hi)
111 vst1.64 {q0},[r0] @ write out Xi
ghash-sparcv9.pl 64 $Xi="%i0"; # input argument block
86 ldub [$Xi+15],$xi0
87 ldub [$Xi+14],$xi1
136 ldub [$Xi+$cnt],$xi1
188 stx $Zlo,[$Xi+8]
190 stx $Zhi,[$Xi]
208 stx $Zlo,[$Xi+8]
210 stx $Zhi,[$Xi]
226 ldub [$Xi+15],$nlo
238 ldub [$Xi+14],$nl
    [all...]
ghash-parisc.pl 53 $Xi="%r26"; # argument block
125 ldb 15($Xi),$nlo
137 ldb 14($Xi),$nlo
166 ldbx $cnt($Xi),$nlo
206 std $Zll,8($Xi)
207 std $Zhh,0($Xi)
216 ldb 15($Xi),$nlo
229 ldb 14($Xi),$nlo
260 ldbx $cnt($Xi),$nlo
315 stw $Zll,12($Xi)
    [all...]
ghash-armv4.pl 77 $Xi="r0"; # argument block
105 str $_,[$Xi,#$i]
107 str $_,[$Xi,#$i]
110 strb $_,[$Xi,#$i+3]
112 strb $Tlh,[$Xi,#$i+2]
114 strb $Thl,[$Xi,#$i+1]
115 strb $Thh,[$Xi,#$i]
158 ldrb $nhi,[$Xi,#15]
176 ldrb $nhi,[$Xi,#14]
209 ldrplb $Tll,[$Xi,$cnt
    [all...]
  /external/chromium_org/third_party/boringssl/src/crypto/modes/
gcm.c 150 static void gcm_gmult_4bit(uint64_t Xi[2], const u128 Htable[16]) {
159 nlo = ((const uint8_t *)Xi)[15];
183 nlo = ((const uint8_t *)Xi)[cnt];
202 Xi[0] = BSWAP8(Z.hi);
203 Xi[1] = BSWAP8(Z.lo);
205 uint8_t *p = (uint8_t *)Xi;
217 Xi[0] = Z.hi;
218 Xi[1] = Z.lo;
227 static void gcm_ghash_4bit(uint64_t Xi[2], const u128 Htable[16], const uint8_t *inp,
239 nlo = ((const uint8_t *)Xi)[15]
    [all...]
internal.h 159 } Yi, EKi, EK0, len, Xi, H;
161 /* Relative position of Xi, H and pre-computed Htable is used in some
164 void (*gmult)(uint64_t Xi[2], const u128 Htable[16]);
165 void (*ghash)(uint64_t Xi[2], const u128 Htable[16], const uint8_t *inp,
  /external/openssl/crypto/sha/asm/
sha1-sparcv9.pl 32 $Xi="%g4";
54 my $xi=($i&1)?@X[($i/2)%8]:$Xi;
68 add $xi,$e,$e
72 " srlx @X[(($i+1)/2)%8],32,$Xi\n";
91 sllx @X[($j+6)%8],32,$Xi ! Xupdate($i)
96 or $tmp1,$Xi,$Xi
98 xor $Xi,@X[$j%8],@X[$j%8]
99 srlx @X[$j%8],31,$Xi
    [all...]
sha1-586.pl 149 # with xi, also note that e becomes
154 &lea($tmp1,&DWP(0x5a827999,$tmp1,$e)); # tmp1+=K_00_19+xi
170 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd)
180 &mov(&swtmp($n%16),$f); # xi=f
187 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd)
197 &mov(&swtmp($n%16),$f); # xi=f
214 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd)
222 &mov(&swtmp($n%16),$f) if($n<77);# xi=f
232 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd)
242 &mov(&swtmp($n%16),$f) if($n<77);# xi=
    [all...]
sha1-thumb.pl 40 $Xi="r12";
133 mov $Xi,sp
170 mov $t0,$Xi
175 mov $Xi,$t1
181 cmp $Xi,$t0
187 mov $Xi,$t1
193 cmp $Xi,$t0
201 mov $Xi,$t1
207 cmp $Xi,$t0
211 mov $Xi,s
    [all...]
sha1-x86_64.pl 100 @xi=("%edx","%ebp");
113 mov `4*$i`($inp),$xi[0]
114 bswap $xi[0]
115 mov $xi[0],`4*$i`(%rsp)
119 mov `4*$j`($inp),$xi[1]
122 bswap $xi[1]
124 lea 0x5a827999($xi[0],$e),$e
126 mov $xi[1],`4*$j`(%rsp)
133 mov `4*($j%16)`(%rsp),$xi[1]
136 xor `4*(($j+2)%16)`(%rsp),$xi[1
    [all...]
sha1-armv4-large.pl 85 $Xi="r14";
91 ldr $t0,[$Xi,#15*4]
92 ldr $t1,[$Xi,#13*4]
93 ldr $t2,[$Xi,#7*4]
95 ldr $t3,[$Xi,#2*4]
102 str $t0,[$Xi,#-4]!
135 str $t0,[$Xi,#-4]!
191 mov $Xi,sp
202 teq $Xi,sp
221 teq $Xi,sp @ preserve carr
    [all...]
sha1-sparcv9a.pl 58 $Xi="%o7";
152 ld [$Xfer+`4*($i%16)`],$Xi
160 add $Xi,$e,$e
171 ld [$Xfer+`4*($i%16)`],$Xi
180 add $Xi,$e,$e
204 ld [$Xfer+`4*($i%16)`],$Xi
217 add $Xi,$e,$e
222 ld [$Xfer+`4*($i%16)`],$Xi
237 add $Xi,$e,$e
242 ld [$Xfer+`4*($i%16)`],$Xi
    [all...]
  /external/chromium_org/third_party/boringssl/src/crypto/sha/asm/
sha1-586.pl 166 # with xi, also note that e becomes
171 &lea($tmp1,&DWP(0x5a827999,$tmp1,$e)); # tmp1+=K_00_19+xi
187 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd)
197 &mov(&swtmp($n%16),$f); # xi=f
204 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd)
214 &mov(&swtmp($n%16),$f); # xi=f
231 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd)
239 &mov(&swtmp($n%16),$f) if($n<77);# xi=f
249 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd)
259 &mov(&swtmp($n%16),$f) if($n<77);# xi=
    [all...]
sha1-armv4-large.pl 71 $Xi="r14";
77 ldr $t0,[$Xi,#15*4]
78 ldr $t1,[$Xi,#13*4]
79 ldr $t2,[$Xi,#7*4]
81 ldr $t3,[$Xi,#2*4]
88 str $t0,[$Xi,#-4]!
121 str $t0,[$Xi,#-4]!
169 mov $Xi,sp
180 teq $Xi,sp
199 teq $Xi,sp @ preserve carr
    [all...]
sha1-x86_64.pl 132 @xi=("%edx","%ebp","%r14d");
145 mov `4*$i`($inp),$xi[0]
146 bswap $xi[0]
149 mov `4*$j`($inp),$xi[1]
151 mov $xi[0],`4*$i`(%rsp)
153 bswap $xi[1]
157 lea 0x5a827999($xi[0],$e),$e
164 xor `4*($j%16)`(%rsp),$xi[1]
166 mov $xi[0],`4*($i%16)`(%rsp)
168 xor `4*(($j+2)%16)`(%rsp),$xi[1
    [all...]
  /external/openssl/crypto/modes/
gcm128.c 145 static void gcm_gmult_8bit(u64 Xi[2], const u128 Htable[256])
148 const u8 *xi = (const u8 *)Xi+15; local
149 size_t rem, n = *xi;
221 if ((u8 *)Xi==xi) break;
223 n = *(--xi);
236 Xi[0] = BSWAP8(Z.hi);
237 Xi[1] = BSWAP8(Z.lo);
239 u8 *p = (u8 *)Xi;
590 const long *xi = (const long *)Xi; local
    [all...]
  /external/openssl/crypto/aes/asm/
aesni-sha1-x86_64.pl 110 my $Xi=4;
252 sub Xupdate_ssse3_16_31() # recall that $Xi starts wtih 4
285 &movdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU
321 &movdqa (@Tx[2],eval(16*(($Xi)/5))."($K_XX_XX)"); # K_XX_XX
329 $Xi++; push(@X,shift(@X)); # "rotate" X[]
339 &movdqa (@Tx[0],@X[-1&7]) if ($Xi==8);
350 if ($Xi%5) {
353 &movdqa (@Tx[2],eval(16*($Xi/5))."($K_XX_XX)");
366 &movdqa (eval(16*(($Xi-1)&3))."(%rsp)",@Tx[1]); # X[]+K xfer to IALU
386 &movdqa (@Tx[1],@X[0]) if ($Xi<19)
    [all...]

Completed in 462 milliseconds

1 2