/external/boringssl/src/crypto/modes/asm/ |
ghash-x86.pl | 372 &mov ($inp,&wparam(0)); # load Xi 405 &mov ("eax",&wparam(0)); # Xi 420 &mov (&DWP(528+16+0,"esp"),"eax"); # save Xi 462 &mov ("edx",&DWP(12,"eax")); # load Xi 476 #&mov (&DWP(528+12,"esp"),$dat); # save inp^Xi 586 &mov ("eax",&DWP(528+16+0,"esp")); # restore Xi 608 ($Xi,$Xhi)=("xmm0","xmm1"); $Hkey="xmm2"; 615 my ($Xhi,$Xi,$Hkey,$HK)=@_; 617 &movdqa ($Xhi,$Xi); # 618 &pshufd ($T1,$Xi,0b01001110) [all...] |
ghash-x86_64.pl | 116 $Xi="%rdi"; 231 movzb 15($Xi),$Zlo 234 &loop ($Xi); 236 mov $Zlo,8($Xi) 237 mov $Zhi,($Xi) 301 &mov ($Zlo,"8($Xi)"); 302 &mov ($Zhi,"0($Xi)"); 312 &mov ("($Xi)",$Zhi); 313 &mov ("8($Xi)","%rdx"); 348 &mov ($dat,"$j($Xi)") if (--$j%4==0) [all...] |
aesni-gcm-x86_64.pl | 68 $Z0,$Z1,$Z2,$Z3,$Xi) = map("%xmm$_",(0..8)); 142 vpxor $Z0,$Xi,$Xi # modulo-scheduled 153 vpxor 16+8(%rsp),$Xi,$Xi # modulo-scheduled [vpxor $Z3,$Xi,$Xi] 217 vpxor 0x70+8(%rsp),$Xi,$Xi # accumulate I[0] 231 vpclmulqdq \$0x10,$Hkey,$Xi,$Z [all...] |
ghashv8-armx.pl | 45 $Xi="x0"; # argument block 131 # void gcm_gmult_v8(u64 Xi[2],const u128 Htable[16]); 133 # input: Xi - current hash value; 135 # output: Xi - next hash value Xi; 142 vld1.64 {$t1},[$Xi] @ load Xi 151 vpmull.p64 $Xl,$H,$IN @ H.lo·Xi.lo 153 vpmull2.p64 $Xh,$H,$IN @ H.hi·Xi.hi 154 vpmull.p64 $Xm,$Hhl,$t1 @ (H.lo+H.hi)·(Xi.lo+Xi.hi [all...] |
ghash-armv4.pl | 89 $Xi="r0"; # argument block 117 str $_,[$Xi,#$i] 119 str $_,[$Xi,#$i] 122 strb $_,[$Xi,#$i+3] 124 strb $Tlh,[$Xi,#$i+2] 126 strb $Thl,[$Xi,#$i+1] 127 strb $Thh,[$Xi,#$i] 177 ldrb $nhi,[$Xi,#15] 195 ldrb $nhi,[$Xi,#14] 228 ldrbpl $Tll,[$Xi,$cnt [all...] |
/external/boringssl/src/crypto/modes/ |
gcm.c | 131 static void gcm_gmult_4bit(uint64_t Xi[2], const u128 Htable[16]) { 136 nlo = ((const uint8_t *)Xi)[15]; 160 nlo = ((const uint8_t *)Xi)[cnt]; 177 Xi[0] = CRYPTO_bswap8(Z.hi); 178 Xi[1] = CRYPTO_bswap8(Z.lo); 186 static void gcm_ghash_4bit(uint64_t Xi[2], const u128 Htable[16], 194 nlo = ((const uint8_t *)Xi)[15]; 219 nlo = ((const uint8_t *)Xi)[cnt]; 237 Xi[0] = CRYPTO_bswap8(Z.hi); 238 Xi[1] = CRYPTO_bswap8(Z.lo) [all...] |
internal.h | 131 /* gmult_func multiplies |Xi| by the GCM key and writes the result back to 132 * |Xi|. */ 133 typedef void (*gmult_func)(uint64_t Xi[2], const u128 Htable[16]); 135 /* ghash_func repeatedly multiplies |Xi| by the GCM key and adds in blocks from 136 * |inp|. The result is written back to |Xi| and the |len| argument must be a 138 typedef void (*ghash_func)(uint64_t Xi[2], const u128 Htable[16], 151 } Yi, EKi, EK0, len, Xi; 153 /* Note that the order of |Xi|, |H| and |Htable| is fixed by the MOVBE-based,
|
/external/boringssl/src/crypto/sha/asm/ |
sha1-586.pl | 171 # with xi, also note that e becomes 176 &lea($tmp1,&DWP(0x5a827999,$tmp1,$e)); # tmp1+=K_00_19+xi 192 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd) 202 &mov(&swtmp($n%16),$f); # xi=f 209 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd) 219 &mov(&swtmp($n%16),$f); # xi=f 236 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd) 244 &mov(&swtmp($n%16),$f) if($n<77);# xi=f 254 &xor($f,&swtmp(($n+2)%16)); # f to hold Xupdate(xi,xa,xb,xc,xd) 264 &mov(&swtmp($n%16),$f) if($n<77);# xi= [all...] |
sha1-armv4-large.pl | 99 $Xi="r14"; 105 ldr $t0,[$Xi,#15*4] 106 ldr $t1,[$Xi,#13*4] 107 ldr $t2,[$Xi,#7*4] 109 ldr $t3,[$Xi,#2*4] 116 str $t0,[$Xi,#-4]! 149 str $t0,[$Xi,#-4]! 208 mov $Xi,sp 219 teq $Xi,sp 238 teq $Xi,sp @ preserve carr [all...] |
sha1-x86_64.pl | 127 @xi=("%edx","%ebp","%r14d"); 140 mov `4*$i`($inp),$xi[0] 141 bswap $xi[0] 144 mov `4*$j`($inp),$xi[1] 146 mov $xi[0],`4*$i`(%rsp) 148 bswap $xi[1] 152 lea 0x5a827999($xi[0],$e),$e 159 xor `4*($j%16)`(%rsp),$xi[1] 161 mov $xi[0],`4*($i%16)`(%rsp) 163 xor `4*(($j+2)%16)`(%rsp),$xi[1 [all...] |
/external/boringssl/linux-arm/crypto/modes/ |
ghashv8-armx32.S | 63 vld1.64 {q9},[r0] @ load Xi 72 .byte 0x86,0x0e,0xa8,0xf2 @ pmull q0,q12,q3 @ H.lo·Xi.lo 74 .byte 0x87,0x4e,0xa9,0xf2 @ pmull2 q2,q12,q3 @ H.hi·Xi.hi 75 .byte 0xa2,0x2e,0xaa,0xf2 @ pmull q1,q13,q9 @ (H.lo+H.hi)·(Xi.lo+Xi.hi) 96 vst1.64 {q0},[r0] @ write out Xi 106 vld1.64 {q0},[r0] @ load [rotated] Xi 126 vext.8 q0,q0,q0,#8 @ rotate Xi 140 veor q3,q3,q0 @ I[i]^=Xi 150 .byte 0x86,0x0e,0xac,0xf2 @ pmull q0,q14,q3 @ H^2.lo·Xi.l [all...] |
ghash-armv4.S | 350 vld1.64 d7,[r0]! @ load Xi 369 vld1.64 d1,[r0]! @ load Xi 386 veor q3,q0 @ inp^=Xi 529 vst1.64 d1,[r0]! @ write out Xi
|
/external/eigen/Eigen/src/Core/products/ |
SelfadjointMatrixVector.h | 121 Packet Xi = pload <Packet>(resIt); 123 Xi = pcj0.pmadd(A0i,ptmp0, pcj0.pmadd(A1i,ptmp1,Xi)); 126 pstore(resIt,Xi); resIt += PacketSize;
|
/packages/apps/UnifiedEmail/src/com/google/android/mail/common/base/ |
CharEscapers.java | 278 .addEscape('\u039E', "Ξ") 302 .addEscape('\u03BE', "ξ") [all...] |
/external/webrtc/data/voice_engine/stereo_rtp_files/ |
toggling_stereo_g729_pt18_pt125.rtp | 26 ?on? ??Y ?.eI??Xu?? 9s@W< |