/art/compiler/dex/quick/ |
mir_to_lir.cc | 141 StoreBaseDisp(TargetPtrReg(kSp), offset, reg_arg, k32, kNotVolatile); local 147 LoadBaseDisp(TargetPtrReg(kSp), offset, reg_arg, wide ? k64 : k32, kNotVolatile); 191 StoreBaseDisp(TargetPtrReg(kSp), offset, reg_arg, k32, kNotVolatile); local 196 OpSize op_size = rl_dest.wide ? k64 : (rl_dest.ref ? kReference : k32); 214 (arg.IsWide() && reg_arg.GetWideKind() == kWide) ? k64 : k32; 226 (arg.IsWide() && reg_arg.GetWideKind() == kWide) ? k64 : k32; 253 size = in_to_reg_storage_mapping_.GetShorty(data.src_arg).IsFP() ? kSingle : k32; 325 size = in_to_reg_storage_mapping_.GetShorty(data.src_arg).IsFP() ? kSingle : k32; 732 GenArrayGet(opt_flags, rl_dest.fp ? kSingle : k32, rl_src[0], rl_src[1], rl_dest, 2); 750 GenArrayPut(opt_flags, rl_src[0].fp ? kSingle : k32, rl_src[1], rl_src[2], rl_src[0], 2, false) [all...] |
gen_invoke.cc | 471 StoreBaseDisp(TargetPtrReg(kSp), offset, reg, t_loc->wide ? k64 : k32, kNotVolatile); 481 LoadBaseDisp(TargetPtrReg(kSp), offset, t_loc->reg, t_loc->wide ? k64 : k32, 715 StoreBaseDisp(TargetPtrReg(kSp), SRegOffset(loc.s_reg_low), loc.reg, k32, local 815 StoreBaseDisp(TargetPtrReg(kSp), out_offset, rl_arg.reg, k32, kNotVolatile); local 823 StoreBaseDisp(TargetPtrReg(kSp), out_offset, regSingle, k32, kNotVolatile); local 851 LoadBaseDisp(TargetPtrReg(kSp), out_offset, reg, k32, kNotVolatile); local 855 LoadBaseDisp(TargetPtrReg(kSp), SRegOffset(rl_arg.s_reg_low + 1), reg, k32, local 857 StoreBaseDisp(TargetPtrReg(kSp), high_offset, reg, k32, kNotVolatile); local 859 LoadBaseDisp(TargetPtrReg(kSp), SRegOffset(rl_arg.s_reg_low), reg, k32, kNotVolatile); local 862 StoreBaseDisp(TargetPtrReg(kSp), low_offset, reg, k32, kNotVolatile); local [all...] |
dex_file_method_inliner.cc | 365 INTRINSIC(JavaLangInteger, ReverseBytes, I_I, kIntrinsicReverseBytes, k32), 368 INTRINSIC(JavaLangInteger, Reverse, I_I, kIntrinsicReverseBits, k32), 423 INTRINSIC(LibcoreIoMemory, PeekIntNative, J_I, kIntrinsicPeek, k32), 427 INTRINSIC(LibcoreIoMemory, PokeIntNative, JI_V, kIntrinsicPoke, k32), [all...] |
gen_loadstore.cc | 62 op_size = k32;
|
gen_common.cc | 614 LoadBaseIndexed(r_src, r_idx, r_val, 2, k32); 618 StoreBaseIndexed(r_dst, r_idx, r_val, 2, k32); 741 case k32: [all...] |
mir_to_lir.h | [all...] |
/art/compiler/dex/quick/arm/ |
utility_arm.cc | 758 DCHECK((size == k32) || (size == kSingle) || (size == kReference)); 769 size = k32; 786 case k32: 824 DCHECK((size == k32) || (size == kSingle) || (size == kReference)); 836 size = k32; 853 case k32: 936 case k32: [all...] |
int_arm.cc | 802 DCHECK(size == kSignedByte || size == kSignedHalf || size == k32); 818 StoreBaseDisp(rl_address.reg, 0, rl_value.reg.GetLow(), k32, kNotVolatile); 819 StoreBaseDisp(rl_address.reg, 4, rl_value.reg.GetHigh(), k32, kNotVolatile); 821 DCHECK(size == kSignedByte || size == kSignedHalf || size == k32); [all...] |
call_arm.cc | 132 LoadBaseIndexed(table_base, keyReg, disp_reg, 2, k32);
|
/art/compiler/dex/quick/mips/ |
utility_mips.cc | 579 DCHECK((size == k32) || (size == kSingle) || (size == kReference)); 583 size = k32; 617 case k32: 651 DCHECK((size == k32) || (size == kSingle) || (size == kReference)); 655 size = k32; 670 case k32: 727 case k32: 862 size = cu_->target64 ? k64 : k32; 904 case k32: 1020 size = cu_->target64 ? k64 : k32; [all...] |
call_mips.cc | 202 LoadBaseIndexed(r_base, r_key, r_disp, 2, k32);
|
/external/bouncycastle/bcprov/src/main/java/org/bouncycastle/crypto/engines/ |
TwofishEngine.java | 411 // fall through, having pre-processed b[0]..b[3] with k32[3] 417 // fall through, having pre-processed b[0]..b[3] with k32[2] 523 private int F32(int x, int[] k32) 529 int k0 = k32[0]; 530 int k1 = k32[1]; 531 int k2 = k32[2]; 532 int k3 = k32[3];
|
/external/boringssl/src/crypto/modes/asm/ |
ghash-armv4.pl | 344 my ($Hlo,$Hhi,$Hhl,$k48,$k32,$k16)=map("d$_",(26..31)); 367 vand $t1#hi, $t1#hi, $k32 425 vmov.i64 $k32,#0x00000000ffffffff 444 vmov.i64 $k32,#0x00000000ffffffff
|
/art/compiler/dex/quick/arm64/ |
utility_arm64.cc | [all...] |
call_arm64.cc | 135 LoadBaseIndexed(table_base, As64BitReg(key_reg), disp_reg, 2, k32);
|
int_arm64.cc | 718 DCHECK(size == kSignedByte || size == kSignedHalf || size == k32); 733 DCHECK(size == kSignedByte || size == kSignedHalf || size == k32); [all...] |
/art/compiler/dex/quick/x86/ |
target_x86.cc | 743 OpSize size = cu_->target64 ? k64 : k32; 763 OpSize size = cu_->target64 ? k64 : k32; [all...] |
fp_x86.cc | 387 StoreBaseDisp(rs_rSP, src1_v_reg_offset, rl_src1.reg, is_double ? k64 : k32, 398 StoreBaseDisp(rs_rSP, src2_v_reg_offset, rl_src2.reg, is_double ? k64 : k32,
|
utility_x86.cc | 671 FALLTHROUGH_INTENDED; // else fall-through to k32 case 672 case k32: 825 FALLTHROUGH_INTENDED; // else fall-through to k32 case 826 case k32: [all...] |
int_x86.cc | [all...] |
/art/compiler/dex/ |
compiler_enums.h | 382 k32, enumerator in enum:art::OpSize
|
/art/compiler/optimizing/ |
intrinsics.cc | 54 case k32:
|
/prebuilts/devtools/tools/lib/ |
jobb.jar | |
/external/valgrind/coregrind/ |
m_transtab.c | 1366 UInt k32 = kHi ^ kLo; local [all...] |
/system/keymaster/ |
ocb.c | 206 const vector unsigned k32 = {32, 32, 32, 32}; local 222 vector unsigned rshift = vec_sub(k32, lshift); [all...] |