/art/compiler/utils/arm/ |
jni_macro_assembler_arm_vixl.cc | 37 #define ___ asm_.GetVIXLAssembler()-> 68 asm_.FinalizeCode(); 119 asm_.StoreToOffset(kStoreWord, r0, sp, 0); 129 asm_.StoreToOffset(kStoreWord, AsVIXLRegister(reg), sp, offset); 132 asm_.StoreSToOffset(AsVIXLSRegister(reg), sp, offset); 135 asm_.StoreDToOffset(AsVIXLDRegister(reg), sp, offset); 196 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); 201 asm_.GenerateMarkingRegisterCheck(temp); 217 asm_.AddConstant(sp, -adjust); 222 asm_.AddConstant(sp, adjust) [all...] |
/art/compiler/utils/ |
jni_macro_assembler.h | 263 asm_.FinalizeCode(); 267 return asm_.CodeSize(); 271 asm_.FinalizeInstructions(region); 275 return asm_.cfi(); 279 explicit JNIMacroAssemblerFwd(ArenaAllocator* allocator) : asm_(allocator) {} 281 T asm_; member in class:art::JNIMacroAssemblerFwd
|
assembler_thumb_test.cc | 300 #define __ assembler.asm_. 322 vixl::aarch32::UseScratchRegisterScope temps(assembler.asm_.GetVIXLAssembler()); 355 vixl::aarch32::UseScratchRegisterScope temps(assembler.asm_.GetVIXLAssembler());
|
/external/v8/src/ic/ |
accessor-assembler.h | 325 : asm_(assembler), indirect_return_handler_(indirect_return_handler) {} 340 asm_->TailCallRuntime(function, context, args...); 342 indirect_return_handler_(asm_->CallRuntime(function, context, args...)); 349 asm_->TailCallStub(callable, context, args...); 351 indirect_return_handler_(asm_->CallStub(callable, context, args...)); 359 asm_->TailCallStub(descriptor, target, context, args...); 362 asm_->CallStub(descriptor, target, context, args...)); 368 asm_->Return(result); 377 CodeStubAssembler* const asm_; member in class:v8::internal::ExitPoint
|
/external/compiler-rt/lib/tsan/ |
analyze_libtsan.sh | 38 file=${OUTPUT_DIR}/asm_$f.s
|
/external/v8/src/wasm/baseline/ |
liftoff-assembler.cc | 24 #define __ asm_-> 77 explicit StackTransferRecipe(LiftoffAssembler* wasm_asm) : asm_(wasm_asm) {} 88 asm_->Move(rm.dst, rm.src, rm.type); 103 uint32_t next_spill_slot = asm_->cache_state()->stack_height(); 108 asm_->Move(rm.dst, rm.src, rm.type); 122 asm_->Spill(next_spill_slot, spill_reg, rm.type); 138 asm_->LoadConstant(rl.dst, rl.type == kWasmI64 143 asm_->Fill(rl.dst, rl.value, rl.type); 147 asm_->FillI64Half(rl.dst.gp(), rl.value); 164 asm_->MoveStackValue(dst_index, src_index, src.type()) 262 LiftoffAssembler* const asm_; member in namespace:v8::internal::wasm::__anon48776 [all...] |
liftoff-compiler.cc | 34 #define __ asm_. 151 void GetCode(CodeDesc* desc) { asm_.GetCode(nullptr, desc); } 413 safepoint_table_builder_.DefineSafepoint(&asm_, Safepoint::kSimple, 0, 432 safepoint_table_builder_.Emit(&asm_, __ GetTotalFrameSlotCount()); 440 asm_.AbortCompilation(); 559 if (emit_fn && (asm_.*emit_fn)(dst.gp(), src.gp())) return; 573 if ((asm_.*emit_fn)(dst.fp(), src.fp())) return; 1806 LiftoffAssembler asm_; variable [all...] |
liftoff-assembler.h | 669 explicit LiftoffStackSlots(LiftoffAssembler* wasm_asm) : asm_(wasm_asm) {} 695 LiftoffAssembler* const asm_; member in class:v8::internal::wasm::LiftoffStackSlots
|
/art/compiler/utils/arm64/ |
jni_macro_assembler_arm64.cc | 32 #define ___ asm_.GetVIXLAssembler()-> 86 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); 177 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); 202 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); 308 asm_.MaybeUnpoisonHeapReference(reg_w(ref_reg)); 319 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); 665 UseScratchRegisterScope temps(asm_.GetVIXLAssembler()); 711 asm_.SpillRegisters(core_reg_list, frame_size - core_reg_size); 712 asm_.SpillRegisters(fp_reg_list, frame_size - core_reg_size - fp_reg_size); 771 asm_.UnspillRegisters(core_reg_list, frame_size - core_reg_size) [all...] |
/external/v8/src/wasm/baseline/arm64/ |
liftoff-assembler-arm64.h | 932 asm_->Claim(RoundUp(slot_count, 2)); 938 UseScratchRegisterScope temps(asm_); 940 asm_->Ldr(scratch, liftoff::GetStackSlot(slot.src_index_)); 941 asm_->Poke(scratch, poke_offset); 945 asm_->Poke(liftoff::GetRegFromType(slot.src_.reg(), slot.src_.type()), 952 asm_->Poke(zero_reg, poke_offset); 954 UseScratchRegisterScope temps(asm_); 957 asm_->Mov(scratch, int64_t{slot.src_.i32_const()}); 958 asm_->Poke(scratch, poke_offset);
|
/external/v8/src/wasm/baseline/mips/ |
liftoff-assembler-mips.h | [all...] |
/external/v8/src/wasm/baseline/arm/ |
liftoff-assembler-arm.h | 347 asm_->BAILOUT("LiftoffStackSlots::Construct");
|
/external/v8/src/wasm/baseline/ppc/ |
liftoff-assembler-ppc.h | 360 asm_->BAILOUT("LiftoffStackSlots::Construct");
|
/external/v8/src/wasm/baseline/s390/ |
liftoff-assembler-s390.h | 360 asm_->BAILOUT("LiftoffStackSlots::Construct");
|
/external/v8/src/wasm/baseline/mips64/ |
liftoff-assembler-mips64.h | [all...] |
/external/v8/src/wasm/baseline/x64/ |
liftoff-assembler-x64.h | [all...] |
/external/libvpx/libvpx/build/make/ |
gen_msvs_vcxproj.sh | 96 tag_content Command "$(eval echo \$asm_${cfg}_cmdline) -o \$(IntDir)$objf" \
|
/art/compiler/utils/x86/ |
jni_macro_assembler_x86.cc | 42 #define __ asm_. 120 DecreaseFrameSizeImpl(&asm_, adjust);
|
/art/compiler/utils/x86_64/ |
jni_macro_assembler_x86_64.cc | 36 #define __ asm_. 151 DecreaseFrameSizeImpl(adjust, &asm_);
|
/external/v8/src/wasm/baseline/ia32/ |
liftoff-assembler-ia32.h | [all...] |
/external/u-boot/scripts/ |
checkpatch.pl | [all...] |