/external/v8/src/mips64/ |
code-stubs-mips64.cc | 28 __ sd(a1, MemOperand(t9, 0)); 65 MemOperand(sp, (param_count - 1 - i) * kPointerSize)); 94 __ ldc1(double_scratch, MemOperand(input_reg, double_offset)); 127 MemOperand(input_reg, double_offset + Register::kMantissaOffset)); 129 MemOperand(input_reg, double_offset + Register::kExponentOffset)); [all...] |
assembler-mips64.h | 356 // Class MemOperand represents a memory operand in load and store instructions. 357 class MemOperand : public Operand { 365 explicit MemOperand(Register rn, int32_t offset = 0); 366 explicit MemOperand(Register rn, int32_t unit, int32_t multiplier, 807 void lb(Register rd, const MemOperand& rs); 808 void lbu(Register rd, const MemOperand& rs); 809 void lh(Register rd, const MemOperand& rs); 810 void lhu(Register rd, const MemOperand& rs); 811 void lw(Register rd, const MemOperand& rs); 812 void lwu(Register rd, const MemOperand& rs) [all...] |
/external/v8/src/ppc/ |
code-stubs-ppc.cc | 28 __ StorePX(r4, MemOperand(sp, r0)); 89 __ lfd(double_scratch, MemOperand(input_reg, double_offset)); 112 MemOperand(input_reg, double_offset + Register::kExponentOffset)); 114 MemOperand(input_reg, double_offset + Register::kMantissaOffset)); 945 __ LoadP(ToRegister(ABI_TOC_REGISTER), MemOperand(r15, kPointerSize)); 946 __ LoadP(ip, MemOperand(r15, 0)); // Instruction address [all...] |
assembler-ppc.cc | 198 // Implementation of Operand and MemOperand 217 MemOperand::MemOperand(Register rn, int32_t offset) { 224 MemOperand::MemOperand(Register ra, Register rb) { 1142 void Assembler::lbz(Register dst, const MemOperand& src) { 1148 void Assembler::lbzx(Register rt, const MemOperand& src) { [all...] |
/art/compiler/linker/arm/ |
relative_patcher_thumb2.cc | 182 vixl::aarch32::MemOperand& lock_word, 205 using vixl::aarch32::MemOperand; 217 __ Ldr(entrypoint, MemOperand(tr, entry_point_offset)); 245 MemOperand lock_word(holder_reg, mirror::Object::MonitorOffset().Int32Value()); 256 MemOperand ldr_half_address(lr, ldr_offset + 2); 259 __ Ldr(ip, MemOperand(base_reg, ip)); // Load the reference. 261 MemOperand ldr_address(lr, ldr_offset); 267 __ Ldr(ip, MemOperand(base_reg, ip, LSL, 2)); // Load the reference. 275 __ Ldr(ip, MemOperand(/* Thread* */ vixl::aarch32::r9, offset)); 290 MemOperand lock_word(base_reg, mirror::Object::MonitorOffset().Int32Value() - data_offset) [all...] |
/external/vixl/examples/aarch64/ |
neon-matrix-multiply.cc | 70 __ Ld1(v4.V4S(), v5.V4S(), v6.V4S(), v7.V4S(), MemOperand(x1)); 72 __ Ld1(v16.V4S(), v17.V4S(), v18.V4S(), v19.V4S(), MemOperand(x2)); 89 __ St1(v0.V4S(), v1.V4S(), v2.V4S(), v3.V4S(), MemOperand(x0));
|
crc-checksums.cc | 57 __ Ldrb(w3, MemOperand(x2, 1, PostIndex));
|
sum-array.cc | 49 __ Ldrb(w3, MemOperand(x2, 1, PostIndex)); // w3 = *(x2++)
|
/external/v8/src/mips/ |
macro-assembler-mips.cc | 34 const MemOperand& src, 52 const MemOperand& dst, 71 lw(destination, MemOperand(s6, index << kPointerSizeLog2)); 80 lw(destination, MemOperand(s6, index << kPointerSizeLog2)); 87 sw(source, MemOperand(s6, index << kPointerSizeLog2)); 97 sw(source, MemOperand(s6, index << kPointerSizeLog2)); 168 MemOperand MacroAssembler::SafepointRegisterSlot(Register reg) { 169 return MemOperand(sp, SafepointRegisterStackIndex(reg.code()) * kPointerSize); 173 MemOperand MacroAssembler::SafepointRegistersAndDoublesSlot(Register reg) { 178 return MemOperand(sp, doubles_size + register_offset) [all...] |
macro-assembler-mips.h | 113 inline MemOperand ContextMemOperand(Register context, int index) { 114 return MemOperand(context, Context::SlotOffset(index)); 118 inline MemOperand NativeContextMemOperand() { 123 // Generate a MemOperand for loading a field from an object. 124 inline MemOperand FieldMemOperand(Register object, int offset) { 125 return MemOperand(object, offset - kHeapObjectTag); 129 // Generate a MemOperand for storing arguments 5..N on the stack 131 inline MemOperand CFunctionArgumentOperand(int index) { 135 return MemOperand(sp, offset); 325 void Load(Register dst, const MemOperand& src, Representation r) [all...] |
/art/compiler/optimizing/ |
intrinsics_arm_vixl.cc | 206 __ Ldr(tmp, MemOperand(src_curr_addr, element_size, PostIndex)); 233 __ Str(tmp, MemOperand(dst_curr_addr, element_size, PostIndex)); [all...] |
/external/v8/src/crankshaft/ppc/ |
lithium-gap-resolver-ppc.cc | 202 MemOperand source_operand = cgen_->ToMemOperand(source); 207 MemOperand destination_operand = cgen_->ToMemOperand(destination); 251 MemOperand source_operand = cgen_->ToMemOperand(source); 256 MemOperand destination_operand = cgen_->ToMemOperand(destination); 264 MemOperand source_high_operand = cgen_->ToHighMemOperand(source); 265 MemOperand destination_high_operand =
|
/external/v8/src/crankshaft/s390/ |
lithium-gap-resolver-s390.cc | 196 MemOperand source_operand = cgen_->ToMemOperand(source); 201 MemOperand destination_operand = cgen_->ToMemOperand(destination); 245 MemOperand source_operand = cgen_->ToMemOperand(source); 250 MemOperand destination_operand = cgen_->ToMemOperand(destination); 258 MemOperand source_high_operand = cgen_->ToHighMemOperand(source); 259 MemOperand destination_high_operand =
|
/external/v8/src/arm/ |
code-stubs-arm.cc | 29 __ str(r1, MemOperand(sp, r5)); 93 __ vldr(double_scratch, MemOperand(input_reg, double_offset)); 110 __ ldr(scratch_low, MemOperand(input_reg, double_offset)); 111 __ ldr(scratch_high, MemOperand(input_reg, double_offset + kIntSize)); 887 __ str(lr, MemOperand(sp, result_stack_size)); 893 __ ldr(r2, MemOperand(sp, 2 * kPointerSize)); 894 __ ldr(r1, MemOperand(sp, 1 * kPointerSize)); 895 __ ldr(r0, MemOperand(sp, 0 * kPointerSize)); [all...] |
macro-assembler-arm.h | 35 // Generate a MemOperand for loading a field from an object. 36 inline MemOperand FieldMemOperand(Register object, int offset) { 37 return MemOperand(object, offset - kHeapObjectTag); 192 void Load(Register dst, const MemOperand& src, Representation r); 193 void Store(Register src, const MemOperand& dst, Representation r); 283 // MemOperand(reg, off). 341 str(src1, MemOperand(sp, 4, NegPreIndex), cond); 342 str(src2, MemOperand(sp, 4, NegPreIndex), cond); 353 str(src3, MemOperand(sp, 4, NegPreIndex), cond); 356 str(src1, MemOperand(sp, 4, NegPreIndex), cond) [all...] |
/external/vixl/src/aarch64/ |
macro-assembler-aarch64.cc | [all...] |
macro-assembler-aarch64.h | 687 // Synthesises the address represented by a MemOperand into a register. 688 void ComputeAddress(const Register& dst, const MemOperand& mem_op); 731 void FN(const REGTYPE REG, const MemOperand& addr); 736 const MemOperand& addr, 740 void FN(const REGTYPE REG, const REGTYPE REG2, const MemOperand& addr); 746 const MemOperand& addr, 749 void Prfm(PrefetchOperation op, const MemOperand& addr); 852 LoadCPURegList(registers, MemOperand(StackPointer(), offset)); 855 StoreCPURegList(registers, MemOperand(StackPointer(), offset)); [all...] |
/external/v8/src/arm64/ |
macro-assembler-arm64.cc | 563 const MemOperand& addr, 578 LoadStore(rt, MemOperand(addr.base(), temp), op); 581 LoadStore(rt, MemOperand(addr.base()), op); 586 LoadStore(rt, MemOperand(addr.base()), op); 595 const MemOperand& addr, 614 LoadStorePair(rt, rt2, MemOperand(temp), op); 616 LoadStorePair(rt, rt2, MemOperand(base), op); 621 LoadStorePair(rt, rt2, MemOperand(base), op); 628 const MemOperand& addr, 650 const MemOperand& addr [all...] |
/art/compiler/trampolines/ |
trampoline_compiler.cc | 61 using vixl::aarch32::MemOperand; 68 ___ Ldr(pc, MemOperand(r0, offset.Int32Value())); 76 ___ Ldr(temp_reg, MemOperand(r0, JNIEnvExt::SelfOffset(4).Int32Value())); 77 ___ Ldr(pc, MemOperand(temp_reg, offset.Int32Value())); 81 ___ Ldr(pc, MemOperand(tr, offset.Int32Value()));
|
/external/v8/src/full-codegen/s390/ |
full-codegen-s390.cc | 122 __ LoadP(r4, MemOperand(sp, receiver_offset), r0); 174 __ lay(sp, MemOperand(sp, -kMaxPushes * kPointerSize)); 176 __ StoreP(ip, MemOperand(sp, i * kPointerSize)); 185 __ lay(sp, MemOperand(sp, -remaining * kPointerSize)); 187 __ StoreP(ip, MemOperand(sp, i * kPointerSize)); 235 __ StoreP(r2, MemOperand(fp, StandardFrameConstants::kContextOffset)); 246 __ LoadP(r2, MemOperand(fp, parameter_offset), r0); 248 MemOperand target = ContextMemOperand(cp, var->index()); 282 __ LoadP(r3, MemOperand(fp, JavaScriptFrameConstants::kFunctionOffset)); 453 __ LoadP(cp, MemOperand(fp, StandardFrameConstants::kContextOffset)) [all...] |
/external/v8/src/compiler/s390/ |
code-generator-s390.cc | 78 MemOperand MemoryOperand(AddressingMode* mode, size_t* first_index) { 86 return MemOperand(InputRegister(index + 0), 0); 89 return MemOperand(InputRegister(index + 0), InputInt32(index + 1)); 92 return MemOperand(InputRegister(index + 0), InputRegister(index + 1)); 95 return MemOperand(InputRegister(index + 0), InputRegister(index + 1), 99 return MemOperand(r0); 102 MemOperand MemoryOperand(AddressingMode* mode = NULL, 107 MemOperand ToMemOperand(InstructionOperand* op) const { 113 MemOperand SlotToMemOperand(int slot) const { 115 return MemOperand(offset.from_stack_pointer() ? sp : fp, offset.offset()) [all...] |
/external/v8/src/full-codegen/mips/ |
full-codegen-mips.cc | 126 __ lw(a2, MemOperand(sp, receiver_offset)); 177 __ sw(t5, MemOperand(sp, i * kPointerSize)); 187 __ sw(t5, MemOperand(sp, i * kPointerSize)); 234 __ sw(v0, MemOperand(fp, StandardFrameConstants::kContextOffset)); 245 __ lw(a0, MemOperand(fp, parameter_offset)); 247 MemOperand target = ContextMemOperand(cp, var->index()); 281 __ lw(a1, MemOperand(fp, JavaScriptFrameConstants::kFunctionOffset)); 460 __ lw(cp, MemOperand(fp, StandardFrameConstants::kContextOffset)); 556 __ sw(reg, MemOperand(sp, 0)); 661 MemOperand FullCodeGenerator::StackOperand(Variable* var) [all...] |
/external/v8/src/full-codegen/mips64/ |
full-codegen-mips64.cc | 126 __ ld(a2, MemOperand(sp, receiver_offset)); 176 __ sd(t1, MemOperand(sp, i * kPointerSize)); 186 __ sd(t1, MemOperand(sp, i * kPointerSize)); 233 __ sd(v0, MemOperand(fp, StandardFrameConstants::kContextOffset)); 244 __ ld(a0, MemOperand(fp, parameter_offset)); 246 MemOperand target = ContextMemOperand(cp, var->index()); 280 __ ld(a1, MemOperand(fp, JavaScriptFrameConstants::kFunctionOffset)); 460 __ ld(cp, MemOperand(fp, StandardFrameConstants::kContextOffset)); 556 __ sd(reg, MemOperand(sp, 0)); 661 MemOperand FullCodeGenerator::StackOperand(Variable* var) [all...] |
/external/v8/src/full-codegen/ppc/ |
full-codegen-ppc.cc | 116 __ LoadP(r5, MemOperand(sp, receiver_offset), r0); 231 __ StoreP(r3, MemOperand(fp, StandardFrameConstants::kContextOffset)); 242 __ LoadP(r3, MemOperand(fp, parameter_offset), r0); 244 MemOperand target = ContextMemOperand(cp, var->index()); 278 __ LoadP(r4, MemOperand(fp, JavaScriptFrameConstants::kFunctionOffset)); 454 __ LoadP(cp, MemOperand(fp, StandardFrameConstants::kContextOffset)); 544 __ StoreP(reg, MemOperand(sp, 0)); 635 MemOperand FullCodeGenerator::StackOperand(Variable* var) { 645 return MemOperand(fp, offset); 649 MemOperand FullCodeGenerator::VarOperand(Variable* var, Register scratch) [all...] |
/external/v8/src/compiler/arm64/ |
code-generator-arm64.cc | 162 MemOperand MemoryOperand(size_t* first_index) { 177 return MemOperand(InputRegister(index + 0), InputRegister(index + 1), 181 return MemOperand(InputRegister(index + 0), InputInt32(index + 1)); 184 return MemOperand(InputRegister(index + 0), InputRegister(index + 1)); 187 return MemOperand(no_reg); 190 MemOperand MemoryOperand(size_t first_index = 0) { 242 MemOperand ToMemOperand(InstructionOperand* op, MacroAssembler* masm) const { 248 MemOperand SlotToMemOperand(int slot, MacroAssembler* masm) const { 258 return MemOperand(offset.from_stack_pointer() ? masm->StackPointer() : fp, 438 __ Ldr(result, MemOperand(buffer, offset, UXTW)); [all...] |