/external/vixl/src/aarch64/ |
simulator-constants-aarch64.h | 86 const unsigned kPrintfArgCountOffset = 1 * kInstructionSize; 87 const unsigned kPrintfArgPatternListOffset = 2 * kInstructionSize; 88 const unsigned kPrintfLength = 3 * kInstructionSize; 110 const unsigned kTraceParamsOffset = 1 * kInstructionSize; 111 const unsigned kTraceCommandOffset = 2 * kInstructionSize; 112 const unsigned kTraceLength = 3 * kInstructionSize; 140 const unsigned kLogParamsOffset = 1 * kInstructionSize; 141 const unsigned kLogLength = 2 * kInstructionSize; 144 const unsigned kRuntimeCallWrapperOffset = 1 * kInstructionSize;
|
macro-assembler-aarch64.cc | 117 if (option == kBranchRequired) emit_size += kInstructionSize; 120 VIXL_ASSERT(emit_size % kInstructionSize == 0); 131 ExactAssemblyScopeWithoutPoolsCheck guard(masm_, kInstructionSize); 138 ExactAssemblyScopeWithoutPoolsCheck guard(masm_, kInstructionSize); 264 ExactAssemblyScopeWithoutPoolsCheck guard(masm_, kInstructionSize); 290 ExactAssemblyScopeWithoutPoolsCheck guard(masm_, kInstructionSize); 573 EmissionCheckScope guard(this, 2 * kInstructionSize); 594 EmissionCheckScope guard(this, 2 * kInstructionSize); 615 EmissionCheckScope guard(this, 2 * kInstructionSize); 636 EmissionCheckScope guard(this, 2 * kInstructionSize); [all...] |
instructions-aarch64.cc | 229 VIXL_ASSERT((LSSize_offset + LSSize_width) == (kInstructionSize * 8)); 279 return encoded_max * kInstructionSize; 305 offset = GetImmBranch() * static_cast<int>(kInstructionSize);
|
instructions-aarch64.h | 40 const unsigned kInstructionSize = 4; 408 // target [instr - range - kInstructionSize, instr + range]. 411 (1 << kLoadLiteralImmBitwidth) / 2 - kInstructionSize; 467 return this + kInstructionSize;
|
/external/vixl/benchmarks/aarch64/ |
bench-branch-link.cc | 54 MacroAssembler masm(instructions * kInstructionSize); 55 ExactAssemblyScope scope(&masm, instructions * kInstructionSize);
|
bench-branch.cc | 59 const int buffer_instruction_count = buffer_size / kInstructionSize; 82 ExactAssemblyScope scope(&masm, remaining * kInstructionSize);
|
bench-dataop.cc | 59 const unsigned buffer_instruction_count = buffer_size / kInstructionSize; 79 ExactAssemblyScope scope(&masm, remaining * kInstructionSize);
|
bench-branch-link-masm.cc | 53 size_t base_buf_size = iterations * 4 * kInstructionSize;
|
bench-branch-masm.cc | 60 buffer_size / (instructions_per_iteration * kInstructionSize);
|
/external/vixl/test/aarch64/ |
test-fuzz-aarch64.cc | 50 Instruction buffer[kInstructionSize]; 69 Instruction buffer[kInstructionSize]; 91 Instruction buffer[kInstructionSize]; 107 Instruction buffer[kInstructionSize];
|
/external/v8/src/arm64/ |
instructions-arm64.h | 125 return InstructionAtOffset(count * static_cast<int>(kInstructionSize)); 129 return InstructionAtOffset(count * static_cast<int>(kInstructionSize)); 311 kInstructionSize; 392 DCHECK(check == NO_CHECK || IsAligned(offset, kInstructionSize)); 399 DCHECK(check == NO_CHECK || IsAligned(offset, kInstructionSize)); 465 const unsigned kPrintfArgCountOffset = 1 * kInstructionSize; 466 const unsigned kPrintfArgPatternListOffset = 2 * kInstructionSize; 467 const unsigned kPrintfLength = 3 * kInstructionSize; 488 // kInstructionSize so that subsequent instructions are correctly aligned. 491 const unsigned kDebugCodeOffset = 1 * kInstructionSize; [all...] |
codegen-arm64.cc | 50 young_sequence_.length() / kInstructionSize); 56 const int length = kCodeAgeStubEntryOffset / kInstructionSize; 86 kNoCodeAgeSequenceLength / kInstructionSize);
|
instructions-arm64.cc | 254 DCHECK(IsAligned(DistanceTo(target), kInstructionSize)); 290 DCHECK(IsAligned(DistanceTo(target), kInstructionSize)); 305 DCHECK(IsAligned(DistanceTo(source), kInstructionSize));
|
deoptimizer-arm64.cc | 20 return 4 * kInstructionSize; 53 patch_size() / kInstructionSize); 54 patcher.ldr_pcrel(ip0, (2 * kInstructionSize) >> kLoadLiteralScaleLog2); 298 const int Deoptimizer::table_entry_size_ = 2 * kInstructionSize;
|
assembler-arm64-inl.h | 592 Address candidate = pc - 2 * kInstructionSize; 617 STATIC_ASSERT(Assembler::kCallSizeWithoutRelocation == 4 * kInstructionSize); 618 STATIC_ASSERT(Assembler::kCallSizeWithRelocation == 2 * kInstructionSize); 779 static const int kNoCodeAgeSequenceLength = 5 * kInstructionSize; 780 static const int kCodeAgeStubEntryOffset = 3 * kInstructionSize; 945 DCHECK(IsAligned(offset, kInstructionSize)); 1166 DCHECK((SizeLS_offset + SizeLS_width) == (kInstructionSize * 8)); [all...] |
assembler-arm64.cc | 366 return 4 * kInstructionSize + EntryCount() * kPointerSize; 378 int prologue_size = require_jump ? kInstructionSize : 0; 379 prologue_size += 2 * kInstructionSize; 381 0 : kInstructionSize; 940 return RoundUp(size, kInstructionSize) / kInstructionSize; 945 return kPrintfLength / kInstructionSize; [all...] |
assembler-arm64.h | 854 static const int kCallSizeWithoutRelocation = 4 * kInstructionSize; 855 static const int kCallSizeWithRelocation = 2 * kInstructionSize; 884 return SizeOfCodeGeneratedSince(label) / kInstructionSize; 892 kDebugBreakSlotInstructions * kInstructionSize; [all...] |
code-stubs-arm64.h | 137 Instruction* instr2 = patcher.InstructionAt(kInstructionSize);
|
/external/vixl/test/ |
test-code-generation-scopes.cc | 71 CodeBufferCheckScope scope(&masm, aarch64::kInstructionSize); 100 CodeBufferCheckScope scope(&masm, 2 * aarch64::kInstructionSize); 133 scope.Open(&masm, aarch64::kInstructionSize); 163 CodeBufferCheckScope scope(&masm, aarch64::kInstructionSize); 199 scope.Open(&masm, aarch64::kInstructionSize); 229 EmissionCheckScope scope(&masm, aarch64::kInstructionSize); 261 scope.Open(&masm, aarch64::kInstructionSize); 291 EmissionCheckScope scope(&masm, aarch64::kInstructionSize); 327 scope.Open(&masm, aarch64::kInstructionSize); 382 VIXL_CHECK((expected + aarch64::kInstructionSize) == [all...] |
/art/disassembler/ |
disassembler_arm64.cc | 116 return kInstructionSize; 120 for (const uint8_t* cur = begin; cur < end; cur += kInstructionSize) {
|
/external/vixl/examples/aarch64/ |
non-const-visitor.cc | 97 for (instr = instr_start; instr < instr_end; instr += kInstructionSize) { 122 for (instr = start; instr < end; instr += kInstructionSize) {
|
literal.cc | 46 kInstructionSize + sizeof(int64_t),
|
custom-disassembler.cc | 157 // custom_disasm.MapCodeAddress(0x0, instr_start + 2 * kInstructionSize); 165 for (instr = instr_start; instr < instr_end; instr += kInstructionSize) {
|
/external/v8/src/debug/arm64/ |
debug-arm64.cc | 59 // ldr ip0, [pc, #(2 * kInstructionSize)] 68 patcher.ldr_pcrel(ip0, (2 * kInstructionSize) >> kLoadLiteralScaleLog2);
|
/art/compiler/optimizing/ |
code_generator_arm64.cc | [all...] |