Home | History | Annotate | Download | only in x86_64

Lines Matching refs:scratch

1864   X86_64ManagedRegister scratch = mscratch.AsX86_64();
1865 CHECK(scratch.IsCpuRegister());
1866 leaq(scratch.AsCpuRegister(), Address(CpuRegister(RSP), fr_offs));
1867 gs()->movq(Address::Absolute(thr_offs, true), scratch.AsCpuRegister());
1875 FrameOffset /*in_off*/, ManagedRegister /*scratch*/) {
2013 X86_64ManagedRegister scratch = mscratch.AsX86_64();
2014 CHECK(scratch.IsCpuRegister());
2015 movl(scratch.AsCpuRegister(), Address(CpuRegister(RSP), src));
2016 movl(Address(CpuRegister(RSP), dest), scratch.AsCpuRegister());
2022 X86_64ManagedRegister scratch = mscratch.AsX86_64();
2023 CHECK(scratch.IsCpuRegister());
2024 gs()->movq(scratch.AsCpuRegister(), Address::Absolute(thr_offs, true));
2025 Store(fr_offs, scratch, 8);
2031 X86_64ManagedRegister scratch = mscratch.AsX86_64();
2032 CHECK(scratch.IsCpuRegister());
2033 Load(scratch, fr_offs, 8);
2034 gs()->movq(Address::Absolute(thr_offs, true), scratch.AsCpuRegister());
2040 X86_64ManagedRegister scratch = mscratch.AsX86_64();
2041 if (scratch.IsCpuRegister() && size == 8) {
2042 Load(scratch, src, 4);
2043 Store(dest, scratch, 4);
2044 Load(scratch, FrameOffset(src.Int32Value() + 4), 4);
2045 Store(FrameOffset(dest.Int32Value() + 4), scratch, 4);
2047 Load(scratch, src, size);
2048 Store(dest, scratch, size);
2053 ManagedRegister /*scratch*/, size_t /*size*/) {
2058 ManagedRegister scratch, size_t size) {
2059 CHECK(scratch.IsNoRegister());
2067 CpuRegister scratch = mscratch.AsX86_64().AsCpuRegister();
2069 movq(scratch, Address(CpuRegister(RSP), src_base));
2070 movq(scratch, Address(scratch, src_offset));
2071 movq(Address(CpuRegister(RSP), dest), scratch);
2076 ManagedRegister scratch, size_t size) {
2078 CHECK(scratch.IsNoRegister());
2085 CpuRegister scratch = mscratch.AsX86_64().AsCpuRegister();
2088 movq(scratch, Address(CpuRegister(RSP), src));
2089 pushq(Address(scratch, src_offset));
2090 popq(Address(scratch, dest_offset));
2131 X86_64ManagedRegister scratch = mscratch.AsX86_64();
2132 CHECK(scratch.IsCpuRegister());
2135 movl(scratch.AsCpuRegister(), Address(CpuRegister(RSP), handle_scope_offset));
2136 testl(scratch.AsCpuRegister(), scratch.AsCpuRegister());
2138 leaq(scratch.AsCpuRegister(), Address(CpuRegister(RSP), handle_scope_offset));
2141 leaq(scratch.AsCpuRegister(), Address(CpuRegister(RSP), handle_scope_offset));
2143 Store(out_off, scratch, 8);
2179 CpuRegister scratch = mscratch.AsX86_64().AsCpuRegister();
2180 movl(scratch, Address(CpuRegister(RSP), base));
2181 call(Address(scratch, offset));
2193 X86_64ManagedRegister scratch = mscratch.AsX86_64();
2194 gs()->movq(scratch.AsCpuRegister(), Address::Absolute(Thread::SelfOffset<8>(), true));
2195 movq(Address(CpuRegister(RSP), offset), scratch.AsCpuRegister());
2207 void X86_64Assembler::ExceptionPoll(ManagedRegister /*scratch*/, size_t stack_adjust) {