HomeSort by relevance Sort by last modified time
    Searched full:spill (Results 1 - 25 of 583) sorted by null

1 2 3 4 5 6 7 8 91011>>

  /external/libunwind/tests/
ia64-test-rbs-asm.S 40 #define SPILL(n) \
123 SPILL(2); SPILL(3)
124 SPILL(4); SPILL(5); SPILL(6); SPILL(7)
125 SPILL(8); SPILL(9); SPILL(10); SPILL(11
    [all...]
  /external/libunwind/src/hppa/
getcontext.S 26 #define SPILL(n) stw %r##n, (LINUX_UC_MCONTEXT_OFF+LINUX_SC_GR_OFF+4*(n))(%r26)
36 SPILL (2) /* return-pointer */
37 SPILL (3) /* frame pointer */
38 SPILL (4) /* 2nd-ary frame pointer */
39 SPILL (5) /* preserved register */
40 SPILL (6) /* preserved register */
41 SPILL (7) /* preserved register */
42 SPILL (8) /* preserved register */
43 SPILL (9) /* preserved register */
44 SPILL (10) /* preserved register *
    [all...]
  /external/llvm/lib/CodeGen/
Spiller.h 22 /// Implementations are utility classes which insert spill or remat code on
29 /// spill - Spill the LRE.getParent() live interval.
30 virtual void spill(LiveRangeEdit &LRE) = 0;
34 /// Create and return a spiller that will insert spill code directly instead
RegAllocBasic.cpp 57 /// algorithm. It prioritizes live virtual registers by spill weight and spills
163 // Spill or split all live virtual registers currently unified under PhysReg
189 // Spill each interfering vreg allocated to PhysReg or an alias.
191 LiveInterval &Spill = *Intfs[i];
194 if (!VRM->hasPhys(Spill.reg))
199 Matrix->unassign(Spill);
201 // Spill the extracted interval.
202 LiveRangeEdit LRE(&Spill, SplitVRegs, *MF, *LIS, VRM);
203 spiller().spill(LRE);
222 // Populate a list of physical register spill candidates
    [all...]
  /external/libunwind/src/ia64/
getcontext.S 51 st8.spill [r2] = r1, (SC_FLAGS - GR(1)) // M3
66 st8.spill [r2] = r12, (GR(4) - GR(12)) // M3
70 stf.spill [r3] = f2 // M2
71 stf.spill [r8] = f16 // M3
76 stf.spill [r9] = f24, (FR(31) - FR(24)) // M2
80 stf.spill [r9] = f31 // M2
81 st8.spill [r2] = r4, (GR(5) - GR(4)) // M3, bank 1
85 .mem.offset 0,0; st8.spill [r2] = r5, (GR(6) - GR(5)) // M4, bank 0
86 .mem.offset 8,0; st8.spill [r3] = r7, (BR(0) - GR(7)) // M3, bank 0
90 st8.spill [r2] = r6, (BR(1) - GR(6)) // M2, bank
    [all...]
  /external/v8/test/mjsunit/regress/
regress-3884.js 8 // TurboFan will hoist the CompareIC for x === 'some_string' and spill it.
15 // TurboFan will hoist the CompareIC for x === 1.4 and spill it.
  /external/llvm/include/llvm/CodeGen/
CalcSpillWeights.h 25 /// \brief Normalize the spill weight of a live interval
27 /// The spill weight of a live interval is computed as:
40 // intervals have a spill weight that is mostly proportional to the number
41 // of uses, while large intervals get a spill weight that is closer to a use
47 /// spill weight and allocation hint.
68 /// \brief (re)compute li's spill weight and allocation hint.
72 /// \brief Compute spill weights and allocation hints for all virtual register
  /art/runtime/interpreter/mterp/x86/
op_mul_long_2addr.S 5 * this code. We must spill %edx (rIBASE) because it
6 * is used by imul. We'll also spill rINST (ebx),
8 * temps. On top of that, we'll spill %esi (edi)
op_mul_long.S 5 * this code. We spill rINSTw (ebx),
7 * temps. On top of that, we'll spill edi (rFP)
op_shl_long_2addr.S 7 /* Need to spill rIBASE */
op_shr_long_2addr.S 7 /* Need to spill rIBASE */
op_ushr_long_2addr.S 7 /* Need to spill rIBASE */
  /external/llvm/test/CodeGen/ARM/
varargs-spill-stack-align-nacl.ll 9 ; stack. A varargs function must therefore spill rN-r3 just below the
12 ; This test checks for a bug in which a gap was left between the spill
  /external/llvm/test/CodeGen/MIR/X86/
spill-slot-fixed-stack-objects.mir 22 # CHECK-NEXT: - { id: 0, type: spill-slot, offset: 0, size: 4, alignment: 4 }
24 - { id: 0, type: spill-slot, offset: 0, size: 4, alignment: 4 }
cfi-def-cfa-register.mir 21 - { id: 0, type: spill-slot, offset: -16, size: 8, alignment: 16 }
  /external/llvm/test/CodeGen/X86/
2013-10-14-FastISel-incorrect-vreg.ll 9 ; In this example, this is illustrated by a spill/reload of the
12 ; Before this patch, the compiler was accessing two different spill
19 ; Spill %arg2.
21 ; Spill %loaded_ptr.
58 ; Spill %arg2.
60 ; Spill %loaded_ptr.
97 ; Spill %arg2.
99 ; Spill %loaded_ptr.
avx512-intel-ocl.ll 64 ; WIN64: vmovups %zmm21, {{.*(%rbp).*}} # 64-byte Spill
65 ; WIN64: vmovups %zmm6, {{.*(%rbp).*}} # 64-byte Spill
71 ; X64: kmovw %k7, {{.*}}(%rsp) ## 8-byte Folded Spill
72 ; X64: kmovw %k6, {{.*}}(%rsp) ## 8-byte Folded Spill
73 ; X64: kmovw %k5, {{.*}}(%rsp) ## 8-byte Folded Spill
74 ; X64: kmovw %k4, {{.*}}(%rsp) ## 8-byte Folded Spill
75 ; X64: vmovups %zmm31, {{.*}}(%rsp) ## 64-byte Spill
76 ; X64: vmovups %zmm16, {{.*}}(%rsp) ## 64-byte Spill
sse-intel-ocl.ll 73 ; NOT_WIN: movaps {{%xmm([8-9]|1[0-5])}}, {{.*(%rsp).*}} ## 16-byte Spill
74 ; NOT_WIN: movaps {{%xmm([8-9]|1[0-5])}}, {{.*(%rsp).*}} ## 16-byte Spill
75 ; NOT_WIN: movaps {{%xmm([8-9]|1[0-5])}}, {{.*(%rsp).*}} ## 16-byte Spill
76 ; NOT_WIN: movaps {{%xmm([8-9]|1[0-5])}}, {{.*(%rsp).*}} ## 16-byte Spill
77 ; NOT_WIN: movaps {{%xmm([8-9]|1[0-5])}}, {{.*(%rsp).*}} ## 16-byte Spill
78 ; NOT_WIN: movaps {{%xmm([8-9]|1[0-5])}}, {{.*(%rsp).*}} ## 16-byte Spill
79 ; NOT_WIN: movaps {{%xmm([8-9]|1[0-5])}}, {{.*(%rsp).*}} ## 16-byte Spill
80 ; NOT_WIN: movaps {{%xmm([8-9]|1[0-5])}}, {{.*(%rsp).*}} ## 16-byte Spill
avx-intel-ocl.ll 69 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill
70 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill
71 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill
72 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill
73 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill
74 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill
75 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill
76 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill
77 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill
78 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill
    [all...]
unaligned-spill-folding.ll 33 ; We can't fold the spill into a padd unless the stack is aligned. Just spilling
37 ; UNALIGNED: movdqu {{.*}} # 16-byte Folded Spill
42 ; ALIGNED: movdqa {{.*}} # 16-byte Spill
47 ; FORCEALIGNED: movdqa {{.*}} # 16-byte Spill
  /external/llvm/test/CodeGen/AArch64/
arm64-neon-vector-list-spill.ll 3 ; FIXME: We should not generate ld/st for such register spill/fill, because the
5 ; spill/fill algorithm is optimized, this test case may not be triggered. And
7 define i32 @spill.DPairReg(i32* %arg1, i32 %arg2) {
8 ; CHECK-LABEL: spill.DPairReg:
27 define i16 @spill.DTripleReg(i16* %arg1, i32 %arg2) {
28 ; CHECK-LABEL: spill.DTripleReg:
47 define i16 @spill.DQuadReg(i16* %arg1, i32 %arg2) {
48 ; CHECK-LABEL: spill.DQuadReg:
67 define i32 @spill.QPairReg(i32* %arg1, i32 %arg2) {
68 ; CHECK-LABEL: spill.QPairReg
    [all...]
  /art/compiler/utils/
managed_register.h 119 ManagedRegisterSpill spill(__x);
120 std::vector<ManagedRegisterSpill>::push_back(spill);
124 ManagedRegisterSpill spill(__x, __size);
125 std::vector<ManagedRegisterSpill>::push_back(spill);
  /external/llvm/test/CodeGen/SPARC/
multiple-div.ll 4 ;; llc -O0 used to try to spill Y to the stack, which isn't possible,
  /external/llvm/test/CodeGen/SystemZ/Large/
spill-02.py 1 # Test cases where we spill from one frame index to another, both of which
2 # are out of range of MVC, and both of which need emergency spill slots.
17 # Arrange for %foo's spill slot to be at 8184(%r15) and the alloca area to be at
18 # 8192(%r15). The two emergency spill slots live below that, so this requires
  /external/llvm/test/MC/AArch64/
elf-extern.s 16 str x30, [sp, #8] // 8-byte Folded Spill

Completed in 1039 milliseconds

1 2 3 4 5 6 7 8 91011>>