Home | History | Annotate | Download | only in mips

Lines Matching refs:a3

324     srl       a3, rPC, 12                  #  cheap, but fast hash function
325 xor a3, a3, rPC
326 andi a3, a3, JIT_PROF_SIZE-1 # eliminate excess bits
327 addu t1, a0, a3
419 move a3, rBIX # r3 <- target translation
430 move a3, rSELF # pass self ptr for convenience
506 EAS2(a3, rFP, a1)
511 lw a1, 0(a3)
512 addu a3, a3, 4
553 lw a3, (t2)
555 sw a3, 0(rBIX)
560 lw a3, (t2)
562 sw a3, 0(rBIX)
567 lw a3, (t2)
569 sw a3, 0(rBIX)
574 lw a3, (t2)
576 sw a3, 0(rBIX)
581 lw a3, (t2)
583 sw a3, 0(rBIX)
591 lhu a3, offMethod_outsSize(a0)
600 sll t2, a3, 2
603 lw a3, offMethod_accessFlags(a0) # a3 <- methodToCall->accessFlags
621 and t2, a3, ACC_NATIVE
624 lw a3, offClassObject_pDvmDex(rINST)
627 # a0=methodToCall, a1=newFp, a3=newMethodClass, rOBJ=newINST
629 sw a3, offThread_methodClassDex(rSELF)
654 STACK_STORE(a3, 12)
659 STACK_LOAD(a3, 12) # restore a0-a3
674 move a3, rSELF
700 # a0=newFp, a1=&retval, a2=methodToCall, a3=self, ra=subModes
704 SCRATCH_STORE(a3, 12)
709 SCRATCH_LOAD(a3, 12) # restore a0-a3
720 SCRATCH_LOAD(a3, 12) # restore a0-a3
850 li a3, 0 # a3 <- false
882 lw a3, offMethod_insns(a1)
884 EAS1(rPC, a3, a0)
934 lw a3, offObject_clazz(rOBJ)
935 lw a3, offClassObject_descriptor(a3)
970 lhu a3, offThread_subMode(rSELF)
971 andi a3, kSubModeJitTraceBuild
972 bnez a3, 1f # Not building trace, continue
982 SCRATCH_STORE(a3, 12)
990 SCRATCH_LOAD(a3, 12)
1008 * and a3 because those just happen to be the registers all our callers are
1009 * using. We move a3 before calling the C function, but a1 happens to match.
1011 * a3: size
1015 move a0, a3
1122 MOVE_REG(a3, a1)