Home | History | Annotate | Download | only in codeflinger

Lines Matching refs:mPC

170     mBase = mPC = (uint32_t *)assembly->base();
181 mBase = mPC = (uint32_t *)base;
196 return mPC;
207 mPC = mBase;
209 mBase = mPC = (uint32_t *)mAssembly->base();
258 mComments.add(mPC, string);
264 mLabels.add(theLabel, mPC);
265 mLabelsInverseMapping.add(mPC, theLabel);
271 mBranchTargets.add(branch_target_t(label, mPC));
273 *mPC++ = (0x54 << 24) | cc;
288 mPrologPC = mPC;
289 *mPC++ = A64_MOVZ_X(mZeroReg,0,0);
296 *mPC++ = A64_RET(XLR);
367 *mPC++ = A64_MOVZ_W(mTmpReg2, imm & 0x0000FFFF, 0);
368 *mPC++ = A64_MOVK_W(mTmpReg2, (imm >> 16) & 0x0000FFFF, 16);
398 case opADD: *mPC++ = A64_ADD_W(Rd, Rn, Rm, shift, amount); break;
399 case opAND: *mPC++ = A64_AND_W(Rd, Rn, Rm, shift, amount); break;
400 case opORR: *mPC++ = A64_ORR_W(Rd, Rn, Rm, shift, amount); break;
401 case opMVN: *mPC++ = A64_ORN_W(Rd, Rn, Rm, shift, amount); break;
402 case opSUB: *mPC++ = A64_SUB_W(Rd, Rn, Rm, shift, amount, s);break;
456 *mPC++ = A64_CSEL_W(Rd, mTmpReg1, Rd, cc);
474 *mPC++ = A64_ADD_X_Wm_SXTW(Rd, Rn, Rm, amount);
480 *mPC++ = A64_ADD_X_Wm_SXTW(Rd, Rn, Rm, amount);
485 *mPC++ = A64_MOVZ_W(mTmpReg1, imm & 0x0000FFFF, 0);
486 *mPC++ = A64_MOVK_W(mTmpReg1, (imm >> 16) & 0x0000FFFF, 16);
490 *mPC++ = A64_ADD_X_Wm_SXTW(Rd, Rn, Rm, amount);
506 *mPC++ = A64_ADD_W(mTmpReg1, mZeroReg, mAddrMode.reg_imm_Rm,
508 *mPC++ = A64_SUB_X_Wm_SXTW(Rd, Rn, mTmpReg1, 0);
523 *mPC++ = A64_MADD_W(Rd, Rm, Rs, Rn);
531 *mPC++ = A64_MADD_W(Rd, Rm, Rs, mZeroReg);
589 *mPC++ = A64_ADD_IMM_X(mTmpReg1, mZeroReg, imm, 0);
591 *mPC++ = A64_SUB_IMM_X(mTmpReg1, mZeroReg, -imm, 0);
601 *mPC++ = A64_ADD_X(mTmpReg2, addrReg, mTmpReg1);
607 *mPC++ = A64_B_COND(cc^1, 8);
609 *mPC++ = A64_LDRSTR_Wm_SXTW_0(op, size, Rd, addrReg, mZeroReg);
612 *mPC++ = A64_CSEL_X(Rn, mTmpReg2, Rn, cc);
617 *mPC++ = A64_B_COND(cc^1, 8);
618 *mPC++ = A64_LDRSTR_Wm_SXTW_0(op, size, Rd, Rn, mAddrMode.reg_offset);
624 *mPC++ = A64_B_COND(cc^1, 8);
625 *mPC++ = A64_LDRSTR_Wm_SXTW_0(op, size, Rd, Rn, mZeroReg);
697 *mPC++ = A64_LDR_IMM_PostIndex(reg, XSP, size);
718 *mPC++ = A64_STR_IMM_PreIndex(reg, XSP, size);
780 *mPC++ = A64_SBFM_W(mTmpReg1, Rm, 16, 31);
782 *mPC++ = A64_SBFM_W(mTmpReg1, Rm, 0, 15);
785 *mPC++ = A64_SBFM_W(mTmpReg2, Rs, 16, 31);
787 *mPC++ = A64_SBFM_W(mTmpReg2, Rs, 0, 15);
789 *mPC++ = A64_MADD_W(Rd,mTmpReg1,mTmpReg2, mZeroReg);
799 *mPC++ = A64_SBFM_W(mTmpReg1, Rs, 16, 31);
801 *mPC++ = A64_SBFM_W(mTmpReg1, Rs, 0, 15);
803 *mPC++ = A64_SBFM_W(mTmpReg2, Rm, 0, 31);
804 *mPC++ = A64_SMADDL(mTmpReg3,mTmpReg1,mTmpReg2, mZeroReg);
805 *mPC++ = A64_UBFM_X(Rd,mTmpReg3, 16, 47);
815 *mPC++ = A64_SBFM_W(mTmpReg1, Rm, 0, 15);
816 *mPC++ = A64_SBFM_W(mTmpReg2, Rs, 0, 15);
817 *mPC++ = A64_MADD_W(Rd, mTmpReg1, mTmpReg2, Rn);
841 *mPC++ = A64_EXTR_W(mTmpReg1, Rm, Rm, rotate * 8);
844 *mPC++ = A64_MOVZ_W(mTmpReg2, imm & 0xFFFF, 0);
845 *mPC++ = A64_MOVK_W(mTmpReg2, (imm >> 16) & 0x0000FFFF, 16);
846 *mPC++ = A64_AND_W(Rd,mTmpReg1, mTmpReg2);
855 *mPC++ = A64_UBFM_W(Rd, Rn, lsb, lsb + width - 1);