Home | History | Annotate | Download | only in codeflinger

Lines Matching refs:AL

208             MOV(AL, 0, parts.count.reg,
210 ADD(AL, 0, parts.count.reg, parts.count.reg,
212 MOV(AL, 0, parts.count.reg,
265 AND(AL, 0, parts.dither.reg, parts.count.reg, imm(mask));
266 ADD(AL, 0, parts.dither.reg, parts.dither.reg, ctxtReg);
267 LDRB(AL, parts.dither.reg, parts.dither.reg,
324 SUB(AL, S, parts.count.reg, parts.count.reg, imm(1<<16));
339 ADD(AL, 0, parts.cbPtr.reg, parts.cbPtr.reg, imm(parts.cbPtr.size>>3));
341 SUB(AL, S, parts.count.reg, parts.count.reg, imm(1<<16));
367 SUB(AL, 0, parts.count.reg, parts.count.reg, Rx);
368 SUB(AL, 0, parts.count.reg, parts.count.reg, imm(1));
378 AND(AL, 0, tx, Rx, imm(GGL_DITHER_MASK));
379 AND(AL, 0, ty, Ry, imm(GGL_DITHER_MASK));
380 ADD(AL, 0, tx, tx, reg_imm(ty, LSL, GGL_DITHER_ORDER_SHIFT));
381 ORR(AL, 0, parts.count.reg, tx, reg_imm(parts.count.reg, LSL, 16));
385 MOV(AL, 0, parts.count.reg, reg_imm(parts.count.reg, LSL, 16));
396 SMLABB(AL, Rs, Ry, Rs, Rx); // Rs = Rx + Ry*Rs
411 MLA(AL, 0, f, Rx, dfdx, ydfdy);
424 MLA(AL, 0, parts.z.reg, Rx, dzdx, ydzdy);
432 SMLABB(AL, Rs, Ry, Rs, Rx);
433 ADD(AL, 0, Rs, Rs, reg_imm(parts.count.reg, LSR, 16));
434 ADD(AL, 0, zbase, zbase, reg_imm(Rs, LSL, 1));
449 ADD(AL, 0, parts.covPtr.reg, parts.covPtr.reg, reg_imm(Rx, LSL, 1));
556 MOV(AL, 0, fragment.reg, reg_imm(incoming.reg, LSR, incoming.l));
575 MOV(AL, 0, mAlphaSource.reg,
584 MOV(AL, 0, mAlphaSource.reg,
587 MOV(AL, 0, mAlphaSource.reg, fragment.reg);
654 ADD(AL, 0, c, c, dx);
683 LDRH(AL, cf, parts.covPtr.reg, immed8_post(2));
686 SMULWB(AL, fragment.reg, incoming.reg, cf);
688 MOV(AL, 0, fragment.reg, reg_imm(incoming.reg, LSL, 1));
689 SMULWB(AL, fragment.reg, fragment.reg, cf);
705 if (shift) CMP(AL, fragment.reg, reg_imm(ref, LSR, shift));
706 else CMP(AL, fragment.reg, ref);
732 int cc=AL, ic=AL;
744 B(AL, "discard_before_textures");
769 SUB(AL, 0, zbase, zbase, reg_imm(parts.count.reg, LSR, 15));
773 LDRH(AL, depth, zbase); // stall
774 CMP(AL, depth, reg_imm(z, LSR, 16));
780 ic = AL;
782 MOV(AL, 0, depth, reg_imm(z, LSR, 16));
795 ADD(AL, 0, parts.z.reg, parts.z.reg, dzdx);
808 ADD(AL, 0, f, f, dfdx);
832 case GGL_CLEAR: MOV(AL, 0, pixel.reg, imm(0)); break;
833 case GGL_AND: AND(AL, 0, pixel.reg, s.reg, d.reg); break;
834 case GGL_AND_REVERSE: BIC(AL, 0, pixel.reg, s.reg, d.reg); break;
836 case GGL_AND_INVERTED: BIC(AL, 0, pixel.reg, d.reg, s.reg); break;
837 case GGL_NOOP: MOV(AL, 0, pixel.reg, d.reg); break;
838 case GGL_XOR: EOR(AL, 0, pixel.reg, s.reg, d.reg); break;
839 case GGL_OR: ORR(AL, 0, pixel.reg, s.reg, d.reg); break;
840 case GGL_NOR: ORR(AL, 0, pixel.reg, s.reg, d.reg);
841 MVN(AL, 0, pixel.reg, pixel.reg); break;
842 case GGL_EQUIV: EOR(AL, 0, pixel.reg, s.reg, d.reg);
843 MVN(AL, 0, pixel.reg, pixel.reg); break;
844 case GGL_INVERT: MVN(AL, 0, pixel.reg, d.reg); break;
846 BIC(AL, 0, pixel.reg, d.reg, s.reg);
847 MVN(AL, 0, pixel.reg, pixel.reg); break;
848 case GGL_COPY_INVERTED: MVN(AL, 0, pixel.reg, s.reg); break;
850 BIC(AL, 0, pixel.reg, s.reg, d.reg);
851 MVN(AL, 0, pixel.reg, pixel.reg); break;
852 case GGL_NAND: AND(AL, 0, pixel.reg, s.reg, d.reg);
853 MVN(AL, 0, pixel.reg, pixel.reg); break;
854 case GGL_SET: MVN(AL, 0, pixel.reg, imm(0)); break;
892 MOV( AL, 0, d, s);
901 AND( AL, 0, d, s, imm(mask) );
920 AND( AL, 0, d, s, imm(newMask) );
922 BIC( AL, 0, d, s, imm(newMask) );
927 MOV( AL, 0, d, imm(0));
979 MOV(AL, 0, pixel.reg, fb.reg);
982 ORR(AL, 0, pixel.reg, s.reg, fb.reg);
993 ADD(AL, 0, d.reg, b.reg, reg_imm(o.reg, LSL, 2));
997 ADD(AL, 0, d.reg, b.reg, reg_imm(o.reg, LSL, 1));
998 ADD(AL, 0, d.reg, d.reg, o.reg);
1000 ADD(AL, 0, d.reg, o.reg, reg_imm(o.reg, LSL, 1));
1001 ADD(AL, 0, d.reg, d.reg, b.reg);
1005 ADD(AL, 0, d.reg, b.reg, reg_imm(o.reg, LSL, 1));
1008 ADD(AL, 0, d.reg, b.reg, o.reg);