Lines Matching refs:t9
269 ASSERT(!AreAliased(object, address, value, t9));
1025 ASSERT(!rs.is(t9));
1028 // Save rs's MSB to t9.
1029 Ext(t9, rs, 31, 1);
1042 Branch(&conversion_done, eq, t9, Operand(zero_reg));
1318 ASSERT(!(rd.is(t8) || rd.is(t9)) && !(rs.is(t8) || rs.is(t9)));
1320 Register scratch = t9;
1407 Register scratch2 = t9;
2380 // This is not an issue, t9 is expected to be clobbered anyway.
2381 li(t9, Operand(target, rmode));
2382 Jump(t9, al, zero_reg, Operand(zero_reg), bd);
2430 // Note: To call gcc-compiled C code on mips, you must call thru t9.
2461 int size = CallSize(t9, cond, rs, rt, bd);
2479 li(t9, Operand(target_int, rmode), CONSTANT_SIZE);
2480 Call(t9, cond, rs, rt, bd);
2736 sra(t9, a2, kSmiTagSize);
2737 Addu(t9, t9, a1);
2738 Jump(t9); // Jump.
2841 ASSERT(!scratch1.is(t9));
2842 ASSERT(!scratch2.is(t9));
2843 t9));
2869 // This code stores a temporary value in t9.
2871 // Load allocation top into result and allocation limit into t9.
2873 lw(t9, MemOperand(topaddr, kPointerSize));
2876 // Assert that result actually contains top on entry. t9 is used
2877 // immediately below so this use of t9 does not cause difference with
2879 lw(t9, MemOperand(topaddr));
2880 Check(eq, kUnexpectedAllocationTop, result, Operand(t9));
2882 // Load allocation limit into t9. Result already contains allocation top.
2883 lw(t9, MemOperand(topaddr, limit - top));
2895 Branch(gc_required, Ugreater_equal, result, Operand(t9));
2906 Branch(gc_required, Ugreater, scratch2, Operand(t9));
2936 ASSERT(!object_size.is(t9));
2937 ASSERT(!scratch1.is(t9) && !scratch2.is(t9) && !result.is(t9));
2956 // This code stores a temporary value in t9.
2958 // Load allocation top into result and allocation limit into t9.
2960 lw(t9, MemOperand(topaddr, kPointerSize));
2963 // Assert that result actually contains top on entry. t9 is used
2964 // immediately below so this use of t9 does not cause difference with
2966 lw(t9, MemOperand(topaddr));
2967 Check(eq, kUnexpectedAllocationTop, result, Operand(t9));
2969 // Load allocation limit into t9. Result already contains allocation top.
2970 lw(t9, MemOperand(topaddr, limit - top));
2982 Branch(gc_required, Ugreater_equal, result, Operand(t9));
2999 Branch(gc_required, Ugreater, scratch2, Operand(t9));
3003 And(t9, scratch2, Operand(kObjectAlignmentMask));
3004 Check(eq, kUnalignedAllocationInNewSpace, t9, Operand(zero_reg));
3937 li(t9, reinterpret_cast<int32_t>(is_profiling_flag));
3938 lb(t9, MemOperand(t9, 0));
3939 beq(t9, zero_reg, &profiler_disabled);
3943 li(t9, Operand(thunk_ref));
3947 li(t9, Operand(function));
3955 stub.GenerateCall(this, t9);
4113 ASSERT(!dst.is(t9));
4114 ASSERT(!scratch.is(t9));
4115 ASSERT(!left.is(t9));
4116 ASSERT(!right.is(t9));
4117 ASSERT(!overflow_dst.is(t9));
4118 mov(t9, right);
4119 right = t9;
4260 GetBuiltinEntry(t9, id);
4262 call_wrapper.BeforeCall(CallSize(t9));
4264 Call(t9);
4269 Jump(t9);
4544 // Load the stub address to t9 and call it,
4546 li(t9,
4550 jalr(t9, a0);
4566 li(t9, Operand(CodeObject()), CONSTANT_SIZE);
4571 sw(t9, MemOperand(sp, 0 * kPointerSize));
5174 // PrepareCallCFunction. The C function must be called via t9, for mips ABI.
5197 if (!function.is(t9)) {
5198 mov(t9, function);
5199 function = t9;
5229 srl(t9, new_value, kImm16Bits);
5230 Ins(scratch, t9, 0, kImm16Bits);
5318 ASSERT(!AreAliased(object, bitmap_scratch, mask_scratch, t9));
5323 lw(t9, MemOperand(bitmap_scratch, MemoryChunk::kHeaderSize));
5324 And(t8, t9, Operand(mask_scratch));
5329 And(t8, t9, Operand(mask_scratch));
5334 lw(t9, MemOperand(bitmap_scratch, MemoryChunk::kHeaderSize + kPointerSize));
5335 And(t9, t9, Operand(1));
5336 Branch(has_color, second_bit == 1 ? ne : eq, t9, Operand(zero_reg));
5460 lw(t9, FieldMemOperand(value, String::kLengthOffset));
5465 srl(t9, t9, 1);
5468 Addu(length, t9, Operand(SeqString::kHeaderSize + kObjectAlignmentMask));