Home | History | Annotate | Download | only in mips64

Lines Matching refs:MemOperand

144     __ lwr(t8, MemOperand(a1));
146 __ swr(t8, MemOperand(a0));
169 __ Pref(pref_hint_load, MemOperand(a1, 0 * pref_chunk));
170 __ Pref(pref_hint_load, MemOperand(a1, 1 * pref_chunk));
171 __ Pref(pref_hint_load, MemOperand(a1, 2 * pref_chunk));
172 __ Pref(pref_hint_load, MemOperand(a1, 3 * pref_chunk));
175 __ Pref(pref_hint_store, MemOperand(a0, 1 * pref_chunk));
176 __ Pref(pref_hint_store, MemOperand(a0, 2 * pref_chunk));
177 __ Pref(pref_hint_store, MemOperand(a0, 3 * pref_chunk));
180 __ lw(a4, MemOperand(a1));
186 __ lw(a5, MemOperand(a1, 1, loadstore_chunk)); // Maybe in delay slot.
188 __ Pref(pref_hint_store, MemOperand(a0, 4 * pref_chunk));
189 __ Pref(pref_hint_store, MemOperand(a0, 5 * pref_chunk));
192 __ lw(a6, MemOperand(a1, 2, loadstore_chunk));
193 __ lw(a7, MemOperand(a1, 3, loadstore_chunk));
194 __ lw(t0, MemOperand(a1, 4, loadstore_chunk));
195 __ lw(t1, MemOperand(a1, 5, loadstore_chunk));
196 __ lw(t2, MemOperand(a1, 6, loadstore_chunk));
197 __ lw(t3, MemOperand(a1, 7, loadstore_chunk));
198 __ Pref(pref_hint_load, MemOperand(a1, 4 * pref_chunk));
200 __ sw(a4, MemOperand(a0));
201 __ sw(a5, MemOperand(a0, 1, loadstore_chunk));
202 __ sw(a6, MemOperand(a0, 2, loadstore_chunk));
203 __ sw(a7, MemOperand(a0, 3, loadstore_chunk));
204 __ sw(t0, MemOperand(a0, 4, loadstore_chunk));
205 __ sw(t1, MemOperand(a0, 5, loadstore_chunk));
206 __ sw(t2, MemOperand(a0, 6, loadstore_chunk));
207 __ sw(t3, MemOperand(a0, 7, loadstore_chunk));
209 __ lw(a4, MemOperand(a1, 8, loadstore_chunk));
210 __ lw(a5, MemOperand(a1, 9, loadstore_chunk));
211 __ lw(a6, MemOperand(a1, 10, loadstore_chunk));
212 __ lw(a7, MemOperand(a1, 11, loadstore_chunk));
213 __ lw(t0, MemOperand(a1, 12, loadstore_chunk));
214 __ lw(t1, MemOperand(a1, 13, loadstore_chunk));
215 __ lw(t2, MemOperand(a1, 14, loadstore_chunk));
216 __ lw(t3, MemOperand(a1, 15, loadstore_chunk));
217 __ Pref(pref_hint_load, MemOperand(a1, 5 * pref_chunk));
219 __ sw(a4, MemOperand(a0, 8, loadstore_chunk));
220 __ sw(a5, MemOperand(a0, 9, loadstore_chunk));
221 __ sw(a6, MemOperand(a0, 10, loadstore_chunk));
222 __ sw(a7, MemOperand(a0, 11, loadstore_chunk));
223 __ sw(t0, MemOperand(a0, 12, loadstore_chunk));
224 __ sw(t1, MemOperand(a0, 13, loadstore_chunk));
225 __ sw(t2, MemOperand(a0, 14, loadstore_chunk));
226 __ sw(t3, MemOperand(a0, 15, loadstore_chunk));
236 __ Pref(pref_hint_load, MemOperand(a1, 0 * pref_chunk));
240 __ lw(a4, MemOperand(a1));
241 __ lw(a5, MemOperand(a1, 1, loadstore_chunk));
242 __ lw(a6, MemOperand(a1, 2, loadstore_chunk));
243 __ lw(a7, MemOperand(a1, 3, loadstore_chunk));
244 __ lw(t0, MemOperand(a1, 4, loadstore_chunk));
245 __ lw(t1, MemOperand(a1, 5, loadstore_chunk));
246 __ lw(t2, MemOperand(a1, 6, loadstore_chunk));
247 __ lw(t3, MemOperand(a1, 7, loadstore_chunk));
249 __ sw(a4, MemOperand(a0));
250 __ sw(a5, MemOperand(a0, 1, loadstore_chunk));
251 __ sw(a6, MemOperand(a0, 2, loadstore_chunk));
252 __ sw(a7, MemOperand(a0, 3, loadstore_chunk));
253 __ sw(t0, MemOperand(a0, 4, loadstore_chunk));
254 __ sw(t1, MemOperand(a0, 5, loadstore_chunk));
255 __ sw(t2, MemOperand(a0, 6, loadstore_chunk));
256 __ sw(t3, MemOperand(a0, 7, loadstore_chunk));
271 __ lw(a7, MemOperand(a1));
275 __ sw(a7, MemOperand(a0, -1, loadstore_chunk)); // In delay slot.
282 __ lb(v1, MemOperand(a1));
286 __ sb(v1, MemOperand(a0, -1)); // In delay slot.
300 __ lwr(v1, MemOperand(a1));
302 MemOperand(a1, 1, loadstore_chunk, MemOperand::offset_minus_one));
304 __ swr(v1, MemOperand(a0));
322 __ Pref(pref_hint_load, MemOperand(a1, 0 * pref_chunk));
323 __ Pref(pref_hint_load, MemOperand(a1, 1 * pref_chunk));
324 __ Pref(pref_hint_load, MemOperand(a1, 2 * pref_chunk));
327 __ Pref(pref_hint_store, MemOperand(a0, 1 * pref_chunk));
328 __ Pref(pref_hint_store, MemOperand(a0, 2 * pref_chunk));
329 __ Pref(pref_hint_store, MemOperand(a0, 3 * pref_chunk));
333 __ Pref(pref_hint_load, MemOperand(a1, 3 * pref_chunk));
334 __ lwr(a4, MemOperand(a1));
335 __ lwr(a5, MemOperand(a1, 1, loadstore_chunk));
336 __ lwr(a6, MemOperand(a1, 2, loadstore_chunk));
342 __ lwr(a7, MemOperand(a1, 3, loadstore_chunk)); // Maybe in delay slot.
344 __ Pref(pref_hint_store, MemOperand(a0, 4 * pref_chunk));
345 __ Pref(pref_hint_store, MemOperand(a0, 5 * pref_chunk));
348 __ lwr(t0, MemOperand(a1, 4, loadstore_chunk));
349 __ lwr(t1, MemOperand(a1, 5, loadstore_chunk));
350 __ lwr(t2, MemOperand(a1, 6, loadstore_chunk));
351 __ lwr(t3, MemOperand(a1, 7, loadstore_chunk));
353 MemOperand(a1, 1, loadstore_chunk, MemOperand::offset_minus_one));
355 MemOperand(a1, 2, loadstore_chunk, MemOperand::offset_minus_one));
357 MemOperand(a1, 3, loadstore_chunk, MemOperand::offset_minus_one));
359 MemOperand(a1, 4, loadstore_chunk, MemOperand::offset_minus_one));
361 MemOperand(a1, 5, loadstore_chunk, MemOperand::offset_minus_one));
363 MemOperand(a1, 6, loadstore_chunk, MemOperand::offset_minus_one));
365 MemOperand(a1, 7, loadstore_chunk, MemOperand::offset_minus_one));
367 MemOperand(a1, 8, loadstore_chunk, MemOperand::offset_minus_one));
368 __ Pref(pref_hint_load, MemOperand(a1, 4 * pref_chunk));
369 __ sw(a4, MemOperand(a0));
370 __ sw(a5, MemOperand(a0, 1, loadstore_chunk));
371 __ sw(a6, MemOperand(a0, 2, loadstore_chunk));
372 __ sw(a7, MemOperand(a0, 3, loadstore_chunk));
373 __ sw(t0, MemOperand(a0, 4, loadstore_chunk));
374 __ sw(t1, MemOperand(a0, 5, loadstore_chunk));
375 __ sw(t2, MemOperand(a0, 6, loadstore_chunk));
376 __ sw(t3, MemOperand(a0, 7, loadstore_chunk));
377 __ lwr(a4, MemOperand(a1, 8, loadstore_chunk));
378 __ lwr(a5, MemOperand(a1, 9, loadstore_chunk));
379 __ lwr(a6, MemOperand(a1, 10, loadstore_chunk));
380 __ lwr(a7, MemOperand(a1, 11, loadstore_chunk));
381 __ lwr(t0, MemOperand(a1, 12, loadstore_chunk));
382 __ lwr(t1, MemOperand(a1, 13, loadstore_chunk));
383 __ lwr(t2, MemOperand(a1, 14, loadstore_chunk));
384 __ lwr(t3, MemOperand(a1, 15, loadstore_chunk));
386 MemOperand(a1, 9, loadstore_chunk, MemOperand::offset_minus_one));
388 MemOperand(a1, 10, loadstore_chunk, MemOperand::offset_minus_one));
390 MemOperand(a1, 11, loadstore_chunk, MemOperand::offset_minus_one));
392 MemOperand(a1, 12, loadstore_chunk, MemOperand::offset_minus_one));
394 MemOperand(a1, 13, loadstore_chunk, MemOperand::offset_minus_one));
396 MemOperand(a1, 14, loadstore_chunk, MemOperand::offset_minus_one));
398 MemOperand(a1, 15, loadstore_chunk, MemOperand::offset_minus_one));
400 MemOperand(a1, 16, loadstore_chunk, MemOperand::offset_minus_one));
401 __ Pref(pref_hint_load, MemOperand(a1, 5 * pref_chunk));
402 __ sw(a4, MemOperand(a0, 8, loadstore_chunk));
403 __ sw(a5, MemOperand(a0, 9, loadstore_chunk));
404 __ sw(a6, MemOperand(a0, 10, loadstore_chunk));
405 __ sw(a7, MemOperand(a0, 11, loadstore_chunk));
406 __ sw(t0, MemOperand(a0, 12, loadstore_chunk));
407 __ sw(t1, MemOperand(a0, 13, loadstore_chunk));
408 __ sw(t2, MemOperand(a0, 14, loadstore_chunk));
409 __ sw(t3, MemOperand(a0, 15, loadstore_chunk));
419 __ Pref(pref_hint_load, MemOperand(a1));
424 __ lwr(a4, MemOperand(a1));
425 __ lwr(a5, MemOperand(a1, 1, loadstore_chunk));
426 __ lwr(a6, MemOperand(a1, 2, loadstore_chunk));
427 __ lwr(a7, MemOperand(a1, 3, loadstore_chunk));
428 __ lwr(t0, MemOperand(a1, 4, loadstore_chunk));
429 __ lwr(t1, MemOperand(a1, 5, loadstore_chunk));
430 __ lwr(t2, MemOperand(a1, 6, loadstore_chunk));
431 __ lwr(t3, MemOperand(a1, 7, loadstore_chunk));
433 MemOperand(a1, 1, loadstore_chunk, MemOperand::offset_minus_one));
435 MemOperand(a1, 2, loadstore_chunk, MemOperand::offset_minus_one));
437 MemOperand(a1, 3, loadstore_chunk, MemOperand::offset_minus_one));
439 MemOperand(a1, 4, loadstore_chunk, MemOperand::offset_minus_one));
441 MemOperand(a1, 5, loadstore_chunk, MemOperand::offset_minus_one));
443 MemOperand(a1, 6, loadstore_chunk, MemOperand::offset_minus_one));
445 MemOperand(a1, 7, loadstore_chunk, MemOperand::offset_minus_one));
447 MemOperand(a1, 8, loadstore_chunk, MemOperand::offset_minus_one));
449 __ sw(a4, MemOperand(a0));
450 __ sw(a5, MemOperand(a0, 1, loadstore_chunk));
451 __ sw(a6, MemOperand(a0, 2, loadstore_chunk));
452 __ sw(a7, MemOperand(a0, 3, loadstore_chunk));
453 __ sw(t0, MemOperand(a0, 4, loadstore_chunk));
454 __ sw(t1, MemOperand(a0, 5, loadstore_chunk));
455 __ sw(t2, MemOperand(a0, 6, loadstore_chunk));
456 __ sw(t3, MemOperand(a0, 7, loadstore_chunk));
468 __ lwr(v1, MemOperand(a1));
470 MemOperand(a1, 1, loadstore_chunk, MemOperand::offset_minus_one));
474 __ sw(v1, MemOperand(a0, -1, loadstore_chunk)); // In delay slot.
482 __ lb(v1, MemOperand(a1));
486 __ sb(v1, MemOperand(a0, -1)); // In delay slot.
631 __ sd(length, MemOperand(array, FixedDoubleArray::kLengthOffset));
633 __ sd(scratch2, MemOperand(array, HeapObject::kMapOffset));
690 __ ld(ra, MemOperand(sp, 0));
696 __ ld(scratch2, MemOperand(scratch1));
705 __ sdc1(f0, MemOperand(scratch3));
718 __ sw(hole_lower, MemOperand(scratch3));
720 __ sw(hole_upper, MemOperand(scratch3, kIntSize));
778 __ sd(length, MemOperand(array, FixedDoubleArray::kLengthOffset));
779 __ sd(scratch, MemOperand(array, HeapObject::kMapOffset));
811 __ lw(upper_bits, MemOperand(src_elements));
826 __ lw(scratch2, MemOperand(heap_number, -12));
830 __ sd(heap_number, MemOperand(dst_elements));
844 __ sd(scratch2, MemOperand(dst_elements));
959 __ lhu(result, MemOperand(at));
964 __ lbu(result, MemOperand(at));
969 static MemOperand ExpConstant(int index, Register base) {
970 return MemOperand(base, index * kDoubleSize);
1032 __ lwu(temp2, MemOperand(temp3, 0));
1033 __ lwu(temp3, MemOperand(temp3, kIntSize));