Home | History | Annotate | Download | only in arm

Lines Matching refs:ref_main_idx

271     add         r10,r8,r9                   @(i row)*pu1_ref[ref_main_idx]
273 vld1.8 {d8},[r10],r11 @(i row)ref_main_idx
277 add r12,r8,r9 @(ii)*pu1_ref[ref_main_idx]
281 add r10,r8,r9 @(iii)*pu1_ref[ref_main_idx]
283 vld1.8 {d12},[r12],r11 @(ii)ref_main_idx
284 vmull.u8 q5,d8,d30 @(i row)vmull_u8(ref_main_idx, dup_const_32_fract)
294 vmull.u8 q7,d12,d28 @(ii)vmull_u8(ref_main_idx, dup_const_32_fract)
295 add r12,r8,r9 @(iv)*pu1_ref[ref_main_idx]
297 vld1.8 {d16},[r10],r11 @(iii)ref_main_idx
303 vld1.8 {d20},[r12],r11 @(iv)ref_main_idx
309 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract)
320 add r10,r8,r9 @(v)*pu1_ref[ref_main_idx]
322 vld1.8 {d8},[r10],r11 @(v)ref_main_idx
325 vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract)
334 add r12,r8,r9 @(vi)*pu1_ref[ref_main_idx]
339 add r10,r8,r9 @(vii)*pu1_ref[ref_main_idx]
341 vld1.8 {d12},[r12],r11 @(vi)ref_main_idx
342 vmull.u8 q5,d8,d30 @(v)vmull_u8(ref_main_idx, dup_const_32_fract)
353 vld1.8 {d16},[r10],r11 @(vii)ref_main_idx
357 vmull.u8 q7,d12,d28 @(vi)vmull_u8(ref_main_idx, dup_const_32_fract)
359 add r12,r8,r9 @(viii)*pu1_ref[ref_main_idx]
369 vld1.8 {d20},[r12],r11 @(viii)ref_main_idx
373 vmull.u8 q9,d16,d26 @(vii)vmull_u8(ref_main_idx, dup_const_32_fract)
393 add r10,r8,r9 @(i)*pu1_ref[ref_main_idx]
400 vld1.8 {d8},[r10],r11 @(i)ref_main_idx
404 add r12,r8,r9 @(ii)*pu1_ref[ref_main_idx]
407 vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract)
420 add r10,r8,r9 @(iii)*pu1_ref[ref_main_idx]
422 vld1.8 {d12},[r12],r11 @(ii)ref_main_idx
423 vmull.u8 q5,d8,d30 @(i)vmull_u8(ref_main_idx, dup_const_32_fract)
436 add r12,r8,r9 @(iv)*pu1_ref[ref_main_idx]
438 vld1.8 {d16},[r10],r11 @(iii)ref_main_idx
439 vmull.u8 q7,d12,d28 @(ii)vmull_u8(ref_main_idx, dup_const_32_fract)
453 vld1.8 {d20},[r12],r11 @(iv)ref_main_idx
454 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract)
466 add r10,r8,r9 @(v)*pu1_ref[ref_main_idx]
473 vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract)
478 add r12,r8,r9 @(vi)*pu1_ref[ref_main_idx]
484 vld1.8 {d8},[r10],r11 @(v)ref_main_idx
493 add r10,r8,r9 @(vii)*pu1_ref[ref_main_idx]
496 vld1.8 {d12},[r12],r11 @(vi)ref_main_idx
502 vmull.u8 q5,d8,d30 @(v)vmull_u8(ref_main_idx, dup_const_32_fract)
507 add r12,r8,r9 @(viii)*pu1_ref[ref_main_idx]
510 vld1.8 {d16},[r10],r11 @(vii)ref_main_idx
514 vmull.u8 q7,d12,d28 @(vi)vmull_u8(ref_main_idx, dup_const_32_fract)
519 vld1.8 {d20},[r12],r11 @(viii)ref_main_idx
530 vmull.u8 q9,d16,d26 @(vii)vmull_u8(ref_main_idx, dup_const_32_fract)
543 add r10,r8,r9 @(i)*pu1_ref[ref_main_idx]
552 vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract)