Home | History | Annotate | Download | only in arm

Lines Matching refs:ref_main_idx

274     add         r10,r8,r9                   @(i row)*pu1_ref[ref_main_idx]
276 vld1.8 {d8},[r10],r11 @(i row)ref_main_idx
280 add r12,r8,r9 @(ii)*pu1_ref[ref_main_idx]
284 add r10,r8,r9 @(iii)*pu1_ref[ref_main_idx]
286 vld1.8 {d12},[r12],r11 @(ii)ref_main_idx
287 vmull.u8 q5,d8,d30 @(i row)vmull_u8(ref_main_idx, dup_const_32_fract)
297 vmull.u8 q7,d12,d28 @(ii)vmull_u8(ref_main_idx, dup_const_32_fract)
298 add r12,r8,r9 @(iv)*pu1_ref[ref_main_idx]
300 vld1.8 {d16},[r10],r11 @(iii)ref_main_idx
306 vld1.8 {d20},[r12],r11 @(iv)ref_main_idx
312 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract)
322 add r10,r8,r9 @(v)*pu1_ref[ref_main_idx]
324 vld1.8 {d8},[r10],r11 @(v)ref_main_idx
327 vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract)
336 add r12,r8,r9 @(vi)*pu1_ref[ref_main_idx]
341 add r10,r8,r9 @(vii)*pu1_ref[ref_main_idx]
343 vld1.8 {d12},[r12],r11 @(vi)ref_main_idx
344 vmull.u8 q5,d8,d30 @(v)vmull_u8(ref_main_idx, dup_const_32_fract)
355 vld1.8 {d16},[r10],r11 @(vii)ref_main_idx
359 vmull.u8 q7,d12,d28 @(vi)vmull_u8(ref_main_idx, dup_const_32_fract)
361 add r12,r8,r9 @(viii)*pu1_ref[ref_main_idx]
368 vld1.8 {d20},[r12],r11 @(viii)ref_main_idx
372 vmull.u8 q9,d16,d26 @(vii)vmull_u8(ref_main_idx, dup_const_32_fract)
390 add r10,r8,r9 @(i)*pu1_ref[ref_main_idx]
397 vld1.8 {d8},[r10],r11 @(i)ref_main_idx
401 add r12,r8,r9 @(ii)*pu1_ref[ref_main_idx]
404 vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract)
417 add r10,r8,r9 @(iii)*pu1_ref[ref_main_idx]
419 vld1.8 {d12},[r12],r11 @(ii)ref_main_idx
420 vmull.u8 q5,d8,d30 @(i)vmull_u8(ref_main_idx, dup_const_32_fract)
433 add r12,r8,r9 @(iv)*pu1_ref[ref_main_idx]
435 vld1.8 {d16},[r10],r11 @(iii)ref_main_idx
436 vmull.u8 q7,d12,d28 @(ii)vmull_u8(ref_main_idx, dup_const_32_fract)
450 vld1.8 {d20},[r12],r11 @(iv)ref_main_idx
451 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract)
462 add r10,r8,r9 @(v)*pu1_ref[ref_main_idx]
469 vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract)
474 add r12,r8,r9 @(vi)*pu1_ref[ref_main_idx]
480 vld1.8 {d8},[r10],r11 @(v)ref_main_idx
489 add r10,r8,r9 @(vii)*pu1_ref[ref_main_idx]
492 vld1.8 {d12},[r12],r11 @(vi)ref_main_idx
496 vmull.u8 q5,d8,d30 @(v)vmull_u8(ref_main_idx, dup_const_32_fract)
501 add r12,r8,r9 @(viii)*pu1_ref[ref_main_idx]
504 vld1.8 {d16},[r10],r11 @(vii)ref_main_idx
508 vmull.u8 q7,d12,d28 @(vi)vmull_u8(ref_main_idx, dup_const_32_fract)
513 vld1.8 {d20},[r12],r11 @(viii)ref_main_idx
524 vmull.u8 q9,d16,d26 @(vii)vmull_u8(ref_main_idx, dup_const_32_fract)
536 add r10,r8,r9 @(i)*pu1_ref[ref_main_idx]
545 vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract)
569 vld1.32 {d2[0]},[r10] @ref_main_idx
582 vmull.u8 q2,d2,d1 @vmull_u8(ref_main_idx, dup_const_32_fract)
588 vld1.32 {d8[0]},[r10] @ref_main_idx
603 vmull.u8 q5,d8,d7 @vmull_u8(ref_main_idx, dup_const_32_fract)
609 vld1.32 {d14[0]},[r10] @ref_main_idx
625 vmull.u8 q8,d14,d13 @vmull_u8(ref_main_idx, dup_const_32_fract)
631 vld1.32 {d20[0]},[r10] @ref_main_idx
637 vmull.u8 q11,d20,d19 @vmull_u8(ref_main_idx, dup_const_32_fract)