Home | History | Annotate | Download | only in arm

Lines Matching refs:ref_main_idx

160     add         r10,r8,r9                   @(i row)*pu1_ref[ref_main_idx]
163 vld1.8 {d8},[r10],r11 @(i row)ref_main_idx
168 add r12,r8,r9 @(ii)*pu1_ref[ref_main_idx]
172 add r10,r8,r9 @(iii)*pu1_ref[ref_main_idx]
174 vld1.8 {d12},[r12],r11 @(ii)ref_main_idx
175 vmull.u8 q5,d8,d30 @(i row)vmull_u8(ref_main_idx, dup_const_32_fract)
186 vmull.u8 q7,d12,d28 @(ii)vmull_u8(ref_main_idx, dup_const_32_fract)
187 add r12,r8,r9 @(iv)*pu1_ref[ref_main_idx]
189 vld1.8 {d16},[r10],r11 @(iii)ref_main_idx
195 vld1.8 {d20},[r12],r11 @(iv)ref_main_idx
201 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract)
212 add r10,r8,r9 @(v)*pu1_ref[ref_main_idx]
214 vld1.8 {d8},[r10],r11 @(v)ref_main_idx
218 vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract)
227 add r12,r8,r9 @(vi)*pu1_ref[ref_main_idx]
233 add r10,r8,r9 @(vii)*pu1_ref[ref_main_idx]
235 vld1.8 {d12},[r12],r11 @(vi)ref_main_idx
236 vmull.u8 q5,d8,d30 @(v)vmull_u8(ref_main_idx, dup_const_32_fract)
248 vld1.8 {d16},[r10],r11 @(vii)ref_main_idx
252 vmull.u8 q7,d12,d28 @(vi)vmull_u8(ref_main_idx, dup_const_32_fract)
254 add r12,r8,r9 @(viii)*pu1_ref[ref_main_idx]
261 vld1.8 {d20},[r12],r11 @(viii)ref_main_idx
265 vmull.u8 q9,d16,d26 @(vii)vmull_u8(ref_main_idx, dup_const_32_fract)
284 add r10,r8,r9 @(i)*pu1_ref[ref_main_idx]
291 vld1.8 {d8},[r10],r11 @(i)ref_main_idx
297 vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract)
298 add r12,r8,r9 @(ii)*pu1_ref[ref_main_idx]
310 add r10,r8,r9 @(iii)*pu1_ref[ref_main_idx]
312 vld1.8 {d12},[r12],r11 @(ii)ref_main_idx
313 vmull.u8 q5,d8,d30 @(i)vmull_u8(ref_main_idx, dup_const_32_fract)
327 vld1.8 {d16},[r10],r11 @(iii)ref_main_idx
328 vmull.u8 q7,d12,d28 @(ii)vmull_u8(ref_main_idx, dup_const_32_fract)
329 add r12,r8,r9 @(iv)*pu1_ref[ref_main_idx]
343 vld1.8 {d20},[r12],r11 @(iv)ref_main_idx
344 vmull.u8 q9,d16,d26 @(iii)vmull_u8(ref_main_idx, dup_const_32_fract)
357 add r10,r8,r9 @(v)*pu1_ref[ref_main_idx]
360 vmull.u8 q11,d20,d24 @(iv)vmull_u8(ref_main_idx, dup_const_32_fract)
369 add r12,r8,r9 @(vi)*pu1_ref[ref_main_idx]
371 vld1.8 {d8},[r10],r11 @(v)ref_main_idx
383 vld1.8 {d12},[r12],r11 @(vi)ref_main_idx
385 add r10,r8,r9 @(vii)*pu1_ref[ref_main_idx]
388 vmull.u8 q5,d8,d30 @(v)vmull_u8(ref_main_idx, dup_const_32_fract)
393 add r12,r8,r9 @(viii)*pu1_ref[ref_main_idx]
395 vld1.8 {d16},[r10],r11 @(vii)ref_main_idx
399 vmull.u8 q7,d12,d28 @(vi)vmull_u8(ref_main_idx, dup_const_32_fract)
406 vld1.8 {d20},[r12],r11 @(viii)ref_main_idx
414 vmull.u8 q9,d16,d26 @(vii)vmull_u8(ref_main_idx, dup_const_32_fract)
427 add r10,r8,r9 @(i)*pu1_ref[ref_main_idx]
436 vmull.u8 q11,d20,d24 @(viii)vmull_u8(ref_main_idx, dup_const_32_fract)
461 vld1.8 {d2},[r10] @ref_main_idx
474 vmull.u8 q2,d2,d1 @vmull_u8(ref_main_idx, dup_const_32_fract)
480 vld1.8 {d8},[r10] @ref_main_idx
495 vmull.u8 q5,d8,d7 @vmull_u8(ref_main_idx, dup_const_32_fract)
501 vld1.8 {d14},[r10] @ref_main_idx
517 vmull.u8 q8,d14,d13 @vmull_u8(ref_main_idx, dup_const_32_fract)
523 vld1.8 {d20},[r10] @ref_main_idx
529 vmull.u8 q11,d20,d19 @vmull_u8(ref_main_idx, dup_const_32_fract)