HomeSort by relevance Sort by last modified time
    Searched refs:v26 (Results 1 - 25 of 102) sorted by null

1 2 3 4 5

  /external/libhevc/common/arm64/
ihevc_sao_edge_offset_class1.s 155 Uxtl v26.8h, v18.8b //II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row)))
198 SADDW v26.8h, v26.8h , v24.8b //II pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0], offset)
204 SMAX v26.8h, v26.8h , v2.8h //II pi2_tmp_cur_row.val[0] = vmaxq_s16(pi2_tmp_cur_row.val[0], const_min_clip)
205 UMIN v26.8h, v26.8h , v4.8h //II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vminq_u16(vreinterpretq_u16_s16(pi2_tmp_cur_row.val[0]), const_max_clip))
211 xtn v30.8b, v26.8h //II vmovn_s16(pi2_tmp_cur_row.val[0])
234 Uxtl v26.8h, v3.8b //pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row)))
235 SADDW v26.8h, v26.8h , v24.8b //pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0], offset
    [all...]
ihevc_inter_pred_chroma_vert_w16inp.s 213 smull v26.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0)
215 smlal v26.4s, v3.4h, v17.4h
216 smlal v26.4s, v4.4h, v18.4h
218 smlal v26.4s, v5.4h, v19.4h
234 sqshrn v26.4h, v26.4s,#6 //right shift
249 sqrshrun v26.8b, v26.8h,#6 //rounding shift
256 st1 {v26.s}[0],[x9],x3 //stores the loaded value
266 smull v26.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0
    [all...]
ihevc_sao_edge_offset_class2_chroma.s 552 ADD v26.16b, v0.16b , v17.16b //II edge_idx = vaddq_s8(const_2, sign_up)
557 ADD v26.16b, v26.16b , v24.16b //II edge_idx = vaddq_s8(edge_idx, sign_down)
564 TBL v26.16b, {v22.16b},v26.16b //II vtbl1_s8(edge_idx_tbl, vget_low_s8(edge_idx))
575 AND v26.16b, v26.16b , v1.16b //II edge_idx = vandq_s8(edge_idx, au1_mask)
578 mov v27.d[0],v26.d[1]
579 UZP1 v31.8b, v26.8b, v27.8b
580 UZP2 v27.8b, v26.8b, v27.8b //I
    [all...]
ihevc_sao_edge_offset_class3.s 406 cmhi v26.16b, v18.16b , v5.16b //II vcltq_u8(pu1_cur_row, pu1_next_row_tmp)
410 SUB v24.16b, v26.16b , v24.16b //II sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt))
420 ADD v26.16b, v0.16b , v17.16b //II edge_idx = vaddq_s8(const_2, sign_up)
425 ADD v26.16b, v26.16b , v24.16b //II edge_idx = vaddq_s8(edge_idx, sign_down)
428 TBL v26.16b, {v6.16b},v26.16b //II vtbl1_s8(edge_idx_tbl, vget_low_s8(edge_idx))
436 AND v26.16b, v26.16b , v1.16b //II edge_idx = vandq_s8(edge_idx, au1_mask)
439 TBL v24.16b, {v7.16b},v26.16b //II offset = vtbl1_s8(offset_tbl, vget_low_s8(edge_idx)
    [all...]
ihevc_itrans_recon_32x32.s 217 smull v26.4s, v8.4h, v0.4h[3] //// y1 * cos3(part of b1)
222 smlal v26.4s, v9.4h, v2.4h[1] //// y1 * cos3 - y3 * sin1(part of b1)
257 smlal v26.4s, v14.4h, v3.4h[3]
263 smlal v26.4s, v15.4h, v5.4h[1]
287 smlal v26.4s, v8.4h, v6.4h[3] //// y1 * cos3(part of b1)
292 smlsl v26.4s, v9.4h, v7.4h[3] //// y1 * cos3 - y3 * sin1(part of b1)
331 smlsl v26.4s, v14.4h, v6.4h[1]
337 smlsl v26.4s, v15.4h, v4.4h[3]
363 smlsl v26.4s, v8.4h, v3.4h[1] //// y1 * cos3(part of b1)
368 smlsl v26.4s, v9.4h, v1.4h[3] //// y1 * cos3 - y3 * sin1(part of b1
    [all...]
ihevc_inter_pred_chroma_vert.s 257 umull v26.8h, v7.8b, v1.8b
259 umlsl v26.8h, v6.8b, v0.8b
261 umlal v26.8h, v16.8b, v2.8b
263 umlsl v26.8h, v17.8b, v3.8b
279 sqrshrun v26.8b, v26.8h,#6
299 st1 {v26.8b},[x7],x3 //stores the loaded value
318 umull v26.8h, v7.8b, v1.8b
325 umlsl v26.8h, v6.8b, v0.8b
328 umlal v26.8h, v16.8b, v2.8
    [all...]
ihevc_inter_pred_chroma_vert_w16inp_w16out.s 212 smull v26.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0)
213 smlal v26.4s, v3.4h, v17.4h
214 smlal v26.4s, v4.4h, v18.4h
217 smlal v26.4s, v5.4h, v19.4h
232 sqshrn v26.4h, v26.4s,#6 //right shift
252 st1 {v26.2s},[x9],x3 //stores the loaded value
261 smull v26.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0)
263 smlal v26.4s, v3.4h, v17.4h
266 smlal v26.4s, v4.4h, v18.4
    [all...]
ihevc_sao_edge_offset_class1_chroma.s 184 Uxtl v26.8h, v18.8b //II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row)))
244 SADDW v26.8h, v26.8h , v24.8b //II pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0], offset)
252 SMAX v26.8h, v26.8h , v2.8h //II pi2_tmp_cur_row.val[0] = vmaxq_s16(pi2_tmp_cur_row.val[0], const_min_clip)
253 UMIN v26.8h, v26.8h , v4.8h //II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vminq_u16(vreinterpretq_u16_s16(pi2_tmp_cur_row.val[0]), const_max_clip))
259 xtn v30.8b, v26.8h //II vmovn_s16(pi2_tmp_cur_row.val[0])
294 Uxtl v26.8h, v3.8b //pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row)))
295 SADDW v26.8h, v26.8h , v24.8b //pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0], offset
    [all...]
ihevc_itrans_recon_16x16.s 256 smull v26.4s, v6.4h, v0.4h[3] //// y1 * cos3(part of b1)
261 smlal v26.4s, v7.4h, v2.4h[1] //// y1 * cos3 - y3 * sin1(part of b1)
283 smlal v26.4s, v8.4h, v3.4h[3]
289 smlsl v26.4s, v9.4h, v2.4h[3]
332 smlsl v26.4s, v6.4h, v1.4h[1] //// y1 * cos3(part of b1)
337 smlsl v26.4s, v7.4h, v0.4h[1] //// y1 * cos3 - y3 * sin1(part of b1)
344 smlsl v26.4s, v8.4h, v1.4h[3]
350 smlsl v26.4s, v9.4h, v3.4h[1]
387 add v12.4s, v14.4s , v26.4s
388 sub v24.4s, v14.4s , v26.4
    [all...]
ihevc_intra_pred_luma_horz.s 205 dup v26.8b, v0.16b[15]
206 uxtl v26.8h, v26.8b
215 sqadd v22.8h, v26.8h , v24.8h
229 sqadd v22.8h, v26.8h , v24.8h
284 dup v26.8b, v0.8b[7]
288 uxtl v26.8h, v26.8b
297 sqadd v22.8h, v26.8h , v24.8h
331 dup v26.8b, v0.8b[3
    [all...]
ihevc_itrans_recon_8x8.s 194 smull v26.4s, v6.4h, v0.4h[3] //// y1 * cos3(part of b1)
202 smlsl v26.4s, v7.4h, v1.4h[3] //// y1 * cos3 - y3 * sin1(part of b1)
235 smlsl v26.4s, v14.4h, v0.4h[1] //// y1 * cos3 - y3 * sin1 - y5 * cos1(part of b1)
246 smlsl v26.4s, v15.4h, v1.4h[1] //// b1 = y1 * cos3 - y3 * sin1 - y5 * cos1 - y7 * sin3(part of x1,x6)
261 add v28.4s, v18.4s , v26.4s //// a1 + b1(part of x1)
262 sub v18.4s, v18.4s , v26.4s //// a1 - b1(part of x6)
264 add v26.4s, v10.4s , v30.4s //// a3 + b3(part of x3)
273 sqrshrn v7.4h, v26.4s,#shift_stage1_idct //// x3 = (a3 + b3 + rnd) >> 7(shift_stage1_idct)
305 smull v26.4s, v6.4h, v0.4h[3] //// y1 * cos3(part of b1)
310 smlsl v26.4s, v7.4h, v1.4h[3] //// y1 * cos3 - y3 * sin1(part of b1
    [all...]
ihevc_inter_pred_chroma_horz_w16out.s 135 dup v26.8b, v2.8b[2] //coeffabs_2 = vdup_lane_u8(coeffabs, 2)
215 umlal v30.8h, v4.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
233 umlal v28.8h, v5.8b, v26.8b
275 umlal v22.8h, v12.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
294 umlal v20.8h, v13.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
312 umlal v30.8h, v4.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
328 umlal v28.8h, v5.8b, v26.8b
354 umlal v22.8h, v12.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
376 umlal v20.8h, v13.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
393 umlal v30.8h, v4.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)/
    [all...]
ihevc_sao_edge_offset_class0.s 152 LD1 {v26.16b},[x12] //II Iteration pu1_cur_row = vld1q_u8(pu1_src_cpy)
169 EXT v28.16b, v28.16b , v26.16b,#15 //II Iteration pu1_cur_row_tmp = vextq_u8(pu1_cur_row_tmp, pu1_cur_row, 15)
174 cmhi v30.16b, v26.16b , v28.16b //II vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp)
180 cmhi v0.16b, v28.16b , v26.16b //II vcltq_u8(pu1_cur_row, pu1_cur_row_tmp)
189 EXT v28.16b, v26.16b , v28.16b,#1 //II pu1_cur_row_tmp = vextq_u8(pu1_cur_row, pu1_cur_row_tmp, 1)
203 cmhi v30.16b, v26.16b , v28.16b //II vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp)
205 cmhi v0.16b, v28.16b , v26.16b //II vcltq_u8(pu1_cur_row, pu1_cur_row_tmp)
211 Uxtl v0.8h, v26.8b //II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row)))
240 Uxtl2 v28.8h, v26.16b //II pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(pu1_cur_row)))
318 TBL v26.8b, {v7.16b},v24.8b //offset = vtbl1_s8(offset_tbl, vget_low_s8(edge_idx)
    [all...]
ihevc_sao_edge_offset_class3_chroma.s 548 ADD v26.16b, v0.16b , v17.16b //II edge_idx = vaddq_s8(const_2, sign_up)
554 ADD v26.16b, v26.16b , v24.16b //II edge_idx = vaddq_s8(edge_idx, sign_down)
558 TBL v26.16b, {v21.16b},v26.16b //II vtbl1_s8(edge_idx_tbl, vget_low_s8(edge_idx))
569 AND v26.16b, v26.16b , v1.16b //II edge_idx = vandq_s8(edge_idx, au1_mask)
570 mov v27.d[0],v26.d[1]
573 UZP1 v31.8b, v26.8b, v27.8b
574 UZP2 v27.8b, v26.8b, v27.8b //I
    [all...]
ihevc_intra_pred_chroma_horz.s 287 dup v26.4h, v0.4h[3]
291 uxtl v26.8h, v26.8b
300 sqadd v22.8h, v26.8h , v24.8h
334 dup v26.8b, v0.8b[3]
335 uxtl v26.8h, v26.8b
344 sqadd v22.8h, v26.8h , v24.8h
ihevc_inter_pred_chroma_horz.s 135 dup v26.8b, v2.8b[2] //coeffabs_2 = vdup_lane_u8(coeffabs, 2)
200 umlal v30.8h, v4.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
215 umlal v28.8h, v5.8b, v26.8b
265 umlal v22.8h, v12.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
287 umlal v20.8h, v13.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
305 umlal v30.8h, v4.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
328 umlal v28.8h, v5.8b, v26.8b
361 umlal v22.8h, v12.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
379 umlal v20.8h, v13.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)//
396 umlal v30.8h, v4.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)/
    [all...]
ihevc_itrans_recon_4x4.s 166 trn1 v26.4h, v30.4h, v31.4h
168 trn1 v0.2s, v24.2s, v26.2s
169 trn2 v2.2s, v24.2s, v26.2s
202 trn1 v26.4h, v30.4h, v31.4h
204 trn1 v0.2s, v24.2s, v26.2s
205 trn2 v2.2s, v24.2s, v26.2s
ihevc_inter_pred_chroma_vert_w16out.s 257 umull v26.8h, v7.8b, v1.8b
259 umlsl v26.8h, v6.8b, v0.8b
261 umlal v26.8h, v16.8b, v2.8b
263 umlsl v26.8h, v17.8b, v3.8b
297 st1 { v26.16b},[x7],x3 //stores the loaded value
313 umull v26.8h, v7.8b, v1.8b
318 umlsl v26.8h, v6.8b, v0.8b
322 umlal v26.8h, v16.8b, v2.8b
325 umlsl v26.8h, v17.8b, v3.8b
358 st1 { v26.16b},[x7],x3 //stores the loaded valu
    [all...]
ihevc_sao_edge_offset_class2.s 418 Uxtl v26.8h, v5.8b //II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row)))
422 SADDW v26.8h, v26.8h , v24.8b //II pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0], offset)
429 SMAX v26.8h, v26.8h , v2.8h //II pi2_tmp_cur_row.val[0] = vmaxq_s16(pi2_tmp_cur_row.val[0], const_min_clip)
433 UMIN v26.8h, v26.8h , v4.8h //II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vminq_u16(vreinterpretq_u16_s16(pi2_tmp_cur_row.val[0]), const_max_clip))
448 xtn v26.8b, v26.8h //II vmovn_s16(pi2_tmp_cur_row.val[0])
450 xtn2 v26.16b, v28.8h //II vmovn_s16(pi2_tmp_cur_row.val[1]
    [all...]
  /external/llvm/test/MC/AArch64/
neon-scalar-by-elem-mla.s 27 fmls s7, s17, v26.s[3]
34 // CHECK: fmls s7, s17, v26.s[3] // encoding: [0x27,0x5a,0xba,0x5f]
  /external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/ppc/
platform_altivec.asm 34 W v26, r3
52 R v26, r3
  /external/libvpx/libvpx/vp8/common/ppc/
platform_altivec.asm 34 W v26, r3
52 R v26, r3
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/ppc/
platform_altivec.asm 34 W v26, r3
52 R v26, r3
  /cts/tests/tests/jni/src/android/jni/cts/
InstanceFromNative.java 88 int v25, int v26, int v27, int v28, int v29,
101 (v24 == 24) && (v25 == 25) && (v26 == 26) && (v27 == 27) &&
StaticFromNative.java 85 int v25, int v26, int v27, int v28, int v29,
98 (v24 == 24) && (v25 == 25) && (v26 == 26) && (v27 == 27) &&

Completed in 367 milliseconds

1 2 3 4 5