/external/libhevc/common/arm64/ |
ihevc_sao_edge_offset_class1.s | 155 Uxtl v26.8h, v18.8b //II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 198 SADDW v26.8h, v26.8h , v24.8b //II pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0], offset) 204 SMAX v26.8h, v26.8h , v2.8h //II pi2_tmp_cur_row.val[0] = vmaxq_s16(pi2_tmp_cur_row.val[0], const_min_clip) 205 UMIN v26.8h, v26.8h , v4.8h //II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vminq_u16(vreinterpretq_u16_s16(pi2_tmp_cur_row.val[0]), const_max_clip)) 211 xtn v30.8b, v26.8h //II vmovn_s16(pi2_tmp_cur_row.val[0]) 234 Uxtl v26.8h, v3.8b //pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 235 SADDW v26.8h, v26.8h , v24.8b //pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0], offset [all...] |
ihevc_inter_pred_chroma_vert_w16inp.s | 213 smull v26.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 215 smlal v26.4s, v3.4h, v17.4h 216 smlal v26.4s, v4.4h, v18.4h 218 smlal v26.4s, v5.4h, v19.4h 234 sqshrn v26.4h, v26.4s,#6 //right shift 249 sqrshrun v26.8b, v26.8h,#6 //rounding shift 256 st1 {v26.s}[0],[x9],x3 //stores the loaded value 266 smull v26.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0 [all...] |
ihevc_sao_edge_offset_class2_chroma.s | 552 ADD v26.16b, v0.16b , v17.16b //II edge_idx = vaddq_s8(const_2, sign_up) 557 ADD v26.16b, v26.16b , v24.16b //II edge_idx = vaddq_s8(edge_idx, sign_down) 564 TBL v26.16b, {v22.16b},v26.16b //II vtbl1_s8(edge_idx_tbl, vget_low_s8(edge_idx)) 575 AND v26.16b, v26.16b , v1.16b //II edge_idx = vandq_s8(edge_idx, au1_mask) 578 mov v27.d[0],v26.d[1] 579 UZP1 v31.8b, v26.8b, v27.8b 580 UZP2 v27.8b, v26.8b, v27.8b //I [all...] |
ihevc_sao_edge_offset_class3.s | 406 cmhi v26.16b, v18.16b , v5.16b //II vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 410 SUB v24.16b, v26.16b , v24.16b //II sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 420 ADD v26.16b, v0.16b , v17.16b //II edge_idx = vaddq_s8(const_2, sign_up) 425 ADD v26.16b, v26.16b , v24.16b //II edge_idx = vaddq_s8(edge_idx, sign_down) 428 TBL v26.16b, {v6.16b},v26.16b //II vtbl1_s8(edge_idx_tbl, vget_low_s8(edge_idx)) 436 AND v26.16b, v26.16b , v1.16b //II edge_idx = vandq_s8(edge_idx, au1_mask) 439 TBL v24.16b, {v7.16b},v26.16b //II offset = vtbl1_s8(offset_tbl, vget_low_s8(edge_idx) [all...] |
ihevc_itrans_recon_32x32.s | 217 smull v26.4s, v8.4h, v0.4h[3] //// y1 * cos3(part of b1) 222 smlal v26.4s, v9.4h, v2.4h[1] //// y1 * cos3 - y3 * sin1(part of b1) 257 smlal v26.4s, v14.4h, v3.4h[3] 263 smlal v26.4s, v15.4h, v5.4h[1] 287 smlal v26.4s, v8.4h, v6.4h[3] //// y1 * cos3(part of b1) 292 smlsl v26.4s, v9.4h, v7.4h[3] //// y1 * cos3 - y3 * sin1(part of b1) 331 smlsl v26.4s, v14.4h, v6.4h[1] 337 smlsl v26.4s, v15.4h, v4.4h[3] 363 smlsl v26.4s, v8.4h, v3.4h[1] //// y1 * cos3(part of b1) 368 smlsl v26.4s, v9.4h, v1.4h[3] //// y1 * cos3 - y3 * sin1(part of b1 [all...] |
ihevc_inter_pred_chroma_vert.s | 257 umull v26.8h, v7.8b, v1.8b 259 umlsl v26.8h, v6.8b, v0.8b 261 umlal v26.8h, v16.8b, v2.8b 263 umlsl v26.8h, v17.8b, v3.8b 279 sqrshrun v26.8b, v26.8h,#6 299 st1 {v26.8b},[x7],x3 //stores the loaded value 318 umull v26.8h, v7.8b, v1.8b 325 umlsl v26.8h, v6.8b, v0.8b 328 umlal v26.8h, v16.8b, v2.8 [all...] |
ihevc_inter_pred_chroma_vert_w16inp_w16out.s | 212 smull v26.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 213 smlal v26.4s, v3.4h, v17.4h 214 smlal v26.4s, v4.4h, v18.4h 217 smlal v26.4s, v5.4h, v19.4h 232 sqshrn v26.4h, v26.4s,#6 //right shift 252 st1 {v26.2s},[x9],x3 //stores the loaded value 261 smull v26.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 263 smlal v26.4s, v3.4h, v17.4h 266 smlal v26.4s, v4.4h, v18.4 [all...] |
ihevc_sao_edge_offset_class1_chroma.s | 184 Uxtl v26.8h, v18.8b //II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 244 SADDW v26.8h, v26.8h , v24.8b //II pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0], offset) 252 SMAX v26.8h, v26.8h , v2.8h //II pi2_tmp_cur_row.val[0] = vmaxq_s16(pi2_tmp_cur_row.val[0], const_min_clip) 253 UMIN v26.8h, v26.8h , v4.8h //II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vminq_u16(vreinterpretq_u16_s16(pi2_tmp_cur_row.val[0]), const_max_clip)) 259 xtn v30.8b, v26.8h //II vmovn_s16(pi2_tmp_cur_row.val[0]) 294 Uxtl v26.8h, v3.8b //pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 295 SADDW v26.8h, v26.8h , v24.8b //pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0], offset [all...] |
ihevc_itrans_recon_16x16.s | 256 smull v26.4s, v6.4h, v0.4h[3] //// y1 * cos3(part of b1) 261 smlal v26.4s, v7.4h, v2.4h[1] //// y1 * cos3 - y3 * sin1(part of b1) 283 smlal v26.4s, v8.4h, v3.4h[3] 289 smlsl v26.4s, v9.4h, v2.4h[3] 332 smlsl v26.4s, v6.4h, v1.4h[1] //// y1 * cos3(part of b1) 337 smlsl v26.4s, v7.4h, v0.4h[1] //// y1 * cos3 - y3 * sin1(part of b1) 344 smlsl v26.4s, v8.4h, v1.4h[3] 350 smlsl v26.4s, v9.4h, v3.4h[1] 387 add v12.4s, v14.4s , v26.4s 388 sub v24.4s, v14.4s , v26.4 [all...] |
ihevc_intra_pred_luma_horz.s | 205 dup v26.8b, v0.16b[15] 206 uxtl v26.8h, v26.8b 215 sqadd v22.8h, v26.8h , v24.8h 229 sqadd v22.8h, v26.8h , v24.8h 284 dup v26.8b, v0.8b[7] 288 uxtl v26.8h, v26.8b 297 sqadd v22.8h, v26.8h , v24.8h 331 dup v26.8b, v0.8b[3 [all...] |
ihevc_itrans_recon_8x8.s | 194 smull v26.4s, v6.4h, v0.4h[3] //// y1 * cos3(part of b1) 202 smlsl v26.4s, v7.4h, v1.4h[3] //// y1 * cos3 - y3 * sin1(part of b1) 235 smlsl v26.4s, v14.4h, v0.4h[1] //// y1 * cos3 - y3 * sin1 - y5 * cos1(part of b1) 246 smlsl v26.4s, v15.4h, v1.4h[1] //// b1 = y1 * cos3 - y3 * sin1 - y5 * cos1 - y7 * sin3(part of x1,x6) 261 add v28.4s, v18.4s , v26.4s //// a1 + b1(part of x1) 262 sub v18.4s, v18.4s , v26.4s //// a1 - b1(part of x6) 264 add v26.4s, v10.4s , v30.4s //// a3 + b3(part of x3) 273 sqrshrn v7.4h, v26.4s,#shift_stage1_idct //// x3 = (a3 + b3 + rnd) >> 7(shift_stage1_idct) 305 smull v26.4s, v6.4h, v0.4h[3] //// y1 * cos3(part of b1) 310 smlsl v26.4s, v7.4h, v1.4h[3] //// y1 * cos3 - y3 * sin1(part of b1 [all...] |
ihevc_inter_pred_chroma_horz_w16out.s | 135 dup v26.8b, v2.8b[2] //coeffabs_2 = vdup_lane_u8(coeffabs, 2) 215 umlal v30.8h, v4.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)// 233 umlal v28.8h, v5.8b, v26.8b 275 umlal v22.8h, v12.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)// 294 umlal v20.8h, v13.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)// 312 umlal v30.8h, v4.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)// 328 umlal v28.8h, v5.8b, v26.8b 354 umlal v22.8h, v12.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)// 376 umlal v20.8h, v13.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)// 393 umlal v30.8h, v4.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)/ [all...] |
ihevc_sao_edge_offset_class0.s | 152 LD1 {v26.16b},[x12] //II Iteration pu1_cur_row = vld1q_u8(pu1_src_cpy) 169 EXT v28.16b, v28.16b , v26.16b,#15 //II Iteration pu1_cur_row_tmp = vextq_u8(pu1_cur_row_tmp, pu1_cur_row, 15) 174 cmhi v30.16b, v26.16b , v28.16b //II vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 180 cmhi v0.16b, v28.16b , v26.16b //II vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 189 EXT v28.16b, v26.16b , v28.16b,#1 //II pu1_cur_row_tmp = vextq_u8(pu1_cur_row, pu1_cur_row_tmp, 1) 203 cmhi v30.16b, v26.16b , v28.16b //II vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 205 cmhi v0.16b, v28.16b , v26.16b //II vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 211 Uxtl v0.8h, v26.8b //II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 240 Uxtl2 v28.8h, v26.16b //II pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(pu1_cur_row))) 318 TBL v26.8b, {v7.16b},v24.8b //offset = vtbl1_s8(offset_tbl, vget_low_s8(edge_idx) [all...] |
ihevc_sao_edge_offset_class3_chroma.s | 548 ADD v26.16b, v0.16b , v17.16b //II edge_idx = vaddq_s8(const_2, sign_up) 554 ADD v26.16b, v26.16b , v24.16b //II edge_idx = vaddq_s8(edge_idx, sign_down) 558 TBL v26.16b, {v21.16b},v26.16b //II vtbl1_s8(edge_idx_tbl, vget_low_s8(edge_idx)) 569 AND v26.16b, v26.16b , v1.16b //II edge_idx = vandq_s8(edge_idx, au1_mask) 570 mov v27.d[0],v26.d[1] 573 UZP1 v31.8b, v26.8b, v27.8b 574 UZP2 v27.8b, v26.8b, v27.8b //I [all...] |
ihevc_intra_pred_chroma_horz.s | 287 dup v26.4h, v0.4h[3] 291 uxtl v26.8h, v26.8b 300 sqadd v22.8h, v26.8h , v24.8h 334 dup v26.8b, v0.8b[3] 335 uxtl v26.8h, v26.8b 344 sqadd v22.8h, v26.8h , v24.8h
|
ihevc_inter_pred_chroma_horz.s | 135 dup v26.8b, v2.8b[2] //coeffabs_2 = vdup_lane_u8(coeffabs, 2) 200 umlal v30.8h, v4.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)// 215 umlal v28.8h, v5.8b, v26.8b 265 umlal v22.8h, v12.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)// 287 umlal v20.8h, v13.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)// 305 umlal v30.8h, v4.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)// 328 umlal v28.8h, v5.8b, v26.8b 361 umlal v22.8h, v12.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)// 379 umlal v20.8h, v13.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)// 396 umlal v30.8h, v4.8b, v26.8b //mul_res = vmlsl_u8(src[0_0], coeffabs_0)/ [all...] |
ihevc_itrans_recon_4x4.s | 166 trn1 v26.4h, v30.4h, v31.4h 168 trn1 v0.2s, v24.2s, v26.2s 169 trn2 v2.2s, v24.2s, v26.2s 202 trn1 v26.4h, v30.4h, v31.4h 204 trn1 v0.2s, v24.2s, v26.2s 205 trn2 v2.2s, v24.2s, v26.2s
|
ihevc_inter_pred_chroma_vert_w16out.s | 257 umull v26.8h, v7.8b, v1.8b 259 umlsl v26.8h, v6.8b, v0.8b 261 umlal v26.8h, v16.8b, v2.8b 263 umlsl v26.8h, v17.8b, v3.8b 297 st1 { v26.16b},[x7],x3 //stores the loaded value 313 umull v26.8h, v7.8b, v1.8b 318 umlsl v26.8h, v6.8b, v0.8b 322 umlal v26.8h, v16.8b, v2.8b 325 umlsl v26.8h, v17.8b, v3.8b 358 st1 { v26.16b},[x7],x3 //stores the loaded valu [all...] |
ihevc_sao_edge_offset_class2.s | 418 Uxtl v26.8h, v5.8b //II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 422 SADDW v26.8h, v26.8h , v24.8b //II pi2_tmp_cur_row.val[0] = vaddw_s8(pi2_tmp_cur_row.val[0], offset) 429 SMAX v26.8h, v26.8h , v2.8h //II pi2_tmp_cur_row.val[0] = vmaxq_s16(pi2_tmp_cur_row.val[0], const_min_clip) 433 UMIN v26.8h, v26.8h , v4.8h //II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vminq_u16(vreinterpretq_u16_s16(pi2_tmp_cur_row.val[0]), const_max_clip)) 448 xtn v26.8b, v26.8h //II vmovn_s16(pi2_tmp_cur_row.val[0]) 450 xtn2 v26.16b, v28.8h //II vmovn_s16(pi2_tmp_cur_row.val[1] [all...] |
/external/llvm/test/MC/AArch64/ |
neon-scalar-by-elem-mla.s | 27 fmls s7, s17, v26.s[3] 34 // CHECK: fmls s7, s17, v26.s[3] // encoding: [0x27,0x5a,0xba,0x5f]
|
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/ppc/ |
platform_altivec.asm | 34 W v26, r3 52 R v26, r3
|
/external/libvpx/libvpx/vp8/common/ppc/ |
platform_altivec.asm | 34 W v26, r3 52 R v26, r3
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/ppc/ |
platform_altivec.asm | 34 W v26, r3 52 R v26, r3
|
/cts/tests/tests/jni/src/android/jni/cts/ |
InstanceFromNative.java | 88 int v25, int v26, int v27, int v28, int v29, 101 (v24 == 24) && (v25 == 25) && (v26 == 26) && (v27 == 27) &&
|
StaticFromNative.java | 85 int v25, int v26, int v27, int v28, int v29, 98 (v24 == 24) && (v25 == 25) && (v26 == 26) && (v27 == 27) &&
|