/external/chromium_org/third_party/libvpx/source/libvpx/vp9/encoder/arm/neon/ |
vp9_sad_neon.c | 43 const uint8x16_t vec_src_00 = vld1q_u8(src); 44 const uint8x16_t vec_src_16 = vld1q_u8(src + 16); 45 const uint8x16_t vec_src_32 = vld1q_u8(src + 32); 46 const uint8x16_t vec_src_48 = vld1q_u8(src + 48); 47 const uint8x16_t vec_ref_00 = vld1q_u8(ref); 48 const uint8x16_t vec_ref_16 = vld1q_u8(ref + 16); 49 const uint8x16_t vec_ref_32 = vld1q_u8(ref + 32); 50 const uint8x16_t vec_ref_48 = vld1q_u8(ref + 48); 80 const uint8x16_t vec_src_00 = vld1q_u8(src); 81 const uint8x16_t vec_src_16 = vld1q_u8(src + 16) [all...] |
vp9_subtract_neon.c | 26 const uint8x16_t v_src_00 = vld1q_u8(&src[c + 0]); 27 const uint8x16_t v_src_16 = vld1q_u8(&src[c + 16]); 28 const uint8x16_t v_pred_00 = vld1q_u8(&pred[c + 0]); 29 const uint8x16_t v_pred_16 = vld1q_u8(&pred[c + 16]); 49 const uint8x16_t v_src = vld1q_u8(&src[0]); 50 const uint8x16_t v_pred = vld1q_u8(&pred[0]);
|
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/arm/neon/ |
sad_neon.c | 124 q0 = vld1q_u8(src_ptr); 126 q4 = vld1q_u8(ref_ptr); 132 q0 = vld1q_u8(src_ptr); 134 q4 = vld1q_u8(ref_ptr); 161 q0 = vld1q_u8(src_ptr); 163 q4 = vld1q_u8(ref_ptr); 169 q0 = vld1q_u8(src_ptr); 171 q4 = vld1q_u8(ref_ptr);
|
vp8_subpixelvariance_neon.c | 57 q11u8 = vld1q_u8(src_ptr); 60 q12u8 = vld1q_u8(src_ptr); 62 q13u8 = vld1q_u8(src_ptr); 64 q14u8 = vld1q_u8(src_ptr); 66 q15u8 = vld1q_u8(src_ptr); 366 q11u8 = vld1q_u8(tmpp); 369 q12u8 = vld1q_u8(tmpp); 371 q13u8 = vld1q_u8(tmpp); 373 q14u8 = vld1q_u8(tmpp); 375 q15u8 = vld1q_u8(tmpp) [all...] |
loopfiltersimplehorizontaledge_neon.c | 28 q5u8 = vld1q_u8(sp); 30 q6u8 = vld1q_u8(sp); 32 q7u8 = vld1q_u8(sp); 34 q8u8 = vld1q_u8(sp);
|
variance_neon.c | 34 q0u8 = vld1q_u8(src_ptr); 36 q1u8 = vld1q_u8(src_ptr); 40 q2u8 = vld1q_u8(ref_ptr); 42 q3u8 = vld1q_u8(ref_ptr); 113 q0u8 = vld1q_u8(src_ptr); 115 q1u8 = vld1q_u8(src_ptr); 119 q2u8 = vld1q_u8(ref_ptr); 121 q3u8 = vld1q_u8(ref_ptr);
|
bilinearpredict_neon.c | 152 q1u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 153 q2u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 154 q3u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 155 q4u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 156 q5u8 = vld1q_u8(src_ptr); 243 q1u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 244 q2u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 245 q3u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 246 q4u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 272 q1u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line [all...] |
copymem_neon.c | 54 qtmp = vld1q_u8(src);
|
loopfilter_neon.c | 140 q3 = vld1q_u8(src); 142 q4 = vld1q_u8(src); 144 q5 = vld1q_u8(src); 146 q6 = vld1q_u8(src); 148 q7 = vld1q_u8(src); 150 q8 = vld1q_u8(src); 152 q9 = vld1q_u8(src); 154 q10 = vld1q_u8(src);
|
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/encoder/arm/neon/ |
subtract_neon.c | 71 q0u8 = vld1q_u8(src); 73 q2u8 = vld1q_u8(src); 75 q1u8 = vld1q_u8(pred); 77 q3u8 = vld1q_u8(pred);
|
vp8_mse16x16_neon.c | 33 q0u8 = vld1q_u8(src_ptr); 35 q1u8 = vld1q_u8(src_ptr); 37 q2u8 = vld1q_u8(ref_ptr); 39 q3u8 = vld1q_u8(ref_ptr);
|
/external/scrypt/lib/crypto/ |
crypto_scrypt-neon-salsa208.h | 41 x0x1x2x3 = vld1q_u8((uint8_t *) input); 42 x4x5x6x7 = vld1q_u8(16 + (uint8_t *) input); 43 x8x9x10x11 = vld1q_u8(32 + (uint8_t *) input); 44 x12x13x14x15 = vld1q_u8(48 + (uint8_t *) input);
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
bilinearpredict_neon.c | 149 q1u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 150 q2u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 151 q3u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 152 q4u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 153 q5u8 = vld1q_u8(src_ptr); 240 q1u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 241 q2u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 242 q3u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 243 q4u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 269 q1u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line [all...] |
copymem_neon.c | 54 qtmp = vld1q_u8(src);
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/arm/neon/ |
bilinearpredict_neon.c | 149 q1u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 150 q2u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 151 q3u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 152 q4u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 153 q5u8 = vld1q_u8(src_ptr); 240 q1u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 241 q2u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 242 q3u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 243 q4u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line; 269 q1u8 = vld1q_u8(src_ptr); src_ptr += src_pixels_per_line [all...] |
copymem_neon.c | 54 qtmp = vld1q_u8(src);
|
/external/libhevc/common/arm/ |
ihevc_sao_edge_offset_class1.s | 125 VLD1.8 D8,[r9]! @pu1_top_row = vld1q_u8(pu1_src_top_cpy || pu1_src - src_strd) 126 VLD1.8 D9,[r9]! @pu1_top_row = vld1q_u8(pu1_src_top_cpy || pu1_src - src_strd) 127 VLD1.8 D10,[r0]! @pu1_cur_row = vld1q_u8(pu1_src) 128 VLD1.8 D11,[r0]! @pu1_cur_row = vld1q_u8(pu1_src) 130 VLD1.8 D30,[r12]! @vld1q_u8(pu1_src[(ht - 1) * src_strd]) 131 VLD1.8 D31,[r12]! @vld1q_u8(pu1_src[(ht - 1) * src_strd]) 142 VLD1.8 D18,[r10]! @pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 143 VLD1.8 D19,[r10] @pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 148 VLD1.8 D30,[r6]! @II pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 149 VLD1.8 D31,[r6] @II pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd [all...] |
ihevc_sao_edge_offset_class1_chroma.s | 130 VLD1.8 D28,[r11]! @pu1_top_row = vld1q_u8(pu1_src_top_cpy || pu1_src - src_strd) 131 VLD1.8 D29,[r11]! @pu1_top_row = vld1q_u8(pu1_src_top_cpy || pu1_src - src_strd) 132 VLD1.8 D10,[r0]! @pu1_cur_row = vld1q_u8(pu1_src) 133 VLD1.8 D11,[r0]! @pu1_cur_row = vld1q_u8(pu1_src) 135 VLD1.8 D30,[r12]! @vld1q_u8(pu1_src[(ht - 1) * src_strd]) 136 VLD1.8 D31,[r12]! @vld1q_u8(pu1_src[(ht - 1) * src_strd]) 147 VLD1.8 D18,[r10]! @pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 148 VLD1.8 D19,[r10] @pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 153 VLD1.8 D30,[r6]! @II pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 154 VLD1.8 D31,[r6] @II pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd [all...] |
ihevc_sao_edge_offset_class2.s | 235 VLD1.8 D10,[r8]! @pu1_top_row = vld1q_u8(pu1_src - src_strd - 1) || vld1q_u8(pu1_src_top_cpy - 1) 236 VLD1.8 D11,[r8] @pu1_top_row = vld1q_u8(pu1_src - src_strd - 1) || vld1q_u8(pu1_src_top_cpy - 1) 241 VLD1.8 D12,[r0]! @pu1_cur_row = vld1q_u8(pu1_src) 242 VLD1.8 D13,[r0] @pu1_cur_row = vld1q_u8(pu1_src) 268 VLD1.8 D16,[r8]! @I pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 269 VLD1.8 D17,[r8] @I pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 333 VLD1.8 D16,[r8]! @II pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 334 VLD1.8 D17,[r8] @II pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd [all...] |
ihevc_sao_edge_offset_class2_chroma.s | 318 VLD1.8 D12,[r0]! @pu1_cur_row = vld1q_u8(pu1_src) 319 VLD1.8 D13,[r0] @pu1_cur_row = vld1q_u8(pu1_src) 333 VLD1.8 D10,[r8]! @pu1_top_row = vld1q_u8(pu1_src - src_strd - 2) || vld1q_u8(pu1_src_top_cpy - 2) 334 VLD1.8 D11,[r8] @pu1_top_row = vld1q_u8(pu1_src - src_strd - 2) || vld1q_u8(pu1_src_top_cpy - 2) 358 VLD1.8 D16,[r8]! @I pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 359 VLD1.8 D17,[r8] @I pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 443 VLD1.8 D16,[r8]! @II pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 444 VLD1.8 D17,[r8] @II pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd [all...] |
ihevc_sao_edge_offset_class3.s | 250 VLD1.8 D10,[r8]! @pu1_top_row = vld1q_u8(pu1_src - src_strd + 1) 251 VLD1.8 D11,[r8] @pu1_top_row = vld1q_u8(pu1_src - src_strd + 1) 256 VLD1.8 D12,[r0]! @pu1_cur_row = vld1q_u8(pu1_src) 257 VLD1.8 D13,[r0] @pu1_cur_row = vld1q_u8(pu1_src) 279 VLD1.8 D16,[r8]! @I pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 280 VLD1.8 D17,[r8] @I pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 349 VLD1.8 D16,[r8]! @II pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 350 VLD1.8 D17,[r8] @II pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 355 VLD1.8 D30,[r2]! @III pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 356 VLD1.8 D31,[r2] @III pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd [all...] |
ihevc_sao_edge_offset_class3_chroma.s | 313 VLD1.8 D12,[r0]! @pu1_cur_row = vld1q_u8(pu1_src) 314 VLD1.8 D13,[r0] @pu1_cur_row = vld1q_u8(pu1_src) 323 VLD1.8 D10,[r8]! @pu1_top_row = vld1q_u8(pu1_src - src_strd + 2) 324 VLD1.8 D11,[r8] @pu1_top_row = vld1q_u8(pu1_src - src_strd + 2) 354 VLD1.8 D16,[r11]! @I pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 355 VLD1.8 D17,[r11] @I pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 439 VLD1.8 D16,[r11]! @II pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 440 VLD1.8 D17,[r11] @II pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 449 VLD1.8 D30,[r4]! @III pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 450 VLD1.8 D31,[r4] @III pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd [all...] |
ihevc_sao_band_offset_chroma.s | 286 VLD2.8 {D5,D6},[r4] @vld1q_u8(pu1_src_cpy) 289 VLD2.8 {D13,D14},[r5] @vld1q_u8(pu1_src_cpy) 292 VLD2.8 {D17,D18},[r6] @vld1q_u8(pu1_src_cpy) 295 VLD2.8 {D21,D22},[r7] @vld1q_u8(pu1_src_cpy) 339 VLD2.8 {D5,D6},[r4] @vld1q_u8(pu1_src_cpy) 343 VLD2.8 {D13,D14},[r5] @vld1q_u8(pu1_src_cpy) 346 VLD2.8 {D17,D18},[r6] @vld1q_u8(pu1_src_cpy) 355 VLD2.8 {D21,D22},[r7] @vld1q_u8(pu1_src_cpy)
|
/external/libhevc/common/arm64/ |
ihevc_sao_edge_offset_class1_chroma.s | 152 LD1 {v28.16b},[x11],#16 //pu1_top_row = vld1q_u8(pu1_src_top_cpy || pu1_src - src_strd) 153 //LD1 {v29.8b},[x11],#8 //pu1_top_row = vld1q_u8(pu1_src_top_cpy || pu1_src - src_strd) 154 LD1 {v3.16b},[x0],#16 //pu1_cur_row = vld1q_u8(pu1_src) 155 //LD1 {v11.8b},[x0],#8 //pu1_cur_row = vld1q_u8(pu1_src) 157 LD1 {v30.16b},[x12],#16 //vld1q_u8(pu1_src[(ht - 1) * src_strd]) 158 //LD1 {v31.8b},[x12],#8 //vld1q_u8(pu1_src[(ht - 1) * src_strd]) 169 LD1 {v18.16b},[x10] //pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 170 //LD1 {v19.8b},[x10] //pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 176 LD1 {v30.16b},[x6] //II pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 177 //LD1 {v31.8b},[x6] //II pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd [all...] |
/external/libpng/arm/ |
filter_neon_intrinsics.c | 53 qrp = vld1q_u8(rp); 54 qpp = vld1q_u8(pp); 67 uint8x16_t vtmp = vld1q_u8(rp); 88 vtmp = vld1q_u8(rp + 12); 146 vtmp = vld1q_u8(rp); 159 vtmp = vld1q_u8(pp); 175 vtmp = vld1q_u8(rp + 12); 278 vtmp = vld1q_u8(rp); 289 vtmp = vld1q_u8(pp); 309 vtmp = vld1q_u8(rp + 12) [all...] |