/external/llvm/test/MC/ARM/ |
neon-mul-encoding.s | 106 vmul.i16 d0, d4[2] 110 vmul.s32 d4, d3[1] 111 vmul.u32 d5, d4[0] 114 vmul.i16 q0, d4[2] 119 vmul.u32 q5, d4[0] 122 vmul.i16 d9, d0, d4[2] 126 vmul.s32 d5, d4, d3[1] 127 vmul.u32 d4, d5, d4[0] 130 vmul.i16 q9, q0, d4[2 [all...] |
fp-armv8.s | 38 vcvtp.s32.f64 s0, d4 39 @ CHECK: vcvtp.s32.f64 s0, d4 @ encoding: [0xc4,0x0b,0xbe,0xfe] 55 vcvtp.u32.f64 s0, d4 56 @ CHECK: vcvtp.u32.f64 s0, d4 @ encoding: [0x44,0x0b,0xbe,0xfe] 74 vseleq.f64 d2, d4, d8 75 @ CHECK: vseleq.f64 d2, d4, d8 @ encoding: [0x08,0x2b,0x04,0xfe] 89 vminnm.f64 d4, d6, d9 90 @ CHECK: vminnm.f64 d4, d6, d9 @ encoding: [0x49,0x4b,0x86,0xfe] 109 vrinta.f64 d3, d4 110 @ CHECK: vrinta.f64 d3, d4 @ encoding: [0x44,0x3b,0xb8,0xfe [all...] |
thumb-fp-armv8.s | 41 vcvtp.s32.f64 s0, d4 42 @ CHECK: vcvtp.s32.f64 s0, d4 @ encoding: [0xbe,0xfe,0xc4,0x0b] 58 vcvtp.u32.f64 s0, d4 59 @ CHECK: vcvtp.u32.f64 s0, d4 @ encoding: [0xbe,0xfe,0x44,0x0b] 77 vseleq.f64 d2, d4, d8 78 @ CHECK: vseleq.f64 d2, d4, d8 @ encoding: [0x04,0xfe,0x08,0x2b] 92 vminnm.f64 d4, d6, d9 93 @ CHECK: vminnm.f64 d4, d6, d9 @ encoding: [0x86,0xfe,0x49,0x4b] 115 vrinta.f64 d3, d4 116 @ CHECK: vrinta.f64 d3, d4 @ encoding: [0xb8,0xfe,0x44,0x3b [all...] |
neon-vld-vst-align.s | [all...] |
neon-vld-encoding.s | 12 vld1.16 {d4, d5, d6}, [r3:64] 15 vld1.8 {d1, d2, d3, d4}, [r3] 16 vld1.16 {d4, d5, d6, d7}, [r3:64] 39 vld1.16 {d4, d5, d6}, [r3:64]! 44 vld1.16 {d4, d5, d6}, [r3:64], r6 48 vld1.8 {d1, d2, d3, d4}, [r3]! 49 vld1.16 {d4, d5, d6, d7}, [r3:64]! 53 vld1.8 {d1, d2, d3, d4}, [r3], r8 54 vld1.16 {d4, d5, d6, d7}, [r3:64], r8 67 @ CHECK: vld1.16 {d4, d5, d6}, [r3:64] @ encoding: [0x5f,0x46,0x23,0xf4 [all...] |
single-precision-fp.s | 6 vsub.f64 d2, d3, d4 7 vdiv.f64 d4, d5, d6 13 @ CHECK-ERRORS-NEXT: vsub.f64 d2, d3, d4 15 @ CHECK-ERRORS-NEXT: vdiv.f64 d4, d5, d6 23 vnmla.f64 d5, d4, d3 26 vfms.f64 d4, d5, d6 34 @ CHECK-ERRORS-NEXT: vnmla.f64 d5, d4, d3 40 @ CHECK-ERRORS-NEXT: vfms.f64 d4, d5, d6 58 vabs.f64 d4, d5 66 @ CHECK-ERRORS-NEXT: vabs.f64 d4, d [all...] |
/art/runtime/arch/arm64/ |
jni_entrypoints_arm64.S | 30 stp d4, d5, [sp, #-16]! 48 ldp d4, d5, [sp], #16
|
/external/libhevc/decoder/arm/ |
ihevcd_itrans_recon_dc_luma.s | 95 vld1.8 d4,[r7],r2 108 vaddw.u8 q13,q0,d4 118 vqmovun.s16 d4,q13 128 vst1.u32 {d4},[r11],r3 159 vld1.8 d4,[r0],r2 167 vaddw.u8 q13,q0,d4 174 vqmovun.s16 d4,q13 181 vst1.u32 {d4[0]},[r1],r3
|
/frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src_gcc/ |
omxVCM4P10_FilterDeblockingChroma_VerEdge_I_s.S | 54 VLD1.8 {d4},[r0],lr 59 VZIP.8 d4,d11 62 VZIP.16 d6,d4 65 VTRN.32 d5,d4 69 VABD.U8 d19,d6,d4 70 VABD.U8 d13,d4,d8 72 VABD.U8 d12,d5,d4 96 VBIF d29,d4,d16
|
omxVCM4P10_FilterDeblockingChroma_HorEdge_I_s.S | 50 VLD1.8 {d4},[r0],r1 52 VABD.U8 d19,d6,d4 54 VABD.U8 d13,d4,d8 57 VABD.U8 d12,d5,d4 80 VBIF d29,d4,d16
|
armVCM4P10_DeblockingLuma_unsafe_s.S | 33 VSUBL.U8 q10,d8,d4 40 VRHADD.U8 d24,d4,d8 55 VMOVL.U8 q14,d4 64 VBIF d29,d4,d16 76 VADDL.U8 q10,d8,d4 86 VHADD.U8 d28,d4,d9 101 VBIF d29,d4,d16
|
armVCM4P10_InterpolateLuma_HalfDiagVerHor4x4_unsafe_s.S | 40 VLD1.8 {d4,d5},[r0],r1 45 VMLAL.U8 q9,d4,d31 52 VADDL.U8 q11,d4,d14 61 VMLSL.U8 q10,d4,d30 104 VEXT.8 d29,d23,d4,#2 124 VQRSHRUN.S32 d4,q2,#10 128 VQMOVN.U16 d4,q2
|
omxVCM4P10_FilterDeblockingLuma_VerEdge_I_s.S | 58 VLD1.8 {d4},[r0],r11 62 VZIP.8 d4,d11 66 VZIP.16 d6,d4 68 VTRN.32 d5,d4 71 VABD.U8 d13,d4,d8 72 VABD.U8 d12,d5,d4 74 VABD.U8 d19,d6,d4
|
/external/libhevc/common/arm/ |
ihevc_deblk_chroma_vert.s | 69 vld1.8 {d4},[r8] 73 vtrn.8 d16,d4 86 vtrn.16 d17,d4 94 vtrn.32 d16,d4 114 vsubw.u8 q2,q0,d4
|
ihevc_padding.s | 138 vst1.8 {d4,d5},[r6]! @128/8 = 16 bytes store 139 vst1.8 {d4,d5},[r6]! @128/8 = 16 bytes store 140 vst1.8 {d4,d5},[r6]! @128/8 = 16 bytes store 141 vst1.8 {d4,d5},[r6]! @128/8 = 16 bytes store 142 vst1.8 {d4,d5},[r6] @128/8 = 16 bytes store 257 vst1.8 {d4,d5},[r6]! @128/8 = 16 bytes store 258 vst1.8 {d4,d5},[r6]! @128/8 = 16 bytes store 259 vst1.8 {d4,d5},[r6]! @128/8 = 16 bytes store 260 vst1.8 {d4,d5},[r6]! @128/8 = 16 bytes store 261 vst1.8 {d4,d5},[r6] @128/8 = 16 bytes stor [all...] |
ihevc_itrans_recon_4x4.s | 140 vld1.16 d4,[r8] @loading first row of g_ai2_ihevc_trans_4_transpose 141 @ d4 = {36,64,83,64} 152 vmull.s16 q3,d1,d4[1] @83 * pi2_src[1] 153 vmlal.s16 q3,d3,d4[3] @o[0] = 83 * pi2_src[1] + 36 * pi2_src[3] 154 vmull.s16 q4,d1,d4[3] @36 * pi2_src[1] 156 vmlsl.s16 q4,d3,d4[1] @o[1] = 36 * pi2_src[1] - 83 * pi2_src[3] 181 vmull.s16 q3,d1,d4[1] @83 * pi2_src[1] 183 vmlal.s16 q3,d3,d4[3] @o[0] = 83 * pi2_src[1] + 36 * pi2_src[3] 184 vmull.s16 q4,d1,d4[3] @36 * pi2_src[1] 185 vmlsl.s16 q4,d3,d4[1] @o[1] = 36 * pi2_src[1] - 83 * pi2_src[3 [all...] |
ihevc_intra_pred_luma_mode_3_to_9.s | 206 vsub.s8 d4, d8, d2 @ref_main_idx (row 1) 209 vtbl.8 d16, {d0,d1}, d4 @load from ref_main_idx (row 1) 224 vsub.s8 d4, d4, d3 @ref_main_idx (row 3) 230 vtbl.8 d10, {d0,d1}, d4 @load from ref_main_idx (row 3) 246 vsub.s8 d4, d4, d3 @ref_main_idx (row 5) 252 vtbl.8 d16, {d0,d1}, d4 @load from ref_main_idx (row 5) 268 vsub.s8 d4, d4, d3 @ref_main_idx (row 7 [all...] |
/external/libvpx/libvpx/vp8/common/arm/neon/ |
shortidct4x4llm_neon.asm | 44 vswp d3, d4 ;q2(vp[4] vp[12]) 69 vqsub.s16 d4, d13, d10 72 vtrn.32 d2, d4 75 vtrn.16 d4, d5 77 vswp d3, d4 96 vqsub.s16 d4, d13, d10 101 vrshr.s16 d4, d4, #3 104 vtrn.32 d2, d4 107 vtrn.16 d4, d [all...] |
dequant_idct_neon.c | 23 int16x4_t d2, d3, d4, d5, d10, d11, d12, d13; local 77 d4 = vqsub_s16(d13, d10); 80 d2tmp0 = vtrn_s32(vreinterpret_s32_s16(d2), vreinterpret_s32_s16(d4)); 107 d4 = vqsub_s16(d13, d10); 112 d4 = vrshr_n_s16(d4, 3); 115 d2tmp0 = vtrn_s32(vreinterpret_s32_s16(d2), vreinterpret_s32_s16(d4));
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/arm/neon/ |
shortidct4x4llm_neon.asm | 44 vswp d3, d4 ;q2(vp[4] vp[12]) 69 vqsub.s16 d4, d13, d10 72 vtrn.32 d2, d4 75 vtrn.16 d4, d5 77 vswp d3, d4 96 vqsub.s16 d4, d13, d10 101 vrshr.s16 d4, d4, #3 104 vtrn.32 d2, d4 107 vtrn.16 d4, d [all...] |
/frameworks/rs/cpu_ref/ |
rsCpuIntrinsics_neon_3DLUT.S | 140 2: vld4.u8 {d0,d2,d4,d6}, [r1]! 149 vmovl.u8 q2, d4 189 lanepair dst=d12, src=d12, xr0=d0[0], xr1=d0[1], yr0=d2[0], yr1=d2[1], zr0=d4[0], zr1=d4[1] 192 lanepair dst=d13, src=d13, xr0=d0[2], xr1=d0[3], yr0=d2[2], yr1=d2[3], zr0=d4[2], zr1=d4[3] 220 vmov d4, d0 228 vld1.u32 {d4}, [r1]! 233 vuzp.8 d4, d6 234 vuzp.8 d0, d4 [all...] |
rsCpuIntrinsics_neon_Convolve.S | 64 d4, d5, d6, d7 69 vmull.s16 q8, d4, d0[0] 133 vld1.16 {d4, d5, d6}, [r6] 217 vmlal.s16 q4, d25, d4[0] 218 vmlal.s16 q4, d26, d4[1] 219 vmlal.s16 q4, d27, d4[2] 220 vmlal.s16 q4, d28, d4[3] 229 vmlal.s16 q5, d26, d4[0] 230 vmlal.s16 q5, d27, d4[1] 231 vmlal.s16 q5, d28, d4[2 [all...] |
/external/valgrind/main/none/tests/arm/ |
neon64.c | 735 TESTINSN_bin("vand d4, d6, d5", d4, d6, i8, 0xff, d5, i16, 0x57); 742 TESTINSN_bin("vbic d4, d6, d5", d4, d6, i8, 0xff, d5, i16, 0x57); 750 TESTINSN_bin("vorr d4, d4, d4", d4, d4, i16, 0xff, d4, i16, 0xff) [all...] |
/external/chromium_org/v8/test/mjsunit/ |
generated-transition-stub.js | 193 d4 = [, 2.5, ,]; 195 d4[i] = 0; 197 assertTrue(%HasFastDoubleElements(d4)); 198 assertTrue(%HasFastHoleyElements(d4)); 199 transition4(d4, 0, new Array(5)); 200 assertTrue(%HasFastHoleyElements(d4)); 201 assertTrue(%HasFastObjectElements(d4)); 202 assertEquals(5, d4[0].length); 203 assertEquals(undefined, d4[2]);
|
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/arm/neon/ |
dequant_idct_neon.c | 23 int16x4_t d2, d3, d4, d5, d10, d11, d12, d13; local 77 d4 = vqsub_s16(d13, d10); 80 d2tmp0 = vtrn_s32(vreinterpret_s32_s16(d2), vreinterpret_s32_s16(d4)); 107 d4 = vqsub_s16(d13, d10); 112 d4 = vrshr_n_s16(d4, 3); 115 d2tmp0 = vtrn_s32(vreinterpret_s32_s16(d2), vreinterpret_s32_s16(d4));
|