/frameworks/av/media/libstagefright/codecs/aacenc/src/asm/ARMV5E/ |
PrePostMDCT_v5.s | 43 smull r14, r11, r4, r8 @ MULHIGH(tr1, cosa) 44 smull r10, r12, r7, r8 @ MULHIGH(ti1, cosa) 46 smull r14, r8, r7, r9 @ MULHIGH(ti1, sina) 47 smull r7, r10, r4, r9 @ MULHIGH(tr1, sina) 55 smull r14, r4, r6, r8 @ MULHIGH(tr2, cosa) 56 smull r10, r12, r5, r8 @ MULHIGH(ti2, cosa) 58 smull r14, r8, r5, r9 @ MULHIGH(ti2, sina) 59 smull r5, r10, r6, r9 @ MULHIGH(tr2, sina) 97 smull r14, r11, r4, r8 @ MULHIGH(tr1, cosa) 98 smull r10, r12, r5, r8 @ MULHIGH(ti1, cosa [all...] |
band_nrg_v5.s | 52 smull r11, r7, r11, r11 54 smull r6, r8, r6, r6 58 smull r11, r7, r11, r11 61 smull r6, r8, r6, r6 124 smull r5, r3, r5, r5 126 smull r8, r4, r8, r8 134 smull r5, r3, r5, r5 136 smull r8, r4, r8, r8 152 smull r5, r3, r5, r5 154 smull r8, r4, r8, r [all...] |
/frameworks/av/media/libstagefright/codecs/mp3dec/src/asm/ |
pvmp3_dct_16_gcc.s | 54 smull lr,r2,r12,r2 63 smull r4,r3,lr,r3 70 smull lr,r2,r12,r2 77 smull r4,r5,lr,r4 82 smull r7,r4,r6,r4 88 smull r12,r4,r1,r12 93 smull r8,lr,r6,lr 101 smull r9,r6,r8,r6 106 smull r12,lr,r9,r12 110 smull r11,r9,r10,r [all...] |
pvmp3_mdct_18_gcc.s | 69 smull r10,lr,r8,lr 73 smull r9,r10,r8,r9 84 smull r8,r9,lr,r8 257 smull r4,lr,r12,r1 262 smull r4,lr,r12,r3 266 smull r4,lr,r12,r2 269 smull r4,lr,r12,r0 272 smull lr,r0,r12,r0 275 smull r0,r2,r12,r2 278 smull r3,r2,r0,r [all...] |
pvmp3_polyphase_filter_window_gcc.s | 78 smull r6,r5,r2,r6 84 smull r6,r2,r5,r8 97 smull r6,r5,r2,r6 102 smull r2,r5,r6,r8 115 smull r6,r5,r2,r6 120 smull r2,r5,r6,r8 134 smull r6,r5,r8,r6 141 smull r8,r7,r6,r3
|
/external/llvm/test/CodeGen/ARM/ |
mulhi.ll | 10 ; V4: smull 13 ; M3: smull 45 ; V4: smull 49 ; M3: smull
|
/frameworks/av/media/libstagefright/codecs/mp3dec/src/ |
pv_mp3dec_fxd_op_arm.h | 62 smull result64_lo, result64_hi, L_var2, L_var1 75 smull result64_lo, result64_hi, L_var2, L_var1 93 smull L_var1, result64_hi, L_var2, L_var1 105 smull result64_lo, result64_hi, L_var2, L_var1 120 smull result64_lo, result64_hi, L_var2, L_var1 135 smull result64_lo, result64_hi, L_var2, L_var1 159 smull L_var2, L_var1, L_var2, L_var1 172 smull result64_lo, result64_hi, L_var2, L_var1
|
pv_mp3dec_fxd_op_arm_gcc.h | 64 asm volatile("smull %1, %0, %2, %3\n\t" 83 asm volatile("smull %1, %0, %2, %3\n\t" 104 "smull %1, %0, %2, %3" 120 asm volatile("smull %1, %0, %2, %3\n\t" 138 asm volatile("smull %1, %0, %2, %3\n\t" 156 asm volatile("smull %1, %0, %2, %3\n\t" 174 asm volatile("smull %1, %0, %2, %3\n\t" 196 asm volatile("smull %1, %0, %2, %3\n\t" 215 asm volatile("smull %1, %0, %2, %3\n\t"
|
/external/chromium_org/third_party/opus/src/silk/arm/ |
macros_armv4.h | 39 "smull %0, %1, %2, %3\n\t" 59 "smull %0, %1, %2, %3\n\t" 79 "smull %0, %1, %2, %3\n\t" 95 "smull %0, %1, %2, %3\n\t"
|
/external/libopus/silk/arm/ |
macros_armv4.h | 39 "smull %0, %1, %2, %3\n\t" 59 "smull %0, %1, %2, %3\n\t" 79 "smull %0, %1, %2, %3\n\t" 95 "smull %0, %1, %2, %3\n\t"
|
/external/chromium_org/third_party/opus/src/celt/arm/ |
kiss_fft_armv4.h | 46 "smull %[tt], %[mi], r1, %[br]\n\t" \ 49 "smull %[br], %[mr], r0, %[br]\n\t" \ 74 "smull %[tt], %[mi], r1, %[br]\n\t" \ 77 "smull %[br], %[mr], r0, %[br]\n\t" \ 102 "smull %[tt], %[mr], r0, %[br]\n\t" \ 105 "smull %[br], %[mi], r1, %[br]\n\t" \
|
fixed_armv4.h | 38 "smull %0, %1, %2, %3\n\t" 55 "smull %0, %1, %2, %3\n\t"
|
/external/libopus/celt/arm/ |
kiss_fft_armv4.h | 46 "smull %[tt], %[mi], r1, %[br]\n\t" \ 49 "smull %[br], %[mr], r0, %[br]\n\t" \ 74 "smull %[tt], %[mi], r1, %[br]\n\t" \ 77 "smull %[br], %[mr], r0, %[br]\n\t" \ 102 "smull %[tt], %[mr], r0, %[br]\n\t" \ 105 "smull %[br], %[mi], r1, %[br]\n\t" \
|
fixed_armv4.h | 38 "smull %0, %1, %2, %3\n\t" 55 "smull %0, %1, %2, %3\n\t"
|
/external/libhevc/common/arm64/ |
ihevc_inter_pred_chroma_vert_w16inp.s | 144 smull v0.4s, v0.4h, v16.4h //vmull_s16(src_tmp1, coeff_0) 147 smull v7.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 192 smull v30.4s, v0.4h, v16.4h //vmull_s16(src_tmp1, coeff_0) 200 smull v28.4s, v1.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 213 smull v26.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 224 smull v24.4s, v3.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 241 smull v30.4s, v0.4h, v16.4h //vmull_s16(src_tmp1, coeff_0) 252 smull v28.4s, v1.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 266 smull v26.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 279 smull v24.4s, v3.4h, v16.4h //vmull_s16(src_tmp2, coeff_0 [all...] |
ihevc_inter_pred_chroma_vert_w16inp_w16out.s | 144 smull v0.4s, v0.4h, v16.4h //vmull_s16(src_tmp1, coeff_0) 147 smull v7.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 191 smull v30.4s, v0.4h, v16.4h //vmull_s16(src_tmp1, coeff_0) 199 smull v28.4s, v1.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 212 smull v26.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 222 smull v24.4s, v3.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 238 smull v30.4s, v0.4h, v16.4h //vmull_s16(src_tmp1, coeff_0) 248 smull v28.4s, v1.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 261 smull v26.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 273 smull v24.4s, v3.4h, v16.4h //vmull_s16(src_tmp2, coeff_0 [all...] |
ihevc_itrans_recon_8x8.s | 187 smull v20.4s, v2.4h, v0.4h[0] //// y0 * cos4(part of c0 and c1) 189 smull v18.4s, v3.4h, v1.4h[2] //// y2 * sin2 (q3 is freed by this time)(part of d1) 192 smull v24.4s, v6.4h, v0.4h[1] //// y1 * cos1(part of b0) 194 smull v26.4s, v6.4h, v0.4h[3] //// y1 * cos3(part of b1) 196 smull v28.4s, v6.4h, v1.4h[1] //// y1 * sin3(part of b2) 198 smull v30.4s, v6.4h, v1.4h[3] //// y1 * sin1(part of b3) 208 smull v22.4s, v10.4h, v0.4h[0] //// y4 * cos4(part of c0 and c1) 210 smull v6.4s, v3.4h, v0.4h[2] //// y2 * cos2(part of d0) 304 smull v24.4s, v6.4h, v0.4h[1] //// y1 * cos1(part of b0) 305 smull v26.4s, v6.4h, v0.4h[3] //// y1 * cos3(part of b1 [all...] |
ihevc_itrans_recon_4x4_ttype1.s | 140 smull v6.4s, v1.4h, v4.4h[2] //74 * pi2_src[1] 145 smull v5.4s, v1.4h, v4.4h[2] //74 * pi2_src[1] 150 smull v7.4s, v0.4h, v4.4h[2] // 74 * pi2_src[0] 154 smull v20.4s, v2.4h, v4.4h[1] // 55 * pi2_src[2] 183 smull v6.4s, v22.4h, v4.4h[2] //74 * pi2_src[1] 188 smull v5.4s, v22.4h, v4.4h[2] //74 * pi2_src[1] 193 smull v7.4s, v21.4h, v4.4h[2] // 74 * pi2_src[0] 198 smull v20.4s, v16.4h, v4.4h[1] // 55 * pi2_src[2]
|
ihevc_weighted_pred_bi.s | 222 smull v4.4s, v0.4h, v7.4h[0] //vmull_n_s16(pi2_src1_val1, (int16_t) wgt0) 224 smull v5.4s, v1.4h, v7.4h[1] //vmull_n_s16(pi2_src2_val1, (int16_t) wgt1) 229 smull v6.4s, v2.4h, v7.4h[0] //vmull_n_s16(pi2_src1_val2, (int16_t) wgt0) ii iteration 233 smull v19.4s, v0.4h, v7.4h[0] //vmull_n_s16(pi2_src1_val1, (int16_t) wgt0) iii iteration 236 smull v17.4s, v3.4h, v7.4h[1] //vmull_n_s16(pi2_src2_val2, (int16_t) wgt1) ii iteration 243 smull v16.4s, v1.4h, v7.4h[1] //vmull_n_s16(pi2_src2_val1, (int16_t) wgt1) iii iteration 251 smull v18.4s, v2.4h, v7.4h[0] //vmull_n_s16(pi2_src1_val2, (int16_t) wgt0) iv iteration 257 smull v20.4s, v3.4h, v7.4h[1] //vmull_n_s16(pi2_src2_val2, (int16_t) wgt1) iv iteration
|
ihevc_itrans_recon_16x16.s | 255 smull v24.4s, v6.4h, v0.4h[1] //// y1 * cos1(part of b0) 256 smull v26.4s, v6.4h, v0.4h[3] //// y1 * cos3(part of b1) 257 smull v28.4s, v6.4h, v1.4h[1] //// y1 * sin3(part of b2) 258 smull v30.4s, v6.4h, v1.4h[3] //// y1 * sin1(part of b3) 270 smull v12.4s, v10.4h, v0.4h[0] 272 smull v14.4s, v10.4h, v0.4h[0] 274 smull v16.4s, v10.4h, v0.4h[0] 276 smull v18.4s, v10.4h, v0.4h[0] 433 smull v24.4s, v6.4h, v2.4h[1] //// y1 * cos1(part of b0) 434 smull v26.4s, v6.4h, v2.4h[3] //// y1 * cos3(part of b1 [all...] |
/external/tremolo/Tremolo/ |
mdctARM.s | 203 SMULL r14,r11,r12,r11 @ (r14,r11) = *l * *wL++ 204 SMULL r14,r6, r7, r6 @ (r14,r6) = *--r * *--wR 242 SMULL r14,r11,r12,r11 @ (r14,r11) = *l * *wL++ 243 SMULL r14,r6, r7, r6 @ (r14,r6) = *--r * *--wR 326 SMULL r8, r9, r7, r11 @ (r8, r9) = s2*T[1] 332 SMULL r8, r12,r7, r10 @ (r8, r12) = s2*T[0] 350 SMULL r8, r9, r6, r10 @ (r8, r9) = s0*T[1] 356 SMULL r8, r12,r7, r10 @ (r8, r12) = s2*T[1] 386 SMULL r14,r12,r8, r11 @ (r14,r12) = ro0*T[1] 392 SMULL r14,r3, r9, r11 @ (r14,r3) = ro2*T[1 [all...] |
asm_arm.h | 45 asm volatile("smull\t%0, %1, %2, %3" 58 asm volatile("smull %0, %1, %2, %3\n\t" 74 asm( "smull %0, %1, %4, %6\n\t" 77 "smull %0, %2, %5, %6\n\t" 93 asm( "smull %0, %1, %4, %6\n\t" 96 "smull %0, %2, %5, %6\n\t" 113 "smull %0, %1, %3, %5\n\t" 115 "smull %0, %2, %4, %5\n\t"
|
/external/llvm/test/Transforms/InstCombine/ |
2012-04-23-Neon-Intrinsics.ll | 71 %a = tail call <4 x i32> @llvm.aarch64.neon.smull.v4i32(<4 x i16> %x, <4 x i16> zeroinitializer) nounwind 79 %a = tail call <4 x i32> @llvm.aarch64.neon.smull.v4i32(<4 x i16> %x, <4 x i16> <i16 1, i16 1, i16 1, i16 1>) nounwind 88 %a = tail call <4 x i32> @llvm.aarch64.neon.smull.v4i32(<4 x i16> <i16 3, i16 3, i16 3, i16 3>, <4 x i16> <i16 2, i16 2, i16 2, i16 2>) nounwind 96 %b = tail call <4 x i32> @llvm.aarch64.neon.smull.v4i32(<4 x i16> <i16 -1, i16 -1, i16 -1, i16 -1>, <4 x i16> <i16 1, i16 1, i16 1, i16 1>) nounwind 112 %a = tail call <4 x i32> @llvm.aarch64.neon.smull.v4i32(<4 x i16> <i16 2, i16 2, i16 2, i16 2>, <4 x i16> %x) nounwind 116 ; CHECK-NEXT: %a = tail call <4 x i32> @llvm.aarch64.neon.smull.v4i32(<4 x i16> <i16 2, i16 2, i16 2, i16 2>, <4 x i16> %x) [[NUW:#[0-9]+]] 122 %a = tail call <4 x i32> @llvm.aarch64.neon.smull.v4i32(<4 x i16> <i16 3, i16 3, i16 3, i16 3>, <4 x i16> <i16 2, i16 2, i16 2, i16 2>) nounwind 130 declare <4 x i32> @llvm.aarch64.neon.smull.v4i32(<4 x i16>, <4 x i16>) nounwind readnone
|
/frameworks/native/opengl/libagl/ |
iterators.S | 63 smull r8, lr, r4, r5 65 smull r6, r7, r4, r12
|
/external/llvm/test/MC/ARM/ |
directive-arch-armv4.s | 34 smull r4, r5, r6, r3
|