/system/core/libpixelflinger/ |
rotate90CW_4x4_16v6.S | 38 add r12, r2, r2 47 strd r10, r11, [r0], r12 52 strd r10, r11, [r0], r12 56 strd r10, r11, [r0], r12
|
/frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src_gcc/ |
armVCM4P10_InterpolateLuma_HalfVer4x4_unsafe_s.S | 16 PUSH {r4-r12,lr} 18 ADD r12,r0,r1,LSL #2 20 VLD1.8 {d12},[r12],r1 24 VLD1.8 {d13},[r12],r1 26 VLD1.8 {d14},[r12],r1 29 VLD1.8 {d15},[r12],r1 54 POP {r4-r12,pc}
|
omxVCM4P10_FilterDeblockingChroma_VerEdge_I_s.S | 16 PUSH {r4-r12,lr} 82 ADD r12,r10,r1 87 VST1.8 {d29[1]},[r12],lr 89 VST1.8 {d29[3]},[r12],lr 91 VST1.8 {d29[5]},[r12],lr 93 VST1.8 {d29[7]},[r12],lr 94 ADD r12,r0,#4 95 ADD r10,r12,r1 96 VST1.8 {d24[0]},[r12],lr 98 VST1.8 {d24[2]},[r12],l [all...] |
omxVCM4P10_PredictIntra_4x4_s.S | 25 PUSH {r4-r12,lr} 41 ADD r12,r5,r5 42 VST1.32 {d0[0]},[r3],r12 43 VST1.32 {d1[0]},[r11],r12 50 ADD r12,r5,r5 52 VST1.32 {d0[0]},[r3],r12 53 VST1.32 {d0[0]},[r11],r12 75 ADD r12,r5,r5 84 ADD r12,r5,r5 96 ADD r12,r5,r [all...] |
omxVCM4P10_FilterDeblockingLuma_HorEdge_I_s.S | 16 PUSH {r4-r12,lr} 31 LDRH r12,[r4],#2 33 CMP r12,#0 47 TST r12,#0xff 51 TST r12,#0xff00 58 TST r12,#4 103 POP {r4-r12,pc}
|
/frameworks/av/media/libstagefright/codecs/mp3dec/src/asm/ |
pvmp3_dct_9_arm.s | 57 ldr r12,[r0, #4] 62 add r2,r3,r12 64 sub r3,r3,r12 65 add r12,r4,r5 72 add r7,r1,r12 102 mov r1,r12,lsl #1 106 smlal r12,r2,r9,r1 113 mov r12,r1 ;;;;;; !!!!!! 114 smlal r12,r8,r11,r1 121 smlal r12,r7,r11,r [all...] |
pvmp3_dct_9_gcc.s | 46 ldr r12,[r0, #4] 51 add r2,r3,r12 53 sub r3,r3,r12 54 add r12,r4,r5 61 add r7,r1,r12 86 mov r1,r12,lsl #1 95 smlal r12,r2,r9,r1 102 mov r12,r1 @@@@@@ !!!!!! 103 smlal r12,r8,r11,r1 110 smlal r12,r7,r11,r [all...] |
/external/libvpx/libvpx/vp8/common/arm/armv6/ |
intra4x4_predict_v6.asm | 34 push {r4-r12, lr} 38 pop {r4-r12, pc} ; default 57 usad8 r12, r8, r9 64 add r4, r4, r12 67 mov r12, r4, asr #3 ; (expected_dc + 4) >> 3 69 add r12, r12, r12, lsl #8 71 add r12, r12, r12, lsl #1 [all...] |
iwalsh_v6.asm | 22 stmdb sp!, {r4 - r12, lr} 35 qsub16 r12, r4, r6 ; c1 [5-9 | 4-8] 39 qadd16 r4, r12, lr ; c1 + d1 [5 | 4] 41 qsub16 r8, lr, r12 ; d1 - c1 [13 | 12] 45 qsub16 r12, r5, r7 ; c1 [7-11 | 6-10] 49 qadd16 r5, r12, lr ; c1 + d1 [7 | 6] 51 qsub16 r9, lr, r12 ; d1 - c1 [15 | 14] 57 qsubaddx r12, r4, r5 ; [c1|a1] [5-6 | 4+7] 63 qaddsubx r4, r12, lr ; [b2|c2] [c1+d1 | a1-b1] 64 qaddsubx r5, lr, r12 ; [a2|d2] [b1+a1 | d1-c1 [all...] |
copymem8x4_v6.asm | 39 mov r12, #4 48 subs r12, r12, #1 83 mov r12, #4 86 subs r12, r12, #1 108 mov r12, #4 115 subs r12, r12, #1
|
copymem8x8_v6.asm | 39 mov r12, #8 48 subs r12, r12, #1 83 mov r12, #8 86 subs r12, r12, #1 108 mov r12, #8 115 subs r12, r12, #1
|
dc_only_idct_add_v6.asm | 27 ldr r12, c0x0000FFFF 29 and r0, r12, r0, asr #3 ; input_dc >> 3 + mask 33 ldr r12, [sp, #16] ; dst stride 46 str r5, [r3], r12 48 str r7, [r3], r12 60 str r5, [r3], r12
|
/frameworks/av/media/libstagefright/codecs/amrwbenc/src/asm/ARMV7/ |
Norm_Corr_neon.s | 53 STMFD r13!, {r4 - r12, r14} 113 MOV r12, r1 @copy of xn[] 118 VLD1.S16 {Q4, Q5}, [r12]! @ load 16 x[] 119 VLD1.S16 {Q6, Q7}, [r12]! @ load 16 x[] 139 VLD1.S16 {Q4, Q5}, [r12]! @ load 16 x[] 140 VLD1.S16 {Q6, Q7}, [r12]! @ load 16 x[] 193 MOV r12, r1 195 STMFD sp!, {r0 - r4, r7 - r12, r14} 203 LDMFD sp!, {r0 - r4, r7 - r12, r14} 205 MOV r1, r12 [all...] |
Dot_p_neon.s | 35 STMFD r13!, {r4 - r12, r14} 109 VMOV.S32 r12, D30[0] 111 ADD r12, r12, r12 112 ADD r12, r12, #1 @ L_sum = (L_sum << 1) + 1 113 MOV r4, r12 114 CMP r12, #0 115 RSBLT r4, r12, # [all...] |
pred_lt4_1_neon.s | 35 STMFD r13!, {r4 - r12, r14} 69 LDRSH r12, [r4], #2 75 VDUP.S16 D24, r12 83 VMOV.S32 r12, D30[0] 86 QADD r1, r12, r12 @ L_sum = (L_sum << 2) 96 LDMFD r13!, {r4 - r12, r15}
|
/external/tremolo/Tremolo/ |
mdctARM.s | 61 LDMDB r2!,{r5,r6,r7,r12} 66 MOV r12,r12,ASR #9 @ r12= (*--r)>>9 68 MOV r14,r12,ASR #15 70 EORNE r12,r4, r14,ASR #31 71 STRH r12,[r0], r3 121 LDR r12,[r2],#8 126 RSB r12,r12,# [all...] |
/external/openssl/crypto/bn/asm/ |
x86_64-mont.S | 19 pushq %r12 33 movq %rdx,%r12 35 movq (%r12),%rbx 100 movq (%r12,%r14,8),%rbx 203 movq 24(%rsi),%r12 216 pushq %r12 231 movq %rdx,%r12 233 movq (%r12),%rbx 373 movq (%r12,%r14,8),%rbx 601 movq 24(%rsi),%r12 [all...] |
x86_64-mont5.S | 19 pushq %r12 31 movq %rdx,%r12 38 leaq 96(%r12,%r11,8),%r12 44 movq -96(%r12),%xmm0 45 movq -32(%r12),%xmm1 47 movq 32(%r12),%xmm2 49 movq 96(%r12),%xmm3 54 leaq 256(%r12),%r12 [all...] |
/dalvik/vm/compiler/template/armv5te-vfp/ |
TEMPLATE_RESTORE_STATE.S | 10 ldmia r0, {r0-r12}
|
/external/compiler-rt/lib/arm/ |
comparesf2.S | 58 orrs r12, r2, r3, lsr #1 62 eorsne r12, r0, r1 110 orrs r12, r2, r3, lsr #1 111 eorsne r12, r0, r1
|
/frameworks/av/media/libstagefright/codecs/amrwbenc/src/asm/ARMV5E/ |
syn_filt_opt.s | 36 STMFD r13!, {r4 - r12, r14} 53 LDRH r12, [r4], #2 62 STRH r12, [r5], #2 71 LDRH r12, [r4], #2 80 STRH r12, [r5], #2 97 ORR r12, r9, r11, LSL #16 @ -a[4] -- -a[3] 99 STR r12, [r13, #-8] 108 ORR r12, r9, r11, LSL #16 @ -a[8] -- -a[7] 110 STR r12, [r13, #-16] 119 ORR r12, r9, r11, LSL #16 @ -a[12] -- -a[11 [all...] |
/system/core/libcutils/arch-arm/ |
memset32.S | 58 mov r12, r1 78 stmhsia r0!, {r1,r3,r12,lr} 79 stmhsia r0!, {r1,r3,r12,lr} 85 stmcsia r0!, {r1,r3,r12,lr}
|
/external/openssl/crypto/sha/asm/ |
sha1-armv4-large.S | 10 stmdb sp!,{r4-r12,lr} 26 ldrb r12,[r1],#4 31 orr r9,r9,r12,lsl#24 51 ldrb r12,[r1],#4 56 orr r9,r9,r12,lsl#24 76 ldrb r12,[r1],#4 81 orr r9,r9,r12,lsl#24 101 ldrb r12,[r1],#4 106 orr r9,r9,r12,lsl#24 126 ldrb r12,[r1],# [all...] |
/frameworks/av/media/libstagefright/codecs/aacenc/src/asm/ARMV5E/ |
AutoCorrelation_v5.s | 47 sub r12, r4, #8 75 cmp r8, r12 111 sub r12, r4, #6 138 cmp r3, r12 146 mov r12, r3, lsl #1 147 ldrsh r9, [r6, r12] 148 ldrsh r12, [r5, r12] 151 mul r12, r12, r [all...] |
/external/libvpx/libvpx/vp8/encoder/arm/armv6/ |
vp8_short_fdct4x4_armv6.asm | 21 stmfd sp!, {r4 - r12, lr} 30 ldr r12, c0x22a453a0 ; [2217*4 | 5352*4] 46 smlad r6, r7, r12, r11 ; o1 = (c1 * 2217 + d1 * 5352 + 14500) 47 smlsdx r7, r7, r12, r10 ; o3 = (d1 * 2217 - c1 * 5352 + 7500) 71 smlad r6, r7, r12, r11 ; o5 = (c1 * 2217 + d1 * 5352 + 14500) 72 smlsdx r7, r7, r12, r10 ; o7 = (d1 * 2217 - c1 * 5352 + 7500) 96 smlad r6, r7, r12, r11 ; o9 = (c1 * 2217 + d1 * 5352 + 14500) 97 smlsdx r7, r7, r12, r10 ; o11 = (d1 * 2217 - c1 * 5352 + 7500) 119 smlad r6, r7, r12, r11 ; o13 = (c1 * 2217 + d1 * 5352 + 14500) 120 smlsdx r7, r7, r12, r10 ; o15 = (d1 * 2217 - c1 * 5352 + 7500 [all...] |