HomeSort by relevance Sort by last modified time
    Searched refs:r14 (Results 1 - 25 of 103) sorted by null

1 2 3 4 5

  /frameworks/base/media/libstagefright/codecs/aacenc/src/asm/ARMV5E/
band_nrg_v5.s 44 mov r14, #0
56 qadd r14, r14, r7
60 qadd r14, r14, r8
63 qadd r14, r14, r7
65 qadd r14, r14, r8
69 qadd r14, r14, r14
    [all...]
Radix4FFT_v5.s 35 mov r14, r0 @ xptr = buf@
55 ldrd r0, [r14, #0] @ r0 = xptr[0]@ r1 = xptr[1]@
56 add r14, r14, r9 @ xptr += step@
58 ldrd r10, [r14, #0] @ r2 = xptr[0]@ r3 = xptr[1]@
71 add r14, r14, r9 @ xptr += step@
82 ldrd r10, [r14, #0] @ r4 = xptr[0]@ r5 = xptr[1]@
91 add r14, r14, r9 @ xptr += step@
    [all...]
AutoCorrelation_v5.s 101 mov r14, #0
123 add r7, r14, r7, asr #9
127 ldrsh r14, [r8, #6]
134 mul r14, r11, r14
137 add r14, r7, r14, asr #9
139 add r14, r14, r1, asr #9
152 add r14, r14, r12, asr #9
    [all...]
R4R8First_v5.s 85 mov r14, r0
90 ldrd r0, [r14]
91 ldrd r2, [r14, #8]
92 ldrd r4, [r14, #16]
93 ldrd r6, [r14, #24]
143 ldrd r2, [r14, #32]
144 ldrd r4, [r14, #40]
145 ldrd r6, [r14, #48]
146 ldrd r8, [r14, #56]
181 strd r10, [r14]
    [all...]
  /frameworks/base/media/libstagefright/codecs/amrwbenc/src/asm/ARMV5E/
syn_filt_opt.s 36 STMFD r13!, {r4 - r12, r14}
54 LDRH r14, [r4], #2
63 STRH r14, [r5], #2
72 LDRH r14, [r4], #2
81 STRH r14, [r5], #2
89 LDR r14, =0xffff
94 AND r6, r6, r14
95 AND r9, r9, r14
105 AND r6, r6, r14
106 AND r9, r9, r14
    [all...]
Filt_6k_7k_opt.s 35 STMFD r13!, {r4 - r12, r14}
48 MOV r14, #0
77 ADD r14, r14, #8
78 CMP r14, #80
84 @ not use registers: r4, r10, r12, r14, r5
99 SMULBB r14, r1, r0 @ (x[i] + x[i + 30]) * fir_7k[0]
101 SMLABT r14, r6, r0, r14 @ (x[i + 1] + x[i + 29]) * fir_7k[1]
110 SMLABB r14, r8, r0, r14 @ (x[i + 2] + x[i + 28]) * fir_7k[2]
    [all...]
convolve_opt.s 34 STMFD r13!, {r4 - r12, r14}
53 LDRSH r14, [r4], #-2 @ *tmpH--
55 MLA r8, r12, r14, r8
59 LDRSH r14, [r4], #-2 @ *tmpH--
62 MLA r8, r12, r14, r8
80 LDRSH r14, [r4], #-2
84 MLA r8, r12, r14, r8
92 LDRSH r14, [r4], #-2 @ *tmpH--
94 MLA r8, r12, r14, r8
98 LDRSH r14, [r4], #-2 @ *tmpH-
    [all...]
Deemph_32_opt.s 36 STMFD r13!, {r4 - r12, r14}
55 MOV r14, r10, ASR #16 @y[0] = round(L_tmp)
60 STRH r14, [r2], #2 @update y[0]
62 MUL r9, r14, r8
66 MOV r14, r10, ASR #16 @y[1] = round(L_tmp)
71 STRH r14, [r2], #2
74 MUL r9, r14, r8
81 MOV r14, r10, ASR #16
85 STRH r14, [r2], #2
86 MUL r9, r14, r8
    [all...]
cor_h_vec_opt.s 39 STMFD r13!, {r4 - r12, r14}
56 LDRSH r14, [r9]
59 MLA r6, r12, r14, r6
64 MLA r5, r12, r14, r5
65 MOV r14, #0x8000
67 ADD r10, r6, r14
68 ADD r9, r5, r14
76 MUL r14, r6, r11
78 MOV r6, r14, ASR #15
102 LDRSH r14, [r9]
    [all...]
Syn_filt_32_opt.s 41 STMFD r13!, {r4 - r12, r14}
49 LDR r14, =0xffff
54 AND r6, r6, r14
55 AND r8, r8, r14
65 AND r6, r6, r14
66 AND r8, r8, r14
76 AND r6, r6, r14
77 AND r8, r8, r14
87 AND r6, r6, r14
88 AND r8, r8, r14
    [all...]
Norm_Corr_opt.s 53 STMFD r13!, {r4 - r12, r14}
71 MOV r14, r1 @copy xn[] address
77 LDR r9, [r14], #4
78 LDR r10, [r14], #4
79 LDR r11, [r14], #4
80 LDR r12, [r14], #4
108 ADD r14, r13, #20 @copy of excf[]
112 LDR r11, [r14], #4 @load excf[i], excf[i+1]
118 LDR r11, [r14], #4 @load excf[i+2], excf[i+3]
152 MOV r14, r
    [all...]
  /external/tremolo/Tremolo/
bitwiseARM.s 45 STMFD r13!,{r10,r11,r14}
54 RSB r14,r12,#32 @ r14= 32-bitsLeftInWord
57 MOV r10,r10,LSR r14 @ r10= ptr[0]>>(32-bitsLeftInWord)
59 RSB r14,r14,#32 @ r14= 32-bitsLeftInWord
60 ORRLT r10,r10,r11,LSL r14 @ r10= Next 32 bits.
61 MOV r14,#1
62 RSB r14,r14,r14,LSL r
    [all...]
mdctLARM.s 54 STMFD r13!,{r4-r7,r14}
68 MOV r14,r12,ASR #15
69 TEQ r14,r14,ASR #31 @ if r14==0 || r14==-1 then in range
70 EORNE r12,r4, r14,ASR #31
73 MOV r14,r7, ASR #15
74 TEQ r14,r14,ASR #31 @ if r14==0 || r14==-1 then in rang
    [all...]
mdctARM.s 54 STMFD r13!,{r4-r7,r14}
68 MOV r14,r12,ASR #15
69 TEQ r14,r14,ASR #31 @ if r14==0 || r14==-1 then in range
70 EORNE r12,r4, r14,ASR #31
73 MOV r14,r7, ASR #15
74 TEQ r14,r14,ASR #31 @ if r14==0 || r14==-1 then in rang
    [all...]
dpen.s 62 ADR r14,dpen_read_return
68 STMFD r13!,{r4-r8,r10,r14}
139 LDRNEB r14,[r6, r7] @ r14= t[chase]
140 MOVEQ r14,#128
142 LDRB r14,[r12,r14,LSR #7] @ r14= t[chase+bit+1+(!bit || t[chase]0x0x80)]
145 ORR r0, r14,r10,LSL #8 @ r7 = chase = (next<<8) | r14
    [all...]
  /frameworks/base/media/libstagefright/codecs/amrwbenc/src/asm/ARMV7/
Deemph_32_neon.s 36 STMFD r13!, {r4 - r12, r14}
55 MOV r14, r10, ASR #16 @y[0] = round(L_tmp)
60 STRH r14, [r2], #2 @update y[0]
62 MUL r9, r14, r8
66 MOV r14, r10, ASR #16 @y[1] = round(L_tmp)
71 STRH r14, [r2], #2
74 MUL r9, r14, r8
81 MOV r14, r10, ASR #16
85 STRH r14, [r2], #2
86 MUL r9, r14, r8
    [all...]
cor_h_vec_neon.s 39 STMFD r13!, {r4 - r12, r14}
57 LDRSH r14, [r9]
60 MLA r6, r12, r14, r6
65 MLA r5, r12, r14, r5
66 MOV r14, #0x8000
68 ADD r10, r6, r14
69 ADD r9, r5, r14
77 MUL r14, r6, r11
79 MOV r6, r14, ASR #15
103 LDRSH r14, [r9]
    [all...]
Norm_Corr_neon.s 53 STMFD r13!, {r4 - r12, r14}
71 MOV r14, r1 @copy xn[] address
73 VLD1.S16 {Q0, Q1}, [r14]!
74 VLD1.S16 {Q2, Q3}, [r14]!
75 VLD1.S16 {Q4, Q5}, [r14]!
76 VLD1.S16 {Q6, Q7}, [r14]!
112 ADD r14, r13, #20 @copy of excf[]
116 VLD1.S16 {Q0, Q1}, [r14]! @ load 16 excf[]
117 VLD1.S16 {Q2, Q3}, [r14]! @ load 16 excf[]
137 VLD1.S16 {Q0, Q1}, [r14]! @ load 16 excf[
    [all...]
  /bionic/libc/arch-arm/bionic/
setjmp.S 53 stmfd sp!, {r0, r14}
59 ldmfd sp!, {r0, r14}
69 stmia r1, {r4-r14}
97 stmfd sp!, {r0, r1, r14}
104 ldmfd sp!, {r0, r1, r14}
117 ldmia r2, {r4-r14}
119 /* Validate sp and r14 */
121 teqne r14, #0
131 mov r15, r14
133 mov r15, r14
    [all...]
_setjmp.S 60 stmia r1, {r4-r14}
94 ldmia r2, {r4-r14}
96 /* Validate sp and r14 */
98 teqne r14, #0
  /external/openssl/crypto/sha/asm/
sha1-armv4-large.s 13 mov r14,sp
30 str r9,[r14,#-4]!
45 str r9,[r14,#-4]!
60 str r9,[r14,#-4]!
75 str r9,[r14,#-4]!
90 str r9,[r14,#-4]!
94 teq r14,sp
107 str r9,[r14,#-4]!
111 ldr r9,[r14,#15*4]
112 ldr r10,[r14,#13*4
    [all...]
  /external/openssl/crypto/
s390xcpuid.S 12 br %r14
21 br %r14
34 br %r14
54 br %r14
66 bcr 8,%r14
71 br %r14
86 br %r14
90 brasl %r14,OPENSSL_cpuid_setup
  /system/core/libpixelflinger/
rotate90CW_4x4_16v6.S 37 add r14, r3, r3
40 ldrd r2, r3, [r1], r14
41 ldrd r4, r5, [r1], r14
42 ldrd r6, r7, [r1], r14
  /external/openssl/crypto/bn/asm/
armv4-mont.s 44 mov r14,#0
45 umlal r12,r14,r6,r8 @ np[j]*n0
48 adc r12,r14,#0
53 mov r14,#0
54 adc r14,r14,#0
58 str r14,[r0,#4] @ tp[num]=
85 mov r14,#0
86 umlal r12,r14,r6,r8 @ np[j]*n0
91 adc r12,r14,#
    [all...]
  /external/libffi/src/sh64/
sysv.S 67 st.l r15, 0, r14
69 add.l r15, r63, r14
296 add r14, r63, r15
298 ld.l r15, 0, r14
319 st.l r15, 8, r14
322 add r15, r63, r14
330 4 bytes (r14)
335 fst.d r14, 24, dr0
336 fst.d r14, 32, dr2
337 fst.d r14, 40, dr
    [all...]

Completed in 642 milliseconds

1 2 3 4 5