/frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src_gcc/ |
armVCM4P10_InterpolateLuma_DiagCopy_unsafe_s.S | 19 LDR r12, =0xff00ff 33 AND r11,r12,r11,LSR #5 34 AND r10,r12,r10,LSR #5 35 AND r5,r12,r5,LSR #5 36 AND r4,r12,r4,LSR #5 52 LDR r12, =0xff00ff 67 AND r11,r12,r11,LSR #5 68 AND r10,r12,r10,LSR #5 69 AND r5,r12,r5,LSR #5 70 AND r4,r12,r4,LSR # [all...] |
omxVCM4P10_InterpolateLuma_s.S | 16 PUSH {r4-r12,lr} 48 ADD r12,r0,r1,LSL #1 50 VLD1.8 {d11},[r12],r1 52 VLD1.8 {d12},[r12] 53 ADD r12,r2,r3,LSL #1 55 VST1.32 {d11[0]},[r12],r3 57 VST1.32 {d12[0]},[r12] 67 ADD r12,r2,r3,LSL #1 69 VST1.32 {d26[0]},[r12],r3 71 VST1.32 {d28[0]},[r12] [all...] |
/external/libvpx/libvpx/vp8/common/arm/armv6/ |
loopfilter_v6.asm | 88 ldr r12, [src], pstep ; p0 99 uqsub8 r6, r11, r12 ; p1 - p0 101 uqsub8 r7, r12, r11 ; p0 - p1 111 uqsub8 r11, r12, r9 ; p0 - q0 112 uqsub8 r12, r9, r12 ; q0 - p0 115 orr r12, r11, r12 ; abs (p0-q0) 117 uqadd8 r12, r12, r12 ; abs (p0-q0) * [all...] |
dequant_idct_v6.asm | 29 mov r12, #4 44 subs r12, r12, #1 63 ldr r12, [r0, #24] 70 smulwt r11, r3, r12 73 smulwt r7, r4, r12 74 smulwb r9, r3, r12 75 smulwb r10, r4, r12 80 uadd16 r7, r12, r9 104 smulwt r12, r4, r [all...] |
idct_v6.asm | 39 ldr r12, [r0, #(12*2)] ; i13|i12 47 smulbt r11, r5, r12 ; (ip[13] * cospi8sqrt2minus1) >> 16 52 smulwt r7, r4, r12 ; (ip[13] * sinpi8sqrt2) >> 16 53 smulbb r9, r5, r12 ; (ip[12] * cospi8sqrt2minus1) >> 16 54 smulwb r10, r4, r12 ; (ip[12] * sinpi8sqrt2) >> 16 61 uadd16 r7, r12, r9 ; 13c+13 | 12c+12 87 ldr r12,[r0, #(2*2)] ; i3 | i2 102 pkhbt r9, r14, r12, lsl #16 ; i2 | i6 105 pkhtb r6, r12, r14, asr #16 ; i3 | i7 111 smulbb r12, r5, r6 ; (ip[7] * cospi8sqrt2minus1) >> 1 [all...] |
/hardware/samsung_slsi/exynos5/libswconverter/ |
csc_tiled_to_linear_uv_neon.s | 66 @r12 temp1 69 stmfd sp!, {r4-r12,r14} @ backup registers 86 mul r12, r2, r5 @ temp1 = width * i + j; 88 add r12, r12, r6 91 add r7, r0, r12 @ dst_offset = y_dst + temp1 114 mul r12, r2, r5 @ temp1 = width * i + j; 115 add r12, r12, r6 116 add r7, r0, r12 @ r7 = y_dst + temp [all...] |
csc_tiled_to_linear_y_neon.s | 67 @r12 temp1 70 stmfd sp!, {r4-r12,r14} @ backup registers 91 mul r12, r2, r5 @ temp1 = width * i + j; 94 add r12, r12, r6 96 add r7, r0, r12 @ dst_offset = y_dst + temp1 129 mul r12, r2, r5 @ temp1 = width * i + j; 130 add r12, r12, r6 131 add r7, r0, r12 @ r7 = y_dst + temp [all...] |
/frameworks/av/media/libstagefright/codecs/mp3dec/src/asm/ |
pvmp3_mdct_18_arm.s | 51 add r12,r3,#0x44 72 ldr r8,[r12],#-4 108 ldr r12,[r5,#0x14] 110 str r12,[r5,#0x28] 111 ldr r12,[r5,#0x10] 112 str r12,[r5,#0x20] 113 ldr r12,[r5,#0xc] 114 str r12,[r5,#0x18] 115 ldr r12,[r5,#8] 116 str r12,[r5,#0x10 [all...] |
pvmp3_mdct_18_wm.asm | 49 add r12,r3,#0x44 70 ldr r8,[r12],#-4 106 ldr r12,[r5,#0x14] 108 str r12,[r5,#0x28] 109 ldr r12,[r5,#0x10] 110 str r12,[r5,#0x20] 111 ldr r12,[r5,#0xc] 112 str r12,[r5,#0x18] 113 ldr r12,[r5,#8] 114 str r12,[r5,#0x10 [all...] |
pvmp3_polyphase_filter_window_arm.s | 70 add r12,r0,r2,lsl #2 71 ldr r6,[r12,#0x780] 75 ldr r7,[r12,#0x80] 90 ldr r6,[r12,#0x680] 94 ldr r7,[r12,#0x180] 108 ldr r6,[r12,#0x580] 112 ldr r7,[r12,#0x280] 127 ldr r6,[r12,#0x480] 131 ldr r12,[r12,#0x380 [all...] |
pvmp3_polyphase_filter_window_gcc.s | 70 add r12,r0,r2,lsl #2 71 ldr r6,[r12,#0x780] 75 ldr r7,[r12,#0x80] 90 ldr r6,[r12,#0x680] 94 ldr r7,[r12,#0x180] 108 ldr r6,[r12,#0x580] 112 ldr r7,[r12,#0x280] 127 ldr r6,[r12,#0x480] 131 ldr r12,[r12,#0x380 [all...] |
pvmp3_mdct_18_gcc.s | 49 add r12,r3,#0x44 70 ldr r8,[r12],#-4 106 ldr r12,[r5,#0x14] 108 str r12,[r5,#0x28] 109 ldr r12,[r5,#0x10] 110 str r12,[r5,#0x20] 111 ldr r12,[r5,#0xc] 112 str r12,[r5,#0x18] 113 ldr r12,[r5,#8] 114 str r12,[r5,#0x10 [all...] |
pvmp3_polyphase_filter_window_wm.asm | 65 add r12,r0,r2,lsl #2 66 ldr r6,[r12,#0x780] 70 ldr r7,[r12,#0x80] 85 ldr r6,[r12,#0x680] 89 ldr r7,[r12,#0x180] 103 ldr r6,[r12,#0x580] 107 ldr r7,[r12,#0x280] 122 ldr r6,[r12,#0x480] 126 ldr r12,[r12,#0x380 [all...] |
/external/openssl/crypto/bn/asm/ |
ppc.pl | 235 #.set r12,12 841 # r10, r11, r12 are the equivalents of c1, c2, and c3. 855 adde r12,r9,r0 863 adde r12,r9,r12 870 addc r12,r8,r12 878 addc r12,r8,r12 886 addc r12,r8,r1 [all...] |
armv4-mont.S | 15 stmdb sp!,{r4-r12,lr} @ save 10 registers 34 mov r12,#0 35 umlal r10,r12,r6,r8 @ np[0]*n0+"t[0]" 45 umlal r12,r14,r6,r8 @ np[j]*n0 46 adds r12,r12,r10 47 str r12,[r4],#4 @ tp[j-1]=,tp++ 48 adc r12,r14,#0 52 adds r12,r12,r1 [all...] |
/frameworks/av/media/libstagefright/codecs/amrwbenc/src/asm/ARMV5E/ |
pred_lt4_1_opt.s | 38 STMFD r13!, {r4 - r12, r14} 70 SMULBB r12, r9, r3 @x[2] * h[0] 75 SMLABT r12, r4, r3, r12 @x[3] * h[1] 81 SMLABB r12, r6, r3, r12 @x[4] * h[2] 86 SMLABT r12, r9, r3, r12 @x[5] * h[3] 92 SMLABB r12, r4, r3, r12 @x[6] * h[4 [all...] |
Norm_Corr_opt.s | 53 STMFD r13!, {r4 - r12, r14} 80 LDR r12, [r14], #4 87 SMLABB r6, r12, r12, r6 88 SMLATT r6, r12, r12, r6 107 MOV r12, r1 @copy of xn[] 113 LDR r10, [r12], #4 @load xn[i], xn[i+1] 119 LDR r10, [r12], #4 @load xn[i+2], xn[i+3] 153 MOV r12, r [all...] |
/external/openssl/crypto/modes/asm/ |
ghash-armv4.S | 26 sub r12,pc,#8 29 sub r12,r12,#48 @ &rem_4bit 31 ldmia r12,{r4-r11} @ copy rem_4bit ... 34 ldrb r12,[r2,#15] 37 eor r12,r12,r14 38 and r14,r12,#0xf0 39 and r12,r12,#0x0 [all...] |
/frameworks/av/media/libstagefright/codecs/aacenc/src/asm/ARMV5E/ |
band_nrg_v5.s | 32 ldr r12, [r13, #36] 78 str r9, [r12, #0] 89 mov r12, #0 109 str r12, [r13, #12] 119 ldr r12, [r0, +r10, lsl #2] 125 mov r12, r12, asr #1 130 add r5, r12, r14 132 sub r8, r12, r14 147 ldr r12, [r0, +r10, lsl #2 [all...] |
/bionic/libc/arch-arm/generic/bionic/ |
memcpy.S | 71 movs r12, r3, lsl #31 75 ldrcsb r12,[r1], #1 78 strcsb r12,[r0], #1 83 eor r12, r0, r1 84 tst r12, #3 100 movs r12, r3, lsl #28 138 bic r12, r1, #0x1F 139 add r12, r12, #64 142 PLD (r12, #64 [all...] |
/external/tremolo/Tremolo/ |
bitwiseARM.s | 46 LDMIA r0,{r2,r3,r12} 49 @ r12= bitsLeftInWord 54 RSB r14,r12,#32 @ r14= 32-bitsLeftInWord 55 SUBS r12,r12,r1 @ r12= bitsLeftInWord -= bits 58 ADDLE r12,r12,#32 @ r12= bitsLeftInWord += 32 71 @ r12 = bitsLeftInWor [all...] |
/external/openssl/crypto/aes/asm/ |
aes-armv4.S | 118 stmdb sp!,{r1,r4-r12,lr} 119 mov r12,r0 @ inp 123 ldrb r0,[r12,#3] @ load input data in endian-neutral 124 ldrb r4,[r12,#2] @ manner... 125 ldrb r5,[r12,#1] 126 ldrb r6,[r12,#0] 128 ldrb r1,[r12,#7] 130 ldrb r4,[r12,#6] 132 ldrb r5,[r12,#5] 133 ldrb r6,[r12,#4 [all...] |
/dalvik/vm/compiler/template/armv5te/ |
TEMPLATE_MEM_OP_DECODE.S | 10 push {r0-r12,lr} @ save out all registers 15 pop {r0-r12,lr} @ restore all registers
|
/dalvik/vm/compiler/template/armv5te-vfp/ |
TEMPLATE_MEM_OP_DECODE.S | 11 push {r0-r12,lr} @ save out all registers 16 pop {r0-r12,lr} @ restore all registers
|
/system/core/libpixelflinger/ |
col32cb16blend.S | 47 and r12, r9, r1, lsr #8 // extract green 50 mov r12, r12, lsl #6 // prescale green 62 smlabb r7, r7, r5, r12 // dest green * alpha + src green
|