/external/tremolo/Tremolo/ |
mdctLARM.s | 187 LDMFD r12,{r8,r9,r10} @ r8 = step 191 MOV r8, r8, LSL #1 209 STRH r6, [r0], r8 227 LDMFD r12,{r8,r9,r10} @ r8 = step 231 MOV r8, r8, LSL #1 250 STRH r6, [r0], r8 [all...] |
/dalvik/vm/compiler/template/armv5te/ |
TEMPLATE_INVOKE_METHOD_PREDICTED_CHAIN.S | 27 ldr r8, [r2, #4] @ r8 <- predictedChainCell->clazz 30 cmp r3, r8 @ predicted class == actual class? 48 cmp r8, #0 @ initialized class or not
|
header.S | 27 r4-r8 are available for general use 66 r8 rIBASE interpreted instruction base pointer, used for computed goto 78 #define rIBASE r8
|
TEMPLATE_RETURN.S | 18 ldrb r8, [rSELF, #offThread_breakFlags] @ r8<- breakFlags 42 cmp r8, #0 @ check the break flags
|
/frameworks/base/media/libstagefright/codecs/aacenc/src/asm/ARMV7/ |
Radix4FFT_v7.s | 36 mov r8, r0 54 VLD2.I32 {D0, D1, D2, D3}, [r8] 57 add r8, r8, r5 @ xptr += step@ 58 VLD2.I32 {D4, D5, D6,D7}, [r8] @ r2 = xptr[0]@ r3 = xptr[1]@ 68 add r8, r8, r5 @ xptr += step@ 77 VLD2.I32 {D8, D9, D10, D11}, [r8] 79 add r8, r8, r [all...] |
/external/valgrind/main/coregrind/m_dispatch/ |
dispatch-arm-linux.S | 57 push {r0, r1, r4, r5, r6, r7, r8, r9, fp, lr} 65 mov r8, r0 66 ldr r0, [r8, #OFFSET_arm_R15T] 81 /* AT ENTRY: r0 is next guest addr, r8 is possibly 86 cmp r8, r1 90 str r0, [r8, #OFFSET_arm_R15T] 109 // r1: live, next-host r8: live, gsp 126 /* AT ENTRY: r0 is next guest addr, r8 is possibly 131 cmp r8, r1 135 str r0, [r8, #OFFSET_arm_R15T [all...] |
/external/libvpx/vp8/encoder/arm/armv6/ |
vp8_subtract_armv6.asm | 38 ldr r8, [r1, #vp8_blockd_predictor] 45 ldr r1, [r8], r2 ; pred 92 uxtb16 r8, r6 ; [s2 | s0] (A) 97 usub16 r6, r8, r9 ; [d2 | d0] (A) 103 pkhbt r8, r6, r7, lsl #16 ; [d1 | d0] (A) 106 str r8, [r0], #4 ; diff (A) 107 uxtb16 r8, r10 ; [s2 | s0] (B) 114 usub16 r6, r8, r9 ; [d2 | d0] (B) 119 pkhbt r8, r6, r7, lsl #16 ; [d1 | d0] (B) 122 str r8, [r0], #4 ; diff (B [all...] |
vp8_mse16x16_armv6.asm | 41 usub8 r8, r5, r6 ; calculate difference 42 sel r7, r8, lr ; select bytes with positive difference 44 sel r8, r9, lr ; select bytes with negative difference 48 usad8 r6, r8, lr ; calculate sum of negative differences 49 orr r8, r8, r7 ; differences of all 4 pixels 54 uxtb16 r6, r8 ; byte (two pixels) to halfwords 55 uxtb16 r7, r8, ror #8 ; another two pixels to halfwords 62 usub8 r8, r5, r6 ; calculate difference 63 sel r7, r8, lr ; select bytes with positive differenc [all...] |
vp8_variance8x8_armv6.asm | 37 usub8 r8, r6, r7 ; calculate difference 38 sel r10, r8, lr ; select bytes with positive difference 40 sel r8, r9, lr ; select bytes with negative difference 44 usad8 r7, r8, lr ; calculate sum of negative differences 45 orr r8, r8, r10 ; differences of all 4 pixels 51 uxtb16 r7, r8 ; byte (two pixels) to halfwords 52 uxtb16 r10, r8, ror #8 ; another two pixels to halfwords 60 usub8 r8, r6, r7 ; calculate difference 62 sel r10, r8, lr ; select bytes with positive differenc [all...] |
vp8_sad16x16_armv6.asm | 33 ldr r8, [r2, #0x0] ; load 4 ref pixels (1A) 39 usada8 r4, r8, r6, r4 ; calculate sad for 4 pixels 40 usad8 r8, r7, r9 ; calculate sad for 4 pixels 49 usada8 r8, r11, lr, r8 ; calculate sad for 4 pixels 53 add r4, r4, r8 ; add partial sad values 56 ldr r8, [r2, #0x0] ; load 4 ref pixels (2A) 61 usada8 r4, r6, r8, r4 ; calculate sad for 4 pixels 62 usad8 r8, r7, r9 ; calculate sad for 4 pixels 71 usada8 r8, r11, lr, r8 ; calculate sad for 4 pixel [all...] |
/external/libvpx/vp8/common/arm/armv6/ |
idct_v6.asm | 12 ; r0 r1 r2 r3 r4 r5 r6 r7 r8 r9 r10 r11 r12 r14 69 ldrsh r8, [r0, #16] ; input[8] 1, r8 unavail 2 [8] 73 add r9, r7, r8 ; a1 = [0] + [8] 1 a1 74 sub r7, r7, r8 ; b1 = [0] - [8] 1 b1 79 add r8, r7, r11 ; b1 + c1 1 b+c 80 strh r8, [r1, r2] ; out[pitch] = b1+c1 1 86 add r8, r2, r2 ; pitch * 2 1 p*2 87 strh r7, [r1, r8] ; out[pitch*2] = b1-c1 1 100 ldrsh r8, [r1, #6] ; input[3] 1, r8 un 2 [3 [all...] |
simpleloopfilter_v6.asm | 83 uqsub8 r8, r6, r3 ; q1 - p1 86 orr r8, r8, r7 ; abs(p1 - q1) 89 uhadd8 r8, r8, lr ; abs(p1 - q2) >> 1 90 uqadd8 r10, r10, r8 ; abs(p0 - q0)*2 + abs(p1 - q1)/2 91 mvn r8, #0 93 sel r10, r8, lr ; filter mask: F or 0 109 ldr r8, c0x03030303 115 qadd8 r8 , r3 , r8 ; Filter2 = vp8_filter + [all...] |
/external/webrtc/src/common_audio/signal_processing_library/main/source/ |
webrtc_fft_4ofq14_gcc_android.s | 20 ldrsh r8, [r12], +r2
29 sub r12, r8, r10
31 add r10, r8, r10
34 sub r8, r5, r11
60 ldrsh r8, [r4], +r2
65 add r6, r6, r8, asr #3
67 sub r8, r6, r8, asr #2
77 sub r12, r8, r5, asr #3
79 add r8, r8, r5, asr #3 [all...] |
webrtc_fft_4oiq14_gcc_android.s | 20 ldrsh r8, [r12], +r2
29 sub r12, r8, r10
31 add r10, r8, r10
34 sub r8, r5, r11
60 ldrsh r8, [r4], +r2
63 add r6, r6, r8
65 sub r8, r6, r8, lsl #1
75 add r12, r8, r5
77 sub r8, r8, r5 [all...] |
/frameworks/base/media/libstagefright/codecs/amrwbenc/src/asm/ARMV5E/ |
Filt_6k_7k_opt.s | 37 MOV r8, r0 @ copy signal[] address 49 MOV r3, r8 @ change myMemCopy to Copy, due to Copy will change r3 content 53 LDRSH r8, [r7], #2 55 MOV r8, r8, ASR #2 61 STRH r8, [r6], #2 65 LDRSH r8, [r7], #2 67 MOV r8, r8, ASR #2 73 STRH r8, [r6], # [all...] |
residu_asm_opt.s | 44 LDRH r8, [r0], #2 45 ORR r7, r8, r7, LSL #16 @r7 --- a4, a5 47 LDRH r8, [r0], #2 49 ORR r8, r9, r8, LSL #16 @r8 --- a6, a7 68 STMFD r13!, {r8 - r12} @store r8-r12 109 SMLATT r12,r8, r2, r12 @ i4 (6) 112 SMLATB r11,r8, r2, r11 @ i3 (6 [all...] |
/frameworks/base/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src_gcc/ |
armVCM4P10_InterpolateLuma_Align_unsafe_s.S | 16 MOV r12,r8 29 STM r8!,{r7,r10,r11} 41 STM r8!,{r7,r10,r11} 53 STM r8!,{r7,r10,r11} 65 STM r8!,{r7,r10,r11} 87 STR r7,[r8],#4 96 STR r7,[r8],#4 105 STR r7,[r8],#4 114 STR r7,[r8],#4 117 SUB r0,r8,#0x1 [all...] |
/bionic/libdl/arch-sh/ |
sobegin.S | 39 mov.l r8, @-r15 40 mov.l 0f, r8 /* first entry is invalid */ 42 add #4, r8 43 mov.l @r8, r0 51 mov.l @r15+, r8
|
/external/libvpx/vpx_scale/arm/neon/ |
vp8_vpxyv12_copyframe_func_neon.asm | 30 ldr r8, [r0, #yv12_buffer_config_u_buffer] ;srcptr1 42 str r8, [sp] 51 mov r8, r2 58 vld1.8 {q0, q1}, [r8]! 60 vld1.8 {q2, q3}, [r8]! 62 vld1.8 {q4, q5}, [r8]! 64 vld1.8 {q6, q7}, [r8]! 109 mov r8, r2 116 vld1.8 {q0, q1}, [r8]! 118 vld1.8 {q2, q3}, [r8]! [all...] |
/frameworks/base/media/libstagefright/codecs/aacenc/src/asm/ARMV5E/ |
Radix4FFT_v5.s | 59 ldr r8, [r12], #4 @ cosxsinx = csptr[0]@ 61 smulwt r4, r10, r8 @ L_mpy_wx(cosx, t0) 62 smulwt r3, r11, r8 @ L_mpy_wx(cosx, t1) 64 smlawb r2, r11, r8, r4 @ r2 = L_mpy_wx(cosx, t0) + L_mpy_wx(sinx, t1)@ 65 smulwb r5, r10, r8 @ L_mpy_wx(sinx, t0) 83 ldr r8, [r12], #4 @ cosxsinx = csptr[1]@ 85 smulwt r6, r10, r8 @ L_mpy_wx(cosx, t0) 86 smulwt r5, r11, r8 @ L_mpy_wx(cosx, t1) 88 smlawb r4, r11, r8, r6 @ r4 = L_mpy_wx(cosx, t0) + L_mpy_wx(sinx, t1)@ 89 smulwb r7, r10, r8 @ L_mpy_wx(sinx, t0 [all...] |
R4R8First_v5.s | 40 add r8, r0, r2 52 add r6, r8, r2 55 sub r8, r8, r2 66 strd r8, [r11, #16] 95 add r8, r0, r2 @ r0 = buf[0] + buf[2]@ 107 add r6, r8, r2 @ r4 = (r0 + r2) >> 1@ 110 sub r8, r8, r2 @ r5 = (r0 - r2) >> 1@ 122 mov r8, r8, asr # [all...] |
/bootable/bootloader/legacy/nandwrite/ |
init.S | 51 mov r8, r1 72 mov r1, r8
|
/external/valgrind/main/none/tests/amd64/ |
sbbmisc.stdout.exp | 8 r8 = 104 105
|
/external/valgrind/main/none/tests/x86/ |
sbbmisc.stdout.exp | 8 r8 = 104 105
|
/system/core/libpixelflinger/ |
col32cb16blend_neon.S | 130 ldrh r8, [r0] // load dest pixel 132 mov r6, r8, lsr #11 // extract dest red 133 ubfx r7, r8, #5, #6 // extract dest green 134 and r8, r8, #0x1f // extract dest blue 138 smlabb r8, r8, r5, r4 // dest blue * alpha + src blue 144 orr r6, r8, lsr #8 // shift blue into 565
|