HomeSort by relevance Sort by last modified time
    Searched refs:r8 (Results 26 - 50 of 298) sorted by null

12 3 4 5 6 7 8 91011>>

  /external/tremolo/Tremolo/
mdctLARM.s 187 LDMFD r12,{r8,r9,r10} @ r8 = step
191 MOV r8, r8, LSL #1
209 STRH r6, [r0], r8
227 LDMFD r12,{r8,r9,r10} @ r8 = step
231 MOV r8, r8, LSL #1
250 STRH r6, [r0], r8
    [all...]
  /dalvik/vm/compiler/template/armv5te/
TEMPLATE_INVOKE_METHOD_PREDICTED_CHAIN.S 27 ldr r8, [r2, #4] @ r8 <- predictedChainCell->clazz
30 cmp r3, r8 @ predicted class == actual class?
48 cmp r8, #0 @ initialized class or not
header.S 27 r4-r8 are available for general use
66 r8 rIBASE interpreted instruction base pointer, used for computed goto
78 #define rIBASE r8
TEMPLATE_RETURN.S 18 ldrb r8, [rSELF, #offThread_breakFlags] @ r8<- breakFlags
42 cmp r8, #0 @ check the break flags
  /frameworks/base/media/libstagefright/codecs/aacenc/src/asm/ARMV7/
Radix4FFT_v7.s 36 mov r8, r0
54 VLD2.I32 {D0, D1, D2, D3}, [r8]
57 add r8, r8, r5 @ xptr += step@
58 VLD2.I32 {D4, D5, D6,D7}, [r8] @ r2 = xptr[0]@ r3 = xptr[1]@
68 add r8, r8, r5 @ xptr += step@
77 VLD2.I32 {D8, D9, D10, D11}, [r8]
79 add r8, r8, r
    [all...]
  /external/valgrind/main/coregrind/m_dispatch/
dispatch-arm-linux.S 57 push {r0, r1, r4, r5, r6, r7, r8, r9, fp, lr}
65 mov r8, r0
66 ldr r0, [r8, #OFFSET_arm_R15T]
81 /* AT ENTRY: r0 is next guest addr, r8 is possibly
86 cmp r8, r1
90 str r0, [r8, #OFFSET_arm_R15T]
109 // r1: live, next-host r8: live, gsp
126 /* AT ENTRY: r0 is next guest addr, r8 is possibly
131 cmp r8, r1
135 str r0, [r8, #OFFSET_arm_R15T
    [all...]
  /external/libvpx/vp8/encoder/arm/armv6/
vp8_subtract_armv6.asm 38 ldr r8, [r1, #vp8_blockd_predictor]
45 ldr r1, [r8], r2 ; pred
92 uxtb16 r8, r6 ; [s2 | s0] (A)
97 usub16 r6, r8, r9 ; [d2 | d0] (A)
103 pkhbt r8, r6, r7, lsl #16 ; [d1 | d0] (A)
106 str r8, [r0], #4 ; diff (A)
107 uxtb16 r8, r10 ; [s2 | s0] (B)
114 usub16 r6, r8, r9 ; [d2 | d0] (B)
119 pkhbt r8, r6, r7, lsl #16 ; [d1 | d0] (B)
122 str r8, [r0], #4 ; diff (B
    [all...]
vp8_mse16x16_armv6.asm 41 usub8 r8, r5, r6 ; calculate difference
42 sel r7, r8, lr ; select bytes with positive difference
44 sel r8, r9, lr ; select bytes with negative difference
48 usad8 r6, r8, lr ; calculate sum of negative differences
49 orr r8, r8, r7 ; differences of all 4 pixels
54 uxtb16 r6, r8 ; byte (two pixels) to halfwords
55 uxtb16 r7, r8, ror #8 ; another two pixels to halfwords
62 usub8 r8, r5, r6 ; calculate difference
63 sel r7, r8, lr ; select bytes with positive differenc
    [all...]
vp8_variance8x8_armv6.asm 37 usub8 r8, r6, r7 ; calculate difference
38 sel r10, r8, lr ; select bytes with positive difference
40 sel r8, r9, lr ; select bytes with negative difference
44 usad8 r7, r8, lr ; calculate sum of negative differences
45 orr r8, r8, r10 ; differences of all 4 pixels
51 uxtb16 r7, r8 ; byte (two pixels) to halfwords
52 uxtb16 r10, r8, ror #8 ; another two pixels to halfwords
60 usub8 r8, r6, r7 ; calculate difference
62 sel r10, r8, lr ; select bytes with positive differenc
    [all...]
vp8_sad16x16_armv6.asm 33 ldr r8, [r2, #0x0] ; load 4 ref pixels (1A)
39 usada8 r4, r8, r6, r4 ; calculate sad for 4 pixels
40 usad8 r8, r7, r9 ; calculate sad for 4 pixels
49 usada8 r8, r11, lr, r8 ; calculate sad for 4 pixels
53 add r4, r4, r8 ; add partial sad values
56 ldr r8, [r2, #0x0] ; load 4 ref pixels (2A)
61 usada8 r4, r6, r8, r4 ; calculate sad for 4 pixels
62 usad8 r8, r7, r9 ; calculate sad for 4 pixels
71 usada8 r8, r11, lr, r8 ; calculate sad for 4 pixel
    [all...]
  /external/libvpx/vp8/common/arm/armv6/
idct_v6.asm 12 ; r0 r1 r2 r3 r4 r5 r6 r7 r8 r9 r10 r11 r12 r14
69 ldrsh r8, [r0, #16] ; input[8] 1, r8 unavail 2 [8]
73 add r9, r7, r8 ; a1 = [0] + [8] 1 a1
74 sub r7, r7, r8 ; b1 = [0] - [8] 1 b1
79 add r8, r7, r11 ; b1 + c1 1 b+c
80 strh r8, [r1, r2] ; out[pitch] = b1+c1 1
86 add r8, r2, r2 ; pitch * 2 1 p*2
87 strh r7, [r1, r8] ; out[pitch*2] = b1-c1 1
100 ldrsh r8, [r1, #6] ; input[3] 1, r8 un 2 [3
    [all...]
simpleloopfilter_v6.asm 83 uqsub8 r8, r6, r3 ; q1 - p1
86 orr r8, r8, r7 ; abs(p1 - q1)
89 uhadd8 r8, r8, lr ; abs(p1 - q2) >> 1
90 uqadd8 r10, r10, r8 ; abs(p0 - q0)*2 + abs(p1 - q1)/2
91 mvn r8, #0
93 sel r10, r8, lr ; filter mask: F or 0
109 ldr r8, c0x03030303
115 qadd8 r8 , r3 , r8 ; Filter2 = vp8_filter +
    [all...]
  /external/webrtc/src/common_audio/signal_processing_library/main/source/
webrtc_fft_4ofq14_gcc_android.s 20 ldrsh r8, [r12], +r2
29 sub r12, r8, r10
31 add r10, r8, r10
34 sub r8, r5, r11
60 ldrsh r8, [r4], +r2
65 add r6, r6, r8, asr #3
67 sub r8, r6, r8, asr #2
77 sub r12, r8, r5, asr #3
79 add r8, r8, r5, asr #3
    [all...]
webrtc_fft_4oiq14_gcc_android.s 20 ldrsh r8, [r12], +r2
29 sub r12, r8, r10
31 add r10, r8, r10
34 sub r8, r5, r11
60 ldrsh r8, [r4], +r2
63 add r6, r6, r8
65 sub r8, r6, r8, lsl #1
75 add r12, r8, r5
77 sub r8, r8, r5
    [all...]
  /frameworks/base/media/libstagefright/codecs/amrwbenc/src/asm/ARMV5E/
Filt_6k_7k_opt.s 37 MOV r8, r0 @ copy signal[] address
49 MOV r3, r8 @ change myMemCopy to Copy, due to Copy will change r3 content
53 LDRSH r8, [r7], #2
55 MOV r8, r8, ASR #2
61 STRH r8, [r6], #2
65 LDRSH r8, [r7], #2
67 MOV r8, r8, ASR #2
73 STRH r8, [r6], #
    [all...]
residu_asm_opt.s 44 LDRH r8, [r0], #2
45 ORR r7, r8, r7, LSL #16 @r7 --- a4, a5
47 LDRH r8, [r0], #2
49 ORR r8, r9, r8, LSL #16 @r8 --- a6, a7
68 STMFD r13!, {r8 - r12} @store r8-r12
109 SMLATT r12,r8, r2, r12 @ i4 (6)
112 SMLATB r11,r8, r2, r11 @ i3 (6
    [all...]
  /frameworks/base/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src_gcc/
armVCM4P10_InterpolateLuma_Align_unsafe_s.S 16 MOV r12,r8
29 STM r8!,{r7,r10,r11}
41 STM r8!,{r7,r10,r11}
53 STM r8!,{r7,r10,r11}
65 STM r8!,{r7,r10,r11}
87 STR r7,[r8],#4
96 STR r7,[r8],#4
105 STR r7,[r8],#4
114 STR r7,[r8],#4
117 SUB r0,r8,#0x1
    [all...]
  /bionic/libdl/arch-sh/
sobegin.S 39 mov.l r8, @-r15
40 mov.l 0f, r8 /* first entry is invalid */
42 add #4, r8
43 mov.l @r8, r0
51 mov.l @r15+, r8
  /external/libvpx/vpx_scale/arm/neon/
vp8_vpxyv12_copyframe_func_neon.asm 30 ldr r8, [r0, #yv12_buffer_config_u_buffer] ;srcptr1
42 str r8, [sp]
51 mov r8, r2
58 vld1.8 {q0, q1}, [r8]!
60 vld1.8 {q2, q3}, [r8]!
62 vld1.8 {q4, q5}, [r8]!
64 vld1.8 {q6, q7}, [r8]!
109 mov r8, r2
116 vld1.8 {q0, q1}, [r8]!
118 vld1.8 {q2, q3}, [r8]!
    [all...]
  /frameworks/base/media/libstagefright/codecs/aacenc/src/asm/ARMV5E/
Radix4FFT_v5.s 59 ldr r8, [r12], #4 @ cosxsinx = csptr[0]@
61 smulwt r4, r10, r8 @ L_mpy_wx(cosx, t0)
62 smulwt r3, r11, r8 @ L_mpy_wx(cosx, t1)
64 smlawb r2, r11, r8, r4 @ r2 = L_mpy_wx(cosx, t0) + L_mpy_wx(sinx, t1)@
65 smulwb r5, r10, r8 @ L_mpy_wx(sinx, t0)
83 ldr r8, [r12], #4 @ cosxsinx = csptr[1]@
85 smulwt r6, r10, r8 @ L_mpy_wx(cosx, t0)
86 smulwt r5, r11, r8 @ L_mpy_wx(cosx, t1)
88 smlawb r4, r11, r8, r6 @ r4 = L_mpy_wx(cosx, t0) + L_mpy_wx(sinx, t1)@
89 smulwb r7, r10, r8 @ L_mpy_wx(sinx, t0
    [all...]
R4R8First_v5.s 40 add r8, r0, r2
52 add r6, r8, r2
55 sub r8, r8, r2
66 strd r8, [r11, #16]
95 add r8, r0, r2 @ r0 = buf[0] + buf[2]@
107 add r6, r8, r2 @ r4 = (r0 + r2) >> 1@
110 sub r8, r8, r2 @ r5 = (r0 - r2) >> 1@
122 mov r8, r8, asr #
    [all...]
  /bootable/bootloader/legacy/nandwrite/
init.S 51 mov r8, r1
72 mov r1, r8
  /external/valgrind/main/none/tests/amd64/
sbbmisc.stdout.exp 8 r8 = 104 105
  /external/valgrind/main/none/tests/x86/
sbbmisc.stdout.exp 8 r8 = 104 105
  /system/core/libpixelflinger/
col32cb16blend_neon.S 130 ldrh r8, [r0] // load dest pixel
132 mov r6, r8, lsr #11 // extract dest red
133 ubfx r7, r8, #5, #6 // extract dest green
134 and r8, r8, #0x1f // extract dest blue
138 smlabb r8, r8, r5, r4 // dest blue * alpha + src blue
144 orr r6, r8, lsr #8 // shift blue into 565

Completed in 1161 milliseconds

12 3 4 5 6 7 8 91011>>