HomeSort by relevance Sort by last modified time
    Searched refs:r8 (Results 1 - 25 of 210) sorted by null

1 2 3 4 5 6 7 8 9

  /frameworks/base/media/libstagefright/codecs/amrwbenc/src/asm/ARMV5E/
convolve_opt.s 45 MUL r8, r9, r10
54 MLA r8, r9, r10, r8
55 MLA r8, r12, r14, r8
60 MLA r8, r9, r10, r8
62 MLA r8, r12, r14, r8
68 ADD r5, r11, r8, LSL #
    [all...]
Dot_p_opt.s 41 LDR r8, [r0], #4
47 SMLABB r4, r8, r9, r4
50 SMLATT r4, r8, r9, r4
51 LDR r8, [r0], #4
57 SMLABB r4, r8, r9, r4
59 SMLATT r4, r8, r9, r4
  /external/libvpx/vp8/common/arm/armv6/
recon_v6.asm 48 pkhbt r8, r6, r7, lsl #16 ; 2 | 0
51 uxtab16 r8, r8, r4 ; 2 | 0 + 3 | 2 | 2 | 0
54 usat16 r8, #8, r8
57 orr r8, r8, r9, lsl #8
59 str r8, [dst], stride
68 pkhbt r8, r6, r7, lsl #16 ; 2 | 0
71 uxtab16 r8, r8, r4 ; 2 | 0 + 3 | 2 | 2 |
    [all...]
loopfilter_v6.asm 91 uqsub8 r8, r10, r11 ; p2 - p1
95 orr r8, r8, r10 ; abs (p2-p1)
97 uqsub8 r8, r8, r2 ; compare to limit
99 orr lr, lr, r8
105 uqsub8 r8, r6, r3 ; compare to thresh -- save r8 for later
158 orr r10, r6, r8 ; calculate vp8_hevmask
166 ldr r8, [src], pstep ; p
    [all...]
idct_v6.asm 12 ; r0 r1 r2 r3 r4 r5 r6 r7 r8 r9 r10 r11 r12 r14
69 ldrsh r8, [r0, #16] ; input[8] 1, r8 unavail 2 [8]
73 add r9, r7, r8 ; a1 = [0] + [8] 1 a1
74 sub r7, r7, r8 ; b1 = [0] - [8] 1 b1
79 add r8, r7, r11 ; b1 + c1 1 b+c
80 strh r8, [r1, r2] ; out[pitch] = b1+c1 1
86 add r8, r2, r2 ; pitch * 2 1 p*2
87 strh r7, [r1, r8] ; out[pitch*2] = b1-c1 1
100 ldrsh r8, [r1, #6] ; input[3] 1, r8 un 2 [3
    [all...]
  /external/openssl/crypto/bn/asm/
ppc.pl 231 #.set r8,8
275 # Freely use registers r5,r6,r7,r8,r9,r10,r11 as follows:
278 # r7,r8 are the results of the 32x32 giving 64 bit multiply.
298 $UMULH r8,r5,r6
300 addc r7,r7,r7 # compute (r7,r8)=2*(r7,r8)
301 adde r8,r8,r8
306 addze r11,r8 # r8 added to r11 which is 0
    [all...]
  /frameworks/base/media/libstagefright/codecs/aacenc/src/asm/ARMV5E/
band_nrg_v5.s 54 smull r6, r8, r6, r6
60 qadd r14, r14, r8
61 smull r6, r8, r6, r6
65 qadd r14, r14, r8
113 ldr r8, [r0, +r10, lsl #2]
115 mov r8, r8, asr #1
120 add r5, r8, r9
122 sub r8, r8, r9
    [all...]
CalcWindowEnergy_v5.s 42 mov r8, #0 @ w=0
50 str r8, [r13, #4]
61 smull r0, r8, r12, r11 @ accu2 = fixmul( Coeff0, states1 );
64 mov r8, r8, lsl #1
67 sub r8, r0, r8 @ out = accu3 - accu2;
70 mov r11, r8 @ states1 = out;
73 mov r8, r8, asr #16
    [all...]
AutoCorrelation_v5.s 39 mov r8, #0
72 add r8, r8, #6
75 cmp r8, r12
84 add r8, r8, #1
85 cmp r8, r4
94 mov r8, #1
107 addlt r6, r5, r8, lsl #1
110 add r6, r5, r8, lsl #1
    [all...]
PrePostMDCT_v5.s 37 ldr r8, [r2], #4
43 smull r14, r11, r4, r8 @ MULHIGH(tr1, cosa)
44 smull r10, r12, r7, r8 @ MULHIGH(ti1, cosa)
46 smull r14, r8, r7, r9 @ MULHIGH(ti1, sina)
49 add r11, r11, r8 @ MULHIGH(cosa, tr1) + MULHIGH(sina, ti1)@
52 ldr r8, [r2], #4
55 smull r14, r4, r6, r8 @ MULHIGH(tr2, cosa)
56 smull r10, r12, r5, r8 @ MULHIGH(ti2, cosa)
58 smull r14, r8, r5, r9 @ MULHIGH(ti2, sina)
61 add r8, r8, r4
    [all...]
  /frameworks/base/media/libstagefright/codecs/amrwbenc/src/asm/ARMV7/
residu_asm_neon.s 43 MOV r8, r9
44 VLD1.S16 D5, [r8]! @get x[i], x[i+1], x[i+2], x[i+3]
47 SUB r8, r9, #2 @get the x[i-1] address
48 VLD1.S16 D5, [r8]!
51 SUB r8, r9, #4 @load the x[i-2] address
52 VLD1.S16 D5, [r8]!
55 SUB r8, r9, #6 @load the x[i-3] address
56 VLD1.S16 D5, [r8]!
59 SUB r8, r9, #8 @load the x[i-4] address
60 VLD1.S16 D5, [r8]!
    [all...]
convolve_neon.s 39 @MOV r8, #0 @ s = 0
47 MUL r8, r9, r10
64 ADD r5, r5, r8
71 @MOV r8, #0
80 MUL r8, r9, r10
82 MLA r8, r12, r14, r8
100 ADD r8, r8, r5
101 ADD r8, r11, r8, LSL #
    [all...]
  /system/core/libpixelflinger/
col32cb16blend.S 56 ldrh r8, [r0] // load dest pixel
58 mov r6, r8, lsr #11 // extract dest red
59 and r7, r9, r8, lsr #5 // extract dest green
60 and r8, r8, #0x1f // extract dest blue
64 smlabb r8, r8, r5, r4 // dest blue * alpha + src blue
70 orr r6, r8, lsr #8 // shift blue into 565
rotate90CW_4x4_16v6.S 36 stmfd sp!, {r4,r5, r6,r7, r8,r9, r10,r11, lr}
43 ldrd r8, r9, [r1]
45 pkhbt r10, r8, r6, lsl #16
49 pkhtb r10, r6, r8, asr #16
62 ldmfd sp!, {r4,r5, r6,r7, r8,r9, r10,r11, pc}
  /external/libvpx/vp8/decoder/arm/armv6/
dequant_idct_v6.asm 71 smulwb r8, r4, r6
74 pkhbt r8, r8, r10, lsl #16
84 usub16 r7, r8, r7
87 usub16 r8, r11, r14
90 uadd16 r6, r8, r7
91 usub16 r7, r8, r7
104 ldr r8, [r0], #4
108 smulwt lr, r3, r8
109 smulwt r10, r4, r8
    [all...]
dequant_dc_idct_v6.asm 93 smulwb r8, r4, r6
96 pkhbt r8, r8, r10, lsl #16
106 usub16 r7, r8, r7
109 usub16 r8, r11, r14
112 uadd16 r6, r8, r7
113 usub16 r7, r8, r7
126 ldr r8, [r0], #4
130 smulwt lr, r3, r8
131 smulwt r10, r4, r8
    [all...]
  /frameworks/base/media/libstagefright/codecs/m4v_h263/dec/src/
idct.cpp 22 This file contains the functions that transform an 8r8 image block from
44 coefficients of type int for an 8r8 image block;
57 blk points to the found IDCT values for an 8r8 image block.
68 This function transforms an 8r8 image block from dequantized DCT coefficients
82 1. Find horizontal 1-D IDCT values for each row from 8r8 dequantized DCT
131 int32 r0, r1, r2, r3, r4, r5, r6, r7, r8; /* butterfly nodes */ local
182 r8 = W7 * (r4 + r5);
183 r4 = (r8 + (W1 - W7) * r4);
186 r5 = (r8 - (W1 + W7) * r5);
187 r8 = W3 * (r6 + r7)
353 int32 r0, r1, r2, r3, r4, r5, r6, r7, r8; \/* butterfly nodes *\/ local
    [all...]
  /bionic/libc/arch-sh/bionic/
_exit_with_stack_teardown.S 38 mov r6, r8 /* save retCode : breaks r8 value */
42 mov r8, r4 /* restore retCode */
  /dalvik/vm/compiler/template/armv5te/
TEMPLATE_STRING_INDEXOF.S 19 ldr r8, [r0, #STRING_FIELDOFF_COUNT]
28 * r8: string length
41 cmp r2, r8
42 movgt r2, r8
49 sub r8, r2
55 * r8: iteration count
60 subs r8, #4
76 subs r8, #4
80 adds r8, #4
87 subs r8, #
    [all...]
TEMPLATE_INVOKE_METHOD_PREDICTED_CHAIN.S 26 ldr r8, [r2, #4] @ r8 <- predictedChainCell->clazz
29 cmp r3, r8 @ predicted class == actual class?
38 cmp r8, #0 @ initialized class or not
  /frameworks/base/opengl/libagl/
iterators.S 51 stmfd sp!, {r4, r5, r6, r7, r8, lr}
63 smull r8, lr, r4, r5
67 smlal r8, lr, r3, r5
74 str r8, [r1, #8]
77 umull r8, r0, r4, r8
80 adds r6, r6, r8
88 ldmfd sp!, {r4, r5, r6, r7, r8, pc}
  /external/jpeg/
jidctfst.S 67 stmdb sp!, {r4,r5, r6,r7, r8,r9, r10,r11, r12,lr}
91 orreqs r8, r2, r3
92 orreqs r8, r4, r5
93 orreqs r8, r6, r7
105 ldr r8, [r10, #QY(0)]
108 smlabb r0, r8, r0, r4
111 mul r0, r8, r0
115 ldr r8, [r10, #QY(2)]
119 smlabb r2, r8, r2, r6
122 mul r2, r8, r
    [all...]
  /external/libffi/src/ia64/
unix.S 60 mov r8 = in1
125 zxt1 r8 = r8
127 st8 [in1] = r8
131 sxt1 r8 = r8
133 st8 [in1] = r8
137 zxt2 r8 = r8
139 st8 [in1] = r8
    [all...]
  /external/tremolo/Tremolo/
mdctLARM.s 187 LDMFD r12,{r8,r9,r10} @ r8 = step
191 MOV r8, r8, LSL #1
209 STRH r6, [r0], r8
227 LDMFD r12,{r8,r9,r10} @ r8 = step
231 MOV r8, r8, LSL #1
250 STRH r6, [r0], r8
    [all...]
  /frameworks/base/media/libstagefright/codecs/aacenc/src/asm/ARMV7/
Radix4FFT_v7.s 36 mov r8, r0
54 VLD2.I32 {D0, D1, D2, D3}, [r8]
57 add r8, r8, r5 @ xptr += step@
58 VLD2.I32 {D4, D5, D6,D7}, [r8] @ r2 = xptr[0]@ r3 = xptr[1]@
68 add r8, r8, r5 @ xptr += step@
77 VLD2.I32 {D8, D9, D10, D11}, [r8]
79 add r8, r8, r5
    [all...]

Completed in 188 milliseconds

1 2 3 4 5 6 7 8 9