HomeSort by relevance Sort by last modified time
    Searched refs:r10 (Results 1 - 25 of 423) sorted by null

1 2 3 4 5 6 7 8 91011>>

  /frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src_gcc/
armVCM4P10_InterpolateLuma_Align_unsafe_s.S 26 LDM r0,{r7,r10,r11}
29 STM r8!,{r7,r10,r11}
33 LDM r0,{r7,r10,r11}
37 ORR r7,r7,r10,LSL #24
38 LSR r10,r10,#8
39 ORR r10,r10,r11,LSL #24
41 STM r8!,{r7,r10,r11}
45 LDM r0,{r7,r10,r11
    [all...]
armVCM4P10_InterpolateLuma_DiagCopy_unsafe_s.S 22 LDR r10,[r0,#8]
26 UQSUB16 r10,r10,r6
30 USAT16 r10,#13,r10
34 AND r10,r12,r10,LSR #5
37 ORR r11,r10,r11,LSL #8
38 ORR r10,r4,r5,LSL #8
40 STRD r10,r11,[r7],#
    [all...]
omxVCM4P10_PredictIntra_16x16_s.S 46 ADD r10,r5,r5
47 VST1.8 {d0,d1},[r3],r10
48 VST1.8 {d0,d1},[r8],r10
49 VST1.8 {d0,d1},[r3],r10
50 VST1.8 {d0,d1},[r8],r10
51 VST1.8 {d0,d1},[r3],r10
52 VST1.8 {d0,d1},[r8],r10
53 VST1.8 {d0,d1},[r3],r10
54 VST1.8 {d0,d1},[r8],r10
55 VST1.8 {d0,d1},[r3],r10
    [all...]
omxVCM4P10_PredictIntraChroma_8x8_s.S 26 PUSH {r4-r10,lr}
40 ADD r10,r4,r4
41 VLD1.8 {d1[0]},[r0],r10
42 VLD1.8 {d1[1]},[r9],r10
43 VLD1.8 {d1[2]},[r0],r10
44 VLD1.8 {d1[3]},[r9],r10
45 VLD1.8 {d1[4]},[r0],r10
46 VLD1.8 {d1[5]},[r9],r10
47 VLD1.8 {d1[6]},[r0],r10
70 ADD r10,r5,r
    [all...]
armVCM4P10_Average_4x_Align_unsafe_s.S 19 LDR r10,[r0],r1
24 UHSUB8 r5,r10,r12
30 LDR r10,[r0],r1
35 UHSUB8 r5,r10,r12
51 LDR r10,[r0],r1
58 LSR r10,r10,#16
59 ORR r10,r10,r4,LSL #16
62 UHSUB8 r5,r10,r1
    [all...]
  /dalvik/vm/compiler/template/armv5te/
TEMPLATE_MUL_LONG.S 23 umull r9, r10, r2, r0 @ r9/r10 <- ZxX
26 add r1, r2, r10 @ r1<- r10 + low(ZxW + (YxX))
  /external/llvm/test/MC/X86/
x86_64-bmi-encoding.s 7 // CHECK: blsmskq %r11, %r10
9 blsmskq %r11, %r10
15 // CHECK: blsmskq (%rax), %r10
17 blsmskq (%rax), %r10
23 // CHECK: blsiq %r11, %r10
25 blsiq %r11, %r10
31 // CHECK: blsiq (%rax), %r10
33 blsiq (%rax), %r10
39 // CHECK: blsrq %r11, %r10
41 blsrq %r11, %r10
    [all...]
  /frameworks/av/media/libstagefright/codecs/amrwbenc/src/asm/ARMV5E/
Deemph_32_opt.s 44 MOV r10, r6, LSL #16 @L_tmp = x_hi[0]<<16
47 ADD r12, r10, r7, LSL #4 @L_tmp += x_lo[0] << 4
48 MOV r10, r12, LSL #3 @L_tmp <<= 3
51 QDADD r10, r10, r9
53 MOV r12, r10, LSL #1 @L_tmp = L_mac(L_tmp, *mem, fac)
54 QADD r10, r12, r11
55 MOV r14, r10, ASR #16 @y[0] = round(L_tmp)
58 MOV r10, r6, LSL #16
59 ADD r12, r10, r7, LSL #
    [all...]
pred_lt4_1_opt.s 68 SMULBB r10, r4, r3 @x[0] * h[0]
73 SMLABT r10, r6, r3, r10 @x[1] * h[1]
79 SMLABB r10, r9, r3, r10 @x[2] * h[2]
84 SMLABT r10, r4, r3, r10 @x[3] * h[3]
90 SMLABB r10, r6, r3, r10 @x[4] * h[4]
95 SMLABT r10, r9, r3, r10 @x[5] * h[5
    [all...]
  /frameworks/av/media/libstagefright/codecs/amrwbenc/src/asm/ARMV7/
Deemph_32_neon.s 44 MOV r10, r6, LSL #16 @L_tmp = x_hi[0]<<16
47 ADD r12, r10, r7, LSL #4 @L_tmp += x_lo[0] << 4
48 MOV r10, r12, LSL #3 @L_tmp <<= 3
51 QDADD r10, r10, r9
53 MOV r12, r10, LSL #1 @L_tmp = L_mac(L_tmp, *mem, fac)
54 QADD r10, r12, r11
55 MOV r14, r10, ASR #16 @y[0] = round(L_tmp)
58 MOV r10, r6, LSL #16
59 ADD r12, r10, r7, LSL #
    [all...]
  /system/core/libpixelflinger/
rotate90CW_4x4_16v6.S 36 stmfd sp!, {r4,r5, r6,r7, r8,r9, r10,r11, lr}
45 pkhbt r10, r8, r6, lsl #16
47 strd r10, r11, [r0], r12
49 pkhtb r10, r6, r8, asr #16
52 strd r10, r11, [r0], r12
53 pkhbt r10, r9, r7, lsl #16
56 strd r10, r11, [r0], r12
58 pkhtb r10, r7, r9, asr #16
60 strd r10, r11, [r0]
62 ldmfd sp!, {r4,r5, r6,r7, r8,r9, r10,r11, pc
    [all...]
  /external/chromium_org/third_party/openssl/openssl/crypto/sha/asm/
sha1-armv4-large.S 22 ldrb r10,[r1,#2]
27 orr r9,r9,r10,lsl#8
28 eor r10,r5,r6 @ F_xx_xx
35 eor r10,r5,r6 @ F_xx_xx
41 and r10,r4,r10,ror#2
43 eor r10,r10,r6,ror#2 @ F_00_19(B,C,D)
45 add r7,r7,r10 @ E+=F_00_19(B,C,D)
47 ldrb r10,[r1,#2
    [all...]
  /external/openssl/crypto/sha/asm/
sha1-armv4-large.S 22 ldrb r10,[r1,#2]
27 orr r9,r9,r10,lsl#8
28 eor r10,r5,r6 @ F_xx_xx
35 eor r10,r5,r6 @ F_xx_xx
41 and r10,r4,r10,ror#2
43 eor r10,r10,r6,ror#2 @ F_00_19(B,C,D)
45 add r7,r7,r10 @ E+=F_00_19(B,C,D)
47 ldrb r10,[r1,#2
    [all...]
  /dalvik/vm/mterp/armv5te/
OP_MUL_LONG_2ADDR.S 19 umull r9, r10, r2, r0 @ r9/r10 <- ZxX
23 add r10, r2, r10 @ r10<- r10 + low(ZxW + (YxX))
25 stmia r0, {r9-r10} @ vAA/vAA+1<- r9/r10
OP_INVOKE_SUPER_QUICK.S 11 FETCH(r10, 2) @ r10<- GFED or CCCC
14 and r10, r10, #15 @ r10<- D (or stays CCCC)
20 GET_VREG(r9, r10) @ r9<- "this"
OP_INVOKE_SUPER.S 11 FETCH(r10, 2) @ r10<- GFED or CCCC
14 and r10, r10, #15 @ r10<- D (or stays CCCC)
18 GET_VREG(r9, r10) @ r9<- "this" ptr
21 ldr r10, [rSELF, #offThread_method] @ r10<- current method
24 ldr r10, [r10, #offMethod_clazz] @ r10<- method->claz
    [all...]
  /dalvik/vm/mterp/armv6t2/
OP_MUL_LONG_2ADDR.S 18 umull r9, r10, r2, r0 @ r9/r10 <- ZxX
22 add r10, r2, r10 @ r10<- r10 + low(ZxW + (YxX))
24 stmia r0, {r9-r10} @ vAA/vAA+1<- r9/r10
  /external/valgrind/main/coregrind/m_mach/
mach_traps-amd64-darwin.S 40 movq %rcx, %r10
51 // movq %rcx, %r10
61 movq %rcx, %r10
71 movq %rcx, %r10
81 movq %rcx, %r10
91 movq %rcx, %r10
101 movq %rcx, %r10
111 movq %rcx, %r10
121 movq %rcx, %r10
131 movq %rcx, %r10
    [all...]
  /external/chromium_org/third_party/openssl/openssl/crypto/bn/asm/
ppc.pl 233 #.set r10,10
275 # Freely use registers r5,r6,r7,r8,r9,r10,r11 as follows:
279 # r9,r10, r11 are the equivalents of c1,c2, c3.
289 $UMULH r10,r5,r5 #in first iteration. No need
305 addc r10,r7,r10 # now add to temp result.
309 $ST r10,`1*$BNSZ`(r3) #r[1]=c2;
315 addze r10,r0
323 addze r10,r10
    [all...]
  /external/openssl/crypto/bn/asm/
ppc.pl 233 #.set r10,10
275 # Freely use registers r5,r6,r7,r8,r9,r10,r11 as follows:
279 # r9,r10, r11 are the equivalents of c1,c2, c3.
289 $UMULH r10,r5,r5 #in first iteration. No need
305 addc r10,r7,r10 # now add to temp result.
309 $ST r10,`1*$BNSZ`(r3) #r[1]=c2;
315 addze r10,r0
323 addze r10,r10
    [all...]
  /frameworks/av/media/libstagefright/codecs/aacenc/src/asm/ARMV5E/
band_nrg_v5.s 41 ldrsh r10, [r1, r2]
45 cmp r10, r2
49 ldr r11, [r0, +r10, lsl #2]
50 add r10, r10, #1
51 ldr r6, [r0, +r10, lsl #2]
53 add r10, r10, #1
55 ldr r11, [r0, +r10, lsl #2]
57 add r10, r10, #
    [all...]
  /external/libvpx/libvpx/vp8/common/arm/armv6/
loopfilter_v6.asm 73 ldr r10, [src], pstep ; p2
90 uqsub8 r6, r9, r10 ; p3 - p2
91 uqsub8 r7, r10, r9 ; p2 - p3
92 uqsub8 r8, r10, r11 ; p2 - p1
93 uqsub8 r10, r11, r10 ; p1 - p2
96 orr r8, r8, r10 ; abs (p2-p1)
103 ldr r10, [src], pstep ; q1
109 uqsub8 r6, r11, r10 ; p1 - q1
110 uqsub8 r7, r10, r11 ; q1 - p
    [all...]
  /external/tremolo/Tremolo/
bitwiseARM.s 45 STMFD r13!,{r10,r11,r14}
53 LDR r10,[r3] @ r10= ptr[0]
57 MOV r10,r10,LSR r14 @ r10= ptr[0]>>(32-bitsLeftInWord)
60 ORRLT r10,r10,r11,LSL r14 @ r10= Next 32 bits.
63 AND r0,r10,r1
    [all...]
  /hardware/samsung_slsi/exynos5/libswconverter/
csc_interleave_memcpy_neon.s 60 @r10 dest_addr
67 mov r10, r0
93 vst2.8 {q0, q1}, [r10]!
94 vst2.8 {q2, q3}, [r10]!
95 vst2.8 {q4, q5}, [r10]!
96 vst2.8 {q6, q7}, [r10]!
97 vst2.8 {q8, q9}, [r10]!
98 vst2.8 {q10, q11}, [r10]!
99 vst2.8 {q12, q13}, [r10]!
100 vst2.8 {q14, q15}, [r10]!
    [all...]
csc_ARGB8888_to_YUV420SP_NEON.s 19 @r10 temp3, r,g,b
197 mov r10, r9,lsr #16 @copy to r10
198 mov r14, r12 @copy to r10
201 and r10, r10, r6 @R: (rgbIn[k] & 0xF800) >> 10;
204 add r10,r10,r14
207 mul r7, r10, r11
209 mov r10, r9,lsr #8 @copy to r1
    [all...]

Completed in 3352 milliseconds

1 2 3 4 5 6 7 8 91011>>