HomeSort by relevance Sort by last modified time
    Searched refs:r4 (Results 251 - 275 of 775) sorted by null

<<11121314151617181920>>

  /external/libvpx/libvpx/vp8/encoder/arm/neon/
subtract_neon.asm 26 stmfd sp!, {r4-r7}
29 ldr r4, [r0, #vp8_block_src]
33 add r3, r3, r4 ; src = *base_src + src
57 ldmfd sp!, {r4-r7}
67 push {r4-r7}
69 ldr r4, [sp, #16] ; pred_stride
75 vld1.8 {q1}, [r3], r4 ;load pred
77 vld1.8 {q3}, [r3], r4
79 vld1.8 {q5}, [r3], r4
81 vld1.8 {q7}, [r3], r4
    [all...]
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/arm/armv6/
vp8_variance16x16_armv6.asm 27 stmfd sp!, {r4-r12, lr}
38 ldr r4, [r0, #0] ; load 4 src pixels
43 usub8 r6, r4, r5 ; calculate difference
46 usub8 r9, r5, r4 ; calculate difference with reversed operands
51 usad8 r4, r7, lr ; calculate sum of positive differences
55 adds r8, r8, r4 ; add positive differences to sum
64 ldr r4, [r0, #4] ; load 4 src pixels
68 usub8 r6, r4, r5 ; calculate difference
70 usub8 r9, r5, r4 ; calculate difference with reversed operands
74 usad8 r4, r7, lr ; calculate sum of positive difference
    [all...]
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/ppc/
variance_altivec.asm 79 add r3, r3, r4
93 lwz r4, 12(r1)
98 stw r4, 0(r7) ;# sse
102 subf r3, r3, r4 ;# sse - ((sum*sum) >> DS)
112 add r3, r3, r4
120 add r3, r3, r4
137 lwz r4, 12(r1)
142 stw r4, 0(r7) ;# sse
146 subf r3, r3, r4 ;# sse - ((sum*sum) >> 8)
151 ;# r4 int source_strid
    [all...]
filter_altivec.asm 192 ;# r4 int src_pitch
231 Read8x8 v2, r3, r4, 1
232 Read8x8 v3, r3, r4, 1
233 Read8x8 v4, r3, r4, 1
234 Read8x8 v5, r3, r4, 1
252 sub r9, r9, r4
253 sub r9, r9, r4
255 Read8x8 v0, r9, r4, 1
256 Read8x8 v1, r9, r4, 0
257 Read8x8 v6, r3, r4,
    [all...]
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/encoder/arm/armv6/
vp8_subtract_armv6.asm 29 stmfd sp!, {r4-r9}
31 ldr r4, [r0, #vp8_block_base_src]
35 ldr r3, [r4]
47 uxtb16 r4, r0 ; [s2 | s0]
52 usub16 r4, r4, r5 ; [d2 | d0]
57 pkhbt r0, r4, r5, lsl #16 ; [d1 | d0]
58 pkhtb r1, r5, r4, asr #16 ; [d3 | d2]
66 ldmfd sp!, {r4-r9}
81 stmfd sp!, {r4-r11
    [all...]
vp8_mse16x16_armv6.asm 29 push {r4-r9, lr}
35 mov r4, #0 ; initialize sse = 0
61 smlad r4, r6, r6, r4 ; dual signed multiply, add and accumulate (1)
65 smlad r4, r7, r7, r4 ; dual signed multiply, add and accumulate (2)
80 smlad r4, r6, r6, r4 ; dual signed multiply, add and accumulate (1)
84 smlad r4, r7, r7, r4 ; dual signed multiply, add and accumulate (2
    [all...]
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/encoder/arm/neon/
subtract_neon.asm 26 stmfd sp!, {r4-r7}
29 ldr r4, [r0, #vp8_block_src]
33 add r3, r3, r4 ; src = *base_src + src
57 ldmfd sp!, {r4-r7}
67 push {r4-r7}
69 ldr r4, [sp, #16] ; pred_stride
75 vld1.8 {q1}, [r3], r4 ;load pred
77 vld1.8 {q3}, [r3], r4
79 vld1.8 {q5}, [r3], r4
81 vld1.8 {q7}, [r3], r4
    [all...]
  /external/libhevc/common/arm/
ihevc_intra_pred_chroma_mode_18_34.s 107 stmfd sp!, {r4-r12, r14} @stack stores the values of the arguments
110 ldr r4,[sp,#40]
113 cmp r4,#4
116 mov r12,r4
117 mov r11,r4
118 add r0,r0,r4,lsl #2
184 ldmfd sp!,{r4-r12,r15} @reload the registers from sp
ihevc_inter_pred_filters_luma_horz.s 102 @ r4 => *pi1_coeff
118 stmfd sp!, {r4-r12, r14} @stack stores the values of the arguments
125 ldr r4,[sp,#40] @loads pi1_coeff
129 vld1.8 {d0},[r4] @coeff = vld1_s8(pi1_coeff)
141 add r4,r12,r2 @pu1_src_tmp2_8 = pu1_src + src_strd
185 add r4,r12,r2 @pu1_src + src_strd
223 vld1.u32 {d12},[r4],r11 @vector load pu1_src + src_strd
225 vld1.u32 {d13},[r4],r11
227 vld1.u32 {d14},[r4],r11
229 vld1.u32 {d15},[r4],r1
    [all...]
ihevc_itrans_recon_4x4.s 98 @ r4 => src_strd
124 stmfd sp!, {r4-r12, r14} @stack stores the values of the arguments
130 ldr r4,[sp,#40] @loading src_strd
132 add r4,r4,r4 @ src_strd in terms of word16
136 add r9,r0,r4 @ pi2_src[0] + src_strd
143 add r10,r9,r4, lsl #1 @ 3*src_strd
144 add r4,r4,r
    [all...]
  /external/linux-tools-perf/perf-3.12.0/arch/arm/lib/
memset.S 64 stmfd sp!, {r4-r8, lr}
65 mov r4, r1
80 stmcsia ip!, {r4, r5, r6, r7}
81 stmmiia ip!, {r4, r5}
90 ldmeqfd sp!, {r4-r8, pc}
95 stmneia ip!, {r4-r7}
96 ldmfd sp!, {r4-r8, lr}
  /external/tremolo/Tremolo/
floor1LARM.s 48 STMFD r13!,{r4-r6,r11,r14}
53 LDR r4, [r1] @ r4 = *d
56 MOV r4, r4, ASR #6
57 MUL r5, r4, r5 @ r5 = MULT31_SHIFT15
64 LDMFD r13!,{r4-r6,r11,PC}
  /external/webrtc/src/modules/audio_coding/codecs/isac/fix/source/
lattice_armv7.S 22 @ r4: out loop counter
37 .save {r4-r11}
38 push {r4-r11}
42 mov r4, #HALF_SUBFRAMELEN
43 sub r4, #1 @ Outer loop counter = HALF_SUBFRAMELEN - 1
75 subs r4, #1
78 pop {r4-r11}
  /frameworks/av/media/libstagefright/codecs/mp3dec/src/asm/
pvmp3_dct_9_gcc.s 43 stmfd sp!,{r4-r11,lr}
50 ldr r4,[r0, #0x18]
54 add r12,r4,r5
55 sub r4,r4,r5
124 mov r2,r4,lsl #1
133 add r4,r5,r4
135 sub lr,r4,lr
138 ldr r4,cos_pi_
    [all...]
  /external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/ppc/
filter_altivec.asm 192 ;# r4 int src_pitch
231 Read8x8 v2, r3, r4, 1
232 Read8x8 v3, r3, r4, 1
233 Read8x8 v4, r3, r4, 1
234 Read8x8 v5, r3, r4, 1
252 sub r9, r9, r4
253 sub r9, r9, r4
255 Read8x8 v0, r9, r4, 1
256 Read8x8 v1, r9, r4, 0
257 Read8x8 v6, r3, r4,
    [all...]
  /external/chromium_org/third_party/libvpx/source/libvpx/vp8/encoder/arm/armv6/
vp8_mse16x16_armv6.asm 29 push {r4-r9, lr}
35 mov r4, #0 ; initialize sse = 0
61 smlad r4, r6, r6, r4 ; dual signed multiply, add and accumulate (1)
65 smlad r4, r7, r7, r4 ; dual signed multiply, add and accumulate (2)
80 smlad r4, r6, r6, r4 ; dual signed multiply, add and accumulate (1)
84 smlad r4, r7, r7, r4 ; dual signed multiply, add and accumulate (2
    [all...]
  /external/chromium_org/third_party/opus/src/celt/arm/
celt_pitch_xcorr_arm.s 47 ; r4 = opus_val16 *x
79 VLD1.16 {d6, d7}, [r4]!
106 VLD1.16 d6, [r4]!
124 VLD2.16 {d6[],d7[]}, [r4]!
139 VLD1.16 {d6[]}, [r4]!
149 VLD1.16 {d6[]}, [r4]!
165 ; r4 = opus_val16 *x (for xcorr_kernel_neon())
170 STMFD sp!, {r4-r6, lr}
178 ; r3 = len, r4 = _x, r5 = _y, q0 = {0, 0, 0, 0}
179 MOV r4, r
    [all...]
  /external/libopus/celt/arm/
celt_pitch_xcorr_arm.s 47 ; r4 = opus_val16 *x
79 VLD1.16 {d6, d7}, [r4]!
106 VLD1.16 d6, [r4]!
124 VLD2.16 {d6[],d7[]}, [r4]!
139 VLD1.16 {d6[]}, [r4]!
149 VLD1.16 {d6[]}, [r4]!
165 ; r4 = opus_val16 *x (for xcorr_kernel_neon())
170 STMFD sp!, {r4-r6, lr}
178 ; r3 = len, r4 = _x, r5 = _y, q0 = {0, 0, 0, 0}
179 MOV r4, r
    [all...]
  /external/libvpx/libvpx/vp8/common/ppc/
filter_altivec.asm 192 ;# r4 int src_pitch
231 Read8x8 v2, r3, r4, 1
232 Read8x8 v3, r3, r4, 1
233 Read8x8 v4, r3, r4, 1
234 Read8x8 v5, r3, r4, 1
252 sub r9, r9, r4
253 sub r9, r9, r4
255 Read8x8 v0, r9, r4, 1
256 Read8x8 v1, r9, r4, 0
257 Read8x8 v6, r3, r4,
    [all...]
  /external/libvpx/libvpx/vp8/encoder/arm/armv6/
vp8_mse16x16_armv6.asm 29 push {r4-r9, lr}
35 mov r4, #0 ; initialize sse = 0
61 smlad r4, r6, r6, r4 ; dual signed multiply, add and accumulate (1)
65 smlad r4, r7, r7, r4 ; dual signed multiply, add and accumulate (2)
80 smlad r4, r6, r6, r4 ; dual signed multiply, add and accumulate (1)
84 smlad r4, r7, r7, r4 ; dual signed multiply, add and accumulate (2
    [all...]
  /external/chromium_org/v8/test/cctest/
test-disasm-arm.cc 107 COMPARE(and_(r2, r3, Operand(r4), SetCC),
108 "e0132004 ands r2, r3, r4");
109 COMPARE(and_(r3, r4, Operand(r5), LeaveCC, eq),
110 "00043005 andeq r3, r4, r5");
112 COMPARE(eor(r4, r5, Operand(r6, LSL, 0)),
113 "e0254006 eor r4, r5, r6");
114 COMPARE(eor(r4, r5, Operand(r7, LSL, 1), SetCC),
115 "e0354087 eors r4, r5, r7, lsl #1");
116 COMPARE(eor(r4, r5, Operand(r8, LSL, 2), LeaveCC, ne),
117 "10254108 eorne r4, r5, r8, lsl #2")
    [all...]
  /external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/arm/armv6/
vp8_variance8x8_armv6.asm 25 push {r4-r10, lr}
31 mov r4, #0 ; initialize sum = 0
53 add r4, r4, r6 ; add positive differences to sum
54 sub r4, r4, r7 ; subtract negative differences from sum
79 add r4, r4, r6 ; add positive differences to sum
80 sub r4, r4, r7 ; subtract negative differences from su
    [all...]
  /external/chromium_org/third_party/libvpx/source/libvpx/vp8/encoder/ppc/
encodemb_altivec.asm 16 ;# r4 unsigned char *usrc
37 lvsl v5, 0, r4 ;# permutate value for alignment
38 lvx v1, 0, r4 ;# src
41 add r4, r4, r7
49 lvsl v5, 0, r4 ;# permutate value for alignment
50 lvx v1, 0, r4 ;# src
52 add r4, r4, r7
113 ;# r4 unsigned char *sr
    [all...]
  /external/libvpx/libvpx/vp8/common/arm/armv6/
vp8_variance8x8_armv6.asm 25 push {r4-r10, lr}
31 mov r4, #0 ; initialize sum = 0
53 add r4, r4, r6 ; add positive differences to sum
54 sub r4, r4, r7 ; subtract negative differences from sum
79 add r4, r4, r6 ; add positive differences to sum
80 sub r4, r4, r7 ; subtract negative differences from su
    [all...]
  /external/libvpx/libvpx/vp8/encoder/ppc/
encodemb_altivec.asm 16 ;# r4 unsigned char *usrc
37 lvsl v5, 0, r4 ;# permutate value for alignment
38 lvx v1, 0, r4 ;# src
41 add r4, r4, r7
49 lvsl v5, 0, r4 ;# permutate value for alignment
50 lvx v1, 0, r4 ;# src
52 add r4, r4, r7
113 ;# r4 unsigned char *sr
    [all...]

Completed in 1338 milliseconds

<<11121314151617181920>>