/external/libvpx/libvpx/vp8/encoder/arm/neon/ |
subtract_neon.asm | 26 stmfd sp!, {r4-r7} 29 ldr r4, [r0, #vp8_block_src] 33 add r3, r3, r4 ; src = *base_src + src 57 ldmfd sp!, {r4-r7} 67 push {r4-r7} 69 ldr r4, [sp, #16] ; pred_stride 75 vld1.8 {q1}, [r3], r4 ;load pred 77 vld1.8 {q3}, [r3], r4 79 vld1.8 {q5}, [r3], r4 81 vld1.8 {q7}, [r3], r4 [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/arm/armv6/ |
vp8_variance16x16_armv6.asm | 27 stmfd sp!, {r4-r12, lr} 38 ldr r4, [r0, #0] ; load 4 src pixels 43 usub8 r6, r4, r5 ; calculate difference 46 usub8 r9, r5, r4 ; calculate difference with reversed operands 51 usad8 r4, r7, lr ; calculate sum of positive differences 55 adds r8, r8, r4 ; add positive differences to sum 64 ldr r4, [r0, #4] ; load 4 src pixels 68 usub8 r6, r4, r5 ; calculate difference 70 usub8 r9, r5, r4 ; calculate difference with reversed operands 74 usad8 r4, r7, lr ; calculate sum of positive difference [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/ppc/ |
variance_altivec.asm | 79 add r3, r3, r4 93 lwz r4, 12(r1) 98 stw r4, 0(r7) ;# sse 102 subf r3, r3, r4 ;# sse - ((sum*sum) >> DS) 112 add r3, r3, r4 120 add r3, r3, r4 137 lwz r4, 12(r1) 142 stw r4, 0(r7) ;# sse 146 subf r3, r3, r4 ;# sse - ((sum*sum) >> 8) 151 ;# r4 int source_strid [all...] |
filter_altivec.asm | 192 ;# r4 int src_pitch 231 Read8x8 v2, r3, r4, 1 232 Read8x8 v3, r3, r4, 1 233 Read8x8 v4, r3, r4, 1 234 Read8x8 v5, r3, r4, 1 252 sub r9, r9, r4 253 sub r9, r9, r4 255 Read8x8 v0, r9, r4, 1 256 Read8x8 v1, r9, r4, 0 257 Read8x8 v6, r3, r4, [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/encoder/arm/armv6/ |
vp8_subtract_armv6.asm | 29 stmfd sp!, {r4-r9} 31 ldr r4, [r0, #vp8_block_base_src] 35 ldr r3, [r4] 47 uxtb16 r4, r0 ; [s2 | s0] 52 usub16 r4, r4, r5 ; [d2 | d0] 57 pkhbt r0, r4, r5, lsl #16 ; [d1 | d0] 58 pkhtb r1, r5, r4, asr #16 ; [d3 | d2] 66 ldmfd sp!, {r4-r9} 81 stmfd sp!, {r4-r11 [all...] |
vp8_mse16x16_armv6.asm | 29 push {r4-r9, lr} 35 mov r4, #0 ; initialize sse = 0 61 smlad r4, r6, r6, r4 ; dual signed multiply, add and accumulate (1) 65 smlad r4, r7, r7, r4 ; dual signed multiply, add and accumulate (2) 80 smlad r4, r6, r6, r4 ; dual signed multiply, add and accumulate (1) 84 smlad r4, r7, r7, r4 ; dual signed multiply, add and accumulate (2 [all...] |
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/encoder/arm/neon/ |
subtract_neon.asm | 26 stmfd sp!, {r4-r7} 29 ldr r4, [r0, #vp8_block_src] 33 add r3, r3, r4 ; src = *base_src + src 57 ldmfd sp!, {r4-r7} 67 push {r4-r7} 69 ldr r4, [sp, #16] ; pred_stride 75 vld1.8 {q1}, [r3], r4 ;load pred 77 vld1.8 {q3}, [r3], r4 79 vld1.8 {q5}, [r3], r4 81 vld1.8 {q7}, [r3], r4 [all...] |
/external/libhevc/common/arm/ |
ihevc_intra_pred_chroma_mode_18_34.s | 107 stmfd sp!, {r4-r12, r14} @stack stores the values of the arguments 110 ldr r4,[sp,#40] 113 cmp r4,#4 116 mov r12,r4 117 mov r11,r4 118 add r0,r0,r4,lsl #2 184 ldmfd sp!,{r4-r12,r15} @reload the registers from sp
|
ihevc_inter_pred_filters_luma_horz.s | 102 @ r4 => *pi1_coeff 118 stmfd sp!, {r4-r12, r14} @stack stores the values of the arguments 125 ldr r4,[sp,#40] @loads pi1_coeff 129 vld1.8 {d0},[r4] @coeff = vld1_s8(pi1_coeff) 141 add r4,r12,r2 @pu1_src_tmp2_8 = pu1_src + src_strd 185 add r4,r12,r2 @pu1_src + src_strd 223 vld1.u32 {d12},[r4],r11 @vector load pu1_src + src_strd 225 vld1.u32 {d13},[r4],r11 227 vld1.u32 {d14},[r4],r11 229 vld1.u32 {d15},[r4],r1 [all...] |
ihevc_itrans_recon_4x4.s | 98 @ r4 => src_strd 124 stmfd sp!, {r4-r12, r14} @stack stores the values of the arguments 130 ldr r4,[sp,#40] @loading src_strd 132 add r4,r4,r4 @ src_strd in terms of word16 136 add r9,r0,r4 @ pi2_src[0] + src_strd 143 add r10,r9,r4, lsl #1 @ 3*src_strd 144 add r4,r4,r [all...] |
/external/linux-tools-perf/perf-3.12.0/arch/arm/lib/ |
memset.S | 64 stmfd sp!, {r4-r8, lr} 65 mov r4, r1 80 stmcsia ip!, {r4, r5, r6, r7} 81 stmmiia ip!, {r4, r5} 90 ldmeqfd sp!, {r4-r8, pc} 95 stmneia ip!, {r4-r7} 96 ldmfd sp!, {r4-r8, lr}
|
/external/tremolo/Tremolo/ |
floor1LARM.s | 48 STMFD r13!,{r4-r6,r11,r14} 53 LDR r4, [r1] @ r4 = *d 56 MOV r4, r4, ASR #6 57 MUL r5, r4, r5 @ r5 = MULT31_SHIFT15 64 LDMFD r13!,{r4-r6,r11,PC}
|
/external/webrtc/src/modules/audio_coding/codecs/isac/fix/source/ |
lattice_armv7.S | 22 @ r4: out loop counter 37 .save {r4-r11} 38 push {r4-r11} 42 mov r4, #HALF_SUBFRAMELEN 43 sub r4, #1 @ Outer loop counter = HALF_SUBFRAMELEN - 1 75 subs r4, #1 78 pop {r4-r11}
|
/frameworks/av/media/libstagefright/codecs/mp3dec/src/asm/ |
pvmp3_dct_9_gcc.s | 43 stmfd sp!,{r4-r11,lr} 50 ldr r4,[r0, #0x18] 54 add r12,r4,r5 55 sub r4,r4,r5 124 mov r2,r4,lsl #1 133 add r4,r5,r4 135 sub lr,r4,lr 138 ldr r4,cos_pi_ [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/ppc/ |
filter_altivec.asm | 192 ;# r4 int src_pitch 231 Read8x8 v2, r3, r4, 1 232 Read8x8 v3, r3, r4, 1 233 Read8x8 v4, r3, r4, 1 234 Read8x8 v5, r3, r4, 1 252 sub r9, r9, r4 253 sub r9, r9, r4 255 Read8x8 v0, r9, r4, 1 256 Read8x8 v1, r9, r4, 0 257 Read8x8 v6, r3, r4, [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/encoder/arm/armv6/ |
vp8_mse16x16_armv6.asm | 29 push {r4-r9, lr} 35 mov r4, #0 ; initialize sse = 0 61 smlad r4, r6, r6, r4 ; dual signed multiply, add and accumulate (1) 65 smlad r4, r7, r7, r4 ; dual signed multiply, add and accumulate (2) 80 smlad r4, r6, r6, r4 ; dual signed multiply, add and accumulate (1) 84 smlad r4, r7, r7, r4 ; dual signed multiply, add and accumulate (2 [all...] |
/external/chromium_org/third_party/opus/src/celt/arm/ |
celt_pitch_xcorr_arm.s | 47 ; r4 = opus_val16 *x 79 VLD1.16 {d6, d7}, [r4]! 106 VLD1.16 d6, [r4]! 124 VLD2.16 {d6[],d7[]}, [r4]! 139 VLD1.16 {d6[]}, [r4]! 149 VLD1.16 {d6[]}, [r4]! 165 ; r4 = opus_val16 *x (for xcorr_kernel_neon()) 170 STMFD sp!, {r4-r6, lr} 178 ; r3 = len, r4 = _x, r5 = _y, q0 = {0, 0, 0, 0} 179 MOV r4, r [all...] |
/external/libopus/celt/arm/ |
celt_pitch_xcorr_arm.s | 47 ; r4 = opus_val16 *x 79 VLD1.16 {d6, d7}, [r4]! 106 VLD1.16 d6, [r4]! 124 VLD2.16 {d6[],d7[]}, [r4]! 139 VLD1.16 {d6[]}, [r4]! 149 VLD1.16 {d6[]}, [r4]! 165 ; r4 = opus_val16 *x (for xcorr_kernel_neon()) 170 STMFD sp!, {r4-r6, lr} 178 ; r3 = len, r4 = _x, r5 = _y, q0 = {0, 0, 0, 0} 179 MOV r4, r [all...] |
/external/libvpx/libvpx/vp8/common/ppc/ |
filter_altivec.asm | 192 ;# r4 int src_pitch 231 Read8x8 v2, r3, r4, 1 232 Read8x8 v3, r3, r4, 1 233 Read8x8 v4, r3, r4, 1 234 Read8x8 v5, r3, r4, 1 252 sub r9, r9, r4 253 sub r9, r9, r4 255 Read8x8 v0, r9, r4, 1 256 Read8x8 v1, r9, r4, 0 257 Read8x8 v6, r3, r4, [all...] |
/external/libvpx/libvpx/vp8/encoder/arm/armv6/ |
vp8_mse16x16_armv6.asm | 29 push {r4-r9, lr} 35 mov r4, #0 ; initialize sse = 0 61 smlad r4, r6, r6, r4 ; dual signed multiply, add and accumulate (1) 65 smlad r4, r7, r7, r4 ; dual signed multiply, add and accumulate (2) 80 smlad r4, r6, r6, r4 ; dual signed multiply, add and accumulate (1) 84 smlad r4, r7, r7, r4 ; dual signed multiply, add and accumulate (2 [all...] |
/external/chromium_org/v8/test/cctest/ |
test-disasm-arm.cc | 107 COMPARE(and_(r2, r3, Operand(r4), SetCC), 108 "e0132004 ands r2, r3, r4"); 109 COMPARE(and_(r3, r4, Operand(r5), LeaveCC, eq), 110 "00043005 andeq r3, r4, r5"); 112 COMPARE(eor(r4, r5, Operand(r6, LSL, 0)), 113 "e0254006 eor r4, r5, r6"); 114 COMPARE(eor(r4, r5, Operand(r7, LSL, 1), SetCC), 115 "e0354087 eors r4, r5, r7, lsl #1"); 116 COMPARE(eor(r4, r5, Operand(r8, LSL, 2), LeaveCC, ne), 117 "10254108 eorne r4, r5, r8, lsl #2") [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/arm/armv6/ |
vp8_variance8x8_armv6.asm | 25 push {r4-r10, lr} 31 mov r4, #0 ; initialize sum = 0 53 add r4, r4, r6 ; add positive differences to sum 54 sub r4, r4, r7 ; subtract negative differences from sum 79 add r4, r4, r6 ; add positive differences to sum 80 sub r4, r4, r7 ; subtract negative differences from su [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/encoder/ppc/ |
encodemb_altivec.asm | 16 ;# r4 unsigned char *usrc 37 lvsl v5, 0, r4 ;# permutate value for alignment 38 lvx v1, 0, r4 ;# src 41 add r4, r4, r7 49 lvsl v5, 0, r4 ;# permutate value for alignment 50 lvx v1, 0, r4 ;# src 52 add r4, r4, r7 113 ;# r4 unsigned char *sr [all...] |
/external/libvpx/libvpx/vp8/common/arm/armv6/ |
vp8_variance8x8_armv6.asm | 25 push {r4-r10, lr} 31 mov r4, #0 ; initialize sum = 0 53 add r4, r4, r6 ; add positive differences to sum 54 sub r4, r4, r7 ; subtract negative differences from sum 79 add r4, r4, r6 ; add positive differences to sum 80 sub r4, r4, r7 ; subtract negative differences from su [all...] |
/external/libvpx/libvpx/vp8/encoder/ppc/ |
encodemb_altivec.asm | 16 ;# r4 unsigned char *usrc 37 lvsl v5, 0, r4 ;# permutate value for alignment 38 lvx v1, 0, r4 ;# src 41 add r4, r4, r7 49 lvsl v5, 0, r4 ;# permutate value for alignment 50 lvx v1, 0, r4 ;# src 52 add r4, r4, r7 113 ;# r4 unsigned char *sr [all...] |