/external/libvpx/libvpx/vp8/encoder/arm/armv5te/ |
vp8_packtokens_partitions_armv5.asm | 51 ldr r6, _VP8_COMMON_MBrows_ 54 ldr r5, [r4, r6] ; load up mb_rows 110 ldrb r6, [r1, #tokenextra_token] ; t 113 add r4, r4, r6, lsl #3 ; a = vp8_coef_encodings + t 118 ldr r6, [r4, #vp8_token_value] ; v 130 lsl r12, r6, r4 ; r12 = v << 32 - n 141 mul r6, r4, r7 ; ((range-1) * pp[i>>1])) 149 add r4, r7, r6, lsr #8 ; 1 + (((range-1) * pp[i>>1]) >> 8) 155 clz r6, r4 156 sub r6, r6, #24 ; shif [all...] |
vp8_packtokens_armv5.asm | 60 ldrb r6, [r1, #tokenextra_token] ; t 63 add r4, r4, r6, lsl #3 ; a = vp8_coef_encodings + t 68 ldr r6, [r4, #vp8_token_value] ; v 80 lsl r12, r6, r4 ; r12 = v << 32 - n 91 mul r6, r4, r7 ; ((range-1) * pp[i>>1])) 99 add r4, r7, r6, lsr #8 ; 1 + (((range-1) * pp[i>>1]) >> 8) 105 clz r6, r4 106 sub r6, r6, #24 ; shift 110 adds r3, r3, r6 ; count += shif [all...] |
boolhuff_armv5te.asm | 71 mul r6, r4, r7 ; ((range-1) * probability) 74 add r4, r7, r6, lsr #8 ; 1 + (((range-1) * probability) >> 8) 80 clz r6, r4 81 sub r6, r6, #24 ; shift 85 adds r3, r3, r6 ; count += shift 86 lsl r5, r4, r6 ; range <<= shift 89 sub r6, r6, r3 ; offset = shift - count 90 sub r4, r6, #1 ; offset- [all...] |
/frameworks/av/media/libstagefright/codecs/amrwbenc/src/asm/ARMV5E/ |
syn_filt_opt.s | 47 LDRH r6, [r4], #2 56 STRH r6, [r5], #2 65 LDRH r6, [r4], #2 74 STRH r6, [r5], #2 90 LDRSH r6, [r0, #2] @ load a[1] 94 AND r6, r6, r14 96 ORR r10, r6, r7, LSL #16 @ -a[2] -- -a[1] 101 LDRSH r6, [r0, #10] @ load a[5] 105 AND r6, r6, r1 [all...] |
Filt_6k_7k_opt.s | 52 ADD r6, r13, #60 @ get x[L_FIR - 1] address 63 STRH r8, [r6], #2 64 STRH r9, [r6], #2 65 STRH r11, [r6], #2 66 STRH r12, [r6], #2 75 STRH r8, [r6], #2 76 STRH r9, [r6], #2 77 STRH r11, [r6], #2 78 STRH r12, [r6], #2 94 LDRSH r6, [r4, #2] @ load x[i + 1 [all...] |
/external/libvpx/libvpx/vp8/common/arm/armv6/ |
intra4x4_predict_v6.asm | 56 ldrb r6, [r1], r2 ; Left[2] 62 add r4, r4, r6 86 ldrb r6, [r1], r2 ; Left[2] 99 add r6, r6, r6, lsl #16 ; l[2|2] 116 sadd16 r1, r6, r10 ; l[2|2] + a[2|0] - [tl|tl] 117 sadd16 r2, r6, r11 ; l[2|2] + a[3|1] - [tl|tl] 180 ldrb r6, [r1], r2 ; Left[2] 185 add r10, r5, r6 ; l[1] + l[2 [all...] |
dequantize_v6.asm | 26 ldr r6, [r1, #4] 33 smulbb r9, r5, r6 34 smultt lr, r5, r6 39 ldr r6, [r1, #12] 46 smulbb r9, r5, r6 48 smultt lr, r5, r6 61 ldrne r6, [r1, #4]
|
/external/libvpx/libvpx/vp8/encoder/arm/armv6/ |
vp8_short_fdct4x4_armv6.asm | 34 qadd16 r6, r4, r5 ; [i1+i2 | i0+i3] = [b1 | a1] without shift 43 smuad r4, r6, lr ; o0 = (i1+i2)*8 + (i0+i3)*8 44 smusd r5, r6, lr ; o2 = (i1+i2)*8 - (i0+i3)*8 46 smlad r6, r7, r12, r11 ; o1 = (c1 * 2217 + d1 * 5352 + 14500) 51 pkhbt r3, r4, r6, lsl #4 ; [o1 | o0], keep in register for PART 2 52 pkhbt r6, r5, r7, lsl #4 ; [o3 | o2] 54 str r6, [r1, #4] 59 qadd16 r6, r8, r9 ; [i5+i6 | i4+i7] = [b1 | a1] without shift 68 smuad r9, r6, lr ; o4 = (i5+i6)*8 + (i4+i7)*8 69 smusd r8, r6, lr ; o6 = (i5+i6)*8 - (i4+i7)* [all...] |
vp8_mse16x16_armv6.asm | 40 ldr r6, [r2, #0x0] ; load 4 ref pixels 44 usub8 r8, r5, r6 ; calculate difference 47 usub8 r9, r6, r5 ; calculate difference with reversed operands 53 usad8 r6, r8, lr ; calculate sum of negative differences 59 uxtb16 r6, r8 ; byte (two pixels) to halfwords 61 smlad r4, r6, r6, r4 ; dual signed multiply, add and accumulate (1) 64 ldr r6, [r2, #0x4] ; load 4 ref pixels 67 usub8 r8, r5, r6 ; calculate difference 69 usub8 r9, r6, r5 ; calculate difference with reversed operand [all...] |
/dalvik/vm/compiler/template/armv5te/ |
header.S | 58 r6 rSELF thread pointer 76 #define rSELF r6
|
/external/llvm/test/MC/ARM/ |
eh-directive-save-diagnoatics.s | 18 .save {r4, r5, r6, r7} 20 @ CHECK: .save {r4, r5, r6, r7} 37 .save {r4, r5, r6, r7} 39 @ CHECK: .save {r4, r5, r6, r7}
|
eh-directive-save.s | 138 .save {r4, r5, r6, r7, r8, r9, r10, r11} 139 push {r4, r5, r6, r7, r8, r9, r10, r11} 140 pop {r4, r5, r6, r7, r8, r9, r10, r11} 194 .save {r4, r5, r6, r7, r8, r9, r10, r11, r14} 195 push {r4, r5, r6, r7, r8, r9, r10, r11, r14} 196 pop {r4, r5, r6, r7, r8, r9, r10, r11, r14} 224 .save {r4, r5, r6, r7, r8, r9, r10, r11, r12, r14} 225 push {r4, r5, r6, r7, r8, r9, r10, r11, r12, r14} 226 pop {r4, r5, r6, r7, r8, r9, r10, r11, r12, r14} 238 .save {r4, r5, r6, r8, r9, r10, r11 [all...] |
/ndk/tests/build/ssax-instructions/jni/ |
test.S | 14 lsr lr, r6, #2 24 ldrd r6, [r0], ip 32 sadd16 r3, r6, sl 33 ssub16 lr, r6, sl 36 ssax r6, r1, lr 48 strd r6, [r2], ip 53 ldrd r6, [sp]
|
/system/core/libpixelflinger/ |
col32cb16blend_neon.S | 132 mov r6, r8, lsr #11 // extract dest red 136 smlabb r6, r6, r5, r11 // dest red * alpha + src red 140 mov r6, r6, lsr #8 // shift down red 142 mov r6, r6, lsl #11 // shift red into 565 143 orr r6, r7, lsl #5 // shift green into 565 144 orr r6, r8, lsr #8 // shift blue into 565 146 strh r6, [r0], #2 // store pixel to dest, update pt [all...] |
/external/chromium_org/third_party/openssl/openssl/crypto/bn/asm/ |
s390x.S | 26 stmg %r6,%r10,48(%r15) 36 mlgr %r6,%r5 // *=w 38 alcgr %r6,zero 44 alcgr %r9,%r6 50 mlgr %r6,%r5 52 alcgr %r6,zero 58 alcgr %r9,%r6 72 lmg %r6,%r10,48(%r15) 77 mlgr %r6,%r5 // *=w 79 alcgr %r6,zer [all...] |
/external/libvpx/libvpx/vp8/encoder/arm/neon/ |
subtract_neon.asm | 32 ldr r6, [r0, #vp8_block_src_stride] 36 vld1.8 {d0}, [r3], r6 ;load src 38 vld1.8 {d2}, [r3], r6 40 vld1.8 {d4}, [r3], r6 42 vld1.8 {d6}, [r3], r6 70 mov r6, #32 ; "diff" stride x2 92 vst1.16 {q8}, [r0], r6 ;store diff 93 vst1.16 {q9}, [r5], r6 94 vst1.16 {q10}, [r0], r6 95 vst1.16 {q11}, [r5], r6 [all...] |
/external/openssl/crypto/bn/asm/ |
s390x.S | 26 stmg %r6,%r10,48(%r15) 36 mlgr %r6,%r5 // *=w 38 alcgr %r6,zero 44 alcgr %r9,%r6 50 mlgr %r6,%r5 52 alcgr %r6,zero 58 alcgr %r9,%r6 72 lmg %r6,%r10,48(%r15) 77 mlgr %r6,%r5 // *=w 79 alcgr %r6,zer [all...] |
/external/lzma/Asm/x86/ |
AesOpt.asm | 29 push r6
40 pop r6
146 add r1, r6
147 neg r6
148 add r6, 32
154 mov r3, r6
164 movdqa [r1 + r6 - 64], xmm0
191 add r1, r6
192 neg r6
193 add r6, 32 [all...] |
/external/tremolo/Tremolo/ |
mdctARM.s | 61 LDMDB r2!,{r5,r6,r7,r12} 64 MOV r6, r6, ASR #9 @ r6 = (*--r)>>9 78 MOV r14,r6, ASR #15 80 EORNE r6, r4, r14,ASR #31 81 STRH r6, [r0], r3 123 LDR r6, [r2],#8 128 RSB r6, r6, # [all...] |
/bionic/libc/arch-arm/generic/bionic/ |
memcpy.S | 109 ldmcsia r1!, {r4, r5, r6, r7} /* 16 bytes */ 111 stmcsia r0!, {r4, r5, r6, r7} 180 ldmcsia r1!, {r4, r5, r6, r7} /* 16 bytes */ 182 stmcsia r0!, {r4, r5, r6, r7} 270 ldmia r1!, { r5,r6,r7, r8,r9,r10,r11} 278 orr r5, r5, r6, lsl #16 279 mov r6, r6, lsr #16 280 orr r6, r6, r7, lsl #1 [all...] |
/frameworks/av/media/libstagefright/codecs/amrwbenc/src/asm/ARMV7/ |
convolve_neon.s | 42 MOV r6, r0 43 LDRSH r9, [r6], #2 @ *tmpX++ 54 VLD1.S16 D0, [r6]! 74 MOV r6, r0 75 LDRSH r9, [r6], #2 @ *tmpX++ 77 LDRSH r12, [r6], #2 90 VLD1.S16 D0, [r6]! 110 MOV r6, r0 111 LDRSH r9, [r6], #2 113 LDRSH r12, [r6], # [all...] |
/hardware/samsung_slsi/exynos5/libswconverter/ |
csc_tiled_to_linear_uv_deinterleave_neon.s | 63 @r6 j 84 mov r6, #0 92 add r12, r12, r6, asr #1 117 add r6, r6, #16 119 cmp r6, r10 128 add r8, r8, r6, lsl #3 132 add r12, r12, r6, asr #1 135 sub r14, r3, r6 @ r14 = (width - j) / 2 171 mov r6, # [all...] |
csc_tiled_to_linear_y_neon.s | 61 @r6 j 82 mov r6, #0 94 add r12, r12, r6 115 add r6, r6, #16 117 cmp r6, r10 126 add r8, r8, r6, lsl #4 130 add r12, r12, r6 132 sub r14, r2, r6 @ r14 = width - j 164 mov r6, # [all...] |
csc_ARGB8888_to_YUV420SP_NEON.s | 15 @r6 temp7, nWidth16m 200 ldr r6, =0x000000FF 201 and r10, r10, r6 @R: (rgbIn[k] & 0xF800) >> 10; 202 ldr r6, =0x00FF0000 203 and r14, r14, r6 @R: (rgbIn[k] & 0xF800) >> 10; 212 ldr r6, =0x000000FF 213 and r10, r10, r6 @G: 214 ldr r6, =0x00FF0000 215 and r14, r14, r6 @G: 224 ldr r6, =0x000000F [all...] |
/external/libvpx/libvpx/vp8/common/arm/neon/ |
buildintrapredictorsmby_neon.asm | 58 sub r6, r0, r2 59 vld1.8 {q1}, [r6] 65 vmov.32 r6, d9[0] 67 add r12, r4, r6 82 ldrb r6, [r0], r2 87 add r12, r12, r6 92 ldrb r6, [r0], r2 97 add r12, r12, r6 102 ldrb r6, [r0], r2 107 add r12, r12, r6 [all...] |