/external/libvpx/vp8/common/ppc/ |
idctllm_altivec.asm | 138 lwz r6, 0(r1) 139 stw r6, 0(r4) 140 lwz r6, 4(r1) 141 stw r6, 4(r4) 145 lwz r6, 8(r1) 146 stw r6, 0(r4) 147 lwz r6, 12(r1) 148 stw r6, 4(r4) 153 lwz r6, 0(r1) 154 stw r6, 0(r4 [all...] |
recon_altivec.asm | 36 ;# r6 = int stride 46 row_of16 r3, r4, r5, r6 47 row_of16 r3, r4, r5, r6 48 row_of16 r3, r4, r5, r6 49 row_of16 r3, r4, r5, r6 84 ;# r6 = int stride 97 two_rows_of8 r3, r4, r5, r6, 1 102 two_rows_of8 r3, r4, r5, r6, 0 124 ;# r6 = int stride 166 stwux r0, r5, r6 [all...] |
/frameworks/base/media/libstagefright/codecs/amrwbenc/src/asm/ARMV5E/ |
Deemph_32_opt.s | 38 LDRSH r6, [r0], #2 @load x_hi[0]
44 MOV r10, r6, LSL #16 @L_tmp = x_hi[0]<<16
50 LDRSH r6, [r0], #2 @load x_hi[1]
58 MOV r10, r6, LSL #16
69 LDRSH r6, [r0], #2 @load x_hi[]
72 MOV r10, r6, LSL #16
77 LDRSH r6, [r0], #2 @load x_hi[]
83 MOV r10, r6, LSL #16
|
/frameworks/base/media/libstagefright/codecs/amrwbenc/src/asm/ARMV7/ |
Deemph_32_neon.s | 38 LDRSH r6, [r0], #2 @load x_hi[0]
44 MOV r10, r6, LSL #16 @L_tmp = x_hi[0]<<16
50 LDRSH r6, [r0], #2 @load x_hi[1]
58 MOV r10, r6, LSL #16
69 LDRSH r6, [r0], #2 @load x_hi[]
72 MOV r10, r6, LSL #16
77 LDRSH r6, [r0], #2 @load x_hi[]
83 MOV r10, r6, LSL #16
|
/bionic/libc/arch-sh/bionic/ |
_exit_with_stack_teardown.S | 38 mov r6, r8 /* save retCode : breaks r8 value */
|
/bionic/libc/arch-arm/bionic/ |
syscall.S | 43 stmfd sp!, {r4, r5, r6, r7} 48 ldmfd ip, {r3, r4, r5, r6} 50 ldmfd sp!, {r4, r5, r6, r7}
|
memcpy.S | 228 ldmcsia r1!, {r4, r5, r6, r7} /* 16 bytes */ 230 stmcsia r0!, {r4, r5, r6, r7} 299 ldmcsia r1!, {r4, r5, r6, r7} /* 16 bytes */ 301 stmcsia r0!, {r4, r5, r6, r7} 389 ldmia r1!, { r5,r6,r7, r8,r9,r10,r11} 397 orr r5, r5, r6, lsl #16 398 mov r6, r6, lsr #16 399 orr r6, r6, r7, lsl #1 [all...] |
memcmp.S | 243 stmfd sp!, {r5, r6, r7} 246 // r6 = lhs 250 rsb r6, r5, #32 /* r6 = left shift */ 260 orr ip, ip, r7, lsl r6 265 orreq ip, ip, r7, lsl r6 271 sub r1, r1, r6, lsr #3 272 ldmfd sp!, {r5, r6, r7} 284 sub r1, r1, r6, lsr #3 287 ldmfd sp!, {r5, r6, r7 [all...] |
/external/libvpx/vp8/common/arm/armv6/ |
sixtappredict8x4_v6.asm | 51 ldrb r6, [r0, #-5] ; load source data 59 pkhbt r6, r6, r7, lsl #16 ; r7 | r6 66 smuad r11, r6, r3 ; vp8_filter[0], vp8_filter[1] 69 ldrb r6, [r0], #1 75 pkhbt r10, r10, r6, lsl #16 ; r10 | r9 76 pkhbt r6, r6, r7, lsl #16 ; r11 | r10 78 smlad r12, r6, r5, r1 [all...] |
simpleloopfilter_v6.asm | 70 ldr r6, [src, pstep] ; q1 82 uqsub8 r7, r3, r6 ; p1 - q1 83 uqsub8 r8, r6, r3 ; q1 - p1 100 eor r6, r6, r2 ; q1 offset to convert to a signed value 104 qsub8 r3, r3, r6 ; vp8_filter = p1 - q1 105 qsub8 r6, r5, r4 ; q0 - p0 106 qadd8 r3, r3, r6 ; += q0 - p0 108 qadd8 r3, r3, r6 ; += q0 - p0 110 qadd8 r3, r3, r6 ; vp8_filter = p1-q1 + 3*(q0-p0) [all...] |
iwalsh_v6.asm | 29 ldr r6, [r0], #4 ; [9 | 8] 35 qadd16 r11, r4, r6 ; b1 [5+9 | 4+8] 36 qsub16 r12, r4, r6 ; c1 [5-9 | 4-8] 41 qsub16 r6, r10, r11 ; a1 - b1 [9 | 8] 91 qsubaddx r2, r6, r7 ; [c1|a1] [9-10 | 8+11] 92 qaddsubx r3, r6, r7 ; [b1|d1] [9+10 | 8-11] 96 qaddsubx r6, r2, r3 ; [b2|c2] [c1+d1 | a1-b1] 101 qadd16 r6, r6, r10 ; [b2+3|c2+3] 106 asr r2, r6, #3 ; [9 | x [all...] |
/external/openssl/crypto/sha/asm/ |
sha256-armv4.s | 31 ldmia r0,{r4,r5,r6,r7,r8,r9,r10,r11} 59 and r0,r0,r6 93 add r6,r6,r3 105 mov r0,r6,ror#6 106 eor r0,r0,r6,ror#11 107 eor r0,r0,r6,ror#25 @ Sigma1(e) 110 and r2,r2,r6 139 eor r2,r6,r7 169 eor r2,r5,r6 [all...] |
sha512-armv4.s | 69 ldr r6,[r0,#0+0] 133 str r6,[sp,#0+4] 164 mov r10,r6,lsr#28 165 eor r9,r9,r6,lsl#4 167 eor r9,r9,r6,lsr#2 170 eor r10,r10,r6,lsl#30 171 eor r9,r9,r6,lsr#7 174 eor r10,r10,r6,lsl#25 @ Sigma0(a) 184 and r12,r6,r10 185 orr r6,r6,r1 [all...] |
/frameworks/base/media/libstagefright/codecs/aacenc/src/asm/ARMV5E/ |
Radix4FFT_v5.s | 85 smulwt r6, r10, r8 @ L_mpy_wx(cosx, t0)
88 smlawb r4, r11, r8, r6 @ r4 = L_mpy_wx(cosx, t0) + L_mpy_wx(sinx, t1)@
94 ldrd r10, [r14] @ r6 = xptr[0]@ r7 = xptr[1]@
100 smlawb r6, r11, r8, r2 @ r4 = L_mpy_wx(cosx, t0) + L_mpy_wx(sinx, t1)@
109 add r4, r10, r6 @ r4 = t0 + r6@
112 sub r6, r10, r6 @ r6 = t0 - r6@ [all...] |
PrePostMDCT_v5.s | 41 ldrd r6, [r3]
55 smull r14, r4, r6, r8 @ MULHIGH(tr2, cosa)
59 smull r5, r10, r6, r9 @ MULHIGH(tr2, sina)
64 mov r6, r11
66 strd r6, [r0]
95 ldrd r6, [r3]
109 smull r14, r5, r6, r8 @ MULHIGH(tr2, cosa)
113 smull r7, r10, r6, r9 @ MULHIGH(tr2, sina)
115 add r6, r8, r5 @ MULHIGH(cosb, tr2) + MULHIGH(sinb, ti2)@
121 strd r6, [r3] [all...] |
/external/libvpx/vpx_scale/arm/neon/ |
vp8_vpxyv12_copyframe_func_neon.asm | 37 ldr r6, [r0, #yv12_buffer_config_y_stride] 53 add r10, r2, r6 81 add r2, r2, r6, lsl #1 98 mov r6, r6, lsr #1 ;src uv_stride 111 add r10, r2, r6 131 add r2, r2, r6, lsl #1 161 add r0, r8, r6 168 add r0, r8, r6 185 add r2, r2, r6, lsl # [all...] |
/frameworks/base/media/libstagefright/codecs/mp3dec/src/asm/ |
pvmp3_dct_9_arm.s | 70 add r6,r5,r7 74 add r7,r7,r6 120 mov r1,r6,lsl #1 122 and r6,r10,r11,asr #14 123 smlal r12,r8,r6,r1 125 add r12,r11,r6 133 smull r2,r6,r9,r1 137 smlal r12,r6,r10,r2 143 smlal r12,r6,r7,r3 155 smlal r8,r6,r7,r1 [all...] |
pvmp3_dct_9_gcc.s | 59 add r6,r5,r7 63 add r7,r7,r6 109 mov r1,r6,lsl #1 111 and r6,r10,r11,asr #14 112 smlal r12,r8,r6,r1 114 add r12,r11,r6 122 smull r2,r6,r9,r1 126 smlal r12,r6,r10,r2 132 smlal r12,r6,r7,r3 144 smlal r8,r6,r7,r1 [all...] |
pvmp3_dct_16_gcc.s | 64 ldr r6,constant5 82 smull r7,r4,r6,r4 91 ldr r6,constant7 93 smull r8,lr,r6,lr 99 sub r6,r12,lr 100 mov r6,r6,lsl #1 101 smull r9,r6,r8,r6 193 add r4,r4,r6 [all...] |
/external/v8/test/cctest/ |
test-disasm-arm.cc | 116 COMPARE(eor(r4, r5, Operand(r6, LSL, 0)), 117 "e0254006 eor r4, r5, r6"); 125 COMPARE(sub(r5, r6, Operand(r10, LSL, 31), LeaveCC, hs), 126 "20465f8a subcs r5, r6, r10, lsl #31"); 127 COMPARE(sub(r5, r6, Operand(r10, LSL, 30), SetCC, cc), 128 "30565f0a subccs r5, r6, r10, lsl #30"); 129 COMPARE(sub(r5, r6, Operand(r10, LSL, 24), LeaveCC, lo), 130 "30465c0a subcc r5, r6, r10, lsl #24"); 131 COMPARE(sub(r5, r6, Operand(r10, LSL, 16), SetCC, mi), 132 "4056580a submis r5, r6, r10, lsl #16") [all...] |
/external/tremolo/Tremolo/ |
dpen.s | 70 LDMIA r0,{r4,r6,r7} @ r4 = read = book->max_length 71 @ r6 = book->dec_table 85 @ r6 = dec_table 99 ADC r2, r6, r7, LSL #1 @ r8 = &t[chase*2+C] 118 @ r6 = dec_table 124 ADC r8, r6, r7 @ r8 = t+chase+bit 125 LDRB r10,[r8], -r6 @ r10= next=t[chase+bit] r8=chase+bit 139 LDRNEB r14,[r6, r7] @ r14= t[chase] 141 ADC r12,r8, r6 @ r12= chase+bit+1+t 153 MOV r6, r6, LSR # [all...] |
/external/openssl/crypto/aes/asm/ |
aes-armv4.s | 125 ldrb r6,[r12,#0] 128 orr r0,r0,r6,lsl#24 132 ldrb r6,[r12,#4] 135 orr r1,r1,r6,lsl#24 139 ldrb r6,[r12,#8] 142 orr r2,r2,r6,lsl#24 146 ldrb r6,[r12,#12] 149 orr r3,r3,r6,lsl#24 156 mov r6,r0,lsr#8 159 strb r6,[r12,#2 [all...] |
/external/libvpx/vp8/encoder/ppc/ |
encodemb_altivec.asm | 18 ;# r6 unsigned char *pred 28 add r6, r6, r9 39 lvx v2, 0, r6 ;# pred 42 addi r6, r6, 16 76 lvx v2, 0, r6 ;# pred 79 addi r6, r6, 16 115 ;# r6 int strid [all...] |
sad_altivec.asm | 63 add r5, r5, r6 88 add r5, r5, r6 100 add r5, r5, r6 121 add r5, r5, r6 129 add r5, r5, r6 150 ;# r6 int ref_stride 170 ;# r6 int ref_stride 190 ;# r6 int ref_stride 210 ;# r6 int ref_stride 248 ;# r6 int ref_strid [all...] |
/bootable/bootloader/legacy/usbloader/ |
init.S | 58 ldr r6, =BOOTLOADER_END 61 cmp r5, r6
|