/external/openssl/crypto/sha/asm/ |
sha512-x86_64.S | 9 pushq %r12 39 movq 0(%rsi),%r12 42 bswapq %r12 45 movq %r12,0(%rsp) 52 addq %r11,%r12 55 addq (%rbp,%rdi,8),%r12 65 addq %r15,%r12 73 addq %r13,%r12 76 addq %r12,%rdx 77 addq %r12,%r1 [all...] |
/external/skia/legacy/src/opts/ |
memset16_neon.S | 59 ands r12, r0, #0xf 66 rsb r12, r12, #16 67 sub r2, r2, r12 68 lsls r12, r12, #29 72 lsls r12, r12, #2 90 mov r12, r2, lsr #7 96 subs r12, r12, # [all...] |
/external/skia/src/opts/ |
memset16_neon.S | 59 ands r12, r0, #0xf 66 rsb r12, r12, #16 67 sub r2, r2, r12 68 lsls r12, r12, #29 72 lsls r12, r12, #2 90 mov r12, r2, lsr #7 96 subs r12, r12, # [all...] |
/frameworks/av/media/libstagefright/codecs/amrwbenc/src/asm/ARMV5E/ |
Syn_filt_32_opt.s | 41 STMFD r13!, {r4 - r12, r14} 104 SMULBB r12, r6, r11 @ sig_lo[i-1] * Aq[1] 107 SMLABT r12, r7, r11, r12 @ sig_lo[i-2] * Aq[2] 110 SMLABB r12, r9, r11, r12 @ sig_lo[i-3] * Aq[3] 112 SMLABT r12, r10, r11, r12 @ sig_lo[i-4] * Aq[4] 115 SMLABB r12, r6, r11, r12 @ sig_lo[i-5] * Aq[5 [all...] |
residu_asm_opt.s | 33 STMFD r13!, {r4 - r12, r14} 60 LDRH r12, [r0], #2 61 ORR r11, r12, r11, LSL #16 @r11 --- a12, a13 63 LDRH r12, [r0], #2 65 ORR r12, r4, r12, LSL #16 @r12 --- a14, a15 68 STMFD r13!, {r8 - r12} @store r8-r12 83 SMULTT r12, r5, r10 @i4(0) --- r12 = x[3] * a [all...] |
scale_sig_opt.s | 32 STMFD r13!, {r4 - r12, r14} 45 MOV r12, r5, LSL r10 46 TEQ r5, r12, ASR r10 47 EORNE r12, r8, r5, ASR #31 49 QADD r11, r12, r9 50 MOV r12, r11, ASR #16 51 STRH r12, [r4], #-2 61 MOV r12, r11, ASR #16 63 STRH r12, [r4], #-2 67 LDMFD r13!, {r4 - r12, r15 [all...] |
Deemph_32_opt.s | 36 STMFD r13!, {r4 - r12, r14} 47 ADD r12, r10, r7, LSL #4 @L_tmp += x_lo[0] << 4 48 MOV r10, r12, LSL #3 @L_tmp <<= 3 53 MOV r12, r10, LSL #1 @L_tmp = L_mac(L_tmp, *mem, fac) 54 QADD r10, r12, r11 59 ADD r12, r10, r7, LSL #4 61 MOV r10, r12, LSL #3 64 MOV r12, r10, LSL #1 65 QADD r10, r12, r11 73 ADD r12, r10, r7, LSL # [all...] |
Dot_p_opt.s | 34 STMFD r13!, {r4 - r12, r14} 62 MOV r12, r4, LSL #1 63 ADD r12, r12, #1 @ L_sum = (L_sum << 1) + 1 64 MOV r4, r12 66 CMP r12, #0 67 RSBLT r4, r12, #0 70 MOV r0, r12, LSL r10 @ L_sum = L_sum << sft 76 LDMFD r13!, {r4 - r12, r15}
|
cor_h_vec_opt.s | 39 STMFD r13!, {r4 - r12, r14} 54 LDRSH r12, [r10], #2 58 MLA r5, r12, r8, r5 59 MLA r6, r12, r14, r6 62 LDRSH r12, [r10], #2 @*p1++ 64 MLA r5, r12, r14, r5 75 MUL r12, r5, r10 77 MOV r5, r12, ASR #15 80 LDR r12, [r13, #48] 84 ADD r12, r12, r4, LSL # [all...] |
convolve_opt.s | 34 STMFD r13!, {r4 - r12, r14} 52 LDRSH r12, [r6], #2 @ *tmpX++ 55 MLA r8, r12, r14, r8 58 LDRSH r12, [r6], #2 @ *tmpX++ 62 MLA r8, r12, r14, r8 79 LDRSH r12, [r6], #2 84 MLA r8, r12, r14, r8 91 LDRSH r12, [r6], #2 @ *tmpX++ 94 MLA r8, r12, r14, r8 97 LDRSH r12, [r6], #2 @ *tmpX+ [all...] |
/frameworks/av/media/libstagefright/codecs/mp3dec/src/asm/ |
pvmp3_dct_16_gcc.s | 52 ldr r12,constant1 54 smull lr,r2,r12,r2 58 ldr r12,[r0,#0x20] 60 sub r3,r2,r12 66 add r3,r2,r12 68 ldr r12,constant3 70 smull lr,r2,r12,r2 71 ldr r12,[r0,#0x38] 74 sub r4,r1,r12 75 add r1,r1,r12 [all...] |
/frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src_gcc/ |
armVCM4P10_DecodeCoeffsToPair_s.S | 16 PUSH {r4-r12,lr} 19 LDR r12,[r1,#0] 26 ADD r12,r12,#8 30 LSLS r8,r11,r12 33 SUBS r12,r12,#8 39 ADDCC r12,r12,#8 40 ADD r12,r12,# [all...] |
armVCM4P10_Average_4x_Align_unsafe_s.S | 18 LDR r12,[r2,#0] 22 MVN r12,r12 24 UHSUB8 r5,r10,r12 31 LDR r12,[r2,#0] 34 MVN r12,r12 35 UHSUB8 r5,r10,r12 52 LDR r12,[r2,#0] 56 MVN r12,r1 [all...] |
/hardware/samsung_slsi/exynos5/libswconverter/ |
csc_tiled_to_linear_uv_deinterleave_neon.s | 69 @r12 temp1 72 stmfd sp!, {r4-r12,r14} @ backup registers 87 mov r12, r3, asr #1 @ temp1 = (width >> 1) * i + (j >> 1) 88 mul r12, r12, r5 92 add r12, r12, r6, asr #1 94 add r7, r0, r12 @ dst_offset = u_dst + temp1 109 add r7, r1, r12 @ dst_offset = v_dst + temp1 130 mov r12, r3, asr #1 @ temp1 = (width >> 1) * i + (j >> 1 [all...] |
csc_interleave_memcpy_neon.s | 62 @r12 src2_addr 65 stmfd sp!, {r8-r12,r14} @ backup registers 69 mov r12, r2 84 vld1.8 {q1}, [r12]! 85 vld1.8 {q3}, [r12]! 86 vld1.8 {q5}, [r12]! 87 vld1.8 {q7}, [r12]! 88 vld1.8 {q9}, [r12]! 89 vld1.8 {q11}, [r12]! 90 vld1.8 {q13}, [r12]! [all...] |
/dalvik/vm/compiler/template/armv5te/ |
TEMPLATE_RESTORE_STATE.S | 7 ldmia r0, {r0-r12}
|
/frameworks/av/media/libstagefright/codecs/amrwbenc/src/asm/ARMV7/ |
Deemph_32_neon.s | 36 STMFD r13!, {r4 - r12, r14} 47 ADD r12, r10, r7, LSL #4 @L_tmp += x_lo[0] << 4 48 MOV r10, r12, LSL #3 @L_tmp <<= 3 53 MOV r12, r10, LSL #1 @L_tmp = L_mac(L_tmp, *mem, fac) 54 QADD r10, r12, r11 59 ADD r12, r10, r7, LSL #4 61 MOV r10, r12, LSL #3 64 MOV r12, r10, LSL #1 65 QADD r10, r12, r11 73 ADD r12, r10, r7, LSL # [all...] |
cor_h_vec_neon.s | 39 STMFD r13!, {r4 - r12, r14} 55 LDRSH r12, [r10], #2 59 MLA r5, r12, r8, r5 60 MLA r6, r12, r14, r6 63 LDRSH r12, [r10], #2 @*p1++ 65 MLA r5, r12, r14, r5 76 MUL r12, r5, r10 78 MOV r5, r12, ASR #15 81 LDR r12, [r13, #48] 85 ADD r12, r12, r4, LSL # [all...] |
/external/llvm/test/MC/X86/ |
x86_64-bmi-encoding.s | 67 // CHECK: bextrq %r12, (%rax), %r10 69 bextrq %r12, (%rax), %r10 71 // CHECK: bextrq %r12, %r11, %r10 73 bextrq %r12, %r11, %r10 83 // CHECK: bzhiq %r12, (%rax), %r10 85 bzhiq %r12, (%rax), %r10 87 // CHECK: bzhiq %r12, %r11, %r10 89 bzhiq %r12, %r11, %r10 99 // CHECK: pextq %r12, %r11, %r10 101 pextq %r12, %r11, %r1 [all...] |
/frameworks/native/opengl/libagl/ |
iterators.S | 54 ldrb r12, [r0, #29] 57 sub r12, r12, #16 58 mov r3, r3, asr r12 59 mov r4, r4, asr r12 62 ldr r12, [r0, #4] 65 smull r6, r7, r4, r12 66 ldr r12, [r0, #12] 68 smlal r6, r7, r3, r12
|
/external/compiler-rt/lib/arm/ |
aeabi_ldivmod.S | 24 add r12, sp, #8 25 str r12, [sp]
|
aeabi_uldivmod.S | 24 add r12, sp, #8 25 str r12, [sp]
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
buildintrapredictorsmby_neon.asm | 44 mov r12, #128 45 vdup.u8 q0, r12 48 mov r12, #0 67 add r12, r4, r6 84 add r12, r12, r3 85 add r12, r12, r4 86 add r12, r12, r [all...] |
/external/tremolo/Tremolo/ |
mdctLARM.s | 61 LDMDB r2!,{r5,r6,r7,r12} 66 MOV r12,r12,ASR #9 @ r12= (*--r)>>9 68 MOV r14,r12,ASR #15 70 EORNE r12,r4, r14,ASR #31 71 STRH r12,[r0], r3 121 LDR r12,[r2],#8 126 RSB r12,r12,# [all...] |
/external/libvpx/libvpx/vpx_scale/arm/neon/ |
vp8_vpxyv12_copysrcframe_func_neon.asm | 49 mov r12, r5 60 sub r12, r12, #128 61 cmp r12, #128 72 cmp r12, #0 78 sub r12, r12, #8 79 cmp r12, #8 84 cmp r12, #0 89 subs r12, r12, # [all...] |