HomeSort by relevance Sort by last modified time
    Searched full:lsl (Results 26 - 50 of 970) sorted by null

12 3 4 5 6 7 8 91011>>

  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/arm/armv6/
dequant_idct_v6.asm 69 pkhbt r7, r7, r9, lsl #16
71 pkhbt r8, r8, r10, lsl #16
77 pkhbt r9, r9, r11, lsl #16
79 pkhbt r10, r10, r7, lsl #16
107 pkhbt r11, r8, r6, lsl #16
108 pkhbt r1, lr, r1, lsl #16
109 pkhbt r12, r10, r12, lsl #16
112 pkhbt lr, r9, r7, lsl #16
121 pkhbt r1, r7, r1, lsl #16
123 pkhbt r11, r9, r11, lsl #1
    [all...]
dc_only_idct_add_v6.asm 31 orr r0, r0, r0, lsl #16 ; a1 | a1
43 orr r5, r5, r4, lsl #8
44 orr r7, r7, r6, lsl #8
58 orr r5, r5, r4, lsl #8
59 orr r7, r7, r6, lsl #8
  /external/chromium_org/third_party/libvpx/source/libvpx/vp9/common/arm/neon/
vp9_copy_neon.asm 34 pld [r0, r1, lsl #1]
44 pld [r0, r1, lsl #1]
46 pld [r0, r1, lsl #1]
55 pld [r0, r1, lsl #1]
57 pld [r0, r1, lsl #1]
66 pld [r0, r1, lsl #1]
68 pld [r0, r1, lsl #1]
  /external/libvpx/libvpx/vp9/common/arm/neon/
vp9_copy_neon.asm 34 pld [r0, r1, lsl #1]
44 pld [r0, r1, lsl #1]
46 pld [r0, r1, lsl #1]
55 pld [r0, r1, lsl #1]
57 pld [r0, r1, lsl #1]
66 pld [r0, r1, lsl #1]
68 pld [r0, r1, lsl #1]
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/arm/neon/
vp9_copy_neon.asm 34 pld [r0, r1, lsl #1]
44 pld [r0, r1, lsl #1]
46 pld [r0, r1, lsl #1]
55 pld [r0, r1, lsl #1]
57 pld [r0, r1, lsl #1]
66 pld [r0, r1, lsl #1]
68 pld [r0, r1, lsl #1]
  /system/core/libpixelflinger/tests/arch-arm64/disassembler/
arm64_diassembler_test.cpp 41 { 0x91000240, "add x0, x18, #0x0, lsl #0" },
42 { 0x9140041f, "add sp, x0, #0x1, lsl #12" },
43 { 0x917ffff2, "add x18, sp, #0xfff, lsl #12" },
45 { 0xd13ffe40, "sub x0, x18, #0xfff, lsl #0" },
46 { 0xd140001f, "sub sp, x0, #0x0, lsl #12" },
47 { 0xd14007f2, "sub x18, sp, #0x1, lsl #12" },
49 { 0x8b1e0200, "add x0, x16, x30, lsl #0" },
55 { 0x4b1e0200, "sub w0, w16, w30, lsl #0" },
60 { 0x6b1e0200, "subs w0, w16, w30, lsl #0" },
65 { 0x0a1e0200, "and w0, w16, w30, lsl #0" }
    [all...]
  /external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/arm/armv6/
dc_only_idct_add_v6.asm 31 orr r0, r0, r0, lsl #16 ; a1 | a1
43 orr r5, r5, r4, lsl #8
44 orr r7, r7, r6, lsl #8
58 orr r5, r5, r4, lsl #8
59 orr r7, r7, r6, lsl #8
  /external/libvpx/libvpx/vp8/common/arm/armv6/
dc_only_idct_add_v6.asm 31 orr r0, r0, r0, lsl #16 ; a1 | a1
43 orr r5, r5, r4, lsl #8
44 orr r7, r7, r6, lsl #8
58 orr r5, r5, r4, lsl #8
59 orr r7, r7, r6, lsl #8
  /frameworks/av/media/libstagefright/codecs/on2/h264dec/source/arm11_asm/
h264bsd_interpolate_chroma_hor_ver.s 164 ADD count, count, tmp2, LSL #16 ;// chromaPartHeight-1
165 ADD count, count, tmp2, LSL #24 ;// loop_y
166 ADD count, count, tmp1, LSL #20 ;// chromaPartWidth-1
168 PKHBT valY, valY, yFrac, LSL #16 ;// |yFrac|valY |
182 LDRB tmp5, [ptrA, width, LSL #1]
184 PKHBT tmp1, tmp1, tmp3, LSL #16 ;// |t3|t1|
185 PKHBT tmp3, tmp3, tmp5, LSL #16 ;// |t5|t3|
190 ADD count, count, tmp2, LSL #8
195 LDRB tmp6, [ptrA, width, LSL #1]
197 PKHBT tmp2, tmp2, tmp4, LSL #16 ;// |t4|t2
    [all...]
h264bsd_interpolate_chroma_hor.s 158 ADD count, count, tmp2, LSL #16 ;// chromaPartHeight-1
159 ADD count, count, tmp2, LSL #24 ;// loop_y
160 ADD count, count, tmp1, LSL #20 ;// chromaPartWidth-1
162 PKHBT valX, valX, xFrac, LSL #16 ;// |xFrac|valX |
163 MOV valX, valX, LSL #3 ;// multiply by 8 in advance
175 ADD count, count, tmp2, LSL #8
183 PKHBT tmp5, tmp1, tmp3, LSL #16
184 PKHBT tmp6, tmp2, tmp4, LSL #16
192 PKHBT tmp7, tmp3, tmp1, LSL #16
193 PKHBT tmp8, tmp4, tmp2, LSL #1
    [all...]
h264bsd_interpolate_ver_quarter.s 137 ADD count, partW, partH, LSL #8 ;// |xx|xx|partH|partW|
139 RSB count, tmp5, count, LSL #8 ;// |xx|partH-1|partW-1|xx|
148 ADD count, count, tmp1, LSL #16 ;// partWidth-1 to top byte
162 ADD tmpa, tmpa, tmpa, LSL #2 ;// 5(G+M)
165 ADD tmpa, plus16, tmpa, LSL #2 ;// 16+20(G+M)
169 ADD tmpb, tmpb, tmpb, LSL #2 ;// 5(C+R)
178 ADD tmpa, tmpa, tmpa, LSL #2 ;// 5(G+M)
180 ADD tmpa, plus16, tmpa, LSL #2 ;// 16+20(G+M)
185 ADD tmpb, tmpb, tmpb, LSL #2 ;// 5(C+R)
189 MOVS tmp1, count, LSL #31 ;// update flags (verOffset
    [all...]
  /external/tremolo/Tremolo/
mdctARM.s 58 MOV r3, r3, LSL #1
117 MOV r3, r3, LSL #1
193 MOV r8, r8, LSL #1
232 MOV r8, r8, LSL #1
300 MOV r2,r2,LSL #1
304 MOV r2,r2,LSL r3 @ r2 = step = 2<<shift
312 ADD r4, r1, r0, LSL #1 @ r4 = aX = in+(n>>1)
323 LDR r10,[r5],r2,LSL #2 @ r10= T[0] T += step
333 MOV r9, r9, LSL #1
338 MOV r12,r12,LSL #
    [all...]
  /external/valgrind/main/none/tests/arm/
v6intThumb.c 84 "orr " #RD "," #RD "," #RD ", LSL #8" "\n\t" \
85 "orr " #RD "," #RD "," #RD ", LSL #16" "\n\t" \
437 printf("LSL\n");
438 TESTINST3("lsl r0, r1, r2", 0xffffffff, 0, r0, r1, r2, 0);
439 TESTINST3("lsl r0, r1, r2", 0xffffffff, 1, r0, r1, r2, 0);
440 TESTINST3("lsl r0, r1, r2", 0xffffffff, 2, r0, r1, r2, 0);
441 TESTINST3("lsl r0, r1, r2", 0xffffffff, 31, r0, r1, r2, 0);
442 TESTINST3("lsl r0, r1, r2", 0xffffffff, 32, r0, r1, r2, 0);
443 TESTINST3("lsl r0, r1, r2", 0xffffffff, 33, r0, r1, r2, 0);
444 TESTINST3("lsl r0, r1, r2", 0xffffffff, 63, r0, r1, r2, 0)
    [all...]
  /frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm11/vc/m4p10/src/
omxVCM4P10_DequantTransformResidualFromPairAndAdd_s.s 97 ORR rowLuma01,rowLuma01,temp3,LSL #16 ;// rowLuma01 = [0b|0a]
106 ;// So we can shift the packed rowLuma values [0b|0a] with a single LSL operation
110 LSL rowLuma01,rowLuma01,shift
111 LSL rowLuma23,rowLuma23,shift
139 PKHBT SrcDst00,SrcDst00,temp1,LSL #16 ;// Pack the first two product values
146 PKHBT SrcDst02,SrcDst02,temp2,LSL #16 ;// Pack the next two product values
150 PKHBT SrcDst10,SrcDst10,temp1,LSL #16 ;// Pack the next two product values
158 PKHBT SrcDst12,SrcDst12,temp2,LSL #16 ;// Pack the next two product values
162 PKHBT SrcDst20,SrcDst20,temp1,LSL #16 ;// Pack the next two product values
174 PKHBT SrcDst22,SrcDst22,temp2,LSL #16 ;// Pack the remaining product value
    [all...]
armVCM4P10_Interpolate_Chroma_s.s 135 ADD dxEightMinusdx, EightMinusdx, dx, LSL #16
136 ORR iWidth, iWidth, temp, LSL #16
167 ORR x01x00, x00, x01, LSL #16
169 ORR x02x01, x01, x02, LSL #16
171 ORR x11x10, x10, x11, LSL #16
172 ORR x12x11, x11, x12, LSL #16
187 RSB pSrc2, pSrc, pSrc1, LSL #1
194 ORR OutRow0100, OutRow00, OutRow01, LSL #8
204 ORR x21x20, x20, x21, LSL #16
205 ORR x22x21, x21, x22, LSL #16
    [all...]
armVCM4P10_InterpolateLuma_HalfVer4x4_unsafe_s.s 121 SUB pSrc, pSrc, srcStep, LSL #2
131 RSB ValCD0, ValEB0, ValCD0, LSL #2 ;// 4*(Off+C+D) - (Off+B+E)
133 LDR ValD, [pSrc, srcStep, LSL #1] ;// Load [d3 d2 d1 d0]
135 RSB ValCD1, ValEB1, ValCD1, LSL #2
140 LDR ValF, [pSrc, srcStep, LSL #2] ;// Load [f3 f2 f1 f0]
142 ADD ValCD0, ValCD0, ValCD0, LSL #2 ;// 5 * [4*(Off+C+D) - (Off+B+E)]
143 ADD ValCD1, ValCD1, ValCD1, LSL #2
146 RSB ValED1, ValCF1, ValED1, LSL #2
148 SUB ValA, pSrc, srcStep, LSL #1
150 RSB ValED0, ValCF0, ValED0, LSL #2 ;// 4*(Off+E+D) - (Off+C+F)
    [all...]
  /external/chromium_org/third_party/libvpx/source/libvpx/build/make/
thumb.pm 19 # Write additions with shifts, such as "add r10, r11, lsl #8",
20 # in three operand form, "add r10, r10, r11, lsl #8".
21 s/(add\s+)(r\d+),\s*(r\d+),\s*(lsl #\d+)/$1$2, $2, $3, $4/g;
28 # This converts instructions such as "add r12, r12, r5, lsl r4"
29 # into the sequence "lsl r5, r4", "add r12, r12, r5", "lsr r5, r4".
30 s/^(\s*)(add)(\s+)(r\d+),\s*(r\d+),\s*(r\d+),\s*lsl (r\d+)/$1lsl$3$6, $7\n$1$2$3$4, $5, $6\n$1lsr$3$6, $7/g;
44 # This converts "ldrne r4, [src, -pstep, lsl #1]" into
45 # "subne src, src, pstep, lsl #1", "ldrne r4, [src]",
46 # "addne src, src, pstep, lsl #1". In a couple of cases wher
    [all...]
  /external/libvpx/libvpx/build/make/
thumb.pm 19 # Write additions with shifts, such as "add r10, r11, lsl #8",
20 # in three operand form, "add r10, r10, r11, lsl #8".
21 s/(add\s+)(r\d+),\s*(r\d+),\s*(lsl #\d+)/$1$2, $2, $3, $4/g;
28 # This converts instructions such as "add r12, r12, r5, lsl r4"
29 # into the sequence "lsl r5, r4", "add r12, r12, r5", "lsr r5, r4".
30 s/^(\s*)(add)(\s+)(r\d+),\s*(r\d+),\s*(r\d+),\s*lsl (r\d+)/$1lsl$3$6, $7\n$1$2$3$4, $5, $6\n$1lsr$3$6, $7/g;
44 # This converts "ldrne r4, [src, -pstep, lsl #1]" into
45 # "subne src, src, pstep, lsl #1", "ldrne r4, [src]",
46 # "addne src, src, pstep, lsl #1". In a couple of cases wher
    [all...]
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/build/make/
thumb.pm 19 # Write additions with shifts, such as "add r10, r11, lsl #8",
20 # in three operand form, "add r10, r10, r11, lsl #8".
21 s/(add\s+)(r\d+),\s*(r\d+),\s*(lsl #\d+)/$1$2, $2, $3, $4/g;
28 # This converts instructions such as "add r12, r12, r5, lsl r4"
29 # into the sequence "lsl r5, r4", "add r12, r12, r5", "lsr r5, r4".
30 s/^(\s*)(add)(\s+)(r\d+),\s*(r\d+),\s*(r\d+),\s*lsl (r\d+)/$1lsl$3$6, $7\n$1$2$3$4, $5, $6\n$1lsr$3$6, $7/g;
44 # This converts "ldrne r4, [src, -pstep, lsl #1]" into
45 # "subne src, src, pstep, lsl #1", "ldrne r4, [src]",
46 # "addne src, src, pstep, lsl #1". In a couple of cases wher
    [all...]
  /external/chromium_org/third_party/boringssl/linux-arm/crypto/aes/
aes-armv4.S 180 orr r0,r0,r4,lsl#8
182 orr r0,r0,r5,lsl#16
184 orr r0,r0,r6,lsl#24
187 orr r1,r1,r4,lsl#8
189 orr r1,r1,r5,lsl#16
191 orr r1,r1,r6,lsl#24
194 orr r2,r2,r4,lsl#8
196 orr r2,r2,r5,lsl#16
198 orr r2,r2,r6,lsl#24
201 orr r3,r3,r4,lsl#
    [all...]
  /external/chromium_org/third_party/boringssl/src/crypto/aes/asm/
aes-armv4.pl 200 orr $s0,$s0,$t1,lsl#8
202 orr $s0,$s0,$t2,lsl#16
204 orr $s0,$s0,$t3,lsl#24
207 orr $s1,$s1,$t1,lsl#8
209 orr $s1,$s1,$t2,lsl#16
211 orr $s1,$s1,$t3,lsl#24
214 orr $s2,$s2,$t1,lsl#8
216 orr $s2,$s2,$t2,lsl#16
218 orr $s2,$s2,$t3,lsl#24
221 orr $s3,$s3,$t1,lsl#
    [all...]
  /external/openssl/crypto/aes/asm/
aes-armv4.S 178 orr r0,r0,r4,lsl#8
180 orr r0,r0,r5,lsl#16
182 orr r0,r0,r6,lsl#24
185 orr r1,r1,r4,lsl#8
187 orr r1,r1,r5,lsl#16
189 orr r1,r1,r6,lsl#24
192 orr r2,r2,r4,lsl#8
194 orr r2,r2,r5,lsl#16
196 orr r2,r2,r6,lsl#24
199 orr r3,r3,r4,lsl#
    [all...]
aes-armv4.pl 198 orr $s0,$s0,$t1,lsl#8
200 orr $s0,$s0,$t2,lsl#16
202 orr $s0,$s0,$t3,lsl#24
205 orr $s1,$s1,$t1,lsl#8
207 orr $s1,$s1,$t2,lsl#16
209 orr $s1,$s1,$t3,lsl#24
212 orr $s2,$s2,$t1,lsl#8
214 orr $s2,$s2,$t2,lsl#16
216 orr $s2,$s2,$t3,lsl#24
219 orr $s3,$s3,$t1,lsl#
    [all...]
  /bionic/libc/arch-arm/generic/bionic/
memcpy.S 82 movs r12, r3, lsl #31
111 movs r12, r3, lsl #28
179 movs r12, r2, lsl #28
184 movs r12, r2, lsl #30
220 mov r12, r5, lsl #3 /* r12 = right */
230 movs r5, r5, lsl #31
246 orr r4, r3, r5, lsl lr
274 orr r3, r3, r4, lsl #16
276 orr r4, r4, r5, lsl #16
278 orr r5, r5, r6, lsl #1
    [all...]
  /frameworks/av/media/libstagefright/codecs/amrwbenc/src/asm/ARMV5E/
cor_h_vec_opt.s 41 ADD r7, r4, r2, LSL #5 @r7 --- p0 = rrixix[track]
49 ADD r9, r1, r2, LSL #1 @p2 = &vec[pos]
63 MOV r6, r6, LSL #2 @L_sum2 = (L_sum2 << 2)
66 MOV r5, r5, LSL #2 @L_sum1 = (L_sum1 << 2)
71 ADD r9, r3, r2, LSL #1 @address of sign[pos]
83 ADD r9, r9, r4, LSL #1
84 ADD r12, r12, r4, LSL #1
94 ADD r9, r1, r2, LSL #1 @p2 = &vec[pos]
109 MOV r6, r6, LSL #2 @L_sum2 = (L_sum2 << 2)
112 MOV r5, r5, LSL #2 @L_sum1 = (L_sum1 << 2
    [all...]

Completed in 525 milliseconds

12 3 4 5 6 7 8 91011>>