HomeSort by relevance Sort by last modified time
    Searched refs:w8 (Results 1 - 25 of 70) sorted by null

1 2 3

  /bionic/libc/arch-arm64/bionic/
_exit_with_stack_teardown.S 33 mov w8, __NR_munmap
38 mov w8, __NR_exit
  /external/valgrind/main/VEX/priv/
host_generic_simd128.c 214 res->w8[ 0] = max8S(argL->w8[ 0], argR->w8[ 0]);
215 res->w8[ 1] = max8S(argL->w8[ 1], argR->w8[ 1]);
216 res->w8[ 2] = max8S(argL->w8[ 2], argR->w8[ 2]);
217 res->w8[ 3] = max8S(argL->w8[ 3], argR->w8[ 3])
    [all...]
  /system/core/libpixelflinger/arch-arm64/
col32cb16blend.S 65 ldrh w8, [x0] // load dest pixel
67 lsr w6, w8, #11 // extract dest red
68 and w7, w9, w8, lsr #5 // extract dest green
69 and w8, w8, #0x1f // extract dest blue
73 madd w8, w8, w5, w4 // dest blue * alpha + src blue
79 orr w6, w6, w8, lsr #8 // shift blue into 565
  /external/llvm/test/MC/AArch64/
arm64-leaf-compact-unwind.s 132 ldr w8, [x8]
149 sub w8, w8, w9
150 sub w8, w8, w7, lsl #1
151 sub w8, w8, w6, lsl #1
152 sub w8, w8, w5, lsl #1
153 sub w8, w8, w4, lsl #
    [all...]
arm64-arithmetic-encoding.s 346 cmp wsp, w8, uxtw
347 subs wzr, wsp, w8, uxtw
348 cmp sp, w8, uxtw
349 subs xzr, sp, w8, uxtw
353 ; CHECK: cmp wsp, w8 ; encoding: [0xff,0x43,0x28,0x6b]
354 ; CHECK: cmp wsp, w8 ; encoding: [0xff,0x43,0x28,0x6b]
355 ; CHECK: cmp sp, w8, uxtw ; encoding: [0xff,0x43,0x28,0xeb]
356 ; CHECK: cmp sp, w8, uxtw ; encoding: [0xff,0x43,0x28,0xeb]
358 sub wsp, w9, w8, uxtw
359 sub w1, wsp, w8, uxt
    [all...]
tls-relocs.s 31 movn w8, #:dtprel_g1:var
39 // CHECK: movn w8, #:dtprel_g1:var // encoding: [0bAAA01000,A,0b101AAAAA,0x12]
233 movn w8, #:tprel_g1:var
241 // CHECK: movn w8, #:tprel_g1:var // encoding: [0bAAA01000,A,0b101AAAAA,0x12]
  /external/llvm/test/MC/Mips/msa/
test_elm_insert.s 5 # CHECK: insert.w $w8[2], $15 # encoding: [0x79,0x32,0x7a,0x19]
9 insert.w $w8[2], $15
test_i10.s 3 # CHECK: ldi.b $w8, 198 # encoding: [0x7b,0x06,0x32,0x07]
8 ldi.b $w8, 198
test_elm.s 3 # CHECK: copy_s.b $13, $w8[2] # encoding: [0x78,0x82,0x43,0x59]
10 # CHECK: sldi.h $w8, $w17[0] # encoding: [0x78,0x20,0x8a,0x19]
19 copy_s.b $13, $w8[2]
26 sldi.h $w8, $w17[0]
test_3rf.s 8 # CHECK: fceq.d $w0, $w8, $w16 # encoding: [0x78,0xb0,0x40,0x1a]
11 # CHECK: fclt.w $w28, $w8, $w8 # encoding: [0x79,0x08,0x47,0x1a]
22 # CHECK: fcult.d $w27, $w8, $w17 # encoding: [0x79,0x71,0x46,0xda]
23 # CHECK: fcun.w $w4, $w20, $w8 # encoding: [0x78,0x48,0xa1,0x1a]
29 # CHECK: fexdo.h $w8, $w0, $w16 # encoding: [0x7a,0x10,0x02,0x1b]
36 # CHECK: fmax.d $w26, $w18, $w8 # encoding: [0x7b,0xa8,0x96,0x9b]
44 # CHECK: fmsub.d $w8, $w18, $w16 # encoding: [0x79,0x70,0x92,0x1b]
75 # CHECK: maddr_q.h $w8, $w18, $w9 # encoding: [0x7b,0x49,0x92,0x1c]
91 fceq.d $w0, $w8, $w1
    [all...]
test_i8.s 10 # CHECK: shf.h $w17, $w8, 76 # encoding: [0x79,0x4c,0x44,0x42]
21 shf.h $w17, $w8, 76
test_vec.s 6 # CHECK: bsel.v $w8, $w0, $w14 # encoding: [0x78,0xce,0x02,0x1e]
14 bsel.v $w8, $w0, $w14
test_2r.s 15 # CHECK: pcnt.h $w0, $w8 # encoding: [0x7b,0x05,0x40,0x1e]
31 pcnt.h $w0, $w8
test_3r.s 81 # CHECK: clt_s.w $w23, $w8, $w16 # encoding: [0x79,0x50,0x45,0xcf]
105 # CHECK: dpadd_u.w $w24, $w8, $w16 # encoding: [0x79,0xd0,0x46,0x13]
148 # CHECK: max_a.d $w8, $w8, $w31 # encoding: [0x7b,0x7f,0x42,0x0e]
162 # CHECK: min_s.h $w27, $w21, $w8 # encoding: [0x7a,0x28,0xae,0xce]
164 # CHECK: min_s.d $w6, $w8, $w21 # encoding: [0x7a,0x75,0x41,0x8e]
165 # CHECK: min_u.b $w22, $w26, $w8 # encoding: [0x7a,0x88,0xd5,0x8e]
167 # CHECK: min_u.w $w8, $w20, $w14 # encoding: [0x7a,0xce,0xa2,0x0e]
174 # CHECK: mod_u.h $w24, $w8, $w7 # encoding: [0x7b,0xa7,0x46,0x12]
190 # CHECK: pckod.h $w26, $w5, $w8 # encoding: [0x79,0xa8,0x2e,0x94
    [all...]
test_mi10.s 13 # CHECK: ld.w $w8, -2048($9) # encoding: [0x7a,0x00,0x4a,0x22]
40 ld.w $w8, -2048($9)
  /external/libhevc/common/arm64/
ihevc_sao_edge_offset_class3.s 89 LDR w8,[sp,#64] //Loads ht
247 csel w8,w20,w8,EQ
250 mov v1.8b[0], w8 //au1_mask = vsetq_lane_s8(-1, au1_mask, 0)
254 LDRB w8,[x5,#1] //pu1_avail[1]
255 mov v1.16b[15], w8 //au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15)
258 LDRB w8,[x5,#2] //pu1_avail[2]
284 LDRB w8,[x7] //load the value and increment by src_strd
287 STRB w8,[x5,#1]! //store it in the stack pointer
302 LDRB w8,[x8
    [all...]
ihevc_sao_edge_offset_class2_chroma.s 100 mov w8, w25 //Loads ht
337 csel w8,w20,w8,EQ
341 mov v1.8b[0], w8 //au1_mask = vsetq_lane_s8(-1, au1_mask, 0)
344 mov v1.8b[1], w8 //au1_mask = vsetq_lane_s8(-1, au1_mask, 0)
347 LDRB w8,[x5,#1] //pu1_avail[1]
348 mov v1.16b[14], w8 //au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15)
349 mov v1.16b[15], w8 //au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15)
382 LDRH w8,[x7] //load the value and increment by src_strd
385 STRH w8,[x5],#2 //store it in the stack pointe
    [all...]
ihevc_padding.s 101 ldrb w8,[x0]
110 dup v0.16b,w8
219 ldrh w8,[x0]
228 dup v0.8h,w8
347 ldrb w8,[x0, #-1]
360 dup v0.16b,w8
465 ldrh w8,[x0, #-2]
474 dup v0.8h,w8
ihevc_sao_edge_offset_class2.s 77 LDR w8,[sp,#16] //Loads ht
238 csel w8,w20,w8,EQ
242 mov v1.8b[0], w8 //au1_mask = vsetq_lane_s8((-1||pu1_avail[0]), au1_mask, 0)
245 LDRB w8,[x5,#1] //pu1_avail[1]
246 mov v1.16b[15], w8 //au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15)
277 LDRB w8,[x7] //load the value and increment by src_strd
279 STRB w8,[x5,#1]! //store it in the stack pointer
361 LDRB w8,[x11,#16] //III pu1_src_cpy[src_strd + 16]
370 mov v18.8b[0], w8 //III pu1_next_row_tmp = vsetq_lane_u8(pu1_src_cpy[src_strd + 16], pu1_next_row_tmp, (…)
    [all...]
ihevc_sao_edge_offset_class3_chroma.s 99 mov w8, w25 //Loads ht
327 csel w8,w20,w8,EQ
331 mov v1.8b[0], w8 //au1_mask = vsetq_lane_s8(-1, au1_mask, 0)
335 mov v1.8b[1], w8 //au1_mask = vsetq_lane_s8(-1, au1_mask, 0)
338 LDRB w8,[x5,#1] //pu1_avail[1]
339 mov v1.16b[14], w8 //au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15)
340 mov v1.16b[15], w8 //au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15)
373 LDRH w8,[x7] //load the value and increment by src_strd
376 STRH w8,[x5],#2 //store it in the stack pointe
    [all...]
ihevc_weighted_pred_bi.s 150 ldr w8,[sp,#0]
158 sxtw x8,w8
189 mov v7.h[1],w8 //moved for scalar multiplication
  /external/openssl/crypto/sha/asm/
sha256-armv8.S 92 eor w8,w22,w22,ror#14
98 eor w16,w16,w8,ror#11 // Sigma1(e)
99 ror w8,w26,#2
106 eor w17,w8,w17,ror#13 // Sigma0(a)
113 ldp w7,w8,[x1],#2*4
160 rev w8,w8 // 5
169 add w22,w22,w8 // h+=X[i]
362 str w8,[sp,#4]
365 eor w8,w27,w27,ror#1
    [all...]
sha1-armv8.S 99 add w24,w24,w8 // future e+=X[i]
255 eor w6,w6,w8
283 eor w8,w8,w10
287 eor w8,w8,w16
291 eor w8,w8,w5
295 ror w8,w8,#3
    [all...]
  /external/valgrind/main/VEX/pub/
libvex_basictypes.h 71 UChar w8[16]; member in union:__anon35977
81 UChar w8[32]; member in union:__anon35978
  /external/openssl/crypto/aes/asm/
aesv8-armx-64.S 561 ldr w8, [x4, #12]
580 rev w8, w8
583 add w8, w8, #1
585 rev w10, w8
616 add w8,w8,#1
619 rev w9,w8
622 add w8,w8,#
    [all...]

Completed in 499 milliseconds

1 2 3