/bionic/libc/arch-arm64/bionic/ |
_exit_with_stack_teardown.S | 33 mov w8, __NR_munmap 38 mov w8, __NR_exit
|
/external/valgrind/main/VEX/priv/ |
host_generic_simd128.c | 214 res->w8[ 0] = max8S(argL->w8[ 0], argR->w8[ 0]); 215 res->w8[ 1] = max8S(argL->w8[ 1], argR->w8[ 1]); 216 res->w8[ 2] = max8S(argL->w8[ 2], argR->w8[ 2]); 217 res->w8[ 3] = max8S(argL->w8[ 3], argR->w8[ 3]) [all...] |
/system/core/libpixelflinger/arch-arm64/ |
col32cb16blend.S | 65 ldrh w8, [x0] // load dest pixel 67 lsr w6, w8, #11 // extract dest red 68 and w7, w9, w8, lsr #5 // extract dest green 69 and w8, w8, #0x1f // extract dest blue 73 madd w8, w8, w5, w4 // dest blue * alpha + src blue 79 orr w6, w6, w8, lsr #8 // shift blue into 565
|
/external/llvm/test/MC/AArch64/ |
arm64-leaf-compact-unwind.s | 132 ldr w8, [x8] 149 sub w8, w8, w9 150 sub w8, w8, w7, lsl #1 151 sub w8, w8, w6, lsl #1 152 sub w8, w8, w5, lsl #1 153 sub w8, w8, w4, lsl # [all...] |
arm64-arithmetic-encoding.s | 346 cmp wsp, w8, uxtw 347 subs wzr, wsp, w8, uxtw 348 cmp sp, w8, uxtw 349 subs xzr, sp, w8, uxtw 353 ; CHECK: cmp wsp, w8 ; encoding: [0xff,0x43,0x28,0x6b] 354 ; CHECK: cmp wsp, w8 ; encoding: [0xff,0x43,0x28,0x6b] 355 ; CHECK: cmp sp, w8, uxtw ; encoding: [0xff,0x43,0x28,0xeb] 356 ; CHECK: cmp sp, w8, uxtw ; encoding: [0xff,0x43,0x28,0xeb] 358 sub wsp, w9, w8, uxtw 359 sub w1, wsp, w8, uxt [all...] |
tls-relocs.s | 31 movn w8, #:dtprel_g1:var 39 // CHECK: movn w8, #:dtprel_g1:var // encoding: [0bAAA01000,A,0b101AAAAA,0x12] 233 movn w8, #:tprel_g1:var 241 // CHECK: movn w8, #:tprel_g1:var // encoding: [0bAAA01000,A,0b101AAAAA,0x12]
|
/external/llvm/test/MC/Mips/msa/ |
test_elm_insert.s | 5 # CHECK: insert.w $w8[2], $15 # encoding: [0x79,0x32,0x7a,0x19] 9 insert.w $w8[2], $15
|
test_i10.s | 3 # CHECK: ldi.b $w8, 198 # encoding: [0x7b,0x06,0x32,0x07] 8 ldi.b $w8, 198
|
test_elm.s | 3 # CHECK: copy_s.b $13, $w8[2] # encoding: [0x78,0x82,0x43,0x59] 10 # CHECK: sldi.h $w8, $w17[0] # encoding: [0x78,0x20,0x8a,0x19] 19 copy_s.b $13, $w8[2] 26 sldi.h $w8, $w17[0]
|
test_3rf.s | 8 # CHECK: fceq.d $w0, $w8, $w16 # encoding: [0x78,0xb0,0x40,0x1a] 11 # CHECK: fclt.w $w28, $w8, $w8 # encoding: [0x79,0x08,0x47,0x1a] 22 # CHECK: fcult.d $w27, $w8, $w17 # encoding: [0x79,0x71,0x46,0xda] 23 # CHECK: fcun.w $w4, $w20, $w8 # encoding: [0x78,0x48,0xa1,0x1a] 29 # CHECK: fexdo.h $w8, $w0, $w16 # encoding: [0x7a,0x10,0x02,0x1b] 36 # CHECK: fmax.d $w26, $w18, $w8 # encoding: [0x7b,0xa8,0x96,0x9b] 44 # CHECK: fmsub.d $w8, $w18, $w16 # encoding: [0x79,0x70,0x92,0x1b] 75 # CHECK: maddr_q.h $w8, $w18, $w9 # encoding: [0x7b,0x49,0x92,0x1c] 91 fceq.d $w0, $w8, $w1 [all...] |
test_i8.s | 10 # CHECK: shf.h $w17, $w8, 76 # encoding: [0x79,0x4c,0x44,0x42] 21 shf.h $w17, $w8, 76
|
test_vec.s | 6 # CHECK: bsel.v $w8, $w0, $w14 # encoding: [0x78,0xce,0x02,0x1e] 14 bsel.v $w8, $w0, $w14
|
test_2r.s | 15 # CHECK: pcnt.h $w0, $w8 # encoding: [0x7b,0x05,0x40,0x1e] 31 pcnt.h $w0, $w8
|
test_3r.s | 81 # CHECK: clt_s.w $w23, $w8, $w16 # encoding: [0x79,0x50,0x45,0xcf] 105 # CHECK: dpadd_u.w $w24, $w8, $w16 # encoding: [0x79,0xd0,0x46,0x13] 148 # CHECK: max_a.d $w8, $w8, $w31 # encoding: [0x7b,0x7f,0x42,0x0e] 162 # CHECK: min_s.h $w27, $w21, $w8 # encoding: [0x7a,0x28,0xae,0xce] 164 # CHECK: min_s.d $w6, $w8, $w21 # encoding: [0x7a,0x75,0x41,0x8e] 165 # CHECK: min_u.b $w22, $w26, $w8 # encoding: [0x7a,0x88,0xd5,0x8e] 167 # CHECK: min_u.w $w8, $w20, $w14 # encoding: [0x7a,0xce,0xa2,0x0e] 174 # CHECK: mod_u.h $w24, $w8, $w7 # encoding: [0x7b,0xa7,0x46,0x12] 190 # CHECK: pckod.h $w26, $w5, $w8 # encoding: [0x79,0xa8,0x2e,0x94 [all...] |
test_mi10.s | 13 # CHECK: ld.w $w8, -2048($9) # encoding: [0x7a,0x00,0x4a,0x22] 40 ld.w $w8, -2048($9)
|
/external/libhevc/common/arm64/ |
ihevc_sao_edge_offset_class3.s | 89 LDR w8,[sp,#64] //Loads ht 247 csel w8,w20,w8,EQ 250 mov v1.8b[0], w8 //au1_mask = vsetq_lane_s8(-1, au1_mask, 0) 254 LDRB w8,[x5,#1] //pu1_avail[1] 255 mov v1.16b[15], w8 //au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15) 258 LDRB w8,[x5,#2] //pu1_avail[2] 284 LDRB w8,[x7] //load the value and increment by src_strd 287 STRB w8,[x5,#1]! //store it in the stack pointer 302 LDRB w8,[x8 [all...] |
ihevc_sao_edge_offset_class2_chroma.s | 100 mov w8, w25 //Loads ht 337 csel w8,w20,w8,EQ 341 mov v1.8b[0], w8 //au1_mask = vsetq_lane_s8(-1, au1_mask, 0) 344 mov v1.8b[1], w8 //au1_mask = vsetq_lane_s8(-1, au1_mask, 0) 347 LDRB w8,[x5,#1] //pu1_avail[1] 348 mov v1.16b[14], w8 //au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15) 349 mov v1.16b[15], w8 //au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15) 382 LDRH w8,[x7] //load the value and increment by src_strd 385 STRH w8,[x5],#2 //store it in the stack pointe [all...] |
ihevc_padding.s | 101 ldrb w8,[x0] 110 dup v0.16b,w8 219 ldrh w8,[x0] 228 dup v0.8h,w8 347 ldrb w8,[x0, #-1] 360 dup v0.16b,w8 465 ldrh w8,[x0, #-2] 474 dup v0.8h,w8
|
ihevc_sao_edge_offset_class2.s | 77 LDR w8,[sp,#16] //Loads ht 238 csel w8,w20,w8,EQ 242 mov v1.8b[0], w8 //au1_mask = vsetq_lane_s8((-1||pu1_avail[0]), au1_mask, 0) 245 LDRB w8,[x5,#1] //pu1_avail[1] 246 mov v1.16b[15], w8 //au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15) 277 LDRB w8,[x7] //load the value and increment by src_strd 279 STRB w8,[x5,#1]! //store it in the stack pointer 361 LDRB w8,[x11,#16] //III pu1_src_cpy[src_strd + 16] 370 mov v18.8b[0], w8 //III pu1_next_row_tmp = vsetq_lane_u8(pu1_src_cpy[src_strd + 16], pu1_next_row_tmp, (…) [all...] |
ihevc_sao_edge_offset_class3_chroma.s | 99 mov w8, w25 //Loads ht 327 csel w8,w20,w8,EQ 331 mov v1.8b[0], w8 //au1_mask = vsetq_lane_s8(-1, au1_mask, 0) 335 mov v1.8b[1], w8 //au1_mask = vsetq_lane_s8(-1, au1_mask, 0) 338 LDRB w8,[x5,#1] //pu1_avail[1] 339 mov v1.16b[14], w8 //au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15) 340 mov v1.16b[15], w8 //au1_mask = vsetq_lane_s8(pu1_avail[1], au1_mask, 15) 373 LDRH w8,[x7] //load the value and increment by src_strd 376 STRH w8,[x5],#2 //store it in the stack pointe [all...] |
ihevc_weighted_pred_bi.s | 150 ldr w8,[sp,#0] 158 sxtw x8,w8 189 mov v7.h[1],w8 //moved for scalar multiplication
|
/external/openssl/crypto/sha/asm/ |
sha256-armv8.S | 92 eor w8,w22,w22,ror#14 98 eor w16,w16,w8,ror#11 // Sigma1(e) 99 ror w8,w26,#2 106 eor w17,w8,w17,ror#13 // Sigma0(a) 113 ldp w7,w8,[x1],#2*4 160 rev w8,w8 // 5 169 add w22,w22,w8 // h+=X[i] 362 str w8,[sp,#4] 365 eor w8,w27,w27,ror#1 [all...] |
sha1-armv8.S | 99 add w24,w24,w8 // future e+=X[i] 255 eor w6,w6,w8 283 eor w8,w8,w10 287 eor w8,w8,w16 291 eor w8,w8,w5 295 ror w8,w8,#3 [all...] |
/external/valgrind/main/VEX/pub/ |
libvex_basictypes.h | 71 UChar w8[16]; member in union:__anon36864 81 UChar w8[32]; member in union:__anon36865
|
/external/openssl/crypto/aes/asm/ |
aesv8-armx-64.S | 561 ldr w8, [x4, #12] 580 rev w8, w8 583 add w8, w8, #1 585 rev w10, w8 616 add w8,w8,#1 619 rev w9,w8 622 add w8,w8,# [all...] |