/system/core/libpixelflinger/arch-arm64/ |
t32cb16blend.S | 52 * clobbered: w6, w7, w16, w17, w18 69 lsr w16, \DREG, #(16 + 11) 70 mul w16, w7, w16 73 add w16, w6, w16, lsr #8 74 cmp w16, #0x1F 76 orr w18, \FB, w16, lsl #(16 + 11) 82 lsr w16, \SRC, #(8+2) 83 and w16, w16, #0x3 [all...] |
/external/valgrind/main/VEX/priv/ |
host_generic_simd128.c | 186 res->w16[0] = max16U(argL->w16[0], argR->w16[0]); 187 res->w16[1] = max16U(argL->w16[1], argR->w16[1]); 188 res->w16[2] = max16U(argL->w16[2], argR->w16[2]); 189 res->w16[3] = max16U(argL->w16[3], argR->w16[3]) [all...] |
/external/openssl/crypto/sha/asm/ |
sha256-armv8.S | 12 ldr w16,[x16] 13 tst w16,#ARMV8_SHA256 41 ror w16,w24,#6 49 eor w16,w16,w6,ror#11 // Sigma1(e) 53 add w27,w27,w16 // h+=Sigma1(e) 66 ror w16,w23,#6 74 eor w16,w16,w7,ror#11 // Sigma1(e) 78 add w26,w26,w16 // h+=Sigma1(e [all...] |
sha1-armv8.S | 12 ldr w16,[x16] 13 tst w16,#ARMV8_SHA1 199 add w21,w21,w16 // future e+=X[i] 224 eor w3,w3,w16 287 eor w8,w8,w16 356 eor w14,w14,w16 380 eor w16,w16,w19 384 eor w16,w16,w [all...] |
/external/llvm/test/MC/AArch64/ |
arm64-basic-a64-instructions.s | 10 crc32cx w18, w16, xzr 18 // CHECK: crc32cx w18, w16, xzr // encoding: [0x12,0x5e,0xdf,0x9a]
|
arm64-tls-relocs.s | 96 movk w16, #:tprel_g0_nc:var 99 // CHECK: movk w16, #:tprel_g0_nc:var // encoding: [0bAAA10000,A,0b100AAAAA,0x72] 220 movk w16, #:dtprel_g0_nc:var 223 // CHECK: movk w16, #:dtprel_g0_nc:var // encoding: [0bAAA10000,A,0b100AAAAA,0x72]
|
tls-relocs.s | 81 movk w16, #:dtprel_g0_nc:var 85 // CHECK: movk w16, #:dtprel_g0_nc:var // encoding: [0bAAA10000,A,0b100AAAAA,0x72] 283 movk w16, #:tprel_g0_nc:var 287 // CHECK: movk w16, #:tprel_g0_nc:var // encoding: [0bAAA10000,A,0b100AAAAA,0x72]
|
arm64-leaf-compact-unwind.s | 122 ldr w16, [x8] 139 add w9, w9, w16
|
neon-simd-copy.s | 115 dup v11.8h, w16 123 // CHECK: {{mov|dup}} v11.8h, w16 // encoding: [0x0b,0x0e,0x02,0x4e]
|
/external/llvm/test/MC/Mips/msa/ |
test_3rf.s | 7 # CHECK: fceq.w $w1, $w23, $w16 # encoding: [0x78,0x90,0xb8,0x5a] 8 # CHECK: fceq.d $w0, $w8, $w16 # encoding: [0x78,0xb0,0x40,0x1a] 9 # CHECK: fcle.w $w16, $w9, $w24 # encoding: [0x79,0x98,0x4c,0x1a] 26 # CHECK: fcune.d $w16, $w26, $w21 # encoding: [0x78,0xb5,0xd4,0x1c] 29 # CHECK: fexdo.h $w8, $w0, $w16 # encoding: [0x7a,0x10,0x02,0x1b] 37 # CHECK: fmax_a.w $w10, $w16, $w10 # encoding: [0x7b,0xca,0x82,0x9b] 44 # CHECK: fmsub.d $w8, $w18, $w16 # encoding: [0x79,0x70,0x92,0x1b] 54 # CHECK: fslt.d $w16, $w26, $w21 # encoding: [0x7b,0x35,0xd4,0x1a] 61 # CHECK: fsueq.w $w16, $w24, $w25 # encoding: [0x7a,0xd9,0xc4,0x1a] 69 # CHECK: fsune.w $w16, $w31, $w2 # encoding: [0x7a,0x82,0xfc,0x1c [all...] |
test_3r.s | 13 # CHECK: adds_s.w $w16, $w14, $w13 # encoding: [0x79,0x4d,0x74,0x10] 23 # CHECK: asub_s.b $w23, $w16, $w3 # encoding: [0x7a,0x03,0x85,0xd1] 32 # CHECK: ave_s.h $w16, $w19, $w9 # encoding: [0x7a,0x29,0x9c,0x10] 35 # CHECK: ave_u.b $w16, $w19, $w9 # encoding: [0x7a,0x89,0x9c,0x10] 39 # CHECK: aver_s.b $w26, $w16, $w2 # encoding: [0x7b,0x02,0x86,0x90] 48 # CHECK: bclr.h $w16, $w21, $w28 # encoding: [0x79,0xbc,0xac,0x0d] 51 # CHECK: binsl.b $w5, $w16, $w24 # encoding: [0x7b,0x18,0x81,0x4d] 60 # CHECK: bneg.h $w28, $w16, $w4 # encoding: [0x7a,0xa4,0x87,0x0d] 81 # CHECK: clt_s.w $w23, $w8, $w16 # encoding: [0x79,0x50,0x45,0xcf] 84 # CHECK: clt_u.h $w16, $w31, $w23 # encoding: [0x79,0xb7,0xfc,0x0f [all...] |
test_i8.s | 12 # CHECK: xori.b $w16, $w10, 20 # encoding: [0x7b,0x14,0x54,0x00] 23 xori.b $w16, $w10, 20
|
test_i5.s | 6 # CHECK: addvi.d $w16, $w1, 21 # encoding: [0x78,0x75,0x0c,0x06] 11 # CHECK: clei_s.b $w12, $w16, 1 # encoding: [0x7a,0x01,0x83,0x07] 46 # CHECK: subvi.d $w19, $w16, 7 # encoding: [0x78,0xe7,0x84,0xc6] 51 addvi.d $w16, $w1, 21 56 clei_s.b $w12, $w16, 1 91 subvi.d $w19, $w16, 7
|
test_2r.s | 5 # CHECK: fill.w $w16, $24 # encoding: [0x7b,0x02,0xc4,0x1e] 21 fill.w $w16, $24
|
test_mi10.s | 22 # CHECK: ld.d $w16, -1024($17) # encoding: [0x7b,0x80,0x8c,0x23] 49 ld.d $w16, -1024($17)
|
test_2rf.s | 12 # CHECK: ffint_u.d $w19, $w16 # encoding: [0x7b,0x3f,0x84,0xde] 45 ffint_u.d $w19, $w16
|
test_bit.s | 30 # CHECK: sat_u.d $w29, $w16, 5 # encoding: [0x78,0x85,0x87,0x4a] 79 sat_u.d $w29, $w16, 5
|
/art/runtime/arch/arm64/ |
asm_support_arm64.S | 40 #define wIP0 w16
|
/external/valgrind/main/VEX/pub/ |
libvex_basictypes.h | 72 UShort w16[8]; member in union:__anon36864 82 UShort w16[16]; member in union:__anon36865
|
/external/libvpx/libvpx/vp9/common/x86/ |
vp9_copy_sse2.asm | 25 je .w16 74 .w16:
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/x86/ |
vp9_copy_sse2.asm | 25 je .w16 74 .w16:
|
/external/chromium_org/third_party/libvpx/source/libvpx/vp9/common/x86/ |
vp9_copy_sse2.asm | 25 je .w16 74 .w16:
|
/external/vixl/test/ |
test-disasm-a64.cc | 343 COMPARE(adds(w15, w16, Operand(w17, SXTB, 4)), "adds w15, w16, w17, sxtb #4"); 369 COMPARE(subs(w15, w16, Operand(w17, SXTB, 4)), "subs w15, w16, w17, sxtb #4"); 435 COMPARE(madd(w30, w21, w22, w16), "madd w30, w21, w22, w16"); 452 COMPARE(msub(w30, w21, w22, w16), "msub w30, w21, w22, w16"); 614 COMPARE(eor(w15, w16, Operand(0x00000001)), 615 "eor w15, w16, #0x1") [all...] |
/external/chromium_org/v8/test/cctest/ |
test-disasm-arm64.cc | 382 COMPARE(adds(w15, w16, Operand(w17, SXTB, 4)), "adds w15, w16, w17, sxtb #4"); 408 COMPARE(subs(w15, w16, Operand(w17, SXTB, 4)), "subs w15, w16, w17, sxtb #4"); 465 COMPARE(msub(w14, w15, w16, w17), "msub w14, w15, w16, w17"); 645 COMPARE(eor(w15, w16, Operand(0x00000001)), 646 "eor w15, w16, #0x1"); 706 COMPARE(bic(w15, w16, Operand(w17)), "bic w15, w16, w17") [all...] |
test-assembler-arm64.cc | 1101 __ Mul(w0, w16, w16); 1102 __ Mul(w1, w16, w17); 1112 __ Mneg(w11, w16, w16); 1113 __ Mneg(w12, w16, w17); 1182 __ Madd(w0, w16, w16, w16); 1183 __ Madd(w1, w16, w16, w17) [all...] |