/external/llvm/test/CodeGen/AArch64/ |
arm64-neon-add-pairwise.ll | 3 declare <8 x i8> @llvm.aarch64.neon.addp.v8i8(<8 x i8>, <8 x i8>) 8 %tmp1 = call <8 x i8> @llvm.aarch64.neon.addp.v8i8(<8 x i8> %lhs, <8 x i8> %rhs) 9 ; CHECK: addp v0.8b, v0.8b, v1.8b 13 declare <16 x i8> @llvm.aarch64.neon.addp.v16i8(<16 x i8>, <16 x i8>) 17 %tmp1 = call <16 x i8> @llvm.aarch64.neon.addp.v16i8(<16 x i8> %lhs, <16 x i8> %rhs) 18 ; CHECK: addp v0.16b, v0.16b, v1.16b 22 declare <4 x i16> @llvm.aarch64.neon.addp.v4i16(<4 x i16>, <4 x i16>) 26 %tmp1 = call <4 x i16> @llvm.aarch64.neon.addp.v4i16(<4 x i16> %lhs, <4 x i16> %rhs) 27 ; CHECK: addp v0.4h, v0.4h, v1.4h 31 declare <8 x i16> @llvm.aarch64.neon.addp.v8i16(<8 x i16>, <8 x i16> [all...] |
arm64-addp.ll | 15 ; CHECK: addp.2d d0, v0
|
arm64-vadd.ll | 641 ;CHECK: addp.8b 644 %tmp3 = call <8 x i8> @llvm.aarch64.neon.addp.v8i8(<8 x i8> %tmp1, <8 x i8> %tmp2) 650 ;CHECK: addp.16b 653 %tmp3 = call <16 x i8> @llvm.aarch64.neon.addp.v16i8(<16 x i8> %tmp1, <16 x i8> %tmp2) 659 ;CHECK: addp.4h 662 %tmp3 = call <4 x i16> @llvm.aarch64.neon.addp.v4i16(<4 x i16> %tmp1, <4 x i16> %tmp2) 668 ;CHECK: addp.8h 671 %tmp3 = call <8 x i16> @llvm.aarch64.neon.addp.v8i16(<8 x i16> %tmp1, <8 x i16> %tmp2) 677 ;CHECK: addp.2s 680 %tmp3 = call <2 x i32> @llvm.aarch64.neon.addp.v2i32(<2 x i32> %tmp1, <2 x i32> %tmp2 [all...] |
arm64-vaddv.ll | 52 ; CHECK: addp.2s v[[REGNUM:[0-9]+]], v0, v0 62 ; CHECK: addp.2s v[[REGNUM:[0-9]+]], v1, v1 73 ; CHECK: addp.2d [[REGNUM:d[0-9]+]], v0 83 ; CHECK: addp.2d d[[REGNUM:[0-9]+]], v1 163 ; CHECK: addp.2s v[[REGNUM:[0-9]+]], v0, v0 173 ; CHECK: addp.2s v[[REGNUM:[0-9]+]], v1, v1 212 ; CHECK: addp.2d [[REGNUM:d[0-9]+]], v0 222 ; CHECK: addp.2d d[[REGNUM:[0-9]+]], v1 233 ; CHECK: addp.2d d0, v0
|
/external/llvm/test/MC/AArch64/ |
neon-add-pairwise.s | 9 addp v0.8b, v1.8b, v2.8b 10 addp v0.16b, v1.16b, v2.16b 11 addp v0.4h, v1.4h, v2.4h 12 addp v0.8h, v1.8h, v2.8h 13 addp v0.2s, v1.2s, v2.2s 14 addp v0.4s, v1.4s, v2.4s 15 addp v0.2d, v1.2d, v2.2d 17 // CHECK: addp v0.8b, v1.8b, v2.8b // encoding: [0x20,0xbc,0x22,0x0e] 18 // CHECK: addp v0.16b, v1.16b, v2.16b // encoding: [0x20,0xbc,0x22,0x4e] 19 // CHECK: addp v0.4h, v1.4h, v2.4h // encoding: [0x20,0xbc,0x62,0x0e [all...] |
neon-scalar-reduce-pairwise.s | 6 addp d0, v1.2d 8 // CHECK: addp d0, v1.2d // encoding: [0x20,0xb8,0xf1,0x5e]
|
arm64-advsimd.s | 53 addp.8b v0, v0, v0 54 addp.16b v0, v0, v0 55 addp.4h v0, v0, v0 56 addp.8h v0, v0, v0 57 addp.2s v0, v0, v0 58 addp.4s v0, v0, v0 59 addp.2d v0, v0, v0 61 ; CHECK: addp.8b v0, v0, v0 ; encoding: [0x00,0xbc,0x20,0x0e] 62 ; CHECK: addp.16b v0, v0, v0 ; encoding: [0x00,0xbc,0x20,0x4e] 63 ; CHECK: addp.4h v0, v0, v0 ; encoding: [0x00,0xbc,0x60,0x0e [all...] |
/external/clang/test/SemaCXX/ |
borland-extensions.cpp | 20 int __pascal addP(); 31 i = h2<int>(&M::addP);
|
MicrosoftExtensions.cpp | 76 int __stdcall addP(); 88 h1<int>(&M::addP);
|
/external/libavc/common/armv8/ |
ih264_intra_pred_chroma_av8.s | 131 addp v0.4s, v0.4s , v0.4s 132 addp v1.4s, v1.4s , v1.4s 133 addp v0.4s, v0.4s , v0.4s 134 addp v1.4s, v1.4s , v1.4s 137 addp v2.4s, v2.4s , v2.4s 138 addp v3.4s, v3.4s , v3.4s 139 addp v2.4s, v2.4s , v2.4s 140 addp v3.4s, v3.4s , v3.4s 158 addp v0.4s, v0.4s , v0.4s 159 addp v1.4s, v1.4s , v1.4 [all...] |
ih264_resi_trans_quant_av8.s | 242 addp v0.8b, v0.8b, v0.8b //i pair add nnz 1 243 addp v0.8b, v0.8b, v0.8b //i pair add nnz 1 244 addp v0.8b, v0.8b, v0.8b //i pair add nnz 1 465 addp v0.8b, v0.8b, v0.8b //i pair add nnz 1 466 addp v0.8b, v0.8b, v0.8b //i pair add nnz 1 467 addp v0.8b, v0.8b, v0.8b //i pair add nnz 1 616 addp v2.8b, v2.8b, v2.8b 617 addp v2.8b, v2.8b, v2.8b 618 addp v2.8b, v2.8b, v2.8b 717 addp v5.8b, v5.8b, v5.8b //sum up nn [all...] |
ih264_intra_pred_luma_16x16_av8.s | 331 addp v0.4h, v0.4h , v0.4h 332 addp v0.4h, v0.4h , v0.4h 445 addp v0.8h, v0.8h, v1.8h
|
/bionic/libc/arch-arm64/generic/bionic/ |
memchr.S | 102 addp vend.16b, vhas_chr1.16b, vhas_chr2.16b /* 256->128 */ 103 addp vend.16b, vend.16b, vend.16b /* 128->64 */ 123 addp vend.2d, vend.2d, vend.2d 132 addp vend.16b, vhas_chr1.16b, vhas_chr2.16b /* 256->128 */ 133 addp vend.16b, vend.16b, vend.16b /* 128->64 */
|
strchr.S | 107 addp vend1.16b, vend1.16b, vend2.16b // 256->128 109 addp vend1.16b, vend1.16b, vend2.16b // 128->64 126 addp vend1.2d, vend1.2d, vend1.2d 138 addp vend1.16b, vend1.16b, vend2.16b // 256->128 139 addp vend1.16b, vend1.16b, vend2.16b // 128->64
|
/external/libavc/encoder/armv8/ |
ih264e_evaluate_intra_chroma_modes_av8.s | 117 addp v0.4s, v0.4s , v0.4s 118 addp v1.4s, v1.4s , v1.4s 119 addp v0.4s, v0.4s , v0.4s 120 addp v1.4s, v1.4s , v1.4s 123 addp v2.4s, v2.4s , v2.4s 124 addp v3.4s, v3.4s , v3.4s 125 addp v2.4s, v2.4s , v2.4s 126 addp v3.4s, v3.4s , v3.4s 148 addp v0.4s, v0.4s , v0.4s 149 addp v1.4s, v1.4s , v1.4 [all...] |
ime_distortion_metrics_av8.s | 132 addp v30.8h, v30.8h, v30.8h 134 addp v30.2s, v30.2s, v30.2s 213 addp v30.8h, v30.8h, v30.8h 215 addp v30.2s, v30.2s, v30.2s 307 addp v31.8h, v30.8h, v30.8h 309 addp v31.2s, v31.2s, v31.2s 356 addp v31.8h, v30.8h, v30.8h 358 addp v31.2s, v31.2s, v31.2s 434 addp v30.8h, v30.8h, v31.8h 436 addp v30.2s, v30.2s, v30.2 [all...] |
ih264e_evaluate_intra16x16_modes_av8.s | 137 addp v0.4h, v0.4h , v0.4h 138 addp v0.4h, v0.4h , v0.4h 445 addp v16.2s, v16.2s, v16.2s 455 addp v26.2s, v26.2s, v26.2s 464 addp v24.2s, v24.2s, v24.2s ///DC
|
/external/v8/src/x64/ |
deoptimizer-x64.cc | 213 __ addp(rsp, Immediate(1 * kRegisterSize + kPCOnStackSize)); 218 __ addp(rcx, rsp); 229 __ addp(rdx, Immediate(sizeof(intptr_t))); 266 __ addp(rax, Immediate(kPointerSize));
|
regexp-macro-assembler-x64.cc | 146 __ addp(register_location(reg), Immediate(by)); 155 __ addp(rbx, code_object_pointer()); 255 __ addp(rbx, r9); // End of capture 288 __ addp(r11, Immediate(1)); 289 __ addp(r9, Immediate(1)); 393 __ addp(rdx, rsi); // Start of capture. 413 __ addp(rbx, Immediate(char_size())); 414 __ addp(rdx, Immediate(char_size())); 808 __ addp(rcx, rdx); 816 __ addp(rax, rcx); // Convert to index from start, not end [all...] |
macro-assembler-x64.cc | 224 addp(scratch, Immediate(kPointerSize)); 280 addp(scratch, kScratchRegister); 951 addp(rsp, Immediate(kDoubleSize * XMMRegister::kMaxNumRegisters)); [all...] |
code-stubs-x64.cc | 220 __ addp(rsp, Immediate(kDoubleSize)); 434 __ addp(rsp, Immediate(kDoubleSize)); 439 __ addp(rsp, Immediate(kDoubleSize)); 655 __ addp(r8, Immediate(Heap::kSloppyArgumentsObjectSize)); 741 __ addp(r8, args.GetArgumentOperand(2)); 794 __ addp(r8, Immediate(1)); [all...] |
debug-x64.cc | 140 __ addp(rsp, Immediate(kPCOnStackSize));
|
/external/v8/src/ic/x64/ |
stub-cache-x64.cc | 78 __ addp(kScratchRegister, Immediate(Code::kHeaderSize - kHeapObjectTag));
|
/external/valgrind/docs/internals/ |
MERGE_3_10_1.txt | 181 //2993 arm64: implement ADDP etc
|
/external/skia/src/core/ |
SkBlitter.cpp | 644 const uint8_t* addp = mulp + size; 657 unsigned add = addp[i]; 677 unsigned add = addp[i];
|