/external/vixl/examples/aarch64/ |
neon-matrix-multiply.cc | 78 __ Movi(v0.V16B(), 0); 79 __ Movi(v1.V16B(), 0); 80 __ Movi(v2.V16B(), 0); 81 __ Movi(v3.V16B(), 0);
|
/external/vixl/test/aarch64/ |
test-assembler-aarch64.cc | [all...] |
test-disasm-aarch64.cc | [all...] |
test-simulator-aarch64.cc | [all...] |
/art/compiler/optimizing/ |
code_generator_vector_arm64.cc | 91 __ Movi(dst.V16B(), Int64FromLocation(src_loc)); 100 __ Movi(dst.V8H(), Int64FromLocation(src_loc)); 108 __ Movi(dst.V4S(), Int64FromLocation(src_loc)); 116 __ Movi(dst.V2D(), Int64FromLocation(src_loc)); 367 __ Movi(dst.V16B(), 1); [all...] |
/external/vixl/src/aarch64/ |
macro-assembler-aarch64.cc | 952 movi(vd.Is64Bits() ? vd.V8B() : vd.V16B(), byte1); 954 movi(vd, byte2, LSL, 8); 956 movi(vd, byte1); 987 movi(vd.Is64Bits() ? vd.V1D() : vd.V2D(), ((imm << 32) | imm)); 995 movi(vd, bytes[i], LSL, i * 8); 1011 movi(vd, bytes[2], MSL, 16); 1017 movi(vd, bytes[1], MSL, 8); [all...] |
macro-assembler-aarch64.h | [all...] |
/external/v8/src/arm64/ |
macro-assembler-arm64-inl.h | 651 Movi(vd, bits); 678 Movi(vd, bits); [all...] |
macro-assembler-arm64.h | [all...] |
macro-assembler-arm64.cc | 368 movi(vd.Is64Bits() ? vd.V8B() : vd.V16B(), byte1); 370 movi(vd, byte2, LSL, 8); 372 movi(vd, byte1); 402 movi(vd.Is64Bits() ? vd.V1D() : vd.V2D(), ((imm << 32) | imm)); 410 movi(vd, bytes[i], LSL, i * 8); 426 movi(vd, bytes[2], MSL, 16); 432 movi(vd, bytes[1], MSL, 8); 474 movi(vd, imm); 498 void TurboAssembler::Movi(const VRegister& vd, uint64_t imm, Shift shift, 502 movi(vd, imm, shift, shift_amount) [all...] |
/external/v8/src/compiler/arm64/ |
code-generator-arm64.cc | [all...] |