/external/clang/test/CodeGen/ |
arm64_vtst.c | 6 uint64x2_t test_vtstq_s64(int64x2_t a1, int64x2_t a2) { 15 uint64x2_t test_vtstq_u64(uint64x2_t a1, uint64x2_t a2) {
|
arm64_vCMP.c | 49 uint64x2_t test_vceqq_u64(uint64x2_t a1, uint64x2_t a2) { 55 uint64x2_t test_vcgeq_s64(int64x2_t a1, int64x2_t a2) { 61 uint64x2_t test_vcgeq_u64(uint64x2_t a1, uint64x2_t a2) { 67 uint64x2_t test_vcgtq_s64(int64x2_t a1, int64x2_t a2) { 73 uint64x2_t test_vcgtq_u64(uint64x2_t a1, uint64x2_t a2) [all...] |
arm-neon-misc.c | 16 uint64x2_t q = vld1q_u64(src); 24 void t2(uint64_t *src1, uint8_t *src2, uint64x2_t *dst) { 26 uint64x2_t q = vld1q_u64(src1);
|
arm64_vca.c | 33 uint64x2_t test_vcagtq_f64(float64x2_t a1, float64x2_t a2) { 40 uint64x2_t test_vcaltq_f64(float64x2_t a1, float64x2_t a2) { 47 uint64x2_t test_vcageq_f64(float64x2_t a1, float64x2_t a2) { 54 uint64x2_t test_vcaleq_f64(float64x2_t a1, float64x2_t a2) {
|
debug-info.c | 57 // CHECK: uint64x2_t 59 typedef uint64_t uint64x2_t __attribute__((ext_vector_type(2))); typedef 60 uint64x2_t extvectbar[4];
|
arm64_vcopy.c | 52 uint64x2_t test_vcopyq_laneq_u64(uint64x2_t a1, uint64x2_t a2) {
|
aarch64-neon-3v.c | 98 uint64x2_t test_vandq_u64(uint64x2_t a, uint64x2_t b) { 194 uint64x2_t test_vorrq_u64(uint64x2_t a, uint64x2_t b) { 290 uint64x2_t test_veorq_u64(uint64x2_t a, uint64x2_t b) { 386 uint64x2_t test_vbicq_u64(uint64x2_t a, uint64x2_t b) [all...] |
arm64_vqmov.c | 48 uint32x4_t test_vqmovn_high_u64(uint32x2_t Vdlow, uint64x2_t Vn) 72 uint32x4_t test_vqmovun_high_s64(uint32x2_t Vdlow, uint64x2_t Vn)
|
arm64_vshift.c | 95 uint64x2_t test_vqshlq_n_u64(uint64x2_t in) { 192 uint64x2_t test_vrshrq_n_u64(uint64x2_t in) { 352 uint64x2_t test_vrsraq_n_u64(uint64x2_t acc, uint64x2_t in) {
|
aarch64-neon-extract.c | 97 uint64x2_t test_vextq_u64(uint64x2_t a, uint64x2_t b) {
|
arm64_vsli.c | 128 uint64x2_t test_vsliq_n_u64(uint64x2_t a1, uint64x2_t a2) {
|
arm64_vsri.c | 129 uint64x2_t test_vsriq_n_u64(uint64x2_t a1, uint64x2_t a2) {
|
arm64_neon_high_half.c | 31 uint64x2_t test_vaddw_high_u32(uint64x2_t lhs, uint32x4_t rhs) { 61 uint64x2_t test_vsubw_high_u32(uint64x2_t lhs, uint32x4_t rhs) { 91 uint64x2_t test_vabdl_high_u32(uint32x4_t lhs, uint32x4_t rhs) { 121 uint64x2_t test_vabal_high_u32(uint64x2_t accum, uint32x4_t lhs, uint32x4_t rhs) { 241 uint64x2_t test_vsubl_high_u32(uint32x4_t lhs, uint32x4_t rhs) { 271 uint32x4_t test_vrshrn_high_n_u64(uint32x2_t lowpart, uint64x2_t input) { 301 uint32x4_t test_vshrn_high_n_u64(uint32x2_t lowpart, uint64x2_t input) [all...] |
arm-neon-vget.c | 49 uint64x1_t low_u64(uint64x2_t a) { 105 uint64x1_t high_u64(uint64x2_t a) {
|
aarch64-neon-misc.c | 65 uint64x2_t test_vceqzq_s64(int64x2_t a) { 107 uint64x2_t test_vceqzq_u64(uint64x2_t a) { 155 uint64x2_t test_vceqzq_f64(float64x2_t a) { 161 uint64x2_t test_vceqzq_p64(poly64x2_t a) { 209 uint64x2_t test_vcgezq_s64(int64x2_t a) { 233 uint64x2_t test_vcgezq_f64(float64x2_t a) { 281 uint64x2_t test_vclezq_s64(int64x2_t a) { 305 uint64x2_t test_vclezq_f64(float64x2_t a) { 353 uint64x2_t test_vcgtzq_s64(int64x2_t a) [all...] |
aarch64-neon-perm.c | 86 uint64x2_t test_vuzp1q_u64(uint64x2_t a, uint64x2_t b) { 212 uint64x2_t test_vuzp2q_u64(uint64x2_t a, uint64x2_t b) { 338 uint64x2_t test_vzip1q_u64(uint64x2_t a, uint64x2_t b) { 464 uint64x2_t test_vzip2q_u64(uint64x2_t a, uint64x2_t b) [all...] |
aarch64-neon-vget-hilo.c | 51 uint64x1_t test_vget_high_u64(uint64x2_t a) { 135 uint64x1_t test_vget_low_u64(uint64x2_t a) {
|
aarch64-poly128.c | 93 poly128_t test_vreinterpretq_p128_u64(uint64x2_t a) { 171 uint64x2_t test_vreinterpretq_u64_p128(poly128_t a) {
|
/external/chromium_org/third_party/skia/src/opts/ |
SkTextureCompression_opts_neon.cpp | 50 static inline uint64x2_t shift_swap(const uint64x2_t &x, const uint64x2_t &mask) { 51 uint64x2_t t = vandq_u64(mask, veorq_u64(x, vshrq_n_u64(x, shift))); 55 static inline uint64x2_t pack_indices(const uint64x2_t &x) { 58 static const uint64x2_t kMask1 = { 0x3FC0003FC00000ULL, 0x3FC0003FC00000ULL }; 59 uint64x2_t ret = shift_swap<10>(x, kMask1); 62 static const uint64x2_t kMask2 = { (0x3FULL << 52), (0x3FULL << 52) }; 63 static const uint64x2_t kMask3 = { (0x3FULL << 28), (0x3FULL << 28) } [all...] |
/external/clang/test/CodeGenCXX/ |
mangle-neon-vectors.cpp | 20 typedef __attribute__((neon_vector_type(2))) uint64_t uint64x2_t; typedef 45 void f4(uint64x2_t v) { }
|
aarch64-mangle-neon-vectors.cpp | 28 typedef __attribute__((neon_vector_type(2))) uint64_t uint64x2_t; typedef 82 void f22(uint64x2_t) {}
|
/external/clang/test/Sema/ |
neon-vector-types.c | 13 typedef __attribute__((neon_vector_type(2))) uint64_t uint64x2_t; typedef
|
/prebuilts/gcc/darwin-x86/aarch64/aarch64-linux-android-4.8/lib/gcc/aarch64-linux-android/4.8/include/ |
arm_neon.h | 81 typedef __builtin_aarch64_simd_udi uint64x2_t typedef 166 uint64x2_t val[2]; 286 uint64x2_t val[3]; 406 uint64x2_t val[4]; 558 __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) 559 vaddq_u64 (uint64x2_t __a, uint64x2_t __b) 596 __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) 599 return (uint64x2_t) __builtin_aarch64_uaddlv2si ((int32x2_t) __a, 635 __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__) [all...] |
/prebuilts/gcc/linux-x86/aarch64/aarch64-linux-android-4.8/lib/gcc/aarch64-linux-android/4.8/include/ |
arm_neon.h | 81 typedef __builtin_aarch64_simd_udi uint64x2_t typedef 166 uint64x2_t val[2]; 286 uint64x2_t val[3]; 406 uint64x2_t val[4]; 558 __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) 559 vaddq_u64 (uint64x2_t __a, uint64x2_t __b) 596 __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) 599 return (uint64x2_t) __builtin_aarch64_uaddlv2si ((int32x2_t) __a, 635 __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__) [all...] |
/prebuilts/gcc/darwin-x86/aarch64/aarch64-linux-android-4.9/lib/gcc/aarch64-linux-android/4.9/include/ |
arm_neon.h | 86 typedef __builtin_aarch64_simd_udi uint64x2_t typedef 173 uint64x2_t val[2]; 293 uint64x2_t val[3]; 413 uint64x2_t val[4]; 731 __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) 732 vaddq_u64 (uint64x2_t __a, uint64x2_t __b) 769 __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) 772 return (uint64x2_t) __builtin_aarch64_uaddlv2si ((int32x2_t) __a, 808 __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__) [all...] |