HomeSort by relevance Sort by last modified time
    Searched refs:uint64x2_t (Results 1 - 25 of 57) sorted by null

1 2 3

  /external/clang/test/CodeGen/
arm64_vtst.c 6 uint64x2_t test_vtstq_s64(int64x2_t a1, int64x2_t a2) {
15 uint64x2_t test_vtstq_u64(uint64x2_t a1, uint64x2_t a2) {
arm64_vCMP.c 49 uint64x2_t test_vceqq_u64(uint64x2_t a1, uint64x2_t a2) {
55 uint64x2_t test_vcgeq_s64(int64x2_t a1, int64x2_t a2) {
61 uint64x2_t test_vcgeq_u64(uint64x2_t a1, uint64x2_t a2) {
67 uint64x2_t test_vcgtq_s64(int64x2_t a1, int64x2_t a2) {
73 uint64x2_t test_vcgtq_u64(uint64x2_t a1, uint64x2_t a2)
    [all...]
arm-neon-misc.c 16 uint64x2_t q = vld1q_u64(src);
24 void t2(uint64_t *src1, uint8_t *src2, uint64x2_t *dst) {
26 uint64x2_t q = vld1q_u64(src1);
arm64_vca.c 33 uint64x2_t test_vcagtq_f64(float64x2_t a1, float64x2_t a2) {
40 uint64x2_t test_vcaltq_f64(float64x2_t a1, float64x2_t a2) {
47 uint64x2_t test_vcageq_f64(float64x2_t a1, float64x2_t a2) {
54 uint64x2_t test_vcaleq_f64(float64x2_t a1, float64x2_t a2) {
debug-info.c 57 // CHECK: uint64x2_t
59 typedef uint64_t uint64x2_t __attribute__((ext_vector_type(2))); typedef
60 uint64x2_t extvectbar[4];
arm64_vcopy.c 52 uint64x2_t test_vcopyq_laneq_u64(uint64x2_t a1, uint64x2_t a2) {
aarch64-neon-3v.c 98 uint64x2_t test_vandq_u64(uint64x2_t a, uint64x2_t b) {
194 uint64x2_t test_vorrq_u64(uint64x2_t a, uint64x2_t b) {
290 uint64x2_t test_veorq_u64(uint64x2_t a, uint64x2_t b) {
386 uint64x2_t test_vbicq_u64(uint64x2_t a, uint64x2_t b)
    [all...]
arm64_vqmov.c 48 uint32x4_t test_vqmovn_high_u64(uint32x2_t Vdlow, uint64x2_t Vn)
72 uint32x4_t test_vqmovun_high_s64(uint32x2_t Vdlow, uint64x2_t Vn)
arm64_vshift.c 95 uint64x2_t test_vqshlq_n_u64(uint64x2_t in) {
192 uint64x2_t test_vrshrq_n_u64(uint64x2_t in) {
352 uint64x2_t test_vrsraq_n_u64(uint64x2_t acc, uint64x2_t in) {
aarch64-neon-extract.c 97 uint64x2_t test_vextq_u64(uint64x2_t a, uint64x2_t b) {
arm64_vsli.c 128 uint64x2_t test_vsliq_n_u64(uint64x2_t a1, uint64x2_t a2) {
arm64_vsri.c 129 uint64x2_t test_vsriq_n_u64(uint64x2_t a1, uint64x2_t a2) {
arm64_neon_high_half.c 31 uint64x2_t test_vaddw_high_u32(uint64x2_t lhs, uint32x4_t rhs) {
61 uint64x2_t test_vsubw_high_u32(uint64x2_t lhs, uint32x4_t rhs) {
91 uint64x2_t test_vabdl_high_u32(uint32x4_t lhs, uint32x4_t rhs) {
121 uint64x2_t test_vabal_high_u32(uint64x2_t accum, uint32x4_t lhs, uint32x4_t rhs) {
241 uint64x2_t test_vsubl_high_u32(uint32x4_t lhs, uint32x4_t rhs) {
271 uint32x4_t test_vrshrn_high_n_u64(uint32x2_t lowpart, uint64x2_t input) {
301 uint32x4_t test_vshrn_high_n_u64(uint32x2_t lowpart, uint64x2_t input)
    [all...]
arm-neon-vget.c 49 uint64x1_t low_u64(uint64x2_t a) {
105 uint64x1_t high_u64(uint64x2_t a) {
aarch64-neon-misc.c 65 uint64x2_t test_vceqzq_s64(int64x2_t a) {
107 uint64x2_t test_vceqzq_u64(uint64x2_t a) {
155 uint64x2_t test_vceqzq_f64(float64x2_t a) {
161 uint64x2_t test_vceqzq_p64(poly64x2_t a) {
209 uint64x2_t test_vcgezq_s64(int64x2_t a) {
233 uint64x2_t test_vcgezq_f64(float64x2_t a) {
281 uint64x2_t test_vclezq_s64(int64x2_t a) {
305 uint64x2_t test_vclezq_f64(float64x2_t a) {
353 uint64x2_t test_vcgtzq_s64(int64x2_t a)
    [all...]
aarch64-neon-perm.c 86 uint64x2_t test_vuzp1q_u64(uint64x2_t a, uint64x2_t b) {
212 uint64x2_t test_vuzp2q_u64(uint64x2_t a, uint64x2_t b) {
338 uint64x2_t test_vzip1q_u64(uint64x2_t a, uint64x2_t b) {
464 uint64x2_t test_vzip2q_u64(uint64x2_t a, uint64x2_t b)
    [all...]
aarch64-neon-vget-hilo.c 51 uint64x1_t test_vget_high_u64(uint64x2_t a) {
135 uint64x1_t test_vget_low_u64(uint64x2_t a) {
aarch64-poly128.c 93 poly128_t test_vreinterpretq_p128_u64(uint64x2_t a) {
171 uint64x2_t test_vreinterpretq_u64_p128(poly128_t a) {
  /external/chromium_org/third_party/skia/src/opts/
SkTextureCompression_opts_neon.cpp 50 static inline uint64x2_t shift_swap(const uint64x2_t &x, const uint64x2_t &mask) {
51 uint64x2_t t = vandq_u64(mask, veorq_u64(x, vshrq_n_u64(x, shift)));
55 static inline uint64x2_t pack_indices(const uint64x2_t &x) {
58 static const uint64x2_t kMask1 = { 0x3FC0003FC00000ULL, 0x3FC0003FC00000ULL };
59 uint64x2_t ret = shift_swap<10>(x, kMask1);
62 static const uint64x2_t kMask2 = { (0x3FULL << 52), (0x3FULL << 52) };
63 static const uint64x2_t kMask3 = { (0x3FULL << 28), (0x3FULL << 28) }
    [all...]
  /external/clang/test/CodeGenCXX/
mangle-neon-vectors.cpp 20 typedef __attribute__((neon_vector_type(2))) uint64_t uint64x2_t; typedef
45 void f4(uint64x2_t v) { }
aarch64-mangle-neon-vectors.cpp 28 typedef __attribute__((neon_vector_type(2))) uint64_t uint64x2_t; typedef
82 void f22(uint64x2_t) {}
  /external/clang/test/Sema/
neon-vector-types.c 13 typedef __attribute__((neon_vector_type(2))) uint64_t uint64x2_t; typedef
  /prebuilts/gcc/darwin-x86/aarch64/aarch64-linux-android-4.8/lib/gcc/aarch64-linux-android/4.8/include/
arm_neon.h 81 typedef __builtin_aarch64_simd_udi uint64x2_t typedef
166 uint64x2_t val[2];
286 uint64x2_t val[3];
406 uint64x2_t val[4];
558 __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
559 vaddq_u64 (uint64x2_t __a, uint64x2_t __b)
596 __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
599 return (uint64x2_t) __builtin_aarch64_uaddlv2si ((int32x2_t) __a,
635 __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)
    [all...]
  /prebuilts/gcc/linux-x86/aarch64/aarch64-linux-android-4.8/lib/gcc/aarch64-linux-android/4.8/include/
arm_neon.h 81 typedef __builtin_aarch64_simd_udi uint64x2_t typedef
166 uint64x2_t val[2];
286 uint64x2_t val[3];
406 uint64x2_t val[4];
558 __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
559 vaddq_u64 (uint64x2_t __a, uint64x2_t __b)
596 __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
599 return (uint64x2_t) __builtin_aarch64_uaddlv2si ((int32x2_t) __a,
635 __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)
    [all...]
  /prebuilts/gcc/darwin-x86/aarch64/aarch64-linux-android-4.9/lib/gcc/aarch64-linux-android/4.9/include/
arm_neon.h 86 typedef __builtin_aarch64_simd_udi uint64x2_t typedef
173 uint64x2_t val[2];
293 uint64x2_t val[3];
413 uint64x2_t val[4];
731 __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
732 vaddq_u64 (uint64x2_t __a, uint64x2_t __b)
769 __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
772 return (uint64x2_t) __builtin_aarch64_uaddlv2si ((int32x2_t) __a,
808 __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)
    [all...]

Completed in 647 milliseconds

1 2 3