Home | History | Annotate | Download | only in include

Lines Matching defs:uint64x2_t

52 typedef __attribute__((neon_vector_type(2)))  uint64_t uint64x2_t;
123 uint64x2_t val[2];
219 uint64x2_t val[3];
315 uint64x2_t val[4];
362 __ai uint64x2_t vmovl_u32(uint32x2_t __a) {
363 return (uint64x2_t)__builtin_neon_vmovl_v((int8x8_t)__a, 51); }
375 __ai uint64x2_t vmull_u32(uint32x2_t __a, uint32x2_t __b) {
376 return (uint64x2_t)__builtin_neon_vmull_v((int8x8_t)__a, (int8x8_t)__b, 51); }
444 __ai uint64x2_t vabal_u32(uint64x2_t __a, uint32x2_t __b, uint32x2_t __c) {
457 __ai uint64x2_t vabdl_u32(uint32x2_t __a, uint32x2_t __b) {
511 __ai uint64x2_t vaddq_u64(uint64x2_t __a, uint64x2_t __b) {
524 __ai uint32x2_t vaddhn_u64(uint64x2_t __a, uint64x2_t __b) {
537 __ai uint64x2_t vaddl_u32(uint32x2_t __a, uint32x2_t __b) {
550 __ai uint64x2_t vaddw_u32(uint64x2_t __a, uint32x2_t __b) {
583 __ai uint64x2_t vandq_u64(uint64x2_t __a, uint64x2_t __b) {
616 __ai uint64x2_t vbicq_u64(uint64x2_t __a, uint64x2_t __b) {
647 __ai int64x2_t vbslq_s64(uint64x2_t __a, int64x2_t __b, int64x2_t __c) {
655 __ai uint64x2_t vbslq_u64(uint64x2_t __a, uint64x2_t __b, uint64x2_t __c) {
656 return (uint64x2_t)__builtin_neon_vbslq_v((int8x16_t)__a, (int8x16_t)__b, (int8x16_t)__c, 51); }
902 __ai uint64x2_t vcombine_u64(uint64x1_t __a, uint64x1_t __b) {
903 return (uint64x2_t)__builtin_shufflevector((int64x1_t)__a, (int64x1_t)__b, 0, 1); }
1095 __ai uint64x2_t vdupq_n_u64(uint64_t __a) {
1096 return (uint64x2_t){ __a, __a }; }
1128 __ai uint64x2_t veorq_u64(uint64x2_t __a, uint64x2_t __b) {
1192 uint64x2_t __a = (a); uint64x2_t __b = (b); \
1193 (uint64x2_t)__builtin_neon_vextq_v((int8x16_t)__a, (int8x16_t)__b, __c, 51); })
1221 __ai uint64x1_t vget_high_u64(uint64x2_t __a) {
1292 uint64x2_t __a = (a); \
1313 __ai uint64x1_t vget_low_u64(uint64x2_t __a) {
1377 (uint64x2_t)__builtin_neon_vld1q_v(__a, 51); })
1426 (uint64x2_t)__builtin_neon_vld1q_dup_v(__a, 51); })
1478 uint64x2_t __b = (b); \
1479 (uint64x2_t)__builtin_neon_vld1q_lane_v(__a, (int8x16_t)__b, __c, 51); })
2004 __ai uint64x2_t vmlal_u32(uint64x2_t __a, uint32x2_t __b, uint32x2_t __c) {
2017 uint64x2_t __a = (a); uint32x2_t __b = (b); uint32x2_t __c = (c); \
2026 __ai uint64x2_t vmlal_n_u32(uint64x2_t __a, uint32x2_t __b, uint32_t __c) {
2120 __ai uint64x2_t vmlsl_u32(uint64x2_t __a, uint32x2_t __b, uint32x2_t __c) {
2133 uint64x2_t __a = (a); uint32x2_t __b = (b); uint32x2_t __c = (c); \
2142 __ai uint64x2_t vmlsl_n_u32(uint64x2_t __a, uint32x2_t __b, uint32_t __c) {
2207 __ai uint32x2_t vmovn_u64(uint64x2_t __a) {
2252 __ai uint64x2_t vmovq_n_u64(uint64_t __a) {
2253 return (uint64x2_t){ __a, __a }; }
2303 __ai uint64x2_t vmull_n_u32(uint32x2_t __a, uint32_t __b) {
2304 return (uint64x2_t)__builtin_neon_vmull_v((int8x8_t)__a, (int8x8_t)(uint32x2_t){ __b, __b }, 51); }
2439 __ai uint64x2_t vornq_u64(uint64x2_t __a, uint64x2_t __b) {
2472 __ai uint64x2_t vorrq_u64(uint64x2_t __a, uint64x2_t __b) {
2497 __ai uint64x2_t vpadalq_u32(uint64x2_t __a, uint32x4_t __b) {
2498 return (uint64x2_t)__builtin_neon_vpadalq_v((int8x16_t)__a, (int8x16_t)__b, 51); }
2537 __ai uint64x2_t vpaddlq_u32(uint32x4_t __a) {
2538 return (uint64x2_t)__builtin_neon_vpaddlq_v((int8x16_t)__a, 51); }
2613 __ai uint64x2_t vqaddq_u64(uint64x2_t __a, uint64x2_t __b) {
2614 return (uint64x2_t)__builtin_neon_vqaddq_v((int8x16_t)__a, (int8x16_t)__b, 51); }
2708 __ai uint32x2_t vqmovn_u64(uint64x2_t __a) {
2792 __ai uint64x2_t vqrshlq_u64(uint64x2_t __a, int64x2_t __b) {
2793 return (uint64x2_t)__builtin_neon_vqrshlq_v((int8x16_t)__a, (int8x16_t)__b, 51); }
2811 uint64x2_t __a = (a); \
2854 __ai uint64x2_t vqshlq_u64(uint64x2_t __a, int64x2_t __b) {
2855 return (uint64x2_t)__builtin_neon_vqshlq_v((int8x16_t)__a, (int8x16_t)__b, 51); }
2880 (uint64x2_t)__builtin_neon_vqshluq_n_v((int8x16_t)__a, __b, 51); })
2928 uint64x2_t __a = (a); \
2929 (uint64x2_t)__builtin_neon_vqshlq_n_v((int8x16_t)__a, __b, 51); })
2947 uint64x2_t __a = (a); \
2990 __ai uint64x2_t vqsubq_u64(uint64x2_t __a, uint64x2_t __b) {
2991 return (uint64x2_t)__builtin_neon_vqsubq_v((int8x16_t)__a, (int8x16_t)__b, 51); }
3003 __ai uint32x2_t vraddhn_u64(uint64x2_t __a, uint64x2_t __b) {
3296 __ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __a) {
3318 __ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __a) {
3340 __ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __a) {
3362 __ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __a) {
3384 __ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __a) {
3406 __ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __a) {
3428 __ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __a) {
3438 __ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __a) {
3439 return (uint64x2_t)__a; }
3440 __ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __a) {
3441 return (uint64x2_t)__a; }
3442 __ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __a) {
3443 return (uint64x2_t)__a; }
3444 __ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __a) {
3445 return (uint64x2_t)__a; }
3446 __ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __a) {
3447 return (uint64x2_t)__a; }
3448 __ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __a) {
3449 return (uint64x2_t)__a; }
3450 __ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __a) {
3451 return (uint64x2_t)__a; }
3452 __ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __a) {
3453 return (uint64x2_t)__a; }
3454 __ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __a) {
3455 return (uint64x2_t)__a; }
3456 __ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __a) {
3457 return (uint64x2_t)__a; }
3458 __ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __a) {
3459 return (uint64x2_t)__a; }
3474 __ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __a) {
3496 __ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __a) {
3518 __ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __a) {
3540 __ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __a) {
3679 __ai uint64x2_t vrshlq_u64(uint64x2_t __a, int64x2_t __b) {
3680 return (uint64x2_t)__builtin_neon_vrshlq_v((int8x16_t)__a, (int8x16_t)__b, 51); }
3698 uint64x2_t __a = (a); \
3747 uint64x2_t __a = (a); \
3748 (uint64x2_t)__builtin_neon_vrshrq_n_v((int8x16_t)__a, __b, 51); })
3810 uint64x2_t __a = (a); uint64x2_t __b = (b); \
3811 (uint64x2_t)__builtin_neon_vrsraq_n_v((int8x16_t)__a, (int8x16_t)__b, __c, 51); })
3823 __ai uint32x2_t vrsubhn_u64(uint64x2_t __a, uint64x2_t __b) {
3890 uint64_t __a = (a); uint64x2_t __b = (b); \
3891 (uint64x2_t)__builtin_neon_vsetq_lane_i64(__a, (int64x2_t)__b, __c); })
3923 __ai uint64x2_t vshlq_u64(uint64x2_t __a, int64x2_t __b) {
3924 return (uint64x2_t)__builtin_neon_vshlq_v((int8x16_t)__a, (int8x16_t)__b, 51); }
3943 (uint64x2_t)__builtin_neon_vshll_n_v((int8x8_t)__a, __b, 51); })
3991 uint64x2_t __a = (a); \
3992 (uint64x2_t)__builtin_neon_vshlq_n_v((int8x16_t)__a, __b, 51); })
4010 uint64x2_t __a = (a); \
4059 uint64x2_t __a = (a); \
4060 (uint64x2_t)__builtin_neon_vshrq_n_v((int8x16_t)__a, __b, 51); })
4114 uint64x2_t __a = (a); uint64x2_t __b = (b); \
4115 (uint64x2_t)__builtin_neon_vsliq_n_v((int8x16_t)__a, (int8x16_t)__b, __c, 51); })
4169 uint64x2_t __a = (a); uint64x2_t __b = (b); \
4170 (uint64x2_t)__builtin_neon_vsraq_n_v((int8x16_t)__a, (int8x16_t)__b, __c, 51); })
4224 uint64x2_t __a = (a); uint64x2_t __b = (b); \
4225 (uint64x2_t)__builtin_neon_vsriq_n_v((int8x16_t)__a, (int8x16_t)__b, __c, 51); })
4243 uint64x2_t __b = (b); \
4316 uint64x2_t __b = (b); \
4770 __ai uint64x2_t vsubq_u64(uint64x2_t __a, uint64x2_t __b) {
4783 __ai uint32x2_t vsubhn_u64(uint64x2_t __a, uint64x2_t __b) {
4796 __ai uint64x2_t vsubl_u32(uint32x2_t __a, uint32x2_t __b) {
4809 __ai uint64x2_t vsubw_u32(uint64x2_t __a, uint32x2_t __b) {