/external/clang/test/Sema/ |
variadic-promotion.c | 5 void test_floating_promotion(__fp16 *f16, float f32, double f64) { 6 variadic(3, *f16, f32, f64);
|
/external/clang/test/CodeGen/ |
mms-bitfields.c | 5 long long f64; member in struct:s1 12 long long f64[4]; member in struct:s2
|
ms_struct.c | 6 long long f64; member in struct:s1 13 long long f64[4]; member in struct:s2
|
/toolchain/binutils/binutils-2.27/gas/testsuite/gas/arm/ |
ual-vcmp.s | 16 vcmp.f64 d0, #0.0 17 vcmp.f64 d1, #0 18 vcmpe.f64 d2, #0.0 19 vcmpe.f64 d3, #0 20 vcmp.f64 d4, #0.0e5 21 vcmp.f64 d5, #0e-2 22 vcmpe.f64 d6, #0.00 23 vcmpe.f64 d7, #.0 24 vcmp.f64 d8, #0x0 25 vcmpe.f64 d9, #0x [all...] |
vcmp-zero-bad.s | 10 vcmp.f64 d0, #-1 11 vcmpe.f64 d3, #0x35 12 vcmpe.f64 d4, 0xf
|
armv8-a+fp.s | 11 vseleq.f64 d0, d0, d0 12 vselvs.f64 d16, d16, d16 13 vselge.f64 d15, d15, d15 14 vselgt.f64 d31, d31, d31 19 vmaxnm.f64 d0, d0, d0 20 vmaxnm.f64 d16, d16, d16 21 vmaxnm.f64 d15, d15, d15 22 vmaxnm.f64 d31, d31, d31 27 vminnm.f64 d0, d0, d0 28 vminnm.f64 d16, d16, d1 [all...] |
armv7e-m+fpv5-d16.s | 11 vseleq.f64 d0, d0, d0 12 vselvs.f64 d8, d8, d8 13 vselge.f64 d15, d15, d15 14 vselgt.f64 d10, d10, d10 19 vmaxnm.f64 d0, d0, d0 20 vmaxnm.f64 d8, d8, d8 21 vmaxnm.f64 d15, d15, d15 22 vmaxnm.f64 d10, d10, d10 27 vminnm.f64 d0, d0, d0 28 vminnm.f64 d8, d8, d [all...] |
ual-vcmp.d | 19 0+028 <[^>]*> eeb50b40 vcmp.f64 d0, #0.0 20 0+02c <[^>]*> eeb51b40 vcmp.f64 d1, #0.0 21 0+030 <[^>]*> eeb52bc0 vcmpe.f64 d2, #0.0 22 0+034 <[^>]*> eeb53bc0 vcmpe.f64 d3, #0.0 23 0+038 <[^>]*> eeb54b40 vcmp.f64 d4, #0.0 24 0+03c <[^>]*> eeb55b40 vcmp.f64 d5, #0.0 25 0+040 <[^>]*> eeb56bc0 vcmpe.f64 d6, #0.0 26 0+044 <[^>]*> eeb57bc0 vcmpe.f64 d7, #0.0 27 0+048 <[^>]*> eeb58b40 vcmp.f64 d8, #0.0 28 0+04c <[^>]*> eeb59bc0 vcmpe.f64 d9, #0. [all...] |
vfpv3-32drs.d | 8 0[0-9a-f]+ <[^>]+> eeb03b66 (vmov\.f64|fcpyd) d3, d22 9 0[0-9a-f]+ <[^>]+> eef06b43 (vmov\.f64|fcpyd) d22, d3 10 0[0-9a-f]+ <[^>]+> eef76acb (vcvt\.f64\.f32|fcvtds) d22, s22 11 0[0-9a-f]+ <[^>]+> eeb7bbe6 (vcvt\.f32\.f64|fcvtsd) s22, d22 16 0[0-9a-f]+ <[^>]+> eef86bcb (vcvt\.f64\.s32|fsitod) d22, s22 17 0[0-9a-f]+ <[^>]+> eef85b6a (vcvt\.f64\.u32|fuitod) d21, s21 18 0[0-9a-f]+ <[^>]+> eebdab64 (vcvtr\.s32\.f64|ftosid) s20, d20 19 0[0-9a-f]+ <[^>]+> eebdabe4 (vcvt\.s32\.f64|ftosizd) s20, d20 20 0[0-9a-f]+ <[^>]+> eefc9b63 (vcvtr\.u32\.f64|ftouid) s19, d19 21 0[0-9a-f]+ <[^>]+> eefc9be3 (vcvt\.u32\.f64|ftouizd) s19, d1 [all...] |
vfp1.d | 10 0+000 <[^>]*> eeb40bc0 (vcmpe\.f64|fcmped) d0, d0 11 0+004 <[^>]*> eeb50bc0 (vcmpe\.f64 d0, #0.0|fcmpezd d0) 12 0+008 <[^>]*> eeb40b40 (vcmp\.f64|fcmpd) d0, d0 13 0+00c <[^>]*> eeb50b40 (vcmp\.f64 d0, #0.0|fcmpzd d0) 14 0+010 <[^>]*> eeb00bc0 (vabs\.f64|fabsd) d0, d0 15 0+014 <[^>]*> eeb00b40 (vmov\.f64|fcpyd) d0, d0 16 0+018 <[^>]*> eeb10b40 (vneg\.f64|fnegd) d0, d0 17 0+01c <[^>]*> eeb10bc0 (vsqrt\.f64|fsqrtd) d0, d0 18 0+020 <[^>]*> ee300b00 (vadd\.f64|faddd) d0, d0, d0 19 0+024 <[^>]*> ee800b00 (vdiv\.f64|fdivd) d0, d0, d [all...] |
neon-suffix-bad.s | 5 faddd.f64 d0, d0, d0 6 vcvt.f64.s32 d0, s0, #11 11 faddd.f64 d0, d0, d0 12 vcvt.f64.s32 d0, s0, #11
|
armv8-a+fp.d | 12 0[0-9a-f]+ <[^>]+> fe000b00 vseleq.f64 d0, d0, d0 13 0[0-9a-f]+ <[^>]+> fe500ba0 vselvs.f64 d16, d16, d16 14 0[0-9a-f]+ <[^>]+> fe2ffb0f vselge.f64 d15, d15, d15 15 0[0-9a-f]+ <[^>]+> fe7ffbaf vselgt.f64 d31, d31, d31 20 0[0-9a-f]+ <[^>]+> fe800b00 vmaxnm.f64 d0, d0, d0 21 0[0-9a-f]+ <[^>]+> fec00ba0 vmaxnm.f64 d16, d16, d16 22 0[0-9a-f]+ <[^>]+> fe8ffb0f vmaxnm.f64 d15, d15, d15 23 0[0-9a-f]+ <[^>]+> fecffbaf vmaxnm.f64 d31, d31, d31 28 0[0-9a-f]+ <[^>]+> fe800b40 vminnm.f64 d0, d0, d0 29 0[0-9a-f]+ <[^>]+> fec00be0 vminnm.f64 d16, d16, d1 [all...] |
vfp1_t2.d | 10 0+000 <[^>]*> eeb4 0bc0 (vcmpe\.f64|fcmped) d0, d0 11 0+004 <[^>]*> eeb5 0bc0 (vcmpe\.f64 d0, #0.0|fcmpezd d0) 12 0+008 <[^>]*> eeb4 0b40 (vcmp\.f64|fcmpd) d0, d0 13 0+00c <[^>]*> eeb5 0b40 (vcmp\.f64 d0, #0.0|fcmpzd d0) 14 0+010 <[^>]*> eeb0 0bc0 (vabs\.f64|fabsd) d0, d0 15 0+014 <[^>]*> eeb0 0b40 (vmov\.f64|fcpyd) d0, d0 16 0+018 <[^>]*> eeb1 0b40 (vneg\.f64|fnegd) d0, d0 17 0+01c <[^>]*> eeb1 0bc0 (vsqrt\.f64|fsqrtd) d0, d0 18 0+020 <[^>]*> ee30 0b00 (vadd\.f64|faddd) d0, d0, d0 19 0+024 <[^>]*> ee80 0b00 (vdiv\.f64|fdivd) d0, d0, d [all...] |
vfp-fma-arm.d | 9 0[0-9a-f]+ <[^>]+> eea10b02 vfma\.f64 d0, d1, d2 11 0[0-9a-f]+ <[^>]+> 0ea10b02 vfmaeq\.f64 d0, d1, d2 13 0[0-9a-f]+ <[^>]+> eea10b42 vfms\.f64 d0, d1, d2 15 0[0-9a-f]+ <[^>]+> 0ea10b42 vfmseq\.f64 d0, d1, d2 17 0[0-9a-f]+ <[^>]+> ee910b42 vfnma\.f64 d0, d1, d2 19 0[0-9a-f]+ <[^>]+> 0e910b42 vfnmaeq\.f64 d0, d1, d2 21 0[0-9a-f]+ <[^>]+> ee910b02 vfnms\.f64 d0, d1, d2 23 0[0-9a-f]+ <[^>]+> 0e910b02 vfnmseq\.f64 d0, d1, d2
|
/external/vulkan-validation-layers/libs/glm/gtx/ |
number_precision.hpp | 69 typedef f64 f64vec1; //!< \brief Single-precision floating-point scalar. (from GLM_GTX_number_precision extension) 76 typedef f64 f64mat1; //!< \brief Double-precision floating-point scalar. (from GLM_GTX_number_precision extension) 77 typedef f64 f64mat1x1; //!< \brief Double-precision floating-point scalar. (from GLM_GTX_number_precision extension)
|
/prebuilts/ndk/r16/sources/third_party/vulkan/src/libs/glm/gtx/ |
number_precision.hpp | 69 typedef f64 f64vec1; //!< \brief Single-precision floating-point scalar. (from GLM_GTX_number_precision extension) 76 typedef f64 f64mat1; //!< \brief Double-precision floating-point scalar. (from GLM_GTX_number_precision extension) 77 typedef f64 f64mat1x1; //!< \brief Double-precision floating-point scalar. (from GLM_GTX_number_precision extension)
|
/external/llvm/test/MC/ARM/ |
single-precision-fp.s | 5 vadd.f64 d0, d1, d2 6 vsub.f64 d2, d3, d4 7 vdiv.f64 d4, d5, d6 8 vmul.f64 d6, d7, d8 9 vnmul.f64 d8, d9, d10 11 @ CHECK-ERRORS-NEXT: vadd.f64 d0, d1, d2 13 @ CHECK-ERRORS-NEXT: vsub.f64 d2, d3, d4 15 @ CHECK-ERRORS-NEXT: vdiv.f64 d4, d5, d6 17 @ CHECK-ERRORS-NEXT: vmul.f64 d6, d7, d8 19 @ CHECK-ERRORS-NEXT: vnmul.f64 d8, d9, d1 [all...] |
d16.s | 7 @ D16-NEXT: vadd.f64 d1, d2, d16 8 vadd.f64 d1, d2, d16 11 @ D16-NEXT: vadd.f64 d1, d17, d6 12 vadd.f64 d1, d17, d6 15 @ D16-NEXT: vadd.f64 d19, d7, d6 16 vadd.f64 d19, d7, d6 19 @ D16-NEXT: vcvt.f64.f32 d22, s4 20 vcvt.f64.f32 d22, s4 23 @ D16-NEXT: vcvt.f32.f64 s26, d30 24 vcvt.f32.f64 s26, d3 [all...] |
fp-armv8.s | 5 vcvtt.f64.f16 d3, s1 6 @ CHECK: vcvtt.f64.f16 d3, s1 @ encoding: [0xe0,0x3b,0xb2,0xee] 7 vcvtt.f16.f64 s5, d12 8 @ CHECK: vcvtt.f16.f64 s5, d12 @ encoding: [0xcc,0x2b,0xf3,0xee] 10 vcvtb.f64.f16 d3, s1 11 @ CHECK: vcvtb.f64.f16 d3, s1 @ encoding: [0x60,0x3b,0xb2,0xee] 12 vcvtb.f16.f64 s4, d1 13 @ CHECK: vcvtb.f16.f64 s4, d1 @ encoding: [0x41,0x2b,0xb3,0xee] 15 vcvttge.f64.f16 d3, s1 16 @ CHECK: vcvttge.f64.f16 d3, s1 @ encoding: [0xe0,0x3b,0xb2,0xae [all...] |
thumb-fp-armv8.s | 5 vcvtt.f64.f16 d3, s1 6 @ CHECK: vcvtt.f64.f16 d3, s1 @ encoding: [0xb2,0xee,0xe0,0x3b] 7 vcvtt.f16.f64 s5, d12 8 @ CHECK: vcvtt.f16.f64 s5, d12 @ encoding: [0xf3,0xee,0xcc,0x2b] 10 vcvtb.f64.f16 d3, s1 11 @ CHECK: vcvtb.f64.f16 d3, s1 @ encoding: [0xb2,0xee,0x60,0x3b] 12 vcvtb.f16.f64 s4, d1 13 @ CHECK: vcvtb.f16.f64 s4, d1 @ encoding: [0xb3,0xee,0x41,0x2b] 16 vcvttge.f64.f16 d3, s1 17 @ CHECK: vcvttge.f64.f16 d3, s1 @ encoding: [0xb2,0xee,0xe0,0x3b [all...] |
directive-arch_extension-simd.s | 24 vmaxnm.f64 d0, d0, d0 26 vminnm.f64 d0, d0, d0 33 vcvta.s32.f64 s0, d0 35 vcvta.u32.f64 s0, d0 41 vcvtn.s32.f64 s0, d0 43 vcvtn.u32.f64 s0, d0 49 vcvtp.s32.f64 s0, d0 51 vcvtp.u32.f64 s0, d0 57 vcvtm.s32.f64 s0, d0 59 vcvtm.u32.f64 s0, d [all...] |
/external/capstone/suite/MC/ARM/ |
fp-armv8.s.cs | 2 0xe0,0x3b,0xb2,0xee = vcvtt.f64.f16 d3, s1 3 0xcc,0x2b,0xf3,0xee = vcvtt.f16.f64 s5, d12 4 0x60,0x3b,0xb2,0xee = vcvtb.f64.f16 d3, s1 5 0x41,0x2b,0xb3,0xee = vcvtb.f16.f64 s4, d1 6 0xe0,0x3b,0xb2,0xae = vcvttge.f64.f16 d3, s1 7 0xcc,0x2b,0xf3,0xce = vcvttgt.f16.f64 s5, d12 8 0x60,0x3b,0xb2,0x0e = vcvtbeq.f64.f16 d3, s1 9 0x41,0x2b,0xb3,0xbe = vcvtblt.f16.f64 s4, d1 11 0xc3,0x1b,0xbc,0xfe = vcvta.s32.f64 s2, d3 13 0xe7,0x3b,0xbd,0xfe = vcvtn.s32.f64 s6, d2 [all...] |
thumb-fp-armv8.s.cs | 2 0xb2,0xee,0xe0,0x3b = vcvtt.f64.f16 d3, s1 3 0xf3,0xee,0xcc,0x2b = vcvtt.f16.f64 s5, d12 4 0xb2,0xee,0x60,0x3b = vcvtb.f64.f16 d3, s1 5 0xb3,0xee,0x41,0x2b = vcvtb.f16.f64 s4, d1 6 0xb2,0xee,0xe0,0x3b = vcvttge.f64.f16 d3, s1 7 0xf3,0xee,0xcc,0x2b = vcvttgt.f16.f64 s5, d12 8 0xb2,0xee,0x60,0x3b = vcvtbeq.f64.f16 d3, s1 9 0xb3,0xee,0x41,0x2b = vcvtblt.f16.f64 s4, d1 11 0xbc,0xfe,0xc3,0x1b = vcvta.s32.f64 s2, d3 13 0xbd,0xfe,0xe7,0x3b = vcvtn.s32.f64 s6, d2 [all...] |
/bionic/libm/arm/ |
sqrt.S | 36 vmov.f64 d0, r0, r1 37 vsqrt.f64 d0, d0 38 vmov.f64 r0, r1, d0
|
/external/valgrind/VEX/priv/ |
guest_generic_x87.c | 105 void convert_f64le_to_f80le ( /*IN*/UChar* f64, /*OUT*/UChar* f80 ) 111 sign = toUChar( (f64[7] >> 7) & 1 ); 112 bexp = (f64[7] << 4) | ((f64[6] >> 4) & 0x0F); 122 (f64[6] & 0x0F) == 0 123 && f64[5] == 0 && f64[4] == 0 && f64[3] == 0 124 && f64[2] == 0 && f64[1] == 0 && f64[0] == [all...] |