/toolchain/binutils/binutils-2.27/gas/testsuite/gas/arm/ |
vcvt.s | 6 VCVT.F64.U32 d1,d1,#1 7 VCVT.F64.U32 d1,d1,#32 8 VCVT.F64.U32 d1,d1,#17 13 VCVT.F32.U32 s1,s1,#1 14 VCVT.F32.U32 s1,s1,#32 15 VCVT.F32.U32 s1,s1,#16 20 VCVT.U32.F64 d1,d1,#1 21 VCVT.U32.F64 d1,d1,#32 22 VCVT.U32.F64 d1,d1,#16 27 VCVT.U32.F32 s1,s1,# [all...] |
vcvt-bad.s | 6 VCVT.F64.U32 d1,d1,#0 7 VCVT.F64.U32 d1,d1,#33 12 VCVT.F32.U32 s1,s1,#0 13 VCVT.F32.U32 s1,s1,#33 18 VCVT.U32.F64 d1,d1,#0 19 VCVT.U32.F64 d1,d1,#33 24 VCVT.U32.F32 s1,s1,#0 25 VCVT.U32.F32 s1,s1,#33
|
/external/syslinux/gpxe/src/drivers/net/phantom/ |
nxhal_nic_interface.h | 15 typedef U32 nx_reg_addr_t; 186 typedef U32 nx_rx_rule_type_t; 195 typedef U32 nx_rx_rule_cmd_t; 223 U32 id; 224 U32 active; 254 typedef U32 nx_host_ctx_state_t; 300 U32 ring_size; /* Ring entries */ 301 U32 rsvd; /* Padding */ 308 U32 capabilities[4]; /* Flag bit vector */ 309 U32 host_int_crb_mode; /* Interrupt crb usage * [all...] |
/external/lz4/programs/ |
datagen.c | 30 #include "util.h" /* U32 */ 60 static unsigned int RDG_rand(U32* src) 62 U32 rand32 = *src; 76 U32 u = 0; 79 U32 const weight = (U32)((double)(LTSIZE - u) * ld) + 1; 80 U32 const end = MIN(u+weight, LTSIZE); 88 static BYTE RDG_genChar(U32* seed, const litDistribTable lt) 90 U32 id = RDG_rand(seed) & LTMASK; 101 const U32 matchProba32 = (U32)(32768 * matchProba) [all...] |
bench.c | 79 static U32 g_compressibilityDefault = 50; 87 static U32 g_displayLevel = 2; /* 0 : no display; 1: errors; 2 : + result + interaction + warnings; 3 : + progression; 4 : + information */ 117 static U32 g_nbSeconds = NBSECONDS; 160 const size_t* fileSizes, U32 nbFiles) 163 U32 const maxNbBlocks = (U32) ((srcSize + (blockSize-1)) / blockSize) + nbFiles; 168 U32 nbBlocks; 198 U32 fileNb; 201 U32 const nbBlocksforThisFile = (U32)((remaining + (blockSize-1)) / blockSize) [all...] |
/external/tensorflow/tensorflow/compiler/xla/service/ |
hlo_sharding_test.cc | 43 Shape tile_shape = ShapeUtil::MakeShape(U32, {4}); 53 EXPECT_IS_OK(sharding.Validate(ShapeUtil::MakeShape(U32, {4}), 69 EXPECT_IS_OK(sharding.Validate(ShapeUtil::MakeShape(U32, {4}), 72 sharding.Validate(ShapeUtil::MakeShape(U32, {4}), /*num_devices=*/5)); 75 sharding.GetAsShapeTree(ShapeUtil::MakeShape(U32, {4})); 83 Shape tile_shape = ShapeUtil::MakeShape(U32, {2, 3}); 92 Shape tile_shape = ShapeUtil::MakeShape(U32, {2, 3}); 95 EXPECT_IS_NOT_OK(sharding.Validate(ShapeUtil::MakeShape(U32, {4, 6}), 101 Shape tile_shape = ShapeUtil::MakeShape(U32, {2, 3}); 111 Shape tile_shape = ShapeUtil::MakeShape(U32, {2, 3}) [all...] |
/external/libxaac/decoder/armv7/ |
ixheaacd_pre_twiddle_compute.s | 129 VSHR.U32 Q15, Q15, #16 130 VSHR.U32 Q14, Q14, #16 131 VSHR.U32 Q13, Q13, #16 132 VSHR.U32 Q12, Q12, #16 148 VSHR.U32 Q11, Q11, #16 149 VSHR.U32 Q10, Q10, #16 150 VSHR.U32 Q9, Q9, #16 151 VSHR.U32 Q8, Q8, #16 198 VSHR.U32 Q15, Q15, #16 199 VSHR.U32 Q14, Q14, #1 [all...] |
ixheaacd_post_twiddle.s | 117 VSHR.U32 Q15, Q15, #16 118 VSHR.U32 Q14, Q14, #16 119 VSHR.U32 Q13, Q13, #16 120 VSHR.U32 Q12, Q12, #16 139 VSHR.U32 Q11, Q11, #16 142 VSHR.U32 Q10, Q10, #16 145 VSHR.U32 Q9, Q9, #16 148 VSHR.U32 Q8, Q8, #16 173 VSHR.U32 Q0, Q0, #16 176 VSHR.U32 Q1, Q1, #1 [all...] |
ixheaacd_post_twiddle_overlap.s | 212 VSHR.U32 Q15, Q15, #16 215 VSHR.U32 Q14, Q14, #16 224 VSHR.U32 Q13, Q13, #16 227 VSHR.U32 Q12, Q12, #16 253 VSHR.U32 Q11, Q11, #16 256 VSHR.U32 Q1, Q1, #16 257 VSHR.U32 Q0, Q0, #16 310 VSHR.U32 Q4, Q4, #16 311 VSHR.U32 Q13, Q13, #16 316 VSHR.U32 Q2, Q2, #1 [all...] |
/system/tools/hidl/c2hal/ |
Expression.h | 37 U32 = 2, // 0b10 46 case U32: return "U32"; 58 case U32: return "uint32_t";
|
/external/lz4/tests/ |
frametest.c | 38 #include "util.h" /* U32 */ 50 static void FUZ_writeLE32 (void* dstVoidPtr, U32 value32) 69 static const U32 nbTestsDefault = 256 KB; 72 static const U32 prime1 = 2654435761U; 73 static const U32 prime2 = 2246822519U; 93 static U32 no_prompt = 0; 94 static U32 displayLevel = 2; 95 static U32 use_pause = 0; 113 U32 rand32 = *src; 124 static void FUZ_fillCompressibleNoiseBuffer(void* buffer, size_t bufferSize, double proba, U32* seed [all...] |
fuzzer.c | 40 #include "util.h" /* U32 */ 104 static U32 FUZ_rotl32(U32 u32, U32 nbBits) 106 return ((u32 << nbBits) | (u32 >> (32 - nbBits))); 109 static U32 FUZ_rand(U32* src) 111 U32 rand32 = *src [all...] |
datagencli.c | 30 #include "util.h" /* U32 */ 83 U32 seed = SEED_DEFAULT; 166 if (proba!=COMPRESSIBILITY_DEFAULT) DISPLAYLEVEL(3, "Compressibility : %i%%\n", (U32)(proba*100));
|
/external/lz4/lib/ |
xxhash.c | 138 typedef uint32_t U32; 143 typedef unsigned int U32; 151 static U32 XXH_read32(const void* memPtr) { return *(const U32*) memPtr; } 157 typedef union { U32 u32; } __attribute__((packed)) unalign; member in union:__anon28668 158 static U32 XXH_read32(const void* ptr) { return ((const unalign*)ptr)->u32; } 165 static U32 XXH_read32(const void* memPtr) 167 U32 val 545 typedef union { U32 u32; U64 u64; } __attribute__((packed)) unalign64; member in union:__anon28671 [all...] |
lz4.c | 141 typedef uint32_t U32; 148 typedef unsigned int U32; 165 const union { U32 u; BYTE c[4]; } one = { 1 }; /* don't use static : performance detrimental */ 174 static U32 LZ4_read32(const void* memPtr) { return *(const U32*) memPtr; } 178 static void LZ4_write32(void* memPtr, U32 value) { *(U32*)memPtr = value; } 184 typedef union { U16 u16; U32 u32; reg_t uArch; } __attribute__((packed)) unalign; member in union:__anon28641 187 static U32 LZ4_read32(const void* ptr) { return ((const unalign*)ptr)->u32; [all...] |
lz4opt.h | 89 U32* const HashTable = ctx->hashTable; 91 const U32 dictLimit = ctx->dictLimit; 92 const U32 current = (U32)(ip - base); 93 const U32 lowLimit = (ctx->lowLimit + MAX_DISTANCE > current) ? ctx->lowLimit : current - (MAX_DISTANCE - 1); 99 U32 matchIndex; 101 U32* HashPos; 173 const U32 target = (U32)(ip - base); 174 U32 idx = ctx->nextToUpdate [all...] |
lz4hc.c | 90 static U32 LZ4HC_hashPtr(const void* ptr) { return HASH_FUNCTION(LZ4_read32(ptr)); } 114 U32* const hashTable = hc4->hashTable; 116 U32 const target = (U32)(ip - base); 117 U32 idx = hc4->nextToUpdate; 120 U32 const h = LZ4HC_hashPtr(base+idx); 138 U32* const HashTable = hc4->hashTable; 141 const U32 dictLimit = hc4->dictLimit; 142 const U32 lowLimit = (hc4->lowLimit + 64 KB > (U32)(ip-base)) ? hc4->lowLimit : (U32)(ip - base) - (64 KB - 1) [all...] |
/external/vulkan-validation-layers/layers/ |
xxhash.c | 139 typedef uint32_t U32; 143 typedef unsigned int U32; 150 static U32 XXH_read32(const void* memPtr) { return *(const U32*) memPtr; } 156 typedef union { U32 u32; } __attribute__((packed)) unalign; member in union:__anon43580 157 static U32 XXH_read32(const void* ptr) { return ((const unalign*)ptr)->u32; } 164 static U32 XXH_read32(const void* memPtr) 166 U32 val 544 typedef union { U32 u32; U64 u64; } __attribute__((packed)) unalign64; member in union:__anon43583 [all...] |
/external/tensorflow/tensorflow/compiler/xla/ |
primitive_util.cc | 34 return type == U8 || type == U16 || type == U32 || type == U64; 56 case U32:
|
primitive_util.h | 70 return U32; 167 struct PrimitiveTypeToNative<U32> {
|
/frameworks/rs/tests/java_api/LatencyBenchmark/src/com/example/android/rs/computebench/ |
Benchmark.java | 31 ain = Allocation.createSized(rs, Element.U32(mRS), 10000); 32 aout = Allocation.createSized(rs, Element.U32(mRS), 10000);
|
/frameworks/rs/tests/cpp_api/typecheck/ |
typecheck.cpp | 56 TEST_ELEM(ENAME, U32) \ 68 TEST_ELEM_ALL(U32) 97 EXECUTE_TEST_ELEM_ALL(U32);
|
/external/skia/src/jumper/ |
SkJumper_stages.cpp | 64 using U32 = uint32_t; 76 SI U32 round (F v, F scale) { return (uint32_t)(v*scale + 0.5f); } 77 SI U16 pack(U32 v) { return (U16)v; } 83 SI T gather(const T* p, U32 ix) { return p[ix]; } 124 using U32 = V<uint32_t>; 134 SI U16 pack(U32 v) { return __builtin_convertvector(v, U16); } 137 SI F if_then_else(I32 c, F t, F e) { return vbslq_f32((U32)c,t,e); } 143 SI U32 round(F v, F scale) { return vcvtnq_u32_f32(v*scale); } 158 SI U32 round(F v, F scale) { 165 SI V<T> gather(const T* p, U32 ix) [all...] |
SkJumper_stages_lowp.cpp | 40 using U32 = uint32_t __attribute__((ext_vector_type(16))); 47 using U32 = uint32_t __attribute__((ext_vector_type(8))); 148 SI U32 if_then_else(I32 c, U32 t, U32 e) { return (t & c) | (e & ~c); } 195 SI U32 trunc_(F x) { return (U32)cast<I32>(x); } 407 SI U32 ix_and_ptr(T** ptr, const SkJumper_GatherCtx* ctx, F x, F y) { 409 limit = bit_cast<F>( bit_cast<U32>(limit) - 1 ); // Exclusive -> inclusive. 470 SI V gather(const T* ptr, U32 ix) [all...] |
/external/skqp/src/jumper/ |
SkJumper_stages.cpp | 64 using U32 = uint32_t; 76 SI U32 round (F v, F scale) { return (uint32_t)(v*scale + 0.5f); } 77 SI U16 pack(U32 v) { return (U16)v; } 83 SI T gather(const T* p, U32 ix) { return p[ix]; } 124 using U32 = V<uint32_t>; 134 SI U16 pack(U32 v) { return __builtin_convertvector(v, U16); } 137 SI F if_then_else(I32 c, F t, F e) { return vbslq_f32((U32)c,t,e); } 143 SI U32 round(F v, F scale) { return vcvtnq_u32_f32(v*scale); } 158 SI U32 round(F v, F scale) { 165 SI V<T> gather(const T* p, U32 ix) [all...] |