/frameworks/ml/nn/runtime/test/specs/V1_1/ |
space_to_batch_quant8_2.mod.py | 3 block = Parameter("block_size", "TENSOR_INT32", "{2}", [3, 2]) variable 7 model = model.Operation("SPACE_TO_BATCH_ND", i1, block, paddings).To(output)
|
space_to_batch_quant8_3.mod.py | 3 block = Parameter("block_size", "TENSOR_INT32", "{2}", [3, 2]) variable 7 model = model.Operation("SPACE_TO_BATCH_ND", i1, block, paddings).To(output)
|
space_to_batch_relaxed.mod.py | 19 block = Parameter("block_size", "TENSOR_INT32", "{2}", [2, 2]) variable 23 model = model.Operation("SPACE_TO_BATCH_ND", i1, block, paddings).To(output)
|
space_to_depth_float_1_relaxed.mod.py | 19 block = Int32Scalar("block_size", 2) variable 22 model = model.Operation("SPACE_TO_DEPTH", i1, block).To(output)
|
space_to_depth_float_2_relaxed.mod.py | 19 block = Int32Scalar("block_size", 2) variable 22 model = model.Operation("SPACE_TO_DEPTH", i1, block).To(output)
|
space_to_depth_float_3_relaxed.mod.py | 19 block = Int32Scalar("block_size", 2) variable 22 model = model.Operation("SPACE_TO_DEPTH", i1, block).To(output)
|
/prebuilts/misc/common/swig/include/2.0.11/octave/ |
exception.i | 5 %define_as(SWIG_exception(code, msg), %block(%error(code, msg); SWIG_fail; ))
|
/prebuilts/misc/common/swig/include/2.0.11/perl5/ |
exception.i | 4 %define_as(SWIG_exception(code, msg), %block(%error(code, msg); SWIG_fail; ))
|
/prebuilts/misc/common/swig/include/2.0.11/python/ |
exception.i | 5 %define_as(SWIG_exception(code, msg), %block(%error(code, msg); SWIG_fail; ))
|
/prebuilts/misc/common/swig/include/2.0.11/ruby/ |
exception.i | 4 %define_as(SWIG_exception(code, msg), %block(%error(code, msg);))
|
/prebuilts/misc/common/swig/include/2.0.11/tcl/ |
exception.i | 5 %define_as(SWIG_exception(code, msg), %block(%error(code, msg); return TCL_ERROR;))
|
/system/nfc/src/nfa/include/ |
nfa_nv_co.h | 79 extern void nfa_nv_co_read(uint8_t* p_buf, uint16_t nbytes, uint8_t block); 101 uint8_t block);
|
/test/vts/utils/python/coverage/ |
function_summary.py | 22 blocks: list of BlockSummary objects for each block in the function. 34 first line number in the object attributes. Initializes the block 51 """Resolves the block and arc counts. 54 counts are resolved in the unresolved arcs. Then, block 56 the block. 63 for block in self.blocks: 64 for arc in block.exit_arcs: 83 # Resolve the block counts 84 for block in self.blocks: 85 if len(block.entry_arcs) [all...] |
/toolchain/binutils/binutils-2.27/gas/testsuite/gas/arm/ |
arm-it-bad-3.l | 2 [^:]*:4: Error: thumb conditional instruction should be in IT block -- `moveq r1,r8' 3 [^:]*:5: Error: thumb conditional instruction should be in IT block -- `movne r1,r9'
|
arm-it-bad.l | 2 [^:]*:8: Error: incorrect condition in IT block -- `moveq r0,r1' 3 [^:]*:10: Warning: conditional outside an IT block for Thumb.
|
insn-error-t.l | 2 [^:]*:4: Error: thumb conditional instruction should be in IT block -- `movne r1,r9'
|
/external/mesa3d/src/compiler/nir/ |
nir_opt_cse.c | 36 * Visits and CSE's the given block and all its descendants in the dominance 38 * contain instructions that dominate the current block. 42 cse_block(nir_block *block, struct set *instr_set) 46 nir_foreach_instr_safe(instr, block) { 53 for (unsigned i = 0; i < block->num_dom_children; i++) { 54 nir_block *child = block->dom_children[i]; 58 nir_foreach_instr(instr, block)
|
/external/syslinux/core/fs/ufs/ |
bmap.c | 46 * Block spans 8 fragments, then address is interleaved by 8. 83 * The actual indirect block map handling - the block passed in should 84 * be relative to the beginning of the particular block hierarchy. 86 * @shft_per_blk: shift to get nr. of addresses in a block. 87 * @mask_per_blk: mask to limit the max nr. of addresses in a block. 88 * @addr_count: nr. of addresses in a block. 91 bmap_indirect(struct fs_info *fs, uint64_t start, uint32_t block, 108 index = (block >> (levels * shft_per_blk)) & mask_per_blk; 117 * Handle the traditional block map, like indirect, double indirec 178 block_t block; local [all...] |
/external/syslinux/core/fs/ext2/ |
bmap.c | 2 * The logical block -> physical block routine. 17 block_t block) 31 if (block < index[i].ei_block) 43 /* handle the ext4 extents to get the phsical block number */ 46 bmap_extent(struct inode *inode, uint32_t block, size_t *nblocks) 54 leaf = ext4_find_leaf(fs, &PVT(inode)->i_extent_hdr, block); 62 if (block < ext[i].ee_block) 66 printf("ERROR, not find the right block\n"); 71 block -= ext[i].ee_block 215 block_t block; local [all...] |
/system/update_engine/scripts/ |
test_paycheck.sh | 24 # - It performs a random block trace on the delta payload (both kernel and 43 # payload reports and block traces will improve this our confidence and are 85 block=$2 86 time ${paycheck} -B ${block} ${payload_file} 91 block=$2 92 time ${paycheck} -b ${block} ${payload_file} 138 # Trace a random block between 0-1024 on all payloads. 139 block=$((RANDOM * 1024 / 32767)) 140 log "Tracing a random block (${block}) in full/delta payloads... [all...] |
/art/compiler/optimizing/ |
dead_code_elimination.cc | 39 HBasicBlock* block = worklist.back(); local 41 int block_id = block->GetBlockId(); 44 ArrayRef<HBasicBlock* const> live_successors(block->GetSuccessors()); 45 HInstruction* last_instruction = block->GetLastInstruction(); 72 DCHECK_EQ(live_successors[0], block->GetSuccessors()[switch_index]); 90 void HDeadCodeElimination::MaybeRecordDeadBlock(HBasicBlock* block) { 93 block->GetPhis().CountSize() + block->GetInstructions().CountSize()); 149 static bool RemoveNonNullControlDependences(HBasicBlock* block, HBasicBlock* throws) { 151 if (!block->EndsWithIf()) 430 HBasicBlock* block = graph_->GetReversePostOrder()[i]; local [all...] |
/external/fec/ |
rstest.c | 2 * for various block sizes and with random data and random error patterns 69 unsigned char block[nn],tblock[nn]; local 85 /* Load block with random data and encode */ 87 block[i] = random() & nn; 88 memcpy(tblock,block,sizeof(block)); 89 encode_rs_8(block,&block[kk],0); 92 memcpy(tblock,block,sizeof(block)); 141 unsigned char block[nn],tblock[nn]; local 222 int block[nn],tblock[nn]; local [all...] |
/external/v8/src/crankshaft/ |
hydrogen-flow-engine.h | 52 // over a graph of basic blocks, either one block at a time (local analysis) 62 // each block at most twice; once for state, and optionally once for effects. 81 // Local analysis. Iterates over the instructions in the given block. 82 State* AnalyzeOneBlock(HBasicBlock* block, State* state) { 83 // Go through all instructions of the current block, updating the state. 84 for (HInstructionIterator it(block); !it.Done(); it.Advance()) { 91 // block, starting with the initial state. Computes effects for nested loops. 96 // Iterate all dominated blocks starting from the given start block. 98 HBasicBlock* block = graph_->blocks()->at(i); local 101 if (SkipNonDominatedBlock(root, block)) continue [all...] |
/external/valgrind/none/tests/arm/ |
v8crypto_a.c | 90 //static void showBlock ( const char* msg, V128* block, Int nBlock ) 96 // showV128(&block[i]); 120 V128 block[4+1]; \ 121 memset(block, 0x55, sizeof(block)); \ 122 randV128(&block[0], ty); \ 123 randV128(&block[1], ty); \ 124 randV128(&block[2], ty); \ 125 randV128(&block[3], ty); \ 134 : : "r"(&block[0]) [all...] |
v8crypto_t.c | 90 //static void showBlock ( const char* msg, V128* block, Int nBlock ) 96 // showV128(&block[i]); 120 V128 block[4+1]; \ 121 memset(block, 0x55, sizeof(block)); \ 122 randV128(&block[0], ty); \ 123 randV128(&block[1], ty); \ 124 randV128(&block[2], ty); \ 125 randV128(&block[3], ty); \ 134 : : "r"(&block[0]) [all...] |