/frameworks/ml/nn/runtime/test/specs/V1_0/ |
lstm2_state.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, n_cell * 3)) variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 132 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 3) ],
|
lstm.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, (n_cell * 4))) variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 140 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 4) ],
|
lstm2.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, n_cell * 3)) variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 132 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 3) ],
|
lstm2_state2.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, n_cell * 3)) variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 132 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 3) ],
|
lstm3.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, (n_cell * 4))) variable 95 ).To([scratch_buffer, output_state_out, cell_state_out, output]) [all...] |
lstm3_state.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, (n_cell * 4))) variable 95 ).To([scratch_buffer, output_state_out, cell_state_out, output]) [all...] |
lstm3_state2.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, (n_cell * 4))) variable 95 ).To([scratch_buffer, output_state_out, cell_state_out, output]) [all...] |
lstm3_state3.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, (n_cell * 4))) variable 95 ).To([scratch_buffer, output_state_out, cell_state_out, output]) [all...] |
lstm_state.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, (n_cell * 4))) variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 140 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 4) ],
|
lstm_state2.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, (n_cell * 4))) variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 140 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 4) ],
|
/frameworks/ml/nn/runtime/test/specs/V1_1/ |
lstm2_relaxed.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, n_cell * 3)) variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 133 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 3) ],
|
lstm2_state2_relaxed.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, n_cell * 3)) variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 133 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 3) ],
|
lstm2_state_relaxed.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, n_cell * 3)) variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 133 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 3) ],
|
lstm3_relaxed.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, (n_cell * 4))) variable 95 ).To([scratch_buffer, output_state_out, cell_state_out, output]) [all...] |
lstm3_state2_relaxed.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, (n_cell * 4))) variable 95 ).To([scratch_buffer, output_state_out, cell_state_out, output]) [all...] |
lstm3_state3_relaxed.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, (n_cell * 4))) variable 95 ).To([scratch_buffer, output_state_out, cell_state_out, output]) [all...] |
lstm3_state_relaxed.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, (n_cell * 4))) variable 95 ).To([scratch_buffer, output_state_out, cell_state_out, output]) [all...] |
lstm_relaxed.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, (n_cell * 4))) variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 141 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 4) ],
|
lstm_state2_relaxed.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, (n_cell * 4))) variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 141 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 4) ],
|
lstm_state_relaxed.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, (n_cell * 4))) variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 141 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 4) ],
|
/frameworks/ml/nn/tools/test_generator/tests/P_lstm/ |
lstm.mod.py | 58 scratch_buffer = IgnoredOutput("scratch_buffer", "TENSOR_FLOAT32", "{%d, %d}" % (n_batch, (n_cell * 4))) variable 94 ).To([scratch_buffer, output_state_out, cell_state_out, output]) 153 scratch_buffer: [ 0 for x in range(n_batch * n_cell * 4) ],
|
/external/mesa3d/src/gallium/drivers/radeonsi/ |
si_cp_dma.c | 278 if (!sctx->scratch_buffer || 279 sctx->scratch_buffer->b.b.width0 < scratch_size) { 280 r600_resource_reference(&sctx->scratch_buffer, NULL); 281 sctx->scratch_buffer = (struct r600_resource*) 284 if (!sctx->scratch_buffer) 289 si_cp_dma_prepare(sctx, &sctx->scratch_buffer->b.b, 290 &sctx->scratch_buffer->b.b, size, size, user_flags, 293 va = sctx->scratch_buffer->gpu_address;
|
/frameworks/ml/nn/runtime/test/generated/models/ |
lstm.model.cpp | 37 auto scratch_buffer = model->addOperand(&type9); local 48 model->addOperation(ANEURALNETWORKS_LSTM, {input, input_to_input_weights, input_to_forget_weights, input_to_cell_weights, input_to_output_weights, recurrent_to_intput_weights, recurrent_to_forget_weights, recurrent_to_cell_weights, recurrent_to_output_weights, cell_to_input_weights, cell_to_forget_weights, cell_to_output_weights, input_gate_bias, forget_gate_bias, cell_gate_bias, output_gate_bias, projection_weights, projection_bias, output_state_in, cell_state_in, activation_param, cell_clip_param, proj_clip_param}, {scratch_buffer, output_state_out, cell_state_out, output}); 52 {scratch_buffer, output_state_out, cell_state_out, output});
|
lstm2.model.cpp | 37 auto scratch_buffer = model->addOperand(&type9); local 48 model->addOperation(ANEURALNETWORKS_LSTM, {input, input_to_input_weights, input_to_forget_weights, input_to_cell_weights, input_to_output_weights, recurrent_to_intput_weights, recurrent_to_forget_weights, recurrent_to_cell_weights, recurrent_to_output_weights, cell_to_input_weights, cell_to_forget_weights, cell_to_output_weights, input_gate_bias, forget_gate_bias, cell_gate_bias, output_gate_bias, projection_weights, projection_bias, output_state_in, cell_state_in, activation_param, cell_clip_param, proj_clip_param}, {scratch_buffer, output_state_out, cell_state_out, output}); 52 {scratch_buffer, output_state_out, cell_state_out, output});
|
lstm2_relaxed.model.cpp | 37 auto scratch_buffer = model->addOperand(&type9); local 48 model->addOperation(ANEURALNETWORKS_LSTM, {input, input_to_input_weights, input_to_forget_weights, input_to_cell_weights, input_to_output_weights, recurrent_to_intput_weights, recurrent_to_forget_weights, recurrent_to_cell_weights, recurrent_to_output_weights, cell_to_input_weights, cell_to_forget_weights, cell_to_output_weights, input_gate_bias, forget_gate_bias, cell_gate_bias, output_gate_bias, projection_weights, projection_bias, output_state_in, cell_state_in, activation_param, cell_clip_param, proj_clip_param}, {scratch_buffer, output_state_out, cell_state_out, output}); 52 {scratch_buffer, output_state_out, cell_state_out, output});
|