/frameworks/ml/nn/runtime/test/generated/models/ |
rnn.model.cpp | 6 OperandType type4(Type::TENSOR_FLOAT32, {2, 16}); 14 auto hidden_state_in = model->addOperand(&type4); 16 auto hidden_state_out = model->addOperand(&type4); 17 auto output = model->addOperand(&type4);
|
rnn_state.model.cpp | 6 OperandType type4(Type::TENSOR_FLOAT32, {2, 16}); 14 auto hidden_state_in = model->addOperand(&type4); 16 auto hidden_state_out = model->addOperand(&type4); 17 auto output = model->addOperand(&type4);
|
lstm2.model.cpp | 11 OperandType type4(Type::TENSOR_FLOAT32, {4}); 24 auto cell_to_forget_weights = model->addOperand(&type4); 25 auto cell_to_output_weights = model->addOperand(&type4); 26 auto input_gate_bias = model->addOperand(&type4); 27 auto forget_gate_bias = model->addOperand(&type4); 28 auto cell_gate_bias = model->addOperand(&type4); 29 auto output_gate_bias = model->addOperand(&type4);
|
lstm2_state.model.cpp | 11 OperandType type4(Type::TENSOR_FLOAT32, {4}); 24 auto cell_to_forget_weights = model->addOperand(&type4); 25 auto cell_to_output_weights = model->addOperand(&type4); 26 auto input_gate_bias = model->addOperand(&type4); 27 auto forget_gate_bias = model->addOperand(&type4); 28 auto cell_gate_bias = model->addOperand(&type4); 29 auto output_gate_bias = model->addOperand(&type4);
|
lstm2_state2.model.cpp | 11 OperandType type4(Type::TENSOR_FLOAT32, {4}); 24 auto cell_to_forget_weights = model->addOperand(&type4); 25 auto cell_to_output_weights = model->addOperand(&type4); 26 auto input_gate_bias = model->addOperand(&type4); 27 auto forget_gate_bias = model->addOperand(&type4); 28 auto cell_gate_bias = model->addOperand(&type4); 29 auto output_gate_bias = model->addOperand(&type4);
|
fully_connected_quant8_weights_as_inputs.model.cpp | 3 OperandType type4(Type::INT32, {}); 13 auto act = model->addOperand(&type4);
|
hashtable_lookup_float.model.cpp | 7 OperandType type4(Type::TENSOR_QUANT8_ASYMM, {4}, 1.f, 0); 13 auto hits = model->addOperand(&type4);
|
hashtable_lookup_quant8.model.cpp | 7 OperandType type4(Type::TENSOR_QUANT8_ASYMM, {4}, 1.f, 0); 13 auto hits = model->addOperand(&type4);
|
lsh_projection_weights_as_inputs.model.cpp | 7 OperandType type4(Type::TENSOR_INT32, {8}); 13 auto output = model->addOperand(&type4);
|
svdf.model.cpp | 4 OperandType type4(Type::TENSOR_FLOAT32, {2, 40}); 15 auto state_in = model->addOperand(&type4); 18 auto state_out = model->addOperand(&type4);
|
svdf_state.model.cpp | 4 OperandType type4(Type::TENSOR_FLOAT32, {2, 40}); 15 auto state_in = model->addOperand(&type4); 18 auto state_out = model->addOperand(&type4);
|
lstm.model.cpp | 11 OperandType type4(Type::TENSOR_FLOAT32, {4}); 26 auto input_gate_bias = model->addOperand(&type4); 27 auto forget_gate_bias = model->addOperand(&type4); 28 auto cell_gate_bias = model->addOperand(&type4); 29 auto output_gate_bias = model->addOperand(&type4);
|
lstm_state.model.cpp | 11 OperandType type4(Type::TENSOR_FLOAT32, {4}); 26 auto input_gate_bias = model->addOperand(&type4); 27 auto forget_gate_bias = model->addOperand(&type4); 28 auto cell_gate_bias = model->addOperand(&type4); 29 auto output_gate_bias = model->addOperand(&type4);
|
lstm_state2.model.cpp | 11 OperandType type4(Type::TENSOR_FLOAT32, {4}); 26 auto input_gate_bias = model->addOperand(&type4); 27 auto forget_gate_bias = model->addOperand(&type4); 28 auto cell_gate_bias = model->addOperand(&type4); 29 auto output_gate_bias = model->addOperand(&type4);
|
conv_1_h3_w2_SAME.model.cpp | 6 OperandType type4(Type::TENSOR_FLOAT32, {1}); 16 auto op1 = model->addOperand(&type4);
|
conv_1_h3_w2_VALID.model.cpp | 6 OperandType type4(Type::TENSOR_FLOAT32, {1}); 15 auto op1 = model->addOperand(&type4);
|
conv_3_h3_w2_VALID.model.cpp | 6 OperandType type4(Type::TENSOR_FLOAT32, {3}); 15 auto op1 = model->addOperand(&type4);
|
conv_quant8.model.cpp | 6 OperandType type4(Type::TENSOR_QUANT8_ASYMM, {1, 2, 2, 1}, 1.f, 0); 15 auto op4 = model->addOperand(&type4);
|
conv_quant8_channels.model.cpp | 6 OperandType type4(Type::TENSOR_QUANT8_ASYMM, {1, 1, 1, 3}, 1.0, 0); 15 auto op4 = model->addOperand(&type4);
|
conv_quant8_channels_weights_as_inputs.model.cpp | 6 OperandType type4(Type::TENSOR_QUANT8_ASYMM, {1, 1, 1, 3}, 1.0, 0); 15 auto op4 = model->addOperand(&type4);
|
conv_quant8_large.model.cpp | 6 OperandType type4(Type::TENSOR_QUANT8_ASYMM, {1, 2, 3, 3}, 1.0, 0); 15 auto op4 = model->addOperand(&type4);
|
conv_quant8_large_weights_as_inputs.model.cpp | 6 OperandType type4(Type::TENSOR_QUANT8_ASYMM, {1, 2, 3, 3}, 1.0, 0); 15 auto op4 = model->addOperand(&type4);
|
conv_quant8_overflow.model.cpp | 6 OperandType type4(Type::TENSOR_QUANT8_ASYMM, {1, 2, 3, 3}, 1.0, 0); 15 auto op4 = model->addOperand(&type4);
|
/frameworks/base/cmds/interrupter/ |
interrupter.h | 50 #define CALL_FUNCTION_4(sym, ret, type1, type2, type3, type4) \ 51 ret (*real_##sym)(type1, type2, type3, type4) = NULL; \ 52 ret sym(type1 arg1, type2 arg2, type3 arg3, type4 arg4) { \ 57 #define CALL_FUNCTION_5(sym, ret, type1, type2, type3, type4, type5) \ 58 ret (*real_##sym)(type1, type2, type3, type4, type5) = NULL; \ 59 ret sym(type1 arg1, type2 arg2, type3 arg3, type4 arg4, type5 arg5) { \
|
/external/clang/test/SemaTemplate/ |
ms-if-exists.cpp | 22 typedef T::X type4; typedef 29 X<int>::type4 i4; // expected-error{{no type named 'type4' in 'X<int>'}}
|