1 // clang-format off 2 // Generated file (from: softmax_quant8_2.mod.py). Do not edit 3 void CreateModel(Model *model) { 4 OperandType type0(Type::TENSOR_QUANT8_ASYMM, {2, 5}, 0.5f, 0); 5 OperandType type1(Type::FLOAT32, {}); 6 OperandType type2(Type::TENSOR_QUANT8_ASYMM, {2, 5}, 0.00390625f, 0); 7 // Phase 1, operands 8 auto input = model->addOperand(&type0); 9 auto beta = model->addOperand(&type1); 10 auto output = model->addOperand(&type2); 11 // Phase 2, operations 12 static float beta_init[] = {1.0f}; 13 model->setOperandValue(beta, beta_init, sizeof(float) * 1); 14 model->addOperation(ANEURALNETWORKS_SOFTMAX, {input, beta}, {output}); 15 // Phase 3, inputs and outputs 16 model->identifyInputsAndOutputs( 17 {input}, 18 {output}); 19 assert(model->isValid()); 20 } 21 22 inline bool is_ignored(int i) { 23 static std::set<int> ignore = {}; 24 return ignore.find(i) != ignore.end(); 25 } 26 27 void CreateModel_dynamic_output_shape(Model *model) { 28 OperandType type0(Type::TENSOR_QUANT8_ASYMM, {2, 5}, 0.5f, 0); 29 OperandType type1(Type::FLOAT32, {}); 30 OperandType type3(Type::TENSOR_QUANT8_ASYMM, {0, 0}, 0.00390625f, 0); 31 // Phase 1, operands 32 auto input = model->addOperand(&type0); 33 auto beta = model->addOperand(&type1); 34 auto output = model->addOperand(&type3); 35 // Phase 2, operations 36 static float beta_init[] = {1.0f}; 37 model->setOperandValue(beta, beta_init, sizeof(float) * 1); 38 model->addOperation(ANEURALNETWORKS_SOFTMAX, {input, beta}, {output}); 39 // Phase 3, inputs and outputs 40 model->identifyInputsAndOutputs( 41 {input}, 42 {output}); 43 assert(model->isValid()); 44 } 45 46 inline bool is_ignored_dynamic_output_shape(int i) { 47 static std::set<int> ignore = {}; 48 return ignore.find(i) != ignore.end(); 49 } 50 51