Home | History | Annotate | Download | only in models
      1 // clang-format off
      2 // Generated file (from: split_float_4.mod.py). Do not edit
      3 void CreateModel(Model *model) {
      4   OperandType type0(Type::TENSOR_FLOAT32, {2, 2, 2});
      5   OperandType type1(Type::INT32, {});
      6   OperandType type2(Type::TENSOR_FLOAT32, {2, 1, 2});
      7   // Phase 1, operands
      8   auto input0 = model->addOperand(&type0);
      9   auto axis = model->addOperand(&type1);
     10   auto num_splits = model->addOperand(&type1);
     11   auto output0 = model->addOperand(&type2);
     12   auto output1 = model->addOperand(&type2);
     13   // Phase 2, operations
     14   static int32_t axis_init[] = {1};
     15   model->setOperandValue(axis, axis_init, sizeof(int32_t) * 1);
     16   static int32_t num_splits_init[] = {2};
     17   model->setOperandValue(num_splits, num_splits_init, sizeof(int32_t) * 1);
     18   model->addOperation(ANEURALNETWORKS_SPLIT, {input0, axis, num_splits}, {output0, output1});
     19   // Phase 3, inputs and outputs
     20   model->identifyInputsAndOutputs(
     21     {input0},
     22     {output0, output1});
     23   assert(model->isValid());
     24 }
     25 
     26 inline bool is_ignored(int i) {
     27   static std::set<int> ignore = {};
     28   return ignore.find(i) != ignore.end();
     29 }
     30 
     31 void CreateModel_relaxed(Model *model) {
     32   OperandType type0(Type::TENSOR_FLOAT32, {2, 2, 2});
     33   OperandType type1(Type::INT32, {});
     34   OperandType type2(Type::TENSOR_FLOAT32, {2, 1, 2});
     35   // Phase 1, operands
     36   auto input0 = model->addOperand(&type0);
     37   auto axis = model->addOperand(&type1);
     38   auto num_splits = model->addOperand(&type1);
     39   auto output0 = model->addOperand(&type2);
     40   auto output1 = model->addOperand(&type2);
     41   // Phase 2, operations
     42   static int32_t axis_init[] = {1};
     43   model->setOperandValue(axis, axis_init, sizeof(int32_t) * 1);
     44   static int32_t num_splits_init[] = {2};
     45   model->setOperandValue(num_splits, num_splits_init, sizeof(int32_t) * 1);
     46   model->addOperation(ANEURALNETWORKS_SPLIT, {input0, axis, num_splits}, {output0, output1});
     47   // Phase 3, inputs and outputs
     48   model->identifyInputsAndOutputs(
     49     {input0},
     50     {output0, output1});
     51   // Phase 4: set relaxed execution
     52   model->relaxComputationFloat32toFloat16(true);
     53   assert(model->isValid());
     54 }
     55 
     56 inline bool is_ignored_relaxed(int i) {
     57   static std::set<int> ignore = {};
     58   return ignore.find(i) != ignore.end();
     59 }
     60 
     61 void CreateModel_float16(Model *model) {
     62   OperandType type1(Type::INT32, {});
     63   OperandType type3(Type::TENSOR_FLOAT16, {2, 2, 2});
     64   OperandType type4(Type::TENSOR_FLOAT16, {2, 1, 2});
     65   // Phase 1, operands
     66   auto input0 = model->addOperand(&type3);
     67   auto axis = model->addOperand(&type1);
     68   auto num_splits = model->addOperand(&type1);
     69   auto output0 = model->addOperand(&type4);
     70   auto output1 = model->addOperand(&type4);
     71   // Phase 2, operations
     72   static int32_t axis_init[] = {1};
     73   model->setOperandValue(axis, axis_init, sizeof(int32_t) * 1);
     74   static int32_t num_splits_init[] = {2};
     75   model->setOperandValue(num_splits, num_splits_init, sizeof(int32_t) * 1);
     76   model->addOperation(ANEURALNETWORKS_SPLIT, {input0, axis, num_splits}, {output0, output1});
     77   // Phase 3, inputs and outputs
     78   model->identifyInputsAndOutputs(
     79     {input0},
     80     {output0, output1});
     81   assert(model->isValid());
     82 }
     83 
     84 inline bool is_ignored_float16(int i) {
     85   static std::set<int> ignore = {};
     86   return ignore.find(i) != ignore.end();
     87 }
     88 
     89 void CreateModel_dynamic_output_shape(Model *model) {
     90   OperandType type0(Type::TENSOR_FLOAT32, {2, 2, 2});
     91   OperandType type1(Type::INT32, {});
     92   OperandType type5(Type::TENSOR_FLOAT32, {0, 0, 0});
     93   // Phase 1, operands
     94   auto input0 = model->addOperand(&type0);
     95   auto axis = model->addOperand(&type1);
     96   auto num_splits = model->addOperand(&type1);
     97   auto output0 = model->addOperand(&type5);
     98   auto output1 = model->addOperand(&type5);
     99   // Phase 2, operations
    100   static int32_t axis_init[] = {1};
    101   model->setOperandValue(axis, axis_init, sizeof(int32_t) * 1);
    102   static int32_t num_splits_init[] = {2};
    103   model->setOperandValue(num_splits, num_splits_init, sizeof(int32_t) * 1);
    104   model->addOperation(ANEURALNETWORKS_SPLIT, {input0, axis, num_splits}, {output0, output1});
    105   // Phase 3, inputs and outputs
    106   model->identifyInputsAndOutputs(
    107     {input0},
    108     {output0, output1});
    109   assert(model->isValid());
    110 }
    111 
    112 inline bool is_ignored_dynamic_output_shape(int i) {
    113   static std::set<int> ignore = {};
    114   return ignore.find(i) != ignore.end();
    115 }
    116 
    117 void CreateModel_dynamic_output_shape_relaxed(Model *model) {
    118   OperandType type0(Type::TENSOR_FLOAT32, {2, 2, 2});
    119   OperandType type1(Type::INT32, {});
    120   OperandType type5(Type::TENSOR_FLOAT32, {0, 0, 0});
    121   // Phase 1, operands
    122   auto input0 = model->addOperand(&type0);
    123   auto axis = model->addOperand(&type1);
    124   auto num_splits = model->addOperand(&type1);
    125   auto output0 = model->addOperand(&type5);
    126   auto output1 = model->addOperand(&type5);
    127   // Phase 2, operations
    128   static int32_t axis_init[] = {1};
    129   model->setOperandValue(axis, axis_init, sizeof(int32_t) * 1);
    130   static int32_t num_splits_init[] = {2};
    131   model->setOperandValue(num_splits, num_splits_init, sizeof(int32_t) * 1);
    132   model->addOperation(ANEURALNETWORKS_SPLIT, {input0, axis, num_splits}, {output0, output1});
    133   // Phase 3, inputs and outputs
    134   model->identifyInputsAndOutputs(
    135     {input0},
    136     {output0, output1});
    137   // Phase 4: set relaxed execution
    138   model->relaxComputationFloat32toFloat16(true);
    139   assert(model->isValid());
    140 }
    141 
    142 inline bool is_ignored_dynamic_output_shape_relaxed(int i) {
    143   static std::set<int> ignore = {};
    144   return ignore.find(i) != ignore.end();
    145 }
    146 
    147 void CreateModel_dynamic_output_shape_float16(Model *model) {
    148   OperandType type1(Type::INT32, {});
    149   OperandType type3(Type::TENSOR_FLOAT16, {2, 2, 2});
    150   OperandType type6(Type::TENSOR_FLOAT16, {0, 0, 0});
    151   // Phase 1, operands
    152   auto input0 = model->addOperand(&type3);
    153   auto axis = model->addOperand(&type1);
    154   auto num_splits = model->addOperand(&type1);
    155   auto output0 = model->addOperand(&type6);
    156   auto output1 = model->addOperand(&type6);
    157   // Phase 2, operations
    158   static int32_t axis_init[] = {1};
    159   model->setOperandValue(axis, axis_init, sizeof(int32_t) * 1);
    160   static int32_t num_splits_init[] = {2};
    161   model->setOperandValue(num_splits, num_splits_init, sizeof(int32_t) * 1);
    162   model->addOperation(ANEURALNETWORKS_SPLIT, {input0, axis, num_splits}, {output0, output1});
    163   // Phase 3, inputs and outputs
    164   model->identifyInputsAndOutputs(
    165     {input0},
    166     {output0, output1});
    167   assert(model->isValid());
    168 }
    169 
    170 inline bool is_ignored_dynamic_output_shape_float16(int i) {
    171   static std::set<int> ignore = {};
    172   return ignore.find(i) != ignore.end();
    173 }
    174 
    175