23450 lines
693 KiB
XML
23450 lines
693 KiB
XML
<?xml version="1.0"?>
|
|
<net name="Model0" version="11">
|
|
<layers>
|
|
<layer id="0" name="x" type="Parameter" version="opset1">
|
|
<data shape="1,3,640,640" element_type="f32" />
|
|
<output>
|
|
<port id="0" precision="FP32" names="x">
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>640</dim>
|
|
<dim>640</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1" name="__module.model.23/aten::unsqueeze/Unsqueeze" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 2, 8400" offset="0" size="67200" />
|
|
<output>
|
|
<port id="0" precision="FP32" names="1361,anchor_points">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="2" name="x/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="67200" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="3" name="x/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="67204" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="4" name="x/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="67200" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="5" name="x/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="67204" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="6" name="x/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>640</dim>
|
|
<dim>640</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>640</dim>
|
|
<dim>640</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="7" name="Constant_527392" type="Const" version="opset1">
|
|
<data element_type="i8" shape="16, 3, 3, 3" offset="67208" size="432" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="8" name="Convert_527393" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="9" name="Constant_527394" type="Const" version="opset1">
|
|
<data element_type="f32" shape="16, 1, 1, 1" offset="67640" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="10" name="__module.model.0.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="11" name="__module.model.0.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="2, 2" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>640</dim>
|
|
<dim>640</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>320</dim>
|
|
<dim>320</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="12" name="__module.model.0.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 16, 1, 1" offset="67704" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="13" name="__module.model.0.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>320</dim>
|
|
<dim>320</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="92_1">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>320</dim>
|
|
<dim>320</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="14" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>320</dim>
|
|
<dim>320</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="92,input.1">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>320</dim>
|
|
<dim>320</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="15" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="67768" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="16" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="67772" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="17" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="67768" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="18" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="67772" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="19" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>320</dim>
|
|
<dim>320</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>320</dim>
|
|
<dim>320</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="20" name="Constant_527396" type="Const" version="opset1">
|
|
<data element_type="i8" shape="32, 16, 3, 3" offset="67776" size="4608" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="21" name="Convert_527397" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="22" name="Constant_527398" type="Const" version="opset1">
|
|
<data element_type="f32" shape="32, 1, 1, 1" offset="72384" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="23" name="__module.model.1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="24" name="__module.model.1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="2, 2" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>320</dim>
|
|
<dim>320</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="25" name="__module.model.1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 32, 1, 1" offset="72512" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="26" name="__module.model.1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="106_1">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="27" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_1" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="106,input.5">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="28" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_1/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="72640" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="29" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_1/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="72644" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="30" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_1/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="72640" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="31" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_1/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="72644" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="32" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_1/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="33" name="Constant_527400" type="Const" version="opset1">
|
|
<data element_type="i8" shape="32, 32, 1, 1" offset="72648" size="1024" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="34" name="Convert_527401" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="35" name="Constant_527402" type="Const" version="opset1">
|
|
<data element_type="f32" shape="32, 1, 1, 1" offset="73672" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="36" name="__module.model.2.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="37" name="__module.model.2.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="38" name="__module.model.2.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 32, 1, 1" offset="73800" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="39" name="__module.model.2.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="124_1">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="40" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_2" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="124,input.9">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="41" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_2/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="73928" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="42" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_2/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="73932" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="43" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_2/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="73928" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="44" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_2/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="73932" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="45" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_2/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="46" name="112" type="Const" version="opset1">
|
|
<data element_type="i64" shape="" offset="73936" size="8" />
|
|
<output>
|
|
<port id="0" precision="I64" names="112" />
|
|
</output>
|
|
</layer>
|
|
<layer id="47" name="Constant_215" type="Const" version="opset1">
|
|
<data element_type="i64" shape="2" offset="73944" size="16" />
|
|
<output>
|
|
<port id="0" precision="I64" names="126">
|
|
<dim>2</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="48" name="__module.model.2/prim::ListUnpack" type="VariadicSplit" version="opset1">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="1" precision="I64" />
|
|
<port id="2" precision="I64">
|
|
<dim>2</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="128">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="4" precision="FP32" names="129,input.11">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="49" name="Constant_527404" type="Const" version="opset1">
|
|
<data element_type="i8" shape="8, 16, 3, 3" offset="73960" size="1152" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="50" name="Convert_527405" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="51" name="Constant_527406" type="Const" version="opset1">
|
|
<data element_type="f32" shape="8, 1, 1, 1" offset="75112" size="32" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="52" name="__module.model.2.m.0.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="53" name="__module.model.2.m.0.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="54" name="__module.model.2.m.0.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 8, 1, 1" offset="75144" size="32" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="55" name="__module.model.2.m.0.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="139_1">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="56" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_3" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="139,input.13">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="57" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_3/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="58" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_3/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75180" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="59" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_3/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="60" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_3/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75180" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="61" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_3/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="62" name="Constant_527408" type="Const" version="opset1">
|
|
<data element_type="i8" shape="16, 8, 3, 3" offset="75184" size="1152" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>8</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="63" name="Convert_527409" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>8</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>8</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="64" name="Constant_527410" type="Const" version="opset1">
|
|
<data element_type="f32" shape="16, 1, 1, 1" offset="76336" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="65" name="__module.model.2.m.0.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>8</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>8</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="66" name="__module.model.2.m.0.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>8</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="67" name="__module.model.2.m.0.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 16, 1, 1" offset="76400" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="68" name="__module.model.2.m.0.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="148_1">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="69" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_4" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="148,input.17">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="70" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_4/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="76464" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="71" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_4/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="76468" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="72" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_4/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="76464" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="73" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_4/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="76468" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="74" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_4/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="75" name="__module.model.2.m.0/aten::add/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="150">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="76" name="__module.model.2.m.0/aten::add/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="73928" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="77" name="__module.model.2.m.0/aten::add/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="73932" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="78" name="__module.model.2.m.0/aten::add/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="73928" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="79" name="__module.model.2.m.0/aten::add/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="73932" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="80" name="__module.model.2.m.0/aten::add/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="81" name="__module.model.2/aten::cat/Concat" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="152,input.19">
|
|
<dim>1</dim>
|
|
<dim>48</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="82" name="Constant_527412" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 48, 1, 1" offset="76472" size="3072" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>48</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="83" name="Convert_527413" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>48</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>48</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="84" name="Constant_527414" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="79544" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="85" name="__module.model.2.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>48</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>48</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="86" name="__module.model.2.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>48</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>48</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="87" name="__module.model.2.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="79800" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="88" name="__module.model.2.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="160_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="89" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_5" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="160,input.21">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="90" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_5/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="80056" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="91" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_5/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="80060" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="92" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_5/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="80056" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="93" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_5/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="80060" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="94" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_5/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="95" name="Constant_527416" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 3, 3" offset="80064" size="36864" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="96" name="Convert_527417" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="97" name="Constant_527418" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="116928" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="98" name="__module.model.3.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="99" name="__module.model.3.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="2, 2" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>160</dim>
|
|
<dim>160</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="100" name="__module.model.3.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="117184" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="101" name="__module.model.3.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="174_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="102" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_6" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="174,input.25">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="103" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_6/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="104" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_6/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="117440" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="105" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_6/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="106" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_6/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="117440" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="107" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_6/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="108" name="Constant_527420" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 1, 1" offset="117444" size="4096" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="109" name="Convert_527421" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="110" name="Constant_527422" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="121540" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="111" name="__module.model.4.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="112" name="__module.model.4.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="113" name="__module.model.4.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="121796" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="114" name="__module.model.4.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="192_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="115" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_7" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="192,input.29">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="116" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_7/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="122052" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="117" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_7/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="122056" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="118" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_7/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="122052" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="119" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_7/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="122056" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="120" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_7/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="121" name="180" type="Const" version="opset1">
|
|
<data element_type="i64" shape="" offset="73936" size="8" />
|
|
<output>
|
|
<port id="0" precision="I64" names="180" />
|
|
</output>
|
|
</layer>
|
|
<layer id="122" name="Constant_472" type="Const" version="opset1">
|
|
<data element_type="i64" shape="2" offset="122060" size="16" />
|
|
<output>
|
|
<port id="0" precision="I64" names="194">
|
|
<dim>2</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="123" name="__module.model.4/prim::ListUnpack" type="VariadicSplit" version="opset1">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="I64" />
|
|
<port id="2" precision="I64">
|
|
<dim>2</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="196">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="4" precision="FP32" names="197,input.31">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="124" name="Constant_527424" type="Const" version="opset1">
|
|
<data element_type="i8" shape="16, 32, 3, 3" offset="122076" size="4608" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="125" name="Convert_527425" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="126" name="Constant_527426" type="Const" version="opset1">
|
|
<data element_type="f32" shape="16, 1, 1, 1" offset="126684" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="127" name="__module.model.4.m.0.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="128" name="__module.model.4.m.0.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="129" name="__module.model.4.m.0.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 16, 1, 1" offset="126748" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="130" name="__module.model.4.m.0.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="207_1">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="131" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_8" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="207,input.33">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="132" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_8/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="126812" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="133" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_8/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="126816" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="134" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_8/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="126812" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="135" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_8/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="126816" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="136" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_8/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="137" name="Constant_527428" type="Const" version="opset1">
|
|
<data element_type="i8" shape="32, 16, 3, 3" offset="126820" size="4608" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="138" name="Convert_527429" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="139" name="Constant_527430" type="Const" version="opset1">
|
|
<data element_type="f32" shape="32, 1, 1, 1" offset="131428" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="140" name="__module.model.4.m.0.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="141" name="__module.model.4.m.0.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="142" name="__module.model.4.m.0.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 32, 1, 1" offset="131556" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="143" name="__module.model.4.m.0.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="216_1">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="144" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_9" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="216,input.37">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="145" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_9/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="131684" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="146" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_9/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="131688" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="147" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_9/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="131684" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="148" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_9/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="131688" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="149" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_9/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="150" name="__module.model.4.m.0/aten::add/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="218">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="151" name="__module.model.4.m.0/aten::add/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="122052" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="152" name="__module.model.4.m.0/aten::add/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="122056" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="153" name="__module.model.4.m.0/aten::add/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="122052" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="154" name="__module.model.4.m.0/aten::add/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="122056" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="155" name="__module.model.4.m.0/aten::add/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="156" name="__module.model.4/aten::cat/Concat" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="220,input.39">
|
|
<dim>1</dim>
|
|
<dim>96</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="157" name="Constant_527432" type="Const" version="opset1">
|
|
<data element_type="i8" shape="128, 96, 1, 1" offset="131692" size="12288" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>96</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="158" name="Convert_527433" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>96</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>96</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="159" name="Constant_527434" type="Const" version="opset1">
|
|
<data element_type="f32" shape="128, 1, 1, 1" offset="143980" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="160" name="__module.model.4.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>96</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>96</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="161" name="__module.model.4.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>96</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>96</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="162" name="__module.model.4.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 128, 1, 1" offset="144492" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="163" name="__module.model.4.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="228_1">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="164" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_10" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="228,input.41">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="165" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_10/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="145004" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="166" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_10/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="145008" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="167" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_10/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="145004" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="168" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_10/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="145008" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="169" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_10/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="170" name="Constant_527436" type="Const" version="opset1">
|
|
<data element_type="i8" shape="128, 128, 3, 3" offset="145012" size="147456" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="171" name="Convert_527437" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="172" name="Constant_527438" type="Const" version="opset1">
|
|
<data element_type="f32" shape="128, 1, 1, 1" offset="292468" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="173" name="__module.model.5.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="174" name="__module.model.5.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="2, 2" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="175" name="__module.model.5.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 128, 1, 1" offset="292980" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="176" name="__module.model.5.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="242_1">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="177" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_11" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="242,input.45">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="178" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_11/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="179" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_11/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="293492" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="180" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_11/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="181" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_11/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="293492" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="182" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_11/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="183" name="Constant_527440" type="Const" version="opset1">
|
|
<data element_type="i8" shape="128, 128, 1, 1" offset="293496" size="16384" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="184" name="Convert_527441" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="185" name="Constant_527442" type="Const" version="opset1">
|
|
<data element_type="f32" shape="128, 1, 1, 1" offset="309880" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="186" name="__module.model.6.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="187" name="__module.model.6.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="188" name="__module.model.6.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 128, 1, 1" offset="310392" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="189" name="__module.model.6.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="260_1">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="190" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_12" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="260,input.49">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="191" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_12/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="192" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_12/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="310904" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="193" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_12/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="194" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_12/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="310904" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="195" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_12/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="196" name="248" type="Const" version="opset1">
|
|
<data element_type="i64" shape="" offset="73936" size="8" />
|
|
<output>
|
|
<port id="0" precision="I64" names="248" />
|
|
</output>
|
|
</layer>
|
|
<layer id="197" name="Constant_729" type="Const" version="opset1">
|
|
<data element_type="i64" shape="2" offset="310908" size="16" />
|
|
<output>
|
|
<port id="0" precision="I64" names="262">
|
|
<dim>2</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="198" name="__module.model.6/prim::ListUnpack" type="VariadicSplit" version="opset1">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="I64" />
|
|
<port id="2" precision="I64">
|
|
<dim>2</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="264">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="4" precision="FP32" names="265,input.51">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="199" name="Constant_527444" type="Const" version="opset1">
|
|
<data element_type="i8" shape="32, 64, 1, 1" offset="310924" size="2048" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="200" name="Convert_527445" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="201" name="Constant_527446" type="Const" version="opset1">
|
|
<data element_type="f32" shape="32, 1, 1, 1" offset="312972" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="202" name="__module.model.6.m.0.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="203" name="__module.model.6.m.0.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="204" name="__module.model.6.m.0.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 32, 1, 1" offset="313100" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="205" name="__module.model.6.m.0.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="277_1">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="206" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_13" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="277,input.53">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="207" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_13/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="313228" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="208" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_13/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="313232" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="209" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_13/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="313228" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="210" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_13/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="313232" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="211" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_13/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="212" name="Constant_527448" type="Const" version="opset1">
|
|
<data element_type="i8" shape="32, 32, 3, 3" offset="313236" size="9216" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="213" name="Convert_527449" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="214" name="Constant_527450" type="Const" version="opset1">
|
|
<data element_type="f32" shape="32, 1, 1, 1" offset="322452" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="215" name="__module.model.6.m.0.m.0.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="216" name="__module.model.6.m.0.m.0.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="217" name="__module.model.6.m.0.m.0.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 32, 1, 1" offset="322580" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="218" name="__module.model.6.m.0.m.0.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="290_1">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="219" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_14" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="290,input.57">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="220" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_14/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="221" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_14/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="322708" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="222" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_14/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="223" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_14/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="322708" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="224" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_14/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="225" name="Constant_527452" type="Const" version="opset1">
|
|
<data element_type="i8" shape="32, 32, 3, 3" offset="322712" size="9216" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="226" name="Convert_527453" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="227" name="Constant_527454" type="Const" version="opset1">
|
|
<data element_type="f32" shape="32, 1, 1, 1" offset="331928" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="228" name="__module.model.6.m.0.m.0.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="229" name="__module.model.6.m.0.m.0.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="230" name="__module.model.6.m.0.m.0.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 32, 1, 1" offset="332056" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="231" name="__module.model.6.m.0.m.0.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="299_1">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="232" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_15" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="299,input.61">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="233" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_15/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="332184" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="234" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_15/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="332188" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="235" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_15/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="332184" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="236" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_15/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="332188" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="237" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_15/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="238" name="__module.model.6.m.0.m.0/aten::add/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="301,input.63">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="239" name="__module.model.6.m.0.m.0/aten::add/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="332192" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="240" name="__module.model.6.m.0.m.0/aten::add/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="332196" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="241" name="__module.model.6.m.0.m.0/aten::add/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="332192" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="242" name="__module.model.6.m.0.m.0/aten::add/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="332196" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="243" name="__module.model.6.m.0.m.0/aten::add/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="244" name="Constant_527456" type="Const" version="opset1">
|
|
<data element_type="i8" shape="32, 32, 3, 3" offset="332200" size="9216" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="245" name="Convert_527457" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="246" name="Constant_527458" type="Const" version="opset1">
|
|
<data element_type="f32" shape="32, 1, 1, 1" offset="341416" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="247" name="__module.model.6.m.0.m.1.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="248" name="__module.model.6.m.0.m.1.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="249" name="__module.model.6.m.0.m.1.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 32, 1, 1" offset="341544" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="250" name="__module.model.6.m.0.m.1.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="311_1">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="251" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_16" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="311,input.65">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="252" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_16/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="341672" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="253" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_16/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="341676" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="254" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_16/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="341672" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="255" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_16/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="341676" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="256" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_16/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="257" name="Constant_527460" type="Const" version="opset1">
|
|
<data element_type="i8" shape="32, 32, 3, 3" offset="341680" size="9216" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="258" name="Convert_527461" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="259" name="Constant_527462" type="Const" version="opset1">
|
|
<data element_type="f32" shape="32, 1, 1, 1" offset="350896" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="260" name="__module.model.6.m.0.m.1.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="261" name="__module.model.6.m.0.m.1.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="262" name="__module.model.6.m.0.m.1.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 32, 1, 1" offset="351024" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="263" name="__module.model.6.m.0.m.1.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="320_1">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="264" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_17" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="320,input.69">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="265" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_17/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="313228" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="266" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_17/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="351152" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="267" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_17/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="313228" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="268" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_17/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="351152" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="269" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_17/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="270" name="__module.model.6.m.0.m.1/aten::add/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="322">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="271" name="__module.model.6.m.0.m.1/aten::add/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="351156" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="272" name="__module.model.6.m.0.m.1/aten::add/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="351160" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="273" name="__module.model.6.m.0.m.1/aten::add/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="351156" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="274" name="__module.model.6.m.0.m.1/aten::add/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="351160" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="275" name="__module.model.6.m.0.m.1/aten::add/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="276" name="Constant_527464" type="Const" version="opset1">
|
|
<data element_type="i8" shape="32, 64, 1, 1" offset="351164" size="2048" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="277" name="Convert_527465" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="278" name="Constant_527466" type="Const" version="opset1">
|
|
<data element_type="f32" shape="32, 1, 1, 1" offset="353212" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="279" name="__module.model.6.m.0.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="280" name="__module.model.6.m.0.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="281" name="__module.model.6.m.0.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 32, 1, 1" offset="353340" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="282" name="__module.model.6.m.0.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="330_1">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="283" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_18" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="330,input.71">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="284" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_18/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="351156" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="285" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_18/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="351160" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="286" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_18/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="351156" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="287" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_18/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="351160" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="288" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_18/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="289" name="__module.model.6.m.0/aten::cat/Concat" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="333,input.73">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="290" name="Constant_527468" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 1, 1" offset="353468" size="4096" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="291" name="Convert_527469" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="292" name="Constant_527470" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="357564" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="293" name="__module.model.6.m.0.cv3.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="294" name="__module.model.6.m.0.cv3.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="295" name="__module.model.6.m.0.cv3.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="357820" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="296" name="__module.model.6.m.0.cv3.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="341_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="297" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_19" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="341,input.75">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="298" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_19/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="299" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_19/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="310904" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="300" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_19/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="301" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_19/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="310904" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="302" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_19/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="303" name="__module.model.6/aten::cat/Concat" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="344,input.77">
|
|
<dim>1</dim>
|
|
<dim>192</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="304" name="Constant_527472" type="Const" version="opset1">
|
|
<data element_type="i8" shape="128, 192, 1, 1" offset="358076" size="24576" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="305" name="Convert_527473" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="306" name="Constant_527474" type="Const" version="opset1">
|
|
<data element_type="f32" shape="128, 1, 1, 1" offset="382652" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="307" name="__module.model.6.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="308" name="__module.model.6.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>192</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="309" name="__module.model.6.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 128, 1, 1" offset="383164" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="310" name="__module.model.6.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="352_1">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="311" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_20" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="352,input.79">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="312" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_20/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="383676" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="313" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_20/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="383680" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="314" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_20/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="383676" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="315" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_20/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="383680" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="316" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_20/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="317" name="Constant_527476" type="Const" version="opset1">
|
|
<data element_type="i8" shape="256, 128, 3, 3" offset="383684" size="294912" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="318" name="Convert_527477" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="319" name="Constant_527478" type="Const" version="opset1">
|
|
<data element_type="f32" shape="256, 1, 1, 1" offset="678596" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="320" name="__module.model.7.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="321" name="__module.model.7.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="2, 2" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="322" name="__module.model.7.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 256, 1, 1" offset="679620" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="323" name="__module.model.7.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="366_1">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="324" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_21" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="366,input.83">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="325" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_21/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="326" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_21/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="680644" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="327" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_21/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="328" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_21/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="680644" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="329" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_21/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="330" name="Constant_527480" type="Const" version="opset1">
|
|
<data element_type="i8" shape="256, 256, 1, 1" offset="680648" size="65536" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="331" name="Convert_527481" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="332" name="Constant_527482" type="Const" version="opset1">
|
|
<data element_type="f32" shape="256, 1, 1, 1" offset="746184" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="333" name="__module.model.8.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="334" name="__module.model.8.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="335" name="__module.model.8.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 256, 1, 1" offset="747208" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="336" name="__module.model.8.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="384_1">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="337" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_22" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="384,input.87">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="338" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_22/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="339" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_22/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="117440" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="340" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_22/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="341" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_22/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="117440" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="342" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_22/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="343" name="372" type="Const" version="opset1">
|
|
<data element_type="i64" shape="" offset="73936" size="8" />
|
|
<output>
|
|
<port id="0" precision="I64" names="372" />
|
|
</output>
|
|
</layer>
|
|
<layer id="344" name="Constant_1228" type="Const" version="opset1">
|
|
<data element_type="i64" shape="2" offset="748232" size="16" />
|
|
<output>
|
|
<port id="0" precision="I64" names="386">
|
|
<dim>2</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="345" name="__module.model.8/prim::ListUnpack" type="VariadicSplit" version="opset1">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="I64" />
|
|
<port id="2" precision="I64">
|
|
<dim>2</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="388">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="4" precision="FP32" names="389,input.89">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="346" name="Constant_527484" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 128, 1, 1" offset="748248" size="8192" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="347" name="Convert_527485" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="348" name="Constant_527486" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="756440" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="349" name="__module.model.8.m.0.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="350" name="__module.model.8.m.0.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="351" name="__module.model.8.m.0.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="756696" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="352" name="__module.model.8.m.0.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="401_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="353" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_23" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="401,input.91">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="354" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_23/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="756952" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="355" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_23/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="756956" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="356" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_23/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="756952" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="357" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_23/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="756956" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="358" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_23/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="359" name="Constant_527488" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 3, 3" offset="756960" size="36864" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="360" name="Convert_527489" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="361" name="Constant_527490" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="793824" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="362" name="__module.model.8.m.0.m.0.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="363" name="__module.model.8.m.0.m.0.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="364" name="__module.model.8.m.0.m.0.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="794080" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="365" name="__module.model.8.m.0.m.0.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="414_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="366" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_24" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="414,input.95">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="367" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_24/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="794336" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="368" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_24/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="794340" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="369" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_24/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="794336" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="370" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_24/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="794340" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="371" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_24/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="372" name="Constant_527492" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 3, 3" offset="794344" size="36864" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="373" name="Convert_527493" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="374" name="Constant_527494" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="831208" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="375" name="__module.model.8.m.0.m.0.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="376" name="__module.model.8.m.0.m.0.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="377" name="__module.model.8.m.0.m.0.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="831464" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="378" name="__module.model.8.m.0.m.0.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="423_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="379" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_25" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="423,input.99">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="380" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_25/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="831720" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="381" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_25/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="831724" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="382" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_25/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="831720" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="383" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_25/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="831724" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="384" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_25/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="385" name="__module.model.8.m.0.m.0/aten::add/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="425,input.101">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="386" name="__module.model.8.m.0.m.0/aten::add/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="831728" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="387" name="__module.model.8.m.0.m.0/aten::add/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="831732" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="388" name="__module.model.8.m.0.m.0/aten::add/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="831728" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="389" name="__module.model.8.m.0.m.0/aten::add/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="831732" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="390" name="__module.model.8.m.0.m.0/aten::add/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="391" name="Constant_527496" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 3, 3" offset="831736" size="36864" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="392" name="Convert_527497" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="393" name="Constant_527498" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="868600" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="394" name="__module.model.8.m.0.m.1.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="395" name="__module.model.8.m.0.m.1.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="396" name="__module.model.8.m.0.m.1.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="868856" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="397" name="__module.model.8.m.0.m.1.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="435_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="398" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_26" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="435,input.103">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="399" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_26/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="400" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_26/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="869112" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="401" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_26/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="402" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_26/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="869112" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="403" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_26/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="404" name="Constant_527500" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 3, 3" offset="869116" size="36864" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="405" name="Convert_527501" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="406" name="Constant_527502" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="905980" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="407" name="__module.model.8.m.0.m.1.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="408" name="__module.model.8.m.0.m.1.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="409" name="__module.model.8.m.0.m.1.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="906236" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="410" name="__module.model.8.m.0.m.1.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="444_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="411" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_27" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="444,input.107">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="412" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_27/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="906492" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="413" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_27/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="906496" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="414" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_27/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="906492" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="415" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_27/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="906496" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="416" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_27/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="417" name="__module.model.8.m.0.m.1/aten::add/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="446">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="418" name="__module.model.8.m.0.m.1/aten::add/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="906500" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="419" name="__module.model.8.m.0.m.1/aten::add/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="906504" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="420" name="__module.model.8.m.0.m.1/aten::add/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="906500" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="421" name="__module.model.8.m.0.m.1/aten::add/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="906504" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="422" name="__module.model.8.m.0.m.1/aten::add/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="423" name="Constant_527504" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 128, 1, 1" offset="906508" size="8192" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="424" name="Convert_527505" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="425" name="Constant_527506" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="914700" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="426" name="__module.model.8.m.0.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="427" name="__module.model.8.m.0.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="428" name="__module.model.8.m.0.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="914956" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="429" name="__module.model.8.m.0.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="454_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="430" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_28" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="454,input.109">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="431" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_28/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="906500" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="432" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_28/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="906504" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="433" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_28/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="906500" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="434" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_28/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="906504" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="435" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_28/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="436" name="__module.model.8.m.0/aten::cat/Concat" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="457,input.111">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="437" name="Constant_527508" type="Const" version="opset1">
|
|
<data element_type="i8" shape="128, 128, 1, 1" offset="915212" size="16384" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="438" name="Convert_527509" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="439" name="Constant_527510" type="Const" version="opset1">
|
|
<data element_type="f32" shape="128, 1, 1, 1" offset="931596" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="440" name="__module.model.8.m.0.cv3.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="441" name="__module.model.8.m.0.cv3.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="442" name="__module.model.8.m.0.cv3.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 128, 1, 1" offset="932108" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="443" name="__module.model.8.m.0.cv3.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="465_1">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="444" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_29" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="465,input.113">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="445" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_29/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="446" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_29/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="117440" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="447" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_29/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="448" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_29/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="117440" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="449" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_29/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="450" name="__module.model.8/aten::cat/Concat" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="468,input.115">
|
|
<dim>1</dim>
|
|
<dim>384</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="451" name="Constant_527512" type="Const" version="opset1">
|
|
<data element_type="i8" shape="256, 384, 1, 1" offset="932620" size="98304" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="452" name="Convert_527513" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="453" name="Constant_527514" type="Const" version="opset1">
|
|
<data element_type="f32" shape="256, 1, 1, 1" offset="1030924" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="454" name="__module.model.8.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="455" name="__module.model.8.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>384</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="456" name="__module.model.8.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 256, 1, 1" offset="1031948" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="457" name="__module.model.8.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="476_1">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="458" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_30" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="476,input.117">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="459" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_30/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="460" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_30/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1032972" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="461" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_30/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="462" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_30/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1032972" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="463" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_30/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="464" name="Constant_527516" type="Const" version="opset1">
|
|
<data element_type="i8" shape="128, 256, 1, 1" offset="1032976" size="32768" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="465" name="Convert_527517" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="466" name="Constant_527518" type="Const" version="opset1">
|
|
<data element_type="f32" shape="128, 1, 1, 1" offset="1065744" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="467" name="__module.model.9.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="468" name="__module.model.9.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="469" name="__module.model.9.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 128, 1, 1" offset="1066256" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="470" name="__module.model.9.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="494_1">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="471" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_31" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="494,input.121">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="472" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_31/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1066768" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="473" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_31/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1066772" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="474" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_31/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1066768" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="475" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_31/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1066772" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="476" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_31/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="477" name="__module.model.9.m/aten::max_pool2d/MaxPool" type="MaxPool" version="opset8">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="2, 2" pads_end="2, 2" kernel="5, 5" rounding_type="floor" auto_pad="explicit" index_element_type="i64" axis="2" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="500,input.125">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="2" precision="I64">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="478" name="__module.model.9.m/aten::max_pool2d/MaxPool_1" type="MaxPool" version="opset8">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="2, 2" pads_end="2, 2" kernel="5, 5" rounding_type="floor" auto_pad="explicit" index_element_type="i64" axis="2" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="505,input.127">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="2" precision="I64">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="479" name="__module.model.9.m/aten::max_pool2d/MaxPool_2" type="MaxPool" version="opset8">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="2, 2" pads_end="2, 2" kernel="5, 5" rounding_type="floor" auto_pad="explicit" index_element_type="i64" axis="2" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="510">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="2" precision="I64">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="480" name="__module.model.9/aten::cat/Concat" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="3" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="4" precision="FP32" names="512,input.129">
|
|
<dim>1</dim>
|
|
<dim>512</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="481" name="Constant_527520" type="Const" version="opset1">
|
|
<data element_type="i8" shape="256, 512, 1, 1" offset="1066776" size="131072" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>512</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="482" name="Convert_527521" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>512</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>512</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="483" name="Constant_527522" type="Const" version="opset1">
|
|
<data element_type="f32" shape="256, 1, 1, 1" offset="1197848" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="484" name="__module.model.9.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>512</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>512</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="485" name="__module.model.9.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>512</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>512</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="486" name="__module.model.9.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 256, 1, 1" offset="1198872" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="487" name="__module.model.9.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="520_1">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="488" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_32" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="520,input.131">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="489" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_32/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="490" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_32/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1199896" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="491" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_32/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="492" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_32/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1199896" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="493" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_32/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="494" name="Constant_527524" type="Const" version="opset1">
|
|
<data element_type="i8" shape="256, 256, 1, 1" offset="1199900" size="65536" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="495" name="Convert_527525" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="496" name="Constant_527526" type="Const" version="opset1">
|
|
<data element_type="f32" shape="256, 1, 1, 1" offset="1265436" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="497" name="__module.model.10.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="498" name="__module.model.10.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="499" name="__module.model.10.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 256, 1, 1" offset="1266460" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="500" name="__module.model.10.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="545_1">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="501" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_33" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="545,input.135">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="502" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_33/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1267484" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="503" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_33/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1267488" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="504" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_33/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1267484" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="505" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_33/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1267488" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="506" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_33/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="507" name="534" type="Const" version="opset1">
|
|
<data element_type="i64" shape="" offset="73936" size="8" />
|
|
<output>
|
|
<port id="0" precision="I64" names="534" />
|
|
</output>
|
|
</layer>
|
|
<layer id="508" name="Constant_1849" type="Const" version="opset1">
|
|
<data element_type="i64" shape="2" offset="748232" size="16" />
|
|
<output>
|
|
<port id="0" precision="I64" names="547">
|
|
<dim>2</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="509" name="__module.model.10/prim::ListUnpack" type="VariadicSplit" version="opset1">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="I64" />
|
|
<port id="2" precision="I64">
|
|
<dim>2</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="549,a.1">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="4" precision="FP32" names="550,x.3">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="510" name="Constant_527528" type="Const" version="opset1">
|
|
<data element_type="i8" shape="256, 128, 1, 1" offset="1267492" size="32768" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="511" name="Convert_527529" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="512" name="Constant_527530" type="Const" version="opset1">
|
|
<data element_type="f32" shape="256, 1, 1, 1" offset="1300260" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="513" name="__module.model.10.m.0.attn.qkv.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="514" name="__module.model.10.m.0.attn.qkv.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="515" name="__module.model.10.m.0.attn.qkv.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 256, 1, 1" offset="1301284" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="516" name="__module.model.10.m.0.attn.qkv.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="572,qkv.1">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="517" name="__module.model.10.m.0.attn.qkv.conv/aten::_convolution/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1302308" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="518" name="__module.model.10.m.0.attn.qkv.conv/aten::_convolution/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1302312" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="519" name="__module.model.10.m.0.attn.qkv.conv/aten::_convolution/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1302308" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="520" name="__module.model.10.m.0.attn.qkv.conv/aten::_convolution/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1302312" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="521" name="__module.model.10.m.0.attn.qkv.conv/aten::_convolution/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="522" name="Constant_11041" type="Const" version="opset1">
|
|
<data element_type="i64" shape="4" offset="1302316" size="32" />
|
|
<output>
|
|
<port id="0" precision="I64">
|
|
<dim>4</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="523" name="__module.model.10.m.0.attn/aten::view/Reshape" type="Reshape" version="opset1">
|
|
<data special_zero="true" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="I64">
|
|
<dim>4</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="574">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>128</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="524" name="522" type="Const" version="opset1">
|
|
<data element_type="i64" shape="" offset="1302348" size="8" />
|
|
<output>
|
|
<port id="0" precision="I64" names="522" />
|
|
</output>
|
|
</layer>
|
|
<layer id="525" name="Constant_1998" type="Const" version="opset1">
|
|
<data element_type="i64" shape="3" offset="1302356" size="24" />
|
|
<output>
|
|
<port id="0" precision="I64" names="575">
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="526" name="__module.model.10.m.0.attn/prim::ListUnpack" type="VariadicSplit" version="opset1">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>128</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
<port id="1" precision="I64" />
|
|
<port id="2" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="577,q">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>32</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
<port id="4" precision="FP32" names="578,k">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>32</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
<port id="5" precision="FP32" names="579,v">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>64</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="527" name="Constant_10819" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 1, 1, 1" offset="1302380" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="528" name="Multiply_10761" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>32</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>32</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="529" name="Multiply_10761/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1302384" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="530" name="Multiply_10761/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1302388" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="531" name="Multiply_10761/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1302384" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="532" name="Multiply_10761/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1302388" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="533" name="Multiply_10761/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>32</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>32</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="534" name="__module.model.10.m.0.attn/aten::mul/Multiply_1" type="MatMul" version="opset1">
|
|
<data transpose_a="true" transpose_b="false" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>32</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>32</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="582,attn.1">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>400</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="535" name="__module.model.10.m.0.attn/aten::softmax/Softmax" type="SoftMax" version="opset8">
|
|
<data axis="-1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>400</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="583,attn.3">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>400</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="536" name="__module.model.10.m.0.attn/aten::softmax/Softmax/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="67200" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="537" name="__module.model.10.m.0.attn/aten::softmax/Softmax/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1302392" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="538" name="__module.model.10.m.0.attn/aten::softmax/Softmax/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="67200" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="539" name="__module.model.10.m.0.attn/aten::softmax/Softmax/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1302392" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="540" name="__module.model.10.m.0.attn/aten::softmax/Softmax/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>400</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>400</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="541" name="__module.model.10.m.0.attn/aten::matmul/MatMul_1" type="MatMul" version="opset1">
|
|
<data transpose_a="false" transpose_b="true" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>64</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>400</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="585">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>64</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="542" name="__module.model.10.m.0.attn/aten::matmul/MatMul_1/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1302396" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="543" name="__module.model.10.m.0.attn/aten::matmul/MatMul_1/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1302400" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="544" name="__module.model.10.m.0.attn/aten::matmul/MatMul_1/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1302396" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="545" name="__module.model.10.m.0.attn/aten::matmul/MatMul_1/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1302400" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="546" name="__module.model.10.m.0.attn/aten::matmul/MatMul_1/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>64</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>64</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="547" name="__module.model.10.m.0.attn/aten::size/ShapeOf" type="ShapeOf" version="opset3">
|
|
<data output_type="i64" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="I64">
|
|
<dim>4</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="548" name="Constant_10053" type="Const" version="opset1">
|
|
<data element_type="i64" shape="2" offset="1302404" size="16" />
|
|
<output>
|
|
<port id="0" precision="I64">
|
|
<dim>2</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="549" name="Constant_10054" type="Const" version="opset1">
|
|
<data element_type="i64" shape="" offset="1302420" size="8" />
|
|
<output>
|
|
<port id="0" precision="I64" />
|
|
</output>
|
|
</layer>
|
|
<layer id="550" name="Gather_10055" type="Gather" version="opset8">
|
|
<data batch_dims="0" />
|
|
<input>
|
|
<port id="0" precision="I64">
|
|
<dim>4</dim>
|
|
</port>
|
|
<port id="1" precision="I64">
|
|
<dim>2</dim>
|
|
</port>
|
|
<port id="2" precision="I64" />
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="I64">
|
|
<dim>2</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="551" name="ShapeOf_11004" type="ShapeOf" version="opset3">
|
|
<data output_type="i64" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="I64">
|
|
<dim>4</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="552" name="Constant_11136" type="Const" version="opset1">
|
|
<data element_type="i64" shape="2" offset="1302428" size="16" />
|
|
<output>
|
|
<port id="0" precision="I64">
|
|
<dim>2</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="553" name="Constant_11137" type="Const" version="opset1">
|
|
<data element_type="i64" shape="" offset="1302420" size="8" />
|
|
<output>
|
|
<port id="0" precision="I64" />
|
|
</output>
|
|
</layer>
|
|
<layer id="554" name="Gather_11138" type="Gather" version="opset8">
|
|
<data batch_dims="0" />
|
|
<input>
|
|
<port id="0" precision="I64">
|
|
<dim>4</dim>
|
|
</port>
|
|
<port id="1" precision="I64">
|
|
<dim>2</dim>
|
|
</port>
|
|
<port id="2" precision="I64" />
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="I64">
|
|
<dim>2</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="555" name="__module.model.10.m.0.attn/prim::ListConstruct/Concat_1" type="Concat" version="opset1">
|
|
<data axis="0" />
|
|
<input>
|
|
<port id="0" precision="I64">
|
|
<dim>2</dim>
|
|
</port>
|
|
<port id="1" precision="I64">
|
|
<dim>2</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="I64" names="586,588">
|
|
<dim>4</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="556" name="__module.model.10.m.0.attn/aten::view/Reshape_1" type="Reshape" version="opset1">
|
|
<data special_zero="false" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>64</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
<port id="1" precision="I64">
|
|
<dim>4</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="587">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="557" name="__module.model.10.m.0.attn/aten::reshape/Reshape" type="Reshape" version="opset1">
|
|
<data special_zero="false" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>64</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
<port id="1" precision="I64">
|
|
<dim>4</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="589,input.137">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="558" name="Constant_527532" type="Const" version="opset1">
|
|
<data element_type="i8" shape="128, 1, 1, 3, 3" offset="1302444" size="1152" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="559" name="Convert_527533" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="560" name="Constant_527534" type="Const" version="opset1">
|
|
<data element_type="f32" shape="128, 1, 1, 1, 1" offset="1303596" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="561" name="__module.model.10.m.0.attn.pe.conv/aten::_convolution/GroupConvolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="562" name="__module.model.10.m.0.attn.pe.conv/aten::_convolution/GroupConvolution" type="GroupConvolution" version="opset1">
|
|
<data strides="1, 1" pads_begin="1, 1" pads_end="1, 1" dilations="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="563" name="__module.model.10.m.0.attn.pe.conv/aten::_convolution/Reshape_1" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 128, 1, 1" offset="1304108" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="564" name="__module.model.10.m.0.attn.pe.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="597">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="565" name="__module.model.10.m.0.attn.pe.conv/aten::_convolution/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1304620" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="566" name="__module.model.10.m.0.attn.pe.conv/aten::_convolution/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1304624" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="567" name="__module.model.10.m.0.attn.pe.conv/aten::_convolution/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1304620" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="568" name="__module.model.10.m.0.attn.pe.conv/aten::_convolution/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1304624" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="569" name="__module.model.10.m.0.attn.pe.conv/aten::_convolution/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="570" name="__module.model.10.m.0.attn/aten::add/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="598,input.139">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="571" name="__module.model.10.m.0.attn/aten::add/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1304628" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="572" name="__module.model.10.m.0.attn/aten::add/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1304632" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="573" name="__module.model.10.m.0.attn/aten::add/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1304628" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="574" name="__module.model.10.m.0.attn/aten::add/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1304632" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="575" name="__module.model.10.m.0.attn/aten::add/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="576" name="Constant_527536" type="Const" version="opset1">
|
|
<data element_type="i8" shape="128, 128, 1, 1" offset="1304636" size="16384" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="577" name="Convert_527537" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="578" name="Constant_527538" type="Const" version="opset1">
|
|
<data element_type="f32" shape="128, 1, 1, 1" offset="1321020" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="579" name="__module.model.10.m.0.attn.proj.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="580" name="__module.model.10.m.0.attn.proj.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="581" name="__module.model.10.m.0.attn.proj.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 128, 1, 1" offset="1321532" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="582" name="__module.model.10.m.0.attn.proj.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="606">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="583" name="__module.model.10.m.0.attn.proj.conv/aten::_convolution/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1322044" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="584" name="__module.model.10.m.0.attn.proj.conv/aten::_convolution/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1322048" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="585" name="__module.model.10.m.0.attn.proj.conv/aten::_convolution/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1322044" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="586" name="__module.model.10.m.0.attn.proj.conv/aten::_convolution/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1322048" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="587" name="__module.model.10.m.0.attn.proj.conv/aten::_convolution/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="588" name="__module.model.10.m.0/aten::add/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="607,input.141">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="589" name="__module.model.10.m.0/aten::add/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1322052" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="590" name="__module.model.10.m.0/aten::add/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1322056" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="591" name="__module.model.10.m.0/aten::add/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1322052" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="592" name="__module.model.10.m.0/aten::add/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1322056" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="593" name="__module.model.10.m.0/aten::add/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="594" name="Constant_527540" type="Const" version="opset1">
|
|
<data element_type="i8" shape="256, 128, 1, 1" offset="1322060" size="32768" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="595" name="Convert_527541" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="596" name="Constant_527542" type="Const" version="opset1">
|
|
<data element_type="f32" shape="256, 1, 1, 1" offset="1354828" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="597" name="__module.model.10.m.0.ffn.0.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="598" name="__module.model.10.m.0.ffn.0.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="599" name="__module.model.10.m.0.ffn.0.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 256, 1, 1" offset="1355852" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="600" name="__module.model.10.m.0.ffn.0.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="617_1">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="601" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_34" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="617,input.143">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="602" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_34/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="603" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_34/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="310904" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="604" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_34/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="605" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_34/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="310904" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="606" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_34/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="607" name="Constant_527544" type="Const" version="opset1">
|
|
<data element_type="i8" shape="128, 256, 1, 1" offset="1356876" size="32768" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="608" name="Convert_527545" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="609" name="Constant_527546" type="Const" version="opset1">
|
|
<data element_type="f32" shape="128, 1, 1, 1" offset="1389644" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="610" name="__module.model.10.m.0.ffn.1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="611" name="__module.model.10.m.0.ffn.1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="612" name="__module.model.10.m.0.ffn.1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 128, 1, 1" offset="1390156" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="613" name="__module.model.10.m.0.ffn.1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="627">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="614" name="__module.model.10.m.0.ffn.1.conv/aten::_convolution/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1390668" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="615" name="__module.model.10.m.0.ffn.1.conv/aten::_convolution/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1390672" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="616" name="__module.model.10.m.0.ffn.1.conv/aten::_convolution/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1390668" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="617" name="__module.model.10.m.0.ffn.1.conv/aten::_convolution/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1390672" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="618" name="__module.model.10.m.0.ffn.1.conv/aten::_convolution/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="619" name="__module.model.10.m.0/aten::add/Add_1" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="628,b.1">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="620" name="__module.model.10.m.0/aten::add/Add_1/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1267484" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="621" name="__module.model.10.m.0/aten::add/Add_1/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1267488" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="622" name="__module.model.10.m.0/aten::add/Add_1/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1267484" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="623" name="__module.model.10.m.0/aten::add/Add_1/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1267488" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="624" name="__module.model.10.m.0/aten::add/Add_1/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="625" name="__module.model.10/aten::cat/Concat" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="630,input.147">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="626" name="Constant_527548" type="Const" version="opset1">
|
|
<data element_type="i8" shape="256, 256, 1, 1" offset="1390676" size="65536" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="627" name="Convert_527549" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="628" name="Constant_527550" type="Const" version="opset1">
|
|
<data element_type="f32" shape="256, 1, 1, 1" offset="1456212" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="629" name="__module.model.10.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="630" name="__module.model.10.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="631" name="__module.model.10.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 256, 1, 1" offset="1457236" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="632" name="__module.model.10.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="638_1">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="633" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_35" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="638,input.149">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="634" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_35/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1458260" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="635" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_35/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1458264" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="636" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_35/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1458260" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="637" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_35/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1458264" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="638" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_35/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="639" name="__module.model.11/aten::upsample_nearest2d/Multiply" type="Const" version="opset1">
|
|
<data element_type="f32" shape="2" offset="1458268" size="8" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>2</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="640" name="Constant_2402" type="Const" version="opset1">
|
|
<data element_type="i32" shape="2" offset="1458276" size="8" />
|
|
<output>
|
|
<port id="0" precision="I32">
|
|
<dim>2</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="641" name="__module.model.11/aten::upsample_nearest2d/Interpolate" type="Interpolate" version="opset11">
|
|
<data mode="nearest" shape_calculation_mode="scales" coordinate_transformation_mode="asymmetric" nearest_mode="floor" antialias="false" pads_begin="0, 0, 0, 0" pads_end="0, 0, 0, 0" cube_coeff="-0.75" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>2</dim>
|
|
</port>
|
|
<port id="2" precision="I32">
|
|
<dim>2</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="643">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="642" name="__module.model.11/aten::upsample_nearest2d/Interpolate/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="383676" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="643" name="__module.model.11/aten::upsample_nearest2d/Interpolate/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="383680" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="644" name="__module.model.11/aten::upsample_nearest2d/Interpolate/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="383676" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="645" name="__module.model.11/aten::upsample_nearest2d/Interpolate/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="383680" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="646" name="__module.model.11/aten::upsample_nearest2d/Interpolate/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="647" name="__module.model.12/aten::cat/Concat" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="646,input.153">
|
|
<dim>1</dim>
|
|
<dim>384</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="648" name="Constant_527552" type="Const" version="opset1">
|
|
<data element_type="i8" shape="128, 384, 1, 1" offset="1458284" size="49152" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="649" name="Convert_527553" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="650" name="Constant_527554" type="Const" version="opset1">
|
|
<data element_type="f32" shape="128, 1, 1, 1" offset="1507436" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="651" name="__module.model.13.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="652" name="__module.model.13.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>384</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="653" name="__module.model.13.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 128, 1, 1" offset="1507948" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="654" name="__module.model.13.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="663_1">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="655" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_36" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="663,input.155">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="656" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_36/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1508460" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="657" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_36/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1508464" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="658" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_36/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1508460" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="659" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_36/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1508464" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="660" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_36/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="661" name="651" type="Const" version="opset1">
|
|
<data element_type="i64" shape="" offset="73936" size="8" />
|
|
<output>
|
|
<port id="0" precision="I64" names="651" />
|
|
</output>
|
|
</layer>
|
|
<layer id="662" name="Constant_2475" type="Const" version="opset1">
|
|
<data element_type="i64" shape="2" offset="310908" size="16" />
|
|
<output>
|
|
<port id="0" precision="I64" names="665">
|
|
<dim>2</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="663" name="__module.model.13/prim::ListUnpack" type="VariadicSplit" version="opset1">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="I64" />
|
|
<port id="2" precision="I64">
|
|
<dim>2</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="667">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="4" precision="FP32" names="668,input.157">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="664" name="Constant_527556" type="Const" version="opset1">
|
|
<data element_type="i8" shape="32, 64, 3, 3" offset="1508468" size="18432" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="665" name="Convert_527557" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="666" name="Constant_527558" type="Const" version="opset1">
|
|
<data element_type="f32" shape="32, 1, 1, 1" offset="1526900" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="667" name="__module.model.13.m.0.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="668" name="__module.model.13.m.0.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="669" name="__module.model.13.m.0.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 32, 1, 1" offset="1527028" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="670" name="__module.model.13.m.0.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="678_1">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="671" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_37" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="678,input.159">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="672" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_37/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="673" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_37/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="293492" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="674" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_37/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="675" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_37/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="293492" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="676" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_37/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="677" name="Constant_527560" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 32, 3, 3" offset="1527156" size="18432" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="678" name="Convert_527561" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="679" name="Constant_527562" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="1545588" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="680" name="__module.model.13.m.0.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="681" name="__module.model.13.m.0.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="682" name="__module.model.13.m.0.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1545844" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="683" name="__module.model.13.m.0.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="687_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="684" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_38" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="687,input.163">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="685" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_38/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1546100" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="686" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_38/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1546104" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="687" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_38/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1546100" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="688" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_38/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1546104" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="689" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_38/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="690" name="__module.model.13.m.0/aten::add/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="689">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="691" name="__module.model.13.m.0/aten::add/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1508460" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="692" name="__module.model.13.m.0/aten::add/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1508464" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="693" name="__module.model.13.m.0/aten::add/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1508460" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="694" name="__module.model.13.m.0/aten::add/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1508464" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="695" name="__module.model.13.m.0/aten::add/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="696" name="__module.model.13/aten::cat/Concat" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="691,input.165">
|
|
<dim>1</dim>
|
|
<dim>192</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="697" name="Constant_527564" type="Const" version="opset1">
|
|
<data element_type="i8" shape="128, 192, 1, 1" offset="1546108" size="24576" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="698" name="Convert_527565" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="699" name="Constant_527566" type="Const" version="opset1">
|
|
<data element_type="f32" shape="128, 1, 1, 1" offset="1570684" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="700" name="__module.model.13.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="701" name="__module.model.13.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>192</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="702" name="__module.model.13.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 128, 1, 1" offset="1571196" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="703" name="__module.model.13.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="699_1">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="704" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_39" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="699,input.167">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="705" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_39/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="145004" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="706" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_39/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="145008" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="707" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_39/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="145004" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="708" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_39/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="145008" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="709" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_39/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="710" name="__module.model.14/aten::upsample_nearest2d/Multiply" type="Const" version="opset1">
|
|
<data element_type="f32" shape="2" offset="1458268" size="8" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>2</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="711" name="Constant_2631" type="Const" version="opset1">
|
|
<data element_type="i32" shape="2" offset="1458276" size="8" />
|
|
<output>
|
|
<port id="0" precision="I32">
|
|
<dim>2</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="712" name="__module.model.14/aten::upsample_nearest2d/Interpolate" type="Interpolate" version="opset11">
|
|
<data mode="nearest" shape_calculation_mode="scales" coordinate_transformation_mode="asymmetric" nearest_mode="floor" antialias="false" pads_begin="0, 0, 0, 0" pads_end="0, 0, 0, 0" cube_coeff="-0.75" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>2</dim>
|
|
</port>
|
|
<port id="2" precision="I32">
|
|
<dim>2</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="704">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="713" name="__module.model.14/aten::upsample_nearest2d/Interpolate/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="145004" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="714" name="__module.model.14/aten::upsample_nearest2d/Interpolate/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="145008" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="715" name="__module.model.14/aten::upsample_nearest2d/Interpolate/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="145004" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="716" name="__module.model.14/aten::upsample_nearest2d/Interpolate/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="145008" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="717" name="__module.model.14/aten::upsample_nearest2d/Interpolate/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="718" name="__module.model.15/aten::cat/Concat" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="707,input.171">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="719" name="Constant_527568" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 256, 1, 1" offset="1571708" size="16384" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="720" name="Convert_527569" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="721" name="Constant_527570" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="1588092" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="722" name="__module.model.16.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="723" name="__module.model.16.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="724" name="__module.model.16.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1588348" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="725" name="__module.model.16.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="724_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="726" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_40" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="724,input.173">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="727" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_40/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1588604" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="728" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_40/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1588608" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="729" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_40/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1588604" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="730" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_40/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1588608" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="731" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_40/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="732" name="712" type="Const" version="opset1">
|
|
<data element_type="i64" shape="" offset="73936" size="8" />
|
|
<output>
|
|
<port id="0" precision="I64" names="712" />
|
|
</output>
|
|
</layer>
|
|
<layer id="733" name="Constant_2704" type="Const" version="opset1">
|
|
<data element_type="i64" shape="2" offset="122060" size="16" />
|
|
<output>
|
|
<port id="0" precision="I64" names="726">
|
|
<dim>2</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="734" name="__module.model.16/prim::ListUnpack" type="VariadicSplit" version="opset1">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="I64" />
|
|
<port id="2" precision="I64">
|
|
<dim>2</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="728">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="4" precision="FP32" names="729,input.175">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="735" name="Constant_527572" type="Const" version="opset1">
|
|
<data element_type="i8" shape="16, 32, 3, 3" offset="1588612" size="4608" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="736" name="Convert_527573" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="737" name="Constant_527574" type="Const" version="opset1">
|
|
<data element_type="f32" shape="16, 1, 1, 1" offset="1593220" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="738" name="__module.model.16.m.0.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="739" name="__module.model.16.m.0.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="740" name="__module.model.16.m.0.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 16, 1, 1" offset="1593284" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="741" name="__module.model.16.m.0.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="739_1">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="742" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_41" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="739,input.177">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="743" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_41/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1593348" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="744" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_41/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1593352" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="745" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_41/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1593348" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="746" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_41/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1593352" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="747" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_41/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="748" name="Constant_527576" type="Const" version="opset1">
|
|
<data element_type="i8" shape="32, 16, 3, 3" offset="1593356" size="4608" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="749" name="Convert_527577" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="750" name="Constant_527578" type="Const" version="opset1">
|
|
<data element_type="f32" shape="32, 1, 1, 1" offset="1597964" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="751" name="__module.model.16.m.0.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="752" name="__module.model.16.m.0.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="753" name="__module.model.16.m.0.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 32, 1, 1" offset="1598092" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="754" name="__module.model.16.m.0.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="748_1">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="755" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_42" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="748,input.181">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="756" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_42/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="757" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_42/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="117440" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="758" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_42/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="759" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_42/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="117440" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="760" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_42/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="761" name="__module.model.16.m.0/aten::add/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="750">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="762" name="__module.model.16.m.0/aten::add/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1588604" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="763" name="__module.model.16.m.0/aten::add/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1588608" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="764" name="__module.model.16.m.0/aten::add/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1588604" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="765" name="__module.model.16.m.0/aten::add/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1588608" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="766" name="__module.model.16.m.0/aten::add/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="767" name="__module.model.16/aten::cat/Concat" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="752,input.183">
|
|
<dim>1</dim>
|
|
<dim>96</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="768" name="Constant_527580" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 96, 1, 1" offset="1598220" size="6144" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>96</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="769" name="Convert_527581" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>96</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>96</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="770" name="Constant_527582" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="1604364" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="771" name="__module.model.16.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>96</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>96</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="772" name="__module.model.16.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>96</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>96</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="773" name="__module.model.16.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1604620" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="774" name="__module.model.16.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="760_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="775" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_43" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="760,input.185">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="776" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_43/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1604876" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="777" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_43/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1604880" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="778" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_43/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1604876" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="779" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_43/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1604880" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="780" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_43/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="781" name="Constant_527584" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 3, 3" offset="1604884" size="36864" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="782" name="Convert_527585" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="783" name="Constant_527586" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="1641748" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="784" name="__module.model.23.cv2.0.0.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="785" name="__module.model.23.cv2.0.0.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="786" name="__module.model.23.cv2.0.0.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1642004" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="787" name="__module.model.23.cv2.0.0.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1103_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="788" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_44" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1103,input.271">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="789" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_44/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="790" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_44/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1642260" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="791" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_44/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="792" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_44/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1642260" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="793" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_44/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="794" name="Constant_527588" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 3, 3" offset="1642264" size="36864" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="795" name="Convert_527589" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="796" name="Constant_527590" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="1679128" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="797" name="__module.model.23.cv2.0.1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="798" name="__module.model.23.cv2.0.1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="799" name="__module.model.23.cv2.0.1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1679384" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="800" name="__module.model.23.cv2.0.1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1112_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="801" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_45" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1112,input.275">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="802" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_45/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1679640" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="803" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_45/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1679644" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="804" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_45/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1679640" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="805" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_45/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1679644" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="806" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_45/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="807" name="Constant_527592" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 1, 1" offset="1679648" size="4096" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="808" name="Convert_527593" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="809" name="Constant_527594" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="1683744" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="810" name="__module.model.23.cv2.0.2/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="811" name="__module.model.23.cv2.0.2/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="812" name="__module.model.23.cv2.0.2/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1684000" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="813" name="__module.model.23.cv2.0.2/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1120">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="814" name="Constant_527596" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 1, 1, 3, 3" offset="1684256" size="576" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="815" name="Convert_527597" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="816" name="Constant_527598" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1, 1" offset="1684832" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="817" name="__module.model.23.cv3.0.0.0.conv/aten::_convolution/GroupConvolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="818" name="__module.model.23.cv3.0.0.0.conv/aten::_convolution/GroupConvolution" type="GroupConvolution" version="opset1">
|
|
<data strides="1, 1" pads_begin="1, 1" pads_end="1, 1" dilations="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="819" name="__module.model.23.cv3.0.0.0.conv/aten::_convolution/Reshape_1" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1685088" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="820" name="__module.model.23.cv3.0.0.0.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1133_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="821" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_46" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1133,input.279">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="822" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_46/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="823" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_46/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1685344" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="824" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_46/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="825" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_46/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1685344" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="826" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_46/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="827" name="Constant_527600" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 1, 1" offset="1685348" size="4096" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="828" name="Convert_527601" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="829" name="Constant_527602" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="1689444" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="830" name="__module.model.23.cv3.0.0.1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="831" name="__module.model.23.cv3.0.0.1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="832" name="__module.model.23.cv3.0.0.1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1689700" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="833" name="__module.model.23.cv3.0.0.1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1142_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="834" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_47" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1142,input.283">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="835" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_47/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1689956" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="836" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_47/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1690212" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="837" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_47/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1689956" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="838" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_47/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1690212" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="839" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_47/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="3" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="4" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="840" name="Constant_527604" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 1, 1, 3, 3" offset="1690468" size="576" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="841" name="Convert_527605" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="842" name="Constant_527606" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1, 1" offset="1691044" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="843" name="__module.model.23.cv3.0.1.0.conv/aten::_convolution/GroupConvolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="844" name="__module.model.23.cv3.0.1.0.conv/aten::_convolution/GroupConvolution" type="GroupConvolution" version="opset1">
|
|
<data strides="1, 1" pads_begin="1, 1" pads_end="1, 1" dilations="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="845" name="__module.model.23.cv3.0.1.0.conv/aten::_convolution/Reshape_1" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1691300" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="846" name="__module.model.23.cv3.0.1.0.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1153_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="847" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_48" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1153,input.287">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="848" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_48/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1691556" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="849" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_48/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1691560" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="850" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_48/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1691556" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="851" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_48/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1691560" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="852" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_48/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="853" name="Constant_527608" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 1, 1" offset="1691564" size="4096" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="854" name="Convert_527609" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="855" name="Constant_527610" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="1695660" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="856" name="__module.model.23.cv3.0.1.1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="857" name="__module.model.23.cv3.0.1.1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="858" name="__module.model.23.cv3.0.1.1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1695916" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="859" name="__module.model.23.cv3.0.1.1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1162_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="860" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_49" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1162,input.291">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="861" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_49/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1696172" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="862" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_49/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1696176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="863" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_49/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1696172" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="864" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_49/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1696176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="865" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_49/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="866" name="Constant_527612" type="Const" version="opset1">
|
|
<data element_type="i8" shape="38, 64, 1, 1" offset="1696180" size="2432" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>38</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="867" name="Convert_527613" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>38</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>38</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="868" name="Constant_527614" type="Const" version="opset1">
|
|
<data element_type="f32" shape="38, 1, 1, 1" offset="1698612" size="152" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>38</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="869" name="__module.model.23.cv3.0.2/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>38</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>38</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>38</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="870" name="__module.model.23.cv3.0.2/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>38</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="871" name="__module.model.23.cv3.0.2/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 38, 1, 1" offset="1698764" size="152" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="872" name="__module.model.23.cv3.0.2/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1170">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="873" name="__module.model.23/aten::cat/Concat" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1172,xi.1">
|
|
<dim>1</dim>
|
|
<dim>102</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="874" name="__module.model.23/prim::ListConstruct/Concat" type="Const" version="opset1">
|
|
<data element_type="i64" shape="3" offset="1698916" size="24" />
|
|
<output>
|
|
<port id="0" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="875" name="__module.model.23/aten::view/Reshape" type="Reshape" version="opset1">
|
|
<data special_zero="true" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>102</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1335">
|
|
<dim>1</dim>
|
|
<dim>102</dim>
|
|
<dim>6400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="876" name="Constant_527616" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 3, 3" offset="1698940" size="36864" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="877" name="Convert_527617" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="878" name="Constant_527618" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="1735804" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="879" name="__module.model.17.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="880" name="__module.model.17.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="2, 2" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="881" name="__module.model.17.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1736060" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="882" name="__module.model.17.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="774_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="883" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_50" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="774,input.189">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="884" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_50/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="145004" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="885" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_50/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="145008" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="886" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_50/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="145004" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="887" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_50/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="145008" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="888" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_50/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="889" name="__module.model.18/aten::cat/Concat" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="778,input.191">
|
|
<dim>1</dim>
|
|
<dim>192</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="890" name="Constant_527620" type="Const" version="opset1">
|
|
<data element_type="i8" shape="128, 192, 1, 1" offset="1736316" size="24576" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="891" name="Convert_527621" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="892" name="Constant_527622" type="Const" version="opset1">
|
|
<data element_type="f32" shape="128, 1, 1, 1" offset="1760892" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="893" name="__module.model.19.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="894" name="__module.model.19.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>192</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="895" name="__module.model.19.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 128, 1, 1" offset="1761404" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="896" name="__module.model.19.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="795_1">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="897" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_51" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="795,input.193">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="898" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_51/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1761916" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="899" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_51/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1761920" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="900" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_51/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1761916" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="901" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_51/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1761920" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="902" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_51/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="903" name="783" type="Const" version="opset1">
|
|
<data element_type="i64" shape="" offset="73936" size="8" />
|
|
<output>
|
|
<port id="0" precision="I64" names="783" />
|
|
</output>
|
|
</layer>
|
|
<layer id="904" name="Constant_2965" type="Const" version="opset1">
|
|
<data element_type="i64" shape="2" offset="310908" size="16" />
|
|
<output>
|
|
<port id="0" precision="I64" names="797">
|
|
<dim>2</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="905" name="__module.model.19/prim::ListUnpack" type="VariadicSplit" version="opset1">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="I64" />
|
|
<port id="2" precision="I64">
|
|
<dim>2</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="799">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="4" precision="FP32" names="800,input.195">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="906" name="Constant_527624" type="Const" version="opset1">
|
|
<data element_type="i8" shape="32, 64, 3, 3" offset="1761924" size="18432" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="907" name="Convert_527625" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="908" name="Constant_527626" type="Const" version="opset1">
|
|
<data element_type="f32" shape="32, 1, 1, 1" offset="1780356" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="909" name="__module.model.19.m.0.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="910" name="__module.model.19.m.0.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>32</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="911" name="__module.model.19.m.0.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 32, 1, 1" offset="1780484" size="128" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="912" name="__module.model.19.m.0.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="810_1">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="913" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_52" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="810,input.197">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="914" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_52/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="915" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_52/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="310904" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="916" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_52/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="917" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_52/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="310904" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="918" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_52/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="919" name="Constant_527628" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 32, 3, 3" offset="1780612" size="18432" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="920" name="Convert_527629" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="921" name="Constant_527630" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="1799044" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="922" name="__module.model.19.m.0.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="923" name="__module.model.19.m.0.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>32</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>32</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="924" name="__module.model.19.m.0.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1799300" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="925" name="__module.model.19.m.0.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="819_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="926" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_53" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="819,input.201">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="927" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_53/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1799556" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="928" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_53/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1799560" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="929" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_53/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1799556" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="930" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_53/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1799560" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="931" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_53/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="932" name="__module.model.19.m.0/aten::add/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="821">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="933" name="__module.model.19.m.0/aten::add/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1761916" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="934" name="__module.model.19.m.0/aten::add/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1761920" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="935" name="__module.model.19.m.0/aten::add/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1761916" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="936" name="__module.model.19.m.0/aten::add/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1761920" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="937" name="__module.model.19.m.0/aten::add/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="938" name="__module.model.19/aten::cat/Concat" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="823,input.203">
|
|
<dim>1</dim>
|
|
<dim>192</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="939" name="Constant_527632" type="Const" version="opset1">
|
|
<data element_type="i8" shape="128, 192, 1, 1" offset="1799564" size="24576" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="940" name="Convert_527633" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="941" name="Constant_527634" type="Const" version="opset1">
|
|
<data element_type="f32" shape="128, 1, 1, 1" offset="1824140" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="942" name="__module.model.19.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="943" name="__module.model.19.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>192</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>192</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="944" name="__module.model.19.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 128, 1, 1" offset="1824652" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="945" name="__module.model.19.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="831_1">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="946" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_54" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="831,input.205">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="947" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_54/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="948" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_54/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75180" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="949" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_54/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="950" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_54/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75180" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="951" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_54/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="952" name="Constant_527636" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 128, 3, 3" offset="1825164" size="73728" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="953" name="Convert_527637" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="954" name="Constant_527638" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="1898892" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="955" name="__module.model.23.cv2.1.0.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="956" name="__module.model.23.cv2.1.0.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="957" name="__module.model.23.cv2.1.0.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1899148" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="958" name="__module.model.23.cv2.1.0.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1183_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="959" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_55" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1183,input.295">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="960" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_55/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1899404" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="961" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_55/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1899408" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="962" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_55/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1899404" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="963" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_55/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1899408" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="964" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_55/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="965" name="Constant_527640" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 3, 3" offset="1899412" size="36864" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="966" name="Convert_527641" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="967" name="Constant_527642" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="1936276" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="968" name="__module.model.23.cv2.1.1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="969" name="__module.model.23.cv2.1.1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="970" name="__module.model.23.cv2.1.1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1936532" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="971" name="__module.model.23.cv2.1.1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1192_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="972" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_56" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1192,input.299">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="973" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_56/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="974" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_56/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1936788" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="975" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_56/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="976" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_56/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1936788" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="977" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_56/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="978" name="Constant_527644" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 1, 1" offset="1936792" size="4096" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="979" name="Convert_527645" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="980" name="Constant_527646" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="1940888" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="981" name="__module.model.23.cv2.1.2/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="982" name="__module.model.23.cv2.1.2/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="983" name="__module.model.23.cv2.1.2/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1941144" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="984" name="__module.model.23.cv2.1.2/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1200">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="985" name="Constant_527648" type="Const" version="opset1">
|
|
<data element_type="i8" shape="128, 1, 1, 3, 3" offset="1941400" size="1152" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="986" name="Convert_527649" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="987" name="Constant_527650" type="Const" version="opset1">
|
|
<data element_type="f32" shape="128, 1, 1, 1, 1" offset="1942552" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="988" name="__module.model.23.cv3.1.0.0.conv/aten::_convolution/GroupConvolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="989" name="__module.model.23.cv3.1.0.0.conv/aten::_convolution/GroupConvolution" type="GroupConvolution" version="opset1">
|
|
<data strides="1, 1" pads_begin="1, 1" pads_end="1, 1" dilations="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="990" name="__module.model.23.cv3.1.0.0.conv/aten::_convolution/Reshape_1" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 128, 1, 1" offset="1943064" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="991" name="__module.model.23.cv3.1.0.0.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1213_1">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="992" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_57" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1213,input.303">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="993" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_57/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1943576" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="994" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_57/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1943580" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="995" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_57/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1943576" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="996" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_57/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1943580" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="997" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_57/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="998" name="Constant_527652" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 128, 1, 1" offset="1943584" size="8192" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="999" name="Convert_527653" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1000" name="Constant_527654" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="1951776" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1001" name="__module.model.23.cv3.1.0.1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1002" name="__module.model.23.cv3.1.0.1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1003" name="__module.model.23.cv3.1.0.1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1952032" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1004" name="__module.model.23.cv3.1.0.1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1222_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1005" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_58" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1222,input.307">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1006" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_58/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1952288" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1007" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_58/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1952544" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1008" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_58/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1952288" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1009" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_58/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1952544" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1010" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_58/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="3" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="4" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1011" name="Constant_527656" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 1, 1, 3, 3" offset="1952800" size="576" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1012" name="Convert_527657" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1013" name="Constant_527658" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1, 1" offset="1953376" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1014" name="__module.model.23.cv3.1.1.0.conv/aten::_convolution/GroupConvolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1015" name="__module.model.23.cv3.1.1.0.conv/aten::_convolution/GroupConvolution" type="GroupConvolution" version="opset1">
|
|
<data strides="1, 1" pads_begin="1, 1" pads_end="1, 1" dilations="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1016" name="__module.model.23.cv3.1.1.0.conv/aten::_convolution/Reshape_1" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1953632" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1017" name="__module.model.23.cv3.1.1.0.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1233_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1018" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_59" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1233,input.311">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1019" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_59/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1020" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_59/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1953888" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1021" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_59/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1022" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_59/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1953888" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1023" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_59/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1024" name="Constant_527660" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 1, 1" offset="1953892" size="4096" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1025" name="Convert_527661" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1026" name="Constant_527662" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="1957988" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1027" name="__module.model.23.cv3.1.1.1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1028" name="__module.model.23.cv3.1.1.1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1029" name="__module.model.23.cv3.1.1.1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="1958244" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1030" name="__module.model.23.cv3.1.1.1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1242_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1031" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_60" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1242,input.315">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1032" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_60/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1033" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_60/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1642260" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1034" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_60/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1035" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_60/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1642260" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1036" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_60/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1037" name="Constant_527664" type="Const" version="opset1">
|
|
<data element_type="i8" shape="38, 64, 1, 1" offset="1958500" size="2432" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>38</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1038" name="Convert_527665" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>38</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>38</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1039" name="Constant_527666" type="Const" version="opset1">
|
|
<data element_type="f32" shape="38, 1, 1, 1" offset="1960932" size="152" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>38</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1040" name="__module.model.23.cv3.1.2/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>38</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>38</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>38</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1041" name="__module.model.23.cv3.1.2/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>38</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1042" name="__module.model.23.cv3.1.2/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 38, 1, 1" offset="1961084" size="152" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1043" name="__module.model.23.cv3.1.2/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1250">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1044" name="__module.model.23/aten::cat/Concat_1" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1252,xi.3">
|
|
<dim>1</dim>
|
|
<dim>102</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1045" name="Constant_11042" type="Const" version="opset1">
|
|
<data element_type="i64" shape="3" offset="1698916" size="24" />
|
|
<output>
|
|
<port id="0" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1046" name="__module.model.23/aten::view/Reshape_1" type="Reshape" version="opset1">
|
|
<data special_zero="true" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>102</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1337">
|
|
<dim>1</dim>
|
|
<dim>102</dim>
|
|
<dim>1600</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1047" name="Constant_527668" type="Const" version="opset1">
|
|
<data element_type="i8" shape="128, 128, 3, 3" offset="1961236" size="147456" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1048" name="Convert_527669" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1049" name="Constant_527670" type="Const" version="opset1">
|
|
<data element_type="f32" shape="128, 1, 1, 1" offset="2108692" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1050" name="__module.model.20.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1051" name="__module.model.20.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="2, 2" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1052" name="__module.model.20.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 128, 1, 1" offset="2109204" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1053" name="__module.model.20.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="845_1">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1054" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_61" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="845,input.209">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1055" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_61/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1458260" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1056" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_61/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1458264" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1057" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_61/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1458260" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1058" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_61/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1458264" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1059" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_61/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1060" name="__module.model.21/aten::cat/Concat" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="849,input.211">
|
|
<dim>1</dim>
|
|
<dim>384</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1061" name="Constant_527672" type="Const" version="opset1">
|
|
<data element_type="i8" shape="256, 384, 1, 1" offset="2109716" size="98304" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1062" name="Convert_527673" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1063" name="Constant_527674" type="Const" version="opset1">
|
|
<data element_type="f32" shape="256, 1, 1, 1" offset="2208020" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1064" name="__module.model.22.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1065" name="__module.model.22.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>384</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1066" name="__module.model.22.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 256, 1, 1" offset="2209044" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1067" name="__module.model.22.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="866_1">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1068" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_62" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="866,input.213">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1069" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_62/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1070" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_62/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="117440" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1071" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_62/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1072" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_62/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="117440" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1073" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_62/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1074" name="854" type="Const" version="opset1">
|
|
<data element_type="i64" shape="" offset="73936" size="8" />
|
|
<output>
|
|
<port id="0" precision="I64" names="854" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1075" name="Constant_3226" type="Const" version="opset1">
|
|
<data element_type="i64" shape="2" offset="748232" size="16" />
|
|
<output>
|
|
<port id="0" precision="I64" names="868">
|
|
<dim>2</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1076" name="__module.model.22/prim::ListUnpack" type="VariadicSplit" version="opset1">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="I64" />
|
|
<port id="2" precision="I64">
|
|
<dim>2</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="870">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="4" precision="FP32" names="871,input.215">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1077" name="Constant_527676" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 128, 1, 1" offset="2210068" size="8192" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1078" name="Convert_527677" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1079" name="Constant_527678" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="2218260" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1080" name="__module.model.22.m.0.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1081" name="__module.model.22.m.0.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1082" name="__module.model.22.m.0.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="2218516" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1083" name="__module.model.22.m.0.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="883_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1084" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_63" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="883,input.217">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1085" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_63/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2218772" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1086" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_63/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2218776" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1087" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_63/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2218772" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1088" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_63/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2218776" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1089" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_63/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1090" name="Constant_527680" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 3, 3" offset="2218780" size="36864" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1091" name="Convert_527681" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1092" name="Constant_527682" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="2255644" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1093" name="__module.model.22.m.0.m.0.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1094" name="__module.model.22.m.0.m.0.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1095" name="__module.model.22.m.0.m.0.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="2255900" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1096" name="__module.model.22.m.0.m.0.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="896_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1097" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_64" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="896,input.221">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1098" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_64/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1099" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_64/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1032972" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1100" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_64/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1101" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_64/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1032972" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1102" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_64/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1103" name="Constant_527684" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 3, 3" offset="2256156" size="36864" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1104" name="Convert_527685" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1105" name="Constant_527686" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="2293020" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1106" name="__module.model.22.m.0.m.0.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1107" name="__module.model.22.m.0.m.0.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1108" name="__module.model.22.m.0.m.0.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="2293276" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1109" name="__module.model.22.m.0.m.0.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="905_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1110" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_65" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="905,input.225">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1111" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_65/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1112" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_65/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="117440" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1113" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_65/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1114" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_65/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="117440" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1115" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_65/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1116" name="__module.model.22.m.0.m.0/aten::add/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="907,input.227">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1117" name="__module.model.22.m.0.m.0/aten::add/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2293532" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1118" name="__module.model.22.m.0.m.0/aten::add/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2293536" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1119" name="__module.model.22.m.0.m.0/aten::add/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2293532" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1120" name="__module.model.22.m.0.m.0/aten::add/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2293536" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1121" name="__module.model.22.m.0.m.0/aten::add/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1122" name="Constant_527688" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 3, 3" offset="2293540" size="36864" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1123" name="Convert_527689" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1124" name="Constant_527690" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="2330404" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1125" name="__module.model.22.m.0.m.1.cv1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1126" name="__module.model.22.m.0.m.1.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1127" name="__module.model.22.m.0.m.1.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="2330660" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1128" name="__module.model.22.m.0.m.1.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="917_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1129" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_66" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="917,input.229">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1130" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_66/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2330916" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1131" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_66/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2330920" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1132" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_66/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2330916" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1133" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_66/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2330920" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1134" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_66/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1135" name="Constant_527692" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 3, 3" offset="2330924" size="36864" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1136" name="Convert_527693" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1137" name="Constant_527694" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="2367788" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1138" name="__module.model.22.m.0.m.1.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1139" name="__module.model.22.m.0.m.1.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1140" name="__module.model.22.m.0.m.1.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="2368044" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1141" name="__module.model.22.m.0.m.1.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="926_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1142" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_67" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="926,input.233">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1143" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_67/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1144" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_67/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2368300" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1145" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_67/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1146" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_67/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2368300" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1147" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_67/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1148" name="__module.model.22.m.0.m.1/aten::add/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="928">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1149" name="__module.model.22.m.0.m.1/aten::add/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2368304" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1150" name="__module.model.22.m.0.m.1/aten::add/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2368308" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1151" name="__module.model.22.m.0.m.1/aten::add/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2368304" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1152" name="__module.model.22.m.0.m.1/aten::add/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2368308" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1153" name="__module.model.22.m.0.m.1/aten::add/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1154" name="Constant_527696" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 128, 1, 1" offset="2368312" size="8192" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1155" name="Convert_527697" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1156" name="Constant_527698" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="2376504" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1157" name="__module.model.22.m.0.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1158" name="__module.model.22.m.0.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1159" name="__module.model.22.m.0.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="2376760" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1160" name="__module.model.22.m.0.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="936_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1161" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_68" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="936,input.235">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1162" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_68/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2368304" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1163" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_68/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2368308" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1164" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_68/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2368304" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1165" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_68/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2368308" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1166" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_68/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1167" name="__module.model.22.m.0/aten::cat/Concat" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="939,input.237">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1168" name="Constant_527700" type="Const" version="opset1">
|
|
<data element_type="i8" shape="128, 128, 1, 1" offset="2377016" size="16384" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1169" name="Convert_527701" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1170" name="Constant_527702" type="Const" version="opset1">
|
|
<data element_type="f32" shape="128, 1, 1, 1" offset="2393400" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1171" name="__module.model.22.m.0.cv3.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1172" name="__module.model.22.m.0.cv3.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>128</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1173" name="__module.model.22.m.0.cv3.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 128, 1, 1" offset="2393912" size="512" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1174" name="__module.model.22.m.0.cv3.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="947_1">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1175" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_69" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="947,input.239">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1176" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_69/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1177" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_69/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="117440" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1178" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_69/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1179" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_69/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="117440" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1180" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_69/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1181" name="__module.model.22/aten::cat/Concat" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="950,input.241">
|
|
<dim>1</dim>
|
|
<dim>384</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1182" name="Constant_527704" type="Const" version="opset1">
|
|
<data element_type="i8" shape="256, 384, 1, 1" offset="2394424" size="98304" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1183" name="Convert_527705" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1184" name="Constant_527706" type="Const" version="opset1">
|
|
<data element_type="f32" shape="256, 1, 1, 1" offset="2492728" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1185" name="__module.model.22.cv2.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1186" name="__module.model.22.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>384</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>384</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1187" name="__module.model.22.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 256, 1, 1" offset="2493752" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1188" name="__module.model.22.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="958_1">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1189" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_70" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="958,input.243">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1190" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_70/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1191" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_70/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75180" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1192" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_70/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1193" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_70/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75180" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1194" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_70/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1195" name="Constant_527708" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 256, 3, 3" offset="2494776" size="147456" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>256</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1196" name="Convert_527709" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>256</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>256</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1197" name="Constant_527710" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="2642232" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1198" name="__module.model.23.cv2.2.0.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>256</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>256</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1199" name="__module.model.23.cv2.2.0.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>256</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1200" name="__module.model.23.cv2.2.0.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="2642488" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1201" name="__module.model.23.cv2.2.0.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1263_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1202" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_71" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1263,input.319">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1203" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_71/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2642744" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1204" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_71/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2642748" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1205" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_71/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2642744" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1206" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_71/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2642748" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1207" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_71/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1208" name="Constant_527712" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 3, 3" offset="2642752" size="36864" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1209" name="Convert_527713" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1210" name="Constant_527714" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="2679616" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1211" name="__module.model.23.cv2.2.1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1212" name="__module.model.23.cv2.2.1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1213" name="__module.model.23.cv2.2.1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="2679872" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1214" name="__module.model.23.cv2.2.1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1272_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1215" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_72" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1272,input.323">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1216" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_72/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2680128" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1217" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_72/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2680132" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1218" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_72/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2680128" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1219" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_72/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2680132" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1220" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_72/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1221" name="Constant_527716" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 1, 1" offset="2680136" size="4096" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1222" name="Convert_527717" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1223" name="Constant_527718" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="2684232" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1224" name="__module.model.23.cv2.2.2/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1225" name="__module.model.23.cv2.2.2/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1226" name="__module.model.23.cv2.2.2/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="2684488" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1227" name="__module.model.23.cv2.2.2/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1280">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1228" name="Constant_527720" type="Const" version="opset1">
|
|
<data element_type="i8" shape="256, 1, 1, 3, 3" offset="2684744" size="2304" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1229" name="Convert_527721" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1230" name="Constant_527722" type="Const" version="opset1">
|
|
<data element_type="f32" shape="256, 1, 1, 1, 1" offset="2687048" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1231" name="__module.model.23.cv3.2.0.0.conv/aten::_convolution/GroupConvolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1232" name="__module.model.23.cv3.2.0.0.conv/aten::_convolution/GroupConvolution" type="GroupConvolution" version="opset1">
|
|
<data strides="1, 1" pads_begin="1, 1" pads_end="1, 1" dilations="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1233" name="__module.model.23.cv3.2.0.0.conv/aten::_convolution/Reshape_1" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 256, 1, 1" offset="2688072" size="1024" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1234" name="__module.model.23.cv3.2.0.0.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1293_1">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1235" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_73" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1293,input.327">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1236" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_73/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1237" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_73/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75180" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1238" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_73/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1239" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_73/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75180" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1240" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_73/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1241" name="Constant_527724" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 256, 1, 1" offset="2689096" size="16384" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1242" name="Convert_527725" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1243" name="Constant_527726" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="2705480" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1244" name="__module.model.23.cv3.2.0.1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1245" name="__module.model.23.cv3.2.0.1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>256</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1246" name="__module.model.23.cv3.2.0.1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="2705736" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1247" name="__module.model.23.cv3.2.0.1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1302_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1248" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_74" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1302,input.331">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1249" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_74/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="2705992" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1250" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_74/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="2706248" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1251" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_74/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="2705992" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1252" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_74/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="2706248" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1253" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_74/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="3" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="4" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1254" name="Constant_527728" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 1, 1, 3, 3" offset="2706504" size="576" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1255" name="Convert_527729" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1256" name="Constant_527730" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1, 1" offset="2707080" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1257" name="__module.model.23.cv3.2.1.0.conv/aten::_convolution/GroupConvolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1258" name="__module.model.23.cv3.2.1.0.conv/aten::_convolution/GroupConvolution" type="GroupConvolution" version="opset1">
|
|
<data strides="1, 1" pads_begin="1, 1" pads_end="1, 1" dilations="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1259" name="__module.model.23.cv3.2.1.0.conv/aten::_convolution/Reshape_1" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="2707336" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1260" name="__module.model.23.cv3.2.1.0.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1313_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1261" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_75" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1313,input.335">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1262" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_75/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1263" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_75/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1032972" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1264" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_75/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1265" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_75/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="1032972" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1266" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_75/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1267" name="Constant_527732" type="Const" version="opset1">
|
|
<data element_type="i8" shape="64, 64, 1, 1" offset="2707592" size="4096" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1268" name="Convert_527733" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1269" name="Constant_527734" type="Const" version="opset1">
|
|
<data element_type="f32" shape="64, 1, 1, 1" offset="2711688" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1270" name="__module.model.23.cv3.2.1.1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1271" name="__module.model.23.cv3.2.1.1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>64</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1272" name="__module.model.23.cv3.2.1.1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 64, 1, 1" offset="2711944" size="256" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1273" name="__module.model.23.cv3.2.1.1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1322_1">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1274" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_76" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1322,input.339">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1275" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_76/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2712200" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1276" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_76/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2712204" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1277" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_76/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2712200" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1278" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_76/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2712204" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1279" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_76/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1280" name="Constant_527736" type="Const" version="opset1">
|
|
<data element_type="i8" shape="38, 64, 1, 1" offset="2712208" size="2432" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>38</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1281" name="Convert_527737" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>38</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>38</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1282" name="Constant_527738" type="Const" version="opset1">
|
|
<data element_type="f32" shape="38, 1, 1, 1" offset="2714640" size="152" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>38</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1283" name="__module.model.23.cv3.2.2/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>38</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>38</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>38</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1284" name="__module.model.23.cv3.2.2/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>38</dim>
|
|
<dim>64</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1285" name="__module.model.23.cv3.2.2/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 38, 1, 1" offset="2714792" size="152" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1286" name="__module.model.23.cv3.2.2/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1330">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1287" name="__module.model.23/aten::cat/Concat_2" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1332,xi">
|
|
<dim>1</dim>
|
|
<dim>102</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1288" name="Constant_11043" type="Const" version="opset1">
|
|
<data element_type="i64" shape="3" offset="1698916" size="24" />
|
|
<output>
|
|
<port id="0" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1289" name="__module.model.23/aten::view/Reshape_2" type="Reshape" version="opset1">
|
|
<data special_zero="true" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>102</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1339">
|
|
<dim>1</dim>
|
|
<dim>102</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1290" name="__module.model.23/aten::cat/Concat_3" type="Concat" version="opset1">
|
|
<data axis="2" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>102</dim>
|
|
<dim>6400</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>102</dim>
|
|
<dim>1600</dim>
|
|
</port>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>102</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="1341">
|
|
<dim>1</dim>
|
|
<dim>102</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1291" name="977" type="Const" version="opset1">
|
|
<data element_type="i64" shape="" offset="73936" size="8" />
|
|
<output>
|
|
<port id="0" precision="I64" names="977" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1292" name="Constant_5544" type="Const" version="opset1">
|
|
<data element_type="i64" shape="2" offset="2714944" size="16" />
|
|
<output>
|
|
<port id="0" precision="I64" names="1342">
|
|
<dim>2</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1293" name="__module.model.23/prim::ListUnpack" type="VariadicSplit" version="opset1">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>102</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="1" precision="I64" />
|
|
<port id="2" precision="I64">
|
|
<dim>2</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="1344,x.5">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="4" precision="FP32" names="1345,cls">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1294" name="Constant_11044" type="Const" version="opset1">
|
|
<data element_type="i64" shape="4" offset="2714960" size="32" />
|
|
<output>
|
|
<port id="0" precision="I64">
|
|
<dim>4</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1295" name="__module.model.23.dfl/aten::view/Reshape" type="Reshape" version="opset1">
|
|
<data special_zero="true" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="1" precision="I64">
|
|
<dim>4</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1350">
|
|
<dim>1</dim>
|
|
<dim>4</dim>
|
|
<dim>16</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1296" name="__module.model.23.dfl/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
|
|
<data element_type="i32" shape="4" offset="2714992" size="16" />
|
|
<output>
|
|
<port id="0" precision="I32">
|
|
<dim>4</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1297" name="__module.model.23.dfl/aten::transpose/Transpose" type="Transpose" version="opset1">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>4</dim>
|
|
<dim>16</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="1" precision="I32">
|
|
<dim>4</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1351">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>4</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1298" name="__module.model.23.dfl/aten::softmax/Softmax" type="SoftMax" version="opset8">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>4</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1352,input">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>4</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1299" name="self.model.23.dfl.conv.weight" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 16, 1, 1" offset="2715008" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32" names="self.model.23.dfl.conv.weight">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1300" name="__module.model.23.dfl.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>4</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1358">
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>4</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1301" name="Constant_11045" type="Const" version="opset1">
|
|
<data element_type="i64" shape="3" offset="2715072" size="24" />
|
|
<output>
|
|
<port id="0" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1302" name="__module.model.23.dfl/aten::view/Reshape_1" type="Reshape" version="opset1">
|
|
<data special_zero="true" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>4</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="1" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1360,distance">
|
|
<dim>1</dim>
|
|
<dim>4</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1303" name="__module.model.23/prim::ListUnpack/ShapeOf" type="ShapeOf" version="opset3">
|
|
<data output_type="i32" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>4</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="I32">
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1304" name="Constant_7557" type="Const" version="opset1">
|
|
<data element_type="i32" shape="1" offset="67200" size="4" />
|
|
<output>
|
|
<port id="0" precision="I32">
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1305" name="__module.model.23/prim::ListUnpack/Gather" type="Gather" version="opset8">
|
|
<data batch_dims="0" />
|
|
<input>
|
|
<port id="0" precision="I32">
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="I64" />
|
|
<port id="2" precision="I32">
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="I32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1306" name="__module.model.23/prim::ListUnpack/Convert" type="Const" version="opset1">
|
|
<data element_type="i32" shape="" offset="2715096" size="4" />
|
|
<output>
|
|
<port id="0" precision="I32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1307" name="__module.model.23/prim::ListUnpack/Divide" type="Divide" version="opset1">
|
|
<data auto_broadcast="numpy" m_pythondiv="true" />
|
|
<input>
|
|
<port id="0" precision="I32" />
|
|
<port id="1" precision="I32" />
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="I32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1308" name="__module.model.23/prim::ListUnpack/Mod" type="Mod" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="I32" />
|
|
<port id="1" precision="I32" />
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="I32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1309" name="__module.model.23/prim::ListUnpack/Greater" type="Greater" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="I32" />
|
|
<port id="1" precision="I32">
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="BOOL">
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1310" name="Convert_11025" type="Convert" version="opset1">
|
|
<data destination_type="i32" />
|
|
<input>
|
|
<port id="0" precision="BOOL">
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="I32">
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1311" name="__module.model.23/prim::ListUnpack/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="I32" />
|
|
<port id="1" precision="I32">
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="I32">
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1312" name="Constant_7558" type="Const" version="opset1">
|
|
<data element_type="i32" shape="1" offset="2715100" size="4" />
|
|
<output>
|
|
<port id="0" precision="I32">
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1313" name="__module.model.23/prim::ListUnpack/Concat" type="Concat" version="opset1">
|
|
<data axis="0" />
|
|
<input>
|
|
<port id="0" precision="I32">
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="I32">
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="I32">
|
|
<dim>2</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1314" name="__module.model.23/prim::ListUnpack/VariadicSplit" type="VariadicSplit" version="opset1">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>4</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="1" precision="I64" />
|
|
<port id="2" precision="I32">
|
|
<dim>2</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="1363,lt">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="4" precision="FP32" names="1364,rb">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1315" name="__module.model.23/aten::sub/Subtract" type="Subtract" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1365,x1y1">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1316" name="__module.model.23/aten::add/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1366,x2y2">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1317" name="__module.model.23/aten::add/Add_1" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1367">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1318" name="Constant_10820" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 1, 1" offset="2715104" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1319" name="__module.model.23/aten::div/Divide" type="Divide" version="opset1">
|
|
<data auto_broadcast="numpy" m_pythondiv="true" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1368,c_xy">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1320" name="__module.model.23/aten::sub/Subtract_1" type="Subtract" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1369,wh">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1321" name="__module.model.23/aten::cat/Concat_4" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1371">
|
|
<dim>1</dim>
|
|
<dim>4</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1322" name="Constant_10821" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 1, 8400" offset="2715108" size="33600" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1323" name="__module.model.23/aten::mul/Multiply" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>4</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1372,dbox">
|
|
<dim>1</dim>
|
|
<dim>4</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1324" name="__module.model.23/aten::sigmoid/Sigmoid" type="Sigmoid" version="opset1">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1373">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1325" name="Constant_527740" type="Const" version="opset1">
|
|
<data element_type="i8" shape="16, 64, 3, 3" offset="2748708" size="9216" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1326" name="Convert_527741" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1327" name="Constant_527742" type="Const" version="opset1">
|
|
<data element_type="f32" shape="16, 1, 1, 1" offset="2757924" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1328" name="__module.model.23.cv4.0.0.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1329" name="__module.model.23.cv4.0.0.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>64</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>64</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1330" name="__module.model.23.cv4.0.0.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 16, 1, 1" offset="2757988" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1331" name="__module.model.23.cv4.0.0.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1011_1">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1332" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_77" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1011,input.245">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1333" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_77/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2758052" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1334" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_77/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2758056" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1335" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_77/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2758052" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1336" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_77/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2758056" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1337" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_77/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1338" name="Constant_527744" type="Const" version="opset1">
|
|
<data element_type="i8" shape="16, 16, 3, 3" offset="2758060" size="2304" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1339" name="Convert_527745" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1340" name="Constant_527746" type="Const" version="opset1">
|
|
<data element_type="f32" shape="16, 1, 1, 1" offset="2760364" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1341" name="__module.model.23.cv4.0.1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1342" name="__module.model.23.cv4.0.1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1343" name="__module.model.23.cv4.0.1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 16, 1, 1" offset="2760428" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1344" name="__module.model.23.cv4.0.1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1020_1">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1345" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_78" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1020,input.249">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1346" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_78/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1347" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_78/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2760492" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1348" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_78/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="75176" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1349" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_78/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2760492" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1350" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_78/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1351" name="Constant_527748" type="Const" version="opset1">
|
|
<data element_type="i8" shape="8, 16, 1, 1" offset="2760496" size="128" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1352" name="Convert_527749" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1353" name="Constant_527750" type="Const" version="opset1">
|
|
<data element_type="f32" shape="8, 1, 1, 1" offset="2760624" size="32" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1354" name="__module.model.23.cv4.0.2/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1355" name="__module.model.23.cv4.0.2/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1356" name="__module.model.23.cv4.0.2/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 8, 1, 1" offset="2760656" size="32" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1357" name="__module.model.23.cv4.0.2/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1028">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1358" name="__module.model.23.cv4.0.2/aten::_convolution/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2760688" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1359" name="__module.model.23.cv4.0.2/aten::_convolution/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2760692" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1360" name="__module.model.23.cv4.0.2/aten::_convolution/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2760688" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1361" name="__module.model.23.cv4.0.2/aten::_convolution/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2760692" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1362" name="__module.model.23.cv4.0.2/aten::_convolution/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1363" name="Constant_11046" type="Const" version="opset1">
|
|
<data element_type="i64" shape="3" offset="2760696" size="24" />
|
|
<output>
|
|
<port id="0" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1364" name="__module.model.23/aten::view/Reshape_3" type="Reshape" version="opset1">
|
|
<data special_zero="true" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>80</dim>
|
|
<dim>80</dim>
|
|
</port>
|
|
<port id="1" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1030">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>6400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1365" name="Constant_527752" type="Const" version="opset1">
|
|
<data element_type="i8" shape="16, 128, 3, 3" offset="2760720" size="18432" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1366" name="Convert_527753" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1367" name="Constant_527754" type="Const" version="opset1">
|
|
<data element_type="f32" shape="16, 1, 1, 1" offset="2779152" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1368" name="__module.model.23.cv4.1.0.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1369" name="__module.model.23.cv4.1.0.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>128</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>128</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1370" name="__module.model.23.cv4.1.0.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 16, 1, 1" offset="2779216" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1371" name="__module.model.23.cv4.1.0.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1041_1">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1372" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_79" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1041,input.253">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1373" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_79/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2779280" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1374" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_79/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2779284" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1375" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_79/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2779280" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1376" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_79/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2779284" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1377" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_79/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1378" name="Constant_527756" type="Const" version="opset1">
|
|
<data element_type="i8" shape="16, 16, 3, 3" offset="2779288" size="2304" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1379" name="Convert_527757" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1380" name="Constant_527758" type="Const" version="opset1">
|
|
<data element_type="f32" shape="16, 1, 1, 1" offset="2781592" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1381" name="__module.model.23.cv4.1.1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1382" name="__module.model.23.cv4.1.1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1383" name="__module.model.23.cv4.1.1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 16, 1, 1" offset="2781656" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1384" name="__module.model.23.cv4.1.1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1050_1">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1385" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_80" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1050,input.257">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1386" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_80/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="313228" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1387" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_80/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2781720" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1388" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_80/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="313228" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1389" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_80/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2781720" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1390" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_80/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1391" name="Constant_527760" type="Const" version="opset1">
|
|
<data element_type="i8" shape="8, 16, 1, 1" offset="2781724" size="128" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1392" name="Convert_527761" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1393" name="Constant_527762" type="Const" version="opset1">
|
|
<data element_type="f32" shape="8, 1, 1, 1" offset="2781852" size="32" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1394" name="__module.model.23.cv4.1.2/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1395" name="__module.model.23.cv4.1.2/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1396" name="__module.model.23.cv4.1.2/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 8, 1, 1" offset="2781884" size="32" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1397" name="__module.model.23.cv4.1.2/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1058">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1398" name="__module.model.23.cv4.1.2/aten::_convolution/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2781916" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1399" name="__module.model.23.cv4.1.2/aten::_convolution/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2781920" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1400" name="__module.model.23.cv4.1.2/aten::_convolution/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2781916" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1401" name="__module.model.23.cv4.1.2/aten::_convolution/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2781920" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1402" name="__module.model.23.cv4.1.2/aten::_convolution/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1403" name="Constant_11047" type="Const" version="opset1">
|
|
<data element_type="i64" shape="3" offset="2760696" size="24" />
|
|
<output>
|
|
<port id="0" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1404" name="__module.model.23/aten::view/Reshape_4" type="Reshape" version="opset1">
|
|
<data special_zero="true" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>40</dim>
|
|
<dim>40</dim>
|
|
</port>
|
|
<port id="1" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1060">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>1600</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1405" name="Constant_527764" type="Const" version="opset1">
|
|
<data element_type="i8" shape="16, 256, 3, 3" offset="2781924" size="36864" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>256</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1406" name="Convert_527765" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>256</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>256</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1407" name="Constant_527766" type="Const" version="opset1">
|
|
<data element_type="f32" shape="16, 1, 1, 1" offset="2818788" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1408" name="__module.model.23.cv4.2.0.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>256</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>256</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1409" name="__module.model.23.cv4.2.0.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>256</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>256</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1410" name="__module.model.23.cv4.2.0.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 16, 1, 1" offset="2818852" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1411" name="__module.model.23.cv4.2.0.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1071_1">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1412" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_81" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1071,input.263">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1413" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_81/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2818916" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1414" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_81/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2818920" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1415" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_81/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2818916" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1416" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_81/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2818920" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1417" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_81/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1418" name="Constant_527768" type="Const" version="opset1">
|
|
<data element_type="i8" shape="16, 16, 3, 3" offset="2818924" size="2304" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1419" name="Convert_527769" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>16</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1420" name="Constant_527770" type="Const" version="opset1">
|
|
<data element_type="f32" shape="16, 1, 1, 1" offset="2821228" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1421" name="__module.model.23.cv4.2.1.conv/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1422" name="__module.model.23.cv4.2.1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>16</dim>
|
|
<dim>16</dim>
|
|
<dim>3</dim>
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1423" name="__module.model.23.cv4.2.1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 16, 1, 1" offset="2821292" size="64" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1424" name="__module.model.23.cv4.2.1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1080_1">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1425" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_82" type="Swish" version="opset4">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32" names="1080,input.267">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1426" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_82/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2821356" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1427" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_82/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2821360" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1428" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_82/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2821356" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1429" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_82/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2821360" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1430" name="__module.model.23.cv4.2.1.act/aten::silu_/Swish_82/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1431" name="Constant_527772" type="Const" version="opset1">
|
|
<data element_type="i8" shape="8, 16, 1, 1" offset="2821364" size="128" />
|
|
<output>
|
|
<port id="0" precision="I8">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1432" name="Convert_527773" type="Convert" version="opset1">
|
|
<data destination_type="f32" />
|
|
<input>
|
|
<port id="0" precision="I8">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1433" name="Constant_527774" type="Const" version="opset1">
|
|
<data element_type="f32" shape="8, 1, 1, 1" offset="2821492" size="32" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1434" name="__module.model.23.cv4.2.2/aten::_convolution/Convolution/fq_weights_1" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1435" name="__module.model.23.cv4.2.2/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>16</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>8</dim>
|
|
<dim>16</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1436" name="__module.model.23.cv4.2.2/aten::_convolution/Reshape" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 8, 1, 1" offset="2821524" size="32" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1437" name="__module.model.23.cv4.2.2/aten::_convolution/Add" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1088">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1438" name="__module.model.23.cv4.2.2/aten::_convolution/Add/fq_output_0/input_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2821556" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1439" name="__module.model.23.cv4.2.2/aten::_convolution/Add/fq_output_0/input_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2821560" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1440" name="__module.model.23.cv4.2.2/aten::_convolution/Add/fq_output_0/output_low" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2821556" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1441" name="__module.model.23.cv4.2.2/aten::_convolution/Add/fq_output_0/output_high" type="Const" version="opset1">
|
|
<data element_type="f32" shape="" offset="2821560" size="4" />
|
|
<output>
|
|
<port id="0" precision="FP32" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1442" name="__module.model.23.cv4.2.2/aten::_convolution/Add/fq_output_0" type="FakeQuantize" version="opset1">
|
|
<data levels="256" auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="FP32" />
|
|
<port id="2" precision="FP32" />
|
|
<port id="3" precision="FP32" />
|
|
<port id="4" precision="FP32" />
|
|
</input>
|
|
<output>
|
|
<port id="5" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1443" name="Constant_11048" type="Const" version="opset1">
|
|
<data element_type="i64" shape="3" offset="2760696" size="24" />
|
|
<output>
|
|
<port id="0" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1444" name="__module.model.23/aten::view/Reshape_5" type="Reshape" version="opset1">
|
|
<data special_zero="true" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>20</dim>
|
|
<dim>20</dim>
|
|
</port>
|
|
<port id="1" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1090">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1445" name="__module.model.23/aten::cat/Concat_5" type="Concat" version="opset1">
|
|
<data axis="-1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>6400</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>1600</dim>
|
|
</port>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>400</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32" names="1092,kpts">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1446" name="Constant_11049" type="Const" version="opset1">
|
|
<data element_type="i64" shape="4" offset="2821564" size="32" />
|
|
<output>
|
|
<port id="0" precision="I64">
|
|
<dim>4</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1447" name="__module.model.23/aten::view/Reshape_6" type="Reshape" version="opset1">
|
|
<data special_zero="true" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="1" precision="I64">
|
|
<dim>4</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1375,1376,1377,y">
|
|
<dim>1</dim>
|
|
<dim>4</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1448" name="Constant_10333" type="Const" version="opset1">
|
|
<data element_type="i64" shape="3" offset="2821596" size="24" />
|
|
<output>
|
|
<port id="0" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1449" name="Constant_10336" type="Const" version="opset1">
|
|
<data element_type="i64" shape="3" offset="2821620" size="24" />
|
|
<output>
|
|
<port id="0" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1450" name="Constant_10339" type="Const" version="opset1">
|
|
<data element_type="i64" shape="3" offset="2821644" size="24" />
|
|
<output>
|
|
<port id="0" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1451" name="__module.model.23/aten::slice/Slice_2" type="StridedSlice" version="opset1">
|
|
<data begin_mask="1, 1, 0" end_mask="1, 1, 0" new_axis_mask="" shrink_axis_mask="" ellipsis_mask="" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>4</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="1" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="2" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
<port id="3" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="4" precision="FP32" names="1378">
|
|
<dim>1</dim>
|
|
<dim>4</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1452" name="Constant_10822" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 1, 1, 8400" offset="2821668" size="33600" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1453" name="Multiply_10754" type="Multiply" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>4</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>4</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1454" name="Constant_10823" type="Const" version="opset1">
|
|
<data element_type="f32" shape="1, 1, 2, 8400" offset="2855268" size="67200" />
|
|
<output>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1455" name="__module.model.23/aten::mul/Multiply_2" type="Add" version="opset1">
|
|
<data auto_broadcast="numpy" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>4</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>1</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1382,a">
|
|
<dim>1</dim>
|
|
<dim>4</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1456" name="ShapeOf_10999" type="ShapeOf" version="opset3">
|
|
<data output_type="i64" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>3</dim>
|
|
<dim>640</dim>
|
|
<dim>640</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="1" precision="I64">
|
|
<dim>4</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1457" name="Constant_11000" type="Const" version="opset1">
|
|
<data element_type="i64" shape="1" offset="1302420" size="8" />
|
|
<output>
|
|
<port id="0" precision="I64">
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1458" name="Constant_11001" type="Const" version="opset1">
|
|
<data element_type="i64" shape="" offset="1302420" size="8" />
|
|
<output>
|
|
<port id="0" precision="I64" />
|
|
</output>
|
|
</layer>
|
|
<layer id="1459" name="Gather_11002" type="Gather" version="opset8">
|
|
<data batch_dims="0" />
|
|
<input>
|
|
<port id="0" precision="I64">
|
|
<dim>4</dim>
|
|
</port>
|
|
<port id="1" precision="I64">
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="2" precision="I64" />
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="I64" names="1000,1333,1347,557">
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1460" name="__module.model.23/prim::ListConstruct/Reshape_0_3" type="Const" version="opset1">
|
|
<data element_type="i64" shape="1" offset="2922468" size="8" />
|
|
<output>
|
|
<port id="0" precision="I64">
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1461" name="__module.model.23/prim::ListConstruct/Reshape_1" type="Const" version="opset1">
|
|
<data element_type="i64" shape="1" offset="2922476" size="8" />
|
|
<output>
|
|
<port id="0" precision="I64">
|
|
<dim>1</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1462" name="__module.model.23/prim::ListConstruct/Concat_3" type="Concat" version="opset1">
|
|
<data axis="0" />
|
|
<input>
|
|
<port id="0" precision="I64">
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="1" precision="I64">
|
|
<dim>1</dim>
|
|
</port>
|
|
<port id="2" precision="I64">
|
|
<dim>1</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="I64" names="1029,1059,1089,1383">
|
|
<dim>3</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1463" name="__module.model.23/aten::view/Reshape_7" type="Reshape" version="opset1">
|
|
<data special_zero="false" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>4</dim>
|
|
<dim>2</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="1" precision="I64">
|
|
<dim>3</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="2" precision="FP32" names="1384,pred_kpt">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1464" name="__module.model.23/aten::cat/Concat_6" type="Concat" version="opset1">
|
|
<data axis="1" />
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>4</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="1" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>38</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
<port id="2" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>8</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</input>
|
|
<output>
|
|
<port id="3" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>50</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</output>
|
|
</layer>
|
|
<layer id="1465" name="Result_5709" type="Result" version="opset1">
|
|
<input>
|
|
<port id="0" precision="FP32">
|
|
<dim>1</dim>
|
|
<dim>50</dim>
|
|
<dim>8400</dim>
|
|
</port>
|
|
</input>
|
|
</layer>
|
|
</layers>
|
|
<edges>
|
|
<edge from-layer="0" from-port="0" to-layer="6" to-port="0" />
|
|
<edge from-layer="1" from-port="0" to-layer="1315" to-port="0" />
|
|
<edge from-layer="1" from-port="0" to-layer="1316" to-port="0" />
|
|
<edge from-layer="2" from-port="0" to-layer="6" to-port="1" />
|
|
<edge from-layer="3" from-port="0" to-layer="6" to-port="2" />
|
|
<edge from-layer="4" from-port="0" to-layer="6" to-port="3" />
|
|
<edge from-layer="5" from-port="0" to-layer="6" to-port="4" />
|
|
<edge from-layer="6" from-port="5" to-layer="1456" to-port="0" />
|
|
<edge from-layer="6" from-port="5" to-layer="11" to-port="0" />
|
|
<edge from-layer="7" from-port="0" to-layer="8" to-port="0" />
|
|
<edge from-layer="8" from-port="1" to-layer="10" to-port="0" />
|
|
<edge from-layer="9" from-port="0" to-layer="10" to-port="1" />
|
|
<edge from-layer="10" from-port="2" to-layer="11" to-port="1" />
|
|
<edge from-layer="11" from-port="2" to-layer="13" to-port="0" />
|
|
<edge from-layer="12" from-port="0" to-layer="13" to-port="1" />
|
|
<edge from-layer="13" from-port="2" to-layer="14" to-port="0" />
|
|
<edge from-layer="14" from-port="1" to-layer="19" to-port="0" />
|
|
<edge from-layer="15" from-port="0" to-layer="19" to-port="1" />
|
|
<edge from-layer="16" from-port="0" to-layer="19" to-port="2" />
|
|
<edge from-layer="17" from-port="0" to-layer="19" to-port="3" />
|
|
<edge from-layer="18" from-port="0" to-layer="19" to-port="4" />
|
|
<edge from-layer="19" from-port="5" to-layer="24" to-port="0" />
|
|
<edge from-layer="20" from-port="0" to-layer="21" to-port="0" />
|
|
<edge from-layer="21" from-port="1" to-layer="23" to-port="0" />
|
|
<edge from-layer="22" from-port="0" to-layer="23" to-port="1" />
|
|
<edge from-layer="23" from-port="2" to-layer="24" to-port="1" />
|
|
<edge from-layer="24" from-port="2" to-layer="26" to-port="0" />
|
|
<edge from-layer="25" from-port="0" to-layer="26" to-port="1" />
|
|
<edge from-layer="26" from-port="2" to-layer="27" to-port="0" />
|
|
<edge from-layer="27" from-port="1" to-layer="32" to-port="0" />
|
|
<edge from-layer="28" from-port="0" to-layer="32" to-port="1" />
|
|
<edge from-layer="29" from-port="0" to-layer="32" to-port="2" />
|
|
<edge from-layer="30" from-port="0" to-layer="32" to-port="3" />
|
|
<edge from-layer="31" from-port="0" to-layer="32" to-port="4" />
|
|
<edge from-layer="32" from-port="5" to-layer="37" to-port="0" />
|
|
<edge from-layer="33" from-port="0" to-layer="34" to-port="0" />
|
|
<edge from-layer="34" from-port="1" to-layer="36" to-port="0" />
|
|
<edge from-layer="35" from-port="0" to-layer="36" to-port="1" />
|
|
<edge from-layer="36" from-port="2" to-layer="37" to-port="1" />
|
|
<edge from-layer="37" from-port="2" to-layer="39" to-port="0" />
|
|
<edge from-layer="38" from-port="0" to-layer="39" to-port="1" />
|
|
<edge from-layer="39" from-port="2" to-layer="40" to-port="0" />
|
|
<edge from-layer="40" from-port="1" to-layer="45" to-port="0" />
|
|
<edge from-layer="41" from-port="0" to-layer="45" to-port="1" />
|
|
<edge from-layer="42" from-port="0" to-layer="45" to-port="2" />
|
|
<edge from-layer="43" from-port="0" to-layer="45" to-port="3" />
|
|
<edge from-layer="44" from-port="0" to-layer="45" to-port="4" />
|
|
<edge from-layer="45" from-port="5" to-layer="48" to-port="0" />
|
|
<edge from-layer="46" from-port="0" to-layer="48" to-port="1" />
|
|
<edge from-layer="47" from-port="0" to-layer="48" to-port="2" />
|
|
<edge from-layer="48" from-port="4" to-layer="75" to-port="0" />
|
|
<edge from-layer="48" from-port="3" to-layer="81" to-port="0" />
|
|
<edge from-layer="48" from-port="4" to-layer="81" to-port="1" />
|
|
<edge from-layer="48" from-port="4" to-layer="53" to-port="0" />
|
|
<edge from-layer="49" from-port="0" to-layer="50" to-port="0" />
|
|
<edge from-layer="50" from-port="1" to-layer="52" to-port="0" />
|
|
<edge from-layer="51" from-port="0" to-layer="52" to-port="1" />
|
|
<edge from-layer="52" from-port="2" to-layer="53" to-port="1" />
|
|
<edge from-layer="53" from-port="2" to-layer="55" to-port="0" />
|
|
<edge from-layer="54" from-port="0" to-layer="55" to-port="1" />
|
|
<edge from-layer="55" from-port="2" to-layer="56" to-port="0" />
|
|
<edge from-layer="56" from-port="1" to-layer="61" to-port="0" />
|
|
<edge from-layer="57" from-port="0" to-layer="61" to-port="1" />
|
|
<edge from-layer="58" from-port="0" to-layer="61" to-port="2" />
|
|
<edge from-layer="59" from-port="0" to-layer="61" to-port="3" />
|
|
<edge from-layer="60" from-port="0" to-layer="61" to-port="4" />
|
|
<edge from-layer="61" from-port="5" to-layer="66" to-port="0" />
|
|
<edge from-layer="62" from-port="0" to-layer="63" to-port="0" />
|
|
<edge from-layer="63" from-port="1" to-layer="65" to-port="0" />
|
|
<edge from-layer="64" from-port="0" to-layer="65" to-port="1" />
|
|
<edge from-layer="65" from-port="2" to-layer="66" to-port="1" />
|
|
<edge from-layer="66" from-port="2" to-layer="68" to-port="0" />
|
|
<edge from-layer="67" from-port="0" to-layer="68" to-port="1" />
|
|
<edge from-layer="68" from-port="2" to-layer="69" to-port="0" />
|
|
<edge from-layer="69" from-port="1" to-layer="74" to-port="0" />
|
|
<edge from-layer="70" from-port="0" to-layer="74" to-port="1" />
|
|
<edge from-layer="71" from-port="0" to-layer="74" to-port="2" />
|
|
<edge from-layer="72" from-port="0" to-layer="74" to-port="3" />
|
|
<edge from-layer="73" from-port="0" to-layer="74" to-port="4" />
|
|
<edge from-layer="74" from-port="5" to-layer="75" to-port="1" />
|
|
<edge from-layer="75" from-port="2" to-layer="80" to-port="0" />
|
|
<edge from-layer="76" from-port="0" to-layer="80" to-port="1" />
|
|
<edge from-layer="77" from-port="0" to-layer="80" to-port="2" />
|
|
<edge from-layer="78" from-port="0" to-layer="80" to-port="3" />
|
|
<edge from-layer="79" from-port="0" to-layer="80" to-port="4" />
|
|
<edge from-layer="80" from-port="5" to-layer="81" to-port="2" />
|
|
<edge from-layer="81" from-port="3" to-layer="86" to-port="0" />
|
|
<edge from-layer="82" from-port="0" to-layer="83" to-port="0" />
|
|
<edge from-layer="83" from-port="1" to-layer="85" to-port="0" />
|
|
<edge from-layer="84" from-port="0" to-layer="85" to-port="1" />
|
|
<edge from-layer="85" from-port="2" to-layer="86" to-port="1" />
|
|
<edge from-layer="86" from-port="2" to-layer="88" to-port="0" />
|
|
<edge from-layer="87" from-port="0" to-layer="88" to-port="1" />
|
|
<edge from-layer="88" from-port="2" to-layer="89" to-port="0" />
|
|
<edge from-layer="89" from-port="1" to-layer="94" to-port="0" />
|
|
<edge from-layer="90" from-port="0" to-layer="94" to-port="1" />
|
|
<edge from-layer="91" from-port="0" to-layer="94" to-port="2" />
|
|
<edge from-layer="92" from-port="0" to-layer="94" to-port="3" />
|
|
<edge from-layer="93" from-port="0" to-layer="94" to-port="4" />
|
|
<edge from-layer="94" from-port="5" to-layer="99" to-port="0" />
|
|
<edge from-layer="95" from-port="0" to-layer="96" to-port="0" />
|
|
<edge from-layer="96" from-port="1" to-layer="98" to-port="0" />
|
|
<edge from-layer="97" from-port="0" to-layer="98" to-port="1" />
|
|
<edge from-layer="98" from-port="2" to-layer="99" to-port="1" />
|
|
<edge from-layer="99" from-port="2" to-layer="101" to-port="0" />
|
|
<edge from-layer="100" from-port="0" to-layer="101" to-port="1" />
|
|
<edge from-layer="101" from-port="2" to-layer="102" to-port="0" />
|
|
<edge from-layer="102" from-port="1" to-layer="107" to-port="0" />
|
|
<edge from-layer="103" from-port="0" to-layer="107" to-port="1" />
|
|
<edge from-layer="104" from-port="0" to-layer="107" to-port="2" />
|
|
<edge from-layer="105" from-port="0" to-layer="107" to-port="3" />
|
|
<edge from-layer="106" from-port="0" to-layer="107" to-port="4" />
|
|
<edge from-layer="107" from-port="5" to-layer="112" to-port="0" />
|
|
<edge from-layer="108" from-port="0" to-layer="109" to-port="0" />
|
|
<edge from-layer="109" from-port="1" to-layer="111" to-port="0" />
|
|
<edge from-layer="110" from-port="0" to-layer="111" to-port="1" />
|
|
<edge from-layer="111" from-port="2" to-layer="112" to-port="1" />
|
|
<edge from-layer="112" from-port="2" to-layer="114" to-port="0" />
|
|
<edge from-layer="113" from-port="0" to-layer="114" to-port="1" />
|
|
<edge from-layer="114" from-port="2" to-layer="115" to-port="0" />
|
|
<edge from-layer="115" from-port="1" to-layer="120" to-port="0" />
|
|
<edge from-layer="116" from-port="0" to-layer="120" to-port="1" />
|
|
<edge from-layer="117" from-port="0" to-layer="120" to-port="2" />
|
|
<edge from-layer="118" from-port="0" to-layer="120" to-port="3" />
|
|
<edge from-layer="119" from-port="0" to-layer="120" to-port="4" />
|
|
<edge from-layer="120" from-port="5" to-layer="123" to-port="0" />
|
|
<edge from-layer="121" from-port="0" to-layer="123" to-port="1" />
|
|
<edge from-layer="122" from-port="0" to-layer="123" to-port="2" />
|
|
<edge from-layer="123" from-port="3" to-layer="156" to-port="0" />
|
|
<edge from-layer="123" from-port="4" to-layer="156" to-port="1" />
|
|
<edge from-layer="123" from-port="4" to-layer="150" to-port="0" />
|
|
<edge from-layer="123" from-port="4" to-layer="128" to-port="0" />
|
|
<edge from-layer="124" from-port="0" to-layer="125" to-port="0" />
|
|
<edge from-layer="125" from-port="1" to-layer="127" to-port="0" />
|
|
<edge from-layer="126" from-port="0" to-layer="127" to-port="1" />
|
|
<edge from-layer="127" from-port="2" to-layer="128" to-port="1" />
|
|
<edge from-layer="128" from-port="2" to-layer="130" to-port="0" />
|
|
<edge from-layer="129" from-port="0" to-layer="130" to-port="1" />
|
|
<edge from-layer="130" from-port="2" to-layer="131" to-port="0" />
|
|
<edge from-layer="131" from-port="1" to-layer="136" to-port="0" />
|
|
<edge from-layer="132" from-port="0" to-layer="136" to-port="1" />
|
|
<edge from-layer="133" from-port="0" to-layer="136" to-port="2" />
|
|
<edge from-layer="134" from-port="0" to-layer="136" to-port="3" />
|
|
<edge from-layer="135" from-port="0" to-layer="136" to-port="4" />
|
|
<edge from-layer="136" from-port="5" to-layer="141" to-port="0" />
|
|
<edge from-layer="137" from-port="0" to-layer="138" to-port="0" />
|
|
<edge from-layer="138" from-port="1" to-layer="140" to-port="0" />
|
|
<edge from-layer="139" from-port="0" to-layer="140" to-port="1" />
|
|
<edge from-layer="140" from-port="2" to-layer="141" to-port="1" />
|
|
<edge from-layer="141" from-port="2" to-layer="143" to-port="0" />
|
|
<edge from-layer="142" from-port="0" to-layer="143" to-port="1" />
|
|
<edge from-layer="143" from-port="2" to-layer="144" to-port="0" />
|
|
<edge from-layer="144" from-port="1" to-layer="149" to-port="0" />
|
|
<edge from-layer="145" from-port="0" to-layer="149" to-port="1" />
|
|
<edge from-layer="146" from-port="0" to-layer="149" to-port="2" />
|
|
<edge from-layer="147" from-port="0" to-layer="149" to-port="3" />
|
|
<edge from-layer="148" from-port="0" to-layer="149" to-port="4" />
|
|
<edge from-layer="149" from-port="5" to-layer="150" to-port="1" />
|
|
<edge from-layer="150" from-port="2" to-layer="155" to-port="0" />
|
|
<edge from-layer="151" from-port="0" to-layer="155" to-port="1" />
|
|
<edge from-layer="152" from-port="0" to-layer="155" to-port="2" />
|
|
<edge from-layer="153" from-port="0" to-layer="155" to-port="3" />
|
|
<edge from-layer="154" from-port="0" to-layer="155" to-port="4" />
|
|
<edge from-layer="155" from-port="5" to-layer="156" to-port="2" />
|
|
<edge from-layer="156" from-port="3" to-layer="161" to-port="0" />
|
|
<edge from-layer="157" from-port="0" to-layer="158" to-port="0" />
|
|
<edge from-layer="158" from-port="1" to-layer="160" to-port="0" />
|
|
<edge from-layer="159" from-port="0" to-layer="160" to-port="1" />
|
|
<edge from-layer="160" from-port="2" to-layer="161" to-port="1" />
|
|
<edge from-layer="161" from-port="2" to-layer="163" to-port="0" />
|
|
<edge from-layer="162" from-port="0" to-layer="163" to-port="1" />
|
|
<edge from-layer="163" from-port="2" to-layer="164" to-port="0" />
|
|
<edge from-layer="164" from-port="1" to-layer="169" to-port="0" />
|
|
<edge from-layer="165" from-port="0" to-layer="169" to-port="1" />
|
|
<edge from-layer="166" from-port="0" to-layer="169" to-port="2" />
|
|
<edge from-layer="167" from-port="0" to-layer="169" to-port="3" />
|
|
<edge from-layer="168" from-port="0" to-layer="169" to-port="4" />
|
|
<edge from-layer="169" from-port="5" to-layer="174" to-port="0" />
|
|
<edge from-layer="169" from-port="5" to-layer="718" to-port="1" />
|
|
<edge from-layer="170" from-port="0" to-layer="171" to-port="0" />
|
|
<edge from-layer="171" from-port="1" to-layer="173" to-port="0" />
|
|
<edge from-layer="172" from-port="0" to-layer="173" to-port="1" />
|
|
<edge from-layer="173" from-port="2" to-layer="174" to-port="1" />
|
|
<edge from-layer="174" from-port="2" to-layer="176" to-port="0" />
|
|
<edge from-layer="175" from-port="0" to-layer="176" to-port="1" />
|
|
<edge from-layer="176" from-port="2" to-layer="177" to-port="0" />
|
|
<edge from-layer="177" from-port="1" to-layer="182" to-port="0" />
|
|
<edge from-layer="178" from-port="0" to-layer="182" to-port="1" />
|
|
<edge from-layer="179" from-port="0" to-layer="182" to-port="2" />
|
|
<edge from-layer="180" from-port="0" to-layer="182" to-port="3" />
|
|
<edge from-layer="181" from-port="0" to-layer="182" to-port="4" />
|
|
<edge from-layer="182" from-port="5" to-layer="187" to-port="0" />
|
|
<edge from-layer="183" from-port="0" to-layer="184" to-port="0" />
|
|
<edge from-layer="184" from-port="1" to-layer="186" to-port="0" />
|
|
<edge from-layer="185" from-port="0" to-layer="186" to-port="1" />
|
|
<edge from-layer="186" from-port="2" to-layer="187" to-port="1" />
|
|
<edge from-layer="187" from-port="2" to-layer="189" to-port="0" />
|
|
<edge from-layer="188" from-port="0" to-layer="189" to-port="1" />
|
|
<edge from-layer="189" from-port="2" to-layer="190" to-port="0" />
|
|
<edge from-layer="190" from-port="1" to-layer="195" to-port="0" />
|
|
<edge from-layer="191" from-port="0" to-layer="195" to-port="1" />
|
|
<edge from-layer="192" from-port="0" to-layer="195" to-port="2" />
|
|
<edge from-layer="193" from-port="0" to-layer="195" to-port="3" />
|
|
<edge from-layer="194" from-port="0" to-layer="195" to-port="4" />
|
|
<edge from-layer="195" from-port="5" to-layer="198" to-port="0" />
|
|
<edge from-layer="196" from-port="0" to-layer="198" to-port="1" />
|
|
<edge from-layer="197" from-port="0" to-layer="198" to-port="2" />
|
|
<edge from-layer="198" from-port="3" to-layer="303" to-port="0" />
|
|
<edge from-layer="198" from-port="4" to-layer="303" to-port="1" />
|
|
<edge from-layer="198" from-port="4" to-layer="203" to-port="0" />
|
|
<edge from-layer="198" from-port="4" to-layer="280" to-port="0" />
|
|
<edge from-layer="199" from-port="0" to-layer="200" to-port="0" />
|
|
<edge from-layer="200" from-port="1" to-layer="202" to-port="0" />
|
|
<edge from-layer="201" from-port="0" to-layer="202" to-port="1" />
|
|
<edge from-layer="202" from-port="2" to-layer="203" to-port="1" />
|
|
<edge from-layer="203" from-port="2" to-layer="205" to-port="0" />
|
|
<edge from-layer="204" from-port="0" to-layer="205" to-port="1" />
|
|
<edge from-layer="205" from-port="2" to-layer="206" to-port="0" />
|
|
<edge from-layer="206" from-port="1" to-layer="211" to-port="0" />
|
|
<edge from-layer="207" from-port="0" to-layer="211" to-port="1" />
|
|
<edge from-layer="208" from-port="0" to-layer="211" to-port="2" />
|
|
<edge from-layer="209" from-port="0" to-layer="211" to-port="3" />
|
|
<edge from-layer="210" from-port="0" to-layer="211" to-port="4" />
|
|
<edge from-layer="211" from-port="5" to-layer="238" to-port="0" />
|
|
<edge from-layer="211" from-port="5" to-layer="216" to-port="0" />
|
|
<edge from-layer="212" from-port="0" to-layer="213" to-port="0" />
|
|
<edge from-layer="213" from-port="1" to-layer="215" to-port="0" />
|
|
<edge from-layer="214" from-port="0" to-layer="215" to-port="1" />
|
|
<edge from-layer="215" from-port="2" to-layer="216" to-port="1" />
|
|
<edge from-layer="216" from-port="2" to-layer="218" to-port="0" />
|
|
<edge from-layer="217" from-port="0" to-layer="218" to-port="1" />
|
|
<edge from-layer="218" from-port="2" to-layer="219" to-port="0" />
|
|
<edge from-layer="219" from-port="1" to-layer="224" to-port="0" />
|
|
<edge from-layer="220" from-port="0" to-layer="224" to-port="1" />
|
|
<edge from-layer="221" from-port="0" to-layer="224" to-port="2" />
|
|
<edge from-layer="222" from-port="0" to-layer="224" to-port="3" />
|
|
<edge from-layer="223" from-port="0" to-layer="224" to-port="4" />
|
|
<edge from-layer="224" from-port="5" to-layer="229" to-port="0" />
|
|
<edge from-layer="225" from-port="0" to-layer="226" to-port="0" />
|
|
<edge from-layer="226" from-port="1" to-layer="228" to-port="0" />
|
|
<edge from-layer="227" from-port="0" to-layer="228" to-port="1" />
|
|
<edge from-layer="228" from-port="2" to-layer="229" to-port="1" />
|
|
<edge from-layer="229" from-port="2" to-layer="231" to-port="0" />
|
|
<edge from-layer="230" from-port="0" to-layer="231" to-port="1" />
|
|
<edge from-layer="231" from-port="2" to-layer="232" to-port="0" />
|
|
<edge from-layer="232" from-port="1" to-layer="237" to-port="0" />
|
|
<edge from-layer="233" from-port="0" to-layer="237" to-port="1" />
|
|
<edge from-layer="234" from-port="0" to-layer="237" to-port="2" />
|
|
<edge from-layer="235" from-port="0" to-layer="237" to-port="3" />
|
|
<edge from-layer="236" from-port="0" to-layer="237" to-port="4" />
|
|
<edge from-layer="237" from-port="5" to-layer="238" to-port="1" />
|
|
<edge from-layer="238" from-port="2" to-layer="243" to-port="0" />
|
|
<edge from-layer="239" from-port="0" to-layer="243" to-port="1" />
|
|
<edge from-layer="240" from-port="0" to-layer="243" to-port="2" />
|
|
<edge from-layer="241" from-port="0" to-layer="243" to-port="3" />
|
|
<edge from-layer="242" from-port="0" to-layer="243" to-port="4" />
|
|
<edge from-layer="243" from-port="5" to-layer="270" to-port="0" />
|
|
<edge from-layer="243" from-port="5" to-layer="248" to-port="0" />
|
|
<edge from-layer="244" from-port="0" to-layer="245" to-port="0" />
|
|
<edge from-layer="245" from-port="1" to-layer="247" to-port="0" />
|
|
<edge from-layer="246" from-port="0" to-layer="247" to-port="1" />
|
|
<edge from-layer="247" from-port="2" to-layer="248" to-port="1" />
|
|
<edge from-layer="248" from-port="2" to-layer="250" to-port="0" />
|
|
<edge from-layer="249" from-port="0" to-layer="250" to-port="1" />
|
|
<edge from-layer="250" from-port="2" to-layer="251" to-port="0" />
|
|
<edge from-layer="251" from-port="1" to-layer="256" to-port="0" />
|
|
<edge from-layer="252" from-port="0" to-layer="256" to-port="1" />
|
|
<edge from-layer="253" from-port="0" to-layer="256" to-port="2" />
|
|
<edge from-layer="254" from-port="0" to-layer="256" to-port="3" />
|
|
<edge from-layer="255" from-port="0" to-layer="256" to-port="4" />
|
|
<edge from-layer="256" from-port="5" to-layer="261" to-port="0" />
|
|
<edge from-layer="257" from-port="0" to-layer="258" to-port="0" />
|
|
<edge from-layer="258" from-port="1" to-layer="260" to-port="0" />
|
|
<edge from-layer="259" from-port="0" to-layer="260" to-port="1" />
|
|
<edge from-layer="260" from-port="2" to-layer="261" to-port="1" />
|
|
<edge from-layer="261" from-port="2" to-layer="263" to-port="0" />
|
|
<edge from-layer="262" from-port="0" to-layer="263" to-port="1" />
|
|
<edge from-layer="263" from-port="2" to-layer="264" to-port="0" />
|
|
<edge from-layer="264" from-port="1" to-layer="269" to-port="0" />
|
|
<edge from-layer="265" from-port="0" to-layer="269" to-port="1" />
|
|
<edge from-layer="266" from-port="0" to-layer="269" to-port="2" />
|
|
<edge from-layer="267" from-port="0" to-layer="269" to-port="3" />
|
|
<edge from-layer="268" from-port="0" to-layer="269" to-port="4" />
|
|
<edge from-layer="269" from-port="5" to-layer="270" to-port="1" />
|
|
<edge from-layer="270" from-port="2" to-layer="275" to-port="0" />
|
|
<edge from-layer="271" from-port="0" to-layer="275" to-port="1" />
|
|
<edge from-layer="272" from-port="0" to-layer="275" to-port="2" />
|
|
<edge from-layer="273" from-port="0" to-layer="275" to-port="3" />
|
|
<edge from-layer="274" from-port="0" to-layer="275" to-port="4" />
|
|
<edge from-layer="275" from-port="5" to-layer="289" to-port="0" />
|
|
<edge from-layer="276" from-port="0" to-layer="277" to-port="0" />
|
|
<edge from-layer="277" from-port="1" to-layer="279" to-port="0" />
|
|
<edge from-layer="278" from-port="0" to-layer="279" to-port="1" />
|
|
<edge from-layer="279" from-port="2" to-layer="280" to-port="1" />
|
|
<edge from-layer="280" from-port="2" to-layer="282" to-port="0" />
|
|
<edge from-layer="281" from-port="0" to-layer="282" to-port="1" />
|
|
<edge from-layer="282" from-port="2" to-layer="283" to-port="0" />
|
|
<edge from-layer="283" from-port="1" to-layer="288" to-port="0" />
|
|
<edge from-layer="284" from-port="0" to-layer="288" to-port="1" />
|
|
<edge from-layer="285" from-port="0" to-layer="288" to-port="2" />
|
|
<edge from-layer="286" from-port="0" to-layer="288" to-port="3" />
|
|
<edge from-layer="287" from-port="0" to-layer="288" to-port="4" />
|
|
<edge from-layer="288" from-port="5" to-layer="289" to-port="1" />
|
|
<edge from-layer="289" from-port="2" to-layer="294" to-port="0" />
|
|
<edge from-layer="290" from-port="0" to-layer="291" to-port="0" />
|
|
<edge from-layer="291" from-port="1" to-layer="293" to-port="0" />
|
|
<edge from-layer="292" from-port="0" to-layer="293" to-port="1" />
|
|
<edge from-layer="293" from-port="2" to-layer="294" to-port="1" />
|
|
<edge from-layer="294" from-port="2" to-layer="296" to-port="0" />
|
|
<edge from-layer="295" from-port="0" to-layer="296" to-port="1" />
|
|
<edge from-layer="296" from-port="2" to-layer="297" to-port="0" />
|
|
<edge from-layer="297" from-port="1" to-layer="302" to-port="0" />
|
|
<edge from-layer="298" from-port="0" to-layer="302" to-port="1" />
|
|
<edge from-layer="299" from-port="0" to-layer="302" to-port="2" />
|
|
<edge from-layer="300" from-port="0" to-layer="302" to-port="3" />
|
|
<edge from-layer="301" from-port="0" to-layer="302" to-port="4" />
|
|
<edge from-layer="302" from-port="5" to-layer="303" to-port="2" />
|
|
<edge from-layer="303" from-port="3" to-layer="308" to-port="0" />
|
|
<edge from-layer="304" from-port="0" to-layer="305" to-port="0" />
|
|
<edge from-layer="305" from-port="1" to-layer="307" to-port="0" />
|
|
<edge from-layer="306" from-port="0" to-layer="307" to-port="1" />
|
|
<edge from-layer="307" from-port="2" to-layer="308" to-port="1" />
|
|
<edge from-layer="308" from-port="2" to-layer="310" to-port="0" />
|
|
<edge from-layer="309" from-port="0" to-layer="310" to-port="1" />
|
|
<edge from-layer="310" from-port="2" to-layer="311" to-port="0" />
|
|
<edge from-layer="311" from-port="1" to-layer="316" to-port="0" />
|
|
<edge from-layer="312" from-port="0" to-layer="316" to-port="1" />
|
|
<edge from-layer="313" from-port="0" to-layer="316" to-port="2" />
|
|
<edge from-layer="314" from-port="0" to-layer="316" to-port="3" />
|
|
<edge from-layer="315" from-port="0" to-layer="316" to-port="4" />
|
|
<edge from-layer="316" from-port="5" to-layer="321" to-port="0" />
|
|
<edge from-layer="316" from-port="5" to-layer="647" to-port="1" />
|
|
<edge from-layer="317" from-port="0" to-layer="318" to-port="0" />
|
|
<edge from-layer="318" from-port="1" to-layer="320" to-port="0" />
|
|
<edge from-layer="319" from-port="0" to-layer="320" to-port="1" />
|
|
<edge from-layer="320" from-port="2" to-layer="321" to-port="1" />
|
|
<edge from-layer="321" from-port="2" to-layer="323" to-port="0" />
|
|
<edge from-layer="322" from-port="0" to-layer="323" to-port="1" />
|
|
<edge from-layer="323" from-port="2" to-layer="324" to-port="0" />
|
|
<edge from-layer="324" from-port="1" to-layer="329" to-port="0" />
|
|
<edge from-layer="325" from-port="0" to-layer="329" to-port="1" />
|
|
<edge from-layer="326" from-port="0" to-layer="329" to-port="2" />
|
|
<edge from-layer="327" from-port="0" to-layer="329" to-port="3" />
|
|
<edge from-layer="328" from-port="0" to-layer="329" to-port="4" />
|
|
<edge from-layer="329" from-port="5" to-layer="334" to-port="0" />
|
|
<edge from-layer="330" from-port="0" to-layer="331" to-port="0" />
|
|
<edge from-layer="331" from-port="1" to-layer="333" to-port="0" />
|
|
<edge from-layer="332" from-port="0" to-layer="333" to-port="1" />
|
|
<edge from-layer="333" from-port="2" to-layer="334" to-port="1" />
|
|
<edge from-layer="334" from-port="2" to-layer="336" to-port="0" />
|
|
<edge from-layer="335" from-port="0" to-layer="336" to-port="1" />
|
|
<edge from-layer="336" from-port="2" to-layer="337" to-port="0" />
|
|
<edge from-layer="337" from-port="1" to-layer="342" to-port="0" />
|
|
<edge from-layer="338" from-port="0" to-layer="342" to-port="1" />
|
|
<edge from-layer="339" from-port="0" to-layer="342" to-port="2" />
|
|
<edge from-layer="340" from-port="0" to-layer="342" to-port="3" />
|
|
<edge from-layer="341" from-port="0" to-layer="342" to-port="4" />
|
|
<edge from-layer="342" from-port="5" to-layer="345" to-port="0" />
|
|
<edge from-layer="343" from-port="0" to-layer="345" to-port="1" />
|
|
<edge from-layer="344" from-port="0" to-layer="345" to-port="2" />
|
|
<edge from-layer="345" from-port="4" to-layer="350" to-port="0" />
|
|
<edge from-layer="345" from-port="4" to-layer="450" to-port="1" />
|
|
<edge from-layer="345" from-port="3" to-layer="450" to-port="0" />
|
|
<edge from-layer="345" from-port="4" to-layer="427" to-port="0" />
|
|
<edge from-layer="346" from-port="0" to-layer="347" to-port="0" />
|
|
<edge from-layer="347" from-port="1" to-layer="349" to-port="0" />
|
|
<edge from-layer="348" from-port="0" to-layer="349" to-port="1" />
|
|
<edge from-layer="349" from-port="2" to-layer="350" to-port="1" />
|
|
<edge from-layer="350" from-port="2" to-layer="352" to-port="0" />
|
|
<edge from-layer="351" from-port="0" to-layer="352" to-port="1" />
|
|
<edge from-layer="352" from-port="2" to-layer="353" to-port="0" />
|
|
<edge from-layer="353" from-port="1" to-layer="358" to-port="0" />
|
|
<edge from-layer="354" from-port="0" to-layer="358" to-port="1" />
|
|
<edge from-layer="355" from-port="0" to-layer="358" to-port="2" />
|
|
<edge from-layer="356" from-port="0" to-layer="358" to-port="3" />
|
|
<edge from-layer="357" from-port="0" to-layer="358" to-port="4" />
|
|
<edge from-layer="358" from-port="5" to-layer="385" to-port="0" />
|
|
<edge from-layer="358" from-port="5" to-layer="363" to-port="0" />
|
|
<edge from-layer="359" from-port="0" to-layer="360" to-port="0" />
|
|
<edge from-layer="360" from-port="1" to-layer="362" to-port="0" />
|
|
<edge from-layer="361" from-port="0" to-layer="362" to-port="1" />
|
|
<edge from-layer="362" from-port="2" to-layer="363" to-port="1" />
|
|
<edge from-layer="363" from-port="2" to-layer="365" to-port="0" />
|
|
<edge from-layer="364" from-port="0" to-layer="365" to-port="1" />
|
|
<edge from-layer="365" from-port="2" to-layer="366" to-port="0" />
|
|
<edge from-layer="366" from-port="1" to-layer="371" to-port="0" />
|
|
<edge from-layer="367" from-port="0" to-layer="371" to-port="1" />
|
|
<edge from-layer="368" from-port="0" to-layer="371" to-port="2" />
|
|
<edge from-layer="369" from-port="0" to-layer="371" to-port="3" />
|
|
<edge from-layer="370" from-port="0" to-layer="371" to-port="4" />
|
|
<edge from-layer="371" from-port="5" to-layer="376" to-port="0" />
|
|
<edge from-layer="372" from-port="0" to-layer="373" to-port="0" />
|
|
<edge from-layer="373" from-port="1" to-layer="375" to-port="0" />
|
|
<edge from-layer="374" from-port="0" to-layer="375" to-port="1" />
|
|
<edge from-layer="375" from-port="2" to-layer="376" to-port="1" />
|
|
<edge from-layer="376" from-port="2" to-layer="378" to-port="0" />
|
|
<edge from-layer="377" from-port="0" to-layer="378" to-port="1" />
|
|
<edge from-layer="378" from-port="2" to-layer="379" to-port="0" />
|
|
<edge from-layer="379" from-port="1" to-layer="384" to-port="0" />
|
|
<edge from-layer="380" from-port="0" to-layer="384" to-port="1" />
|
|
<edge from-layer="381" from-port="0" to-layer="384" to-port="2" />
|
|
<edge from-layer="382" from-port="0" to-layer="384" to-port="3" />
|
|
<edge from-layer="383" from-port="0" to-layer="384" to-port="4" />
|
|
<edge from-layer="384" from-port="5" to-layer="385" to-port="1" />
|
|
<edge from-layer="385" from-port="2" to-layer="390" to-port="0" />
|
|
<edge from-layer="386" from-port="0" to-layer="390" to-port="1" />
|
|
<edge from-layer="387" from-port="0" to-layer="390" to-port="2" />
|
|
<edge from-layer="388" from-port="0" to-layer="390" to-port="3" />
|
|
<edge from-layer="389" from-port="0" to-layer="390" to-port="4" />
|
|
<edge from-layer="390" from-port="5" to-layer="395" to-port="0" />
|
|
<edge from-layer="390" from-port="5" to-layer="417" to-port="0" />
|
|
<edge from-layer="391" from-port="0" to-layer="392" to-port="0" />
|
|
<edge from-layer="392" from-port="1" to-layer="394" to-port="0" />
|
|
<edge from-layer="393" from-port="0" to-layer="394" to-port="1" />
|
|
<edge from-layer="394" from-port="2" to-layer="395" to-port="1" />
|
|
<edge from-layer="395" from-port="2" to-layer="397" to-port="0" />
|
|
<edge from-layer="396" from-port="0" to-layer="397" to-port="1" />
|
|
<edge from-layer="397" from-port="2" to-layer="398" to-port="0" />
|
|
<edge from-layer="398" from-port="1" to-layer="403" to-port="0" />
|
|
<edge from-layer="399" from-port="0" to-layer="403" to-port="1" />
|
|
<edge from-layer="400" from-port="0" to-layer="403" to-port="2" />
|
|
<edge from-layer="401" from-port="0" to-layer="403" to-port="3" />
|
|
<edge from-layer="402" from-port="0" to-layer="403" to-port="4" />
|
|
<edge from-layer="403" from-port="5" to-layer="408" to-port="0" />
|
|
<edge from-layer="404" from-port="0" to-layer="405" to-port="0" />
|
|
<edge from-layer="405" from-port="1" to-layer="407" to-port="0" />
|
|
<edge from-layer="406" from-port="0" to-layer="407" to-port="1" />
|
|
<edge from-layer="407" from-port="2" to-layer="408" to-port="1" />
|
|
<edge from-layer="408" from-port="2" to-layer="410" to-port="0" />
|
|
<edge from-layer="409" from-port="0" to-layer="410" to-port="1" />
|
|
<edge from-layer="410" from-port="2" to-layer="411" to-port="0" />
|
|
<edge from-layer="411" from-port="1" to-layer="416" to-port="0" />
|
|
<edge from-layer="412" from-port="0" to-layer="416" to-port="1" />
|
|
<edge from-layer="413" from-port="0" to-layer="416" to-port="2" />
|
|
<edge from-layer="414" from-port="0" to-layer="416" to-port="3" />
|
|
<edge from-layer="415" from-port="0" to-layer="416" to-port="4" />
|
|
<edge from-layer="416" from-port="5" to-layer="417" to-port="1" />
|
|
<edge from-layer="417" from-port="2" to-layer="422" to-port="0" />
|
|
<edge from-layer="418" from-port="0" to-layer="422" to-port="1" />
|
|
<edge from-layer="419" from-port="0" to-layer="422" to-port="2" />
|
|
<edge from-layer="420" from-port="0" to-layer="422" to-port="3" />
|
|
<edge from-layer="421" from-port="0" to-layer="422" to-port="4" />
|
|
<edge from-layer="422" from-port="5" to-layer="436" to-port="0" />
|
|
<edge from-layer="423" from-port="0" to-layer="424" to-port="0" />
|
|
<edge from-layer="424" from-port="1" to-layer="426" to-port="0" />
|
|
<edge from-layer="425" from-port="0" to-layer="426" to-port="1" />
|
|
<edge from-layer="426" from-port="2" to-layer="427" to-port="1" />
|
|
<edge from-layer="427" from-port="2" to-layer="429" to-port="0" />
|
|
<edge from-layer="428" from-port="0" to-layer="429" to-port="1" />
|
|
<edge from-layer="429" from-port="2" to-layer="430" to-port="0" />
|
|
<edge from-layer="430" from-port="1" to-layer="435" to-port="0" />
|
|
<edge from-layer="431" from-port="0" to-layer="435" to-port="1" />
|
|
<edge from-layer="432" from-port="0" to-layer="435" to-port="2" />
|
|
<edge from-layer="433" from-port="0" to-layer="435" to-port="3" />
|
|
<edge from-layer="434" from-port="0" to-layer="435" to-port="4" />
|
|
<edge from-layer="435" from-port="5" to-layer="436" to-port="1" />
|
|
<edge from-layer="436" from-port="2" to-layer="441" to-port="0" />
|
|
<edge from-layer="437" from-port="0" to-layer="438" to-port="0" />
|
|
<edge from-layer="438" from-port="1" to-layer="440" to-port="0" />
|
|
<edge from-layer="439" from-port="0" to-layer="440" to-port="1" />
|
|
<edge from-layer="440" from-port="2" to-layer="441" to-port="1" />
|
|
<edge from-layer="441" from-port="2" to-layer="443" to-port="0" />
|
|
<edge from-layer="442" from-port="0" to-layer="443" to-port="1" />
|
|
<edge from-layer="443" from-port="2" to-layer="444" to-port="0" />
|
|
<edge from-layer="444" from-port="1" to-layer="449" to-port="0" />
|
|
<edge from-layer="445" from-port="0" to-layer="449" to-port="1" />
|
|
<edge from-layer="446" from-port="0" to-layer="449" to-port="2" />
|
|
<edge from-layer="447" from-port="0" to-layer="449" to-port="3" />
|
|
<edge from-layer="448" from-port="0" to-layer="449" to-port="4" />
|
|
<edge from-layer="449" from-port="5" to-layer="450" to-port="2" />
|
|
<edge from-layer="450" from-port="3" to-layer="455" to-port="0" />
|
|
<edge from-layer="451" from-port="0" to-layer="452" to-port="0" />
|
|
<edge from-layer="452" from-port="1" to-layer="454" to-port="0" />
|
|
<edge from-layer="453" from-port="0" to-layer="454" to-port="1" />
|
|
<edge from-layer="454" from-port="2" to-layer="455" to-port="1" />
|
|
<edge from-layer="455" from-port="2" to-layer="457" to-port="0" />
|
|
<edge from-layer="456" from-port="0" to-layer="457" to-port="1" />
|
|
<edge from-layer="457" from-port="2" to-layer="458" to-port="0" />
|
|
<edge from-layer="458" from-port="1" to-layer="463" to-port="0" />
|
|
<edge from-layer="459" from-port="0" to-layer="463" to-port="1" />
|
|
<edge from-layer="460" from-port="0" to-layer="463" to-port="2" />
|
|
<edge from-layer="461" from-port="0" to-layer="463" to-port="3" />
|
|
<edge from-layer="462" from-port="0" to-layer="463" to-port="4" />
|
|
<edge from-layer="463" from-port="5" to-layer="468" to-port="0" />
|
|
<edge from-layer="464" from-port="0" to-layer="465" to-port="0" />
|
|
<edge from-layer="465" from-port="1" to-layer="467" to-port="0" />
|
|
<edge from-layer="466" from-port="0" to-layer="467" to-port="1" />
|
|
<edge from-layer="467" from-port="2" to-layer="468" to-port="1" />
|
|
<edge from-layer="468" from-port="2" to-layer="470" to-port="0" />
|
|
<edge from-layer="469" from-port="0" to-layer="470" to-port="1" />
|
|
<edge from-layer="470" from-port="2" to-layer="471" to-port="0" />
|
|
<edge from-layer="471" from-port="1" to-layer="476" to-port="0" />
|
|
<edge from-layer="472" from-port="0" to-layer="476" to-port="1" />
|
|
<edge from-layer="473" from-port="0" to-layer="476" to-port="2" />
|
|
<edge from-layer="474" from-port="0" to-layer="476" to-port="3" />
|
|
<edge from-layer="475" from-port="0" to-layer="476" to-port="4" />
|
|
<edge from-layer="476" from-port="5" to-layer="477" to-port="0" />
|
|
<edge from-layer="476" from-port="5" to-layer="480" to-port="0" />
|
|
<edge from-layer="477" from-port="1" to-layer="478" to-port="0" />
|
|
<edge from-layer="477" from-port="1" to-layer="480" to-port="1" />
|
|
<edge from-layer="478" from-port="1" to-layer="479" to-port="0" />
|
|
<edge from-layer="478" from-port="1" to-layer="480" to-port="2" />
|
|
<edge from-layer="479" from-port="1" to-layer="480" to-port="3" />
|
|
<edge from-layer="480" from-port="4" to-layer="485" to-port="0" />
|
|
<edge from-layer="481" from-port="0" to-layer="482" to-port="0" />
|
|
<edge from-layer="482" from-port="1" to-layer="484" to-port="0" />
|
|
<edge from-layer="483" from-port="0" to-layer="484" to-port="1" />
|
|
<edge from-layer="484" from-port="2" to-layer="485" to-port="1" />
|
|
<edge from-layer="485" from-port="2" to-layer="487" to-port="0" />
|
|
<edge from-layer="486" from-port="0" to-layer="487" to-port="1" />
|
|
<edge from-layer="487" from-port="2" to-layer="488" to-port="0" />
|
|
<edge from-layer="488" from-port="1" to-layer="493" to-port="0" />
|
|
<edge from-layer="489" from-port="0" to-layer="493" to-port="1" />
|
|
<edge from-layer="490" from-port="0" to-layer="493" to-port="2" />
|
|
<edge from-layer="491" from-port="0" to-layer="493" to-port="3" />
|
|
<edge from-layer="492" from-port="0" to-layer="493" to-port="4" />
|
|
<edge from-layer="493" from-port="5" to-layer="498" to-port="0" />
|
|
<edge from-layer="494" from-port="0" to-layer="495" to-port="0" />
|
|
<edge from-layer="495" from-port="1" to-layer="497" to-port="0" />
|
|
<edge from-layer="496" from-port="0" to-layer="497" to-port="1" />
|
|
<edge from-layer="497" from-port="2" to-layer="498" to-port="1" />
|
|
<edge from-layer="498" from-port="2" to-layer="551" to-port="0" />
|
|
<edge from-layer="498" from-port="2" to-layer="500" to-port="0" />
|
|
<edge from-layer="499" from-port="0" to-layer="500" to-port="1" />
|
|
<edge from-layer="500" from-port="2" to-layer="501" to-port="0" />
|
|
<edge from-layer="501" from-port="1" to-layer="506" to-port="0" />
|
|
<edge from-layer="502" from-port="0" to-layer="506" to-port="1" />
|
|
<edge from-layer="503" from-port="0" to-layer="506" to-port="2" />
|
|
<edge from-layer="504" from-port="0" to-layer="506" to-port="3" />
|
|
<edge from-layer="505" from-port="0" to-layer="506" to-port="4" />
|
|
<edge from-layer="506" from-port="5" to-layer="509" to-port="0" />
|
|
<edge from-layer="507" from-port="0" to-layer="509" to-port="1" />
|
|
<edge from-layer="508" from-port="0" to-layer="509" to-port="2" />
|
|
<edge from-layer="509" from-port="4" to-layer="547" to-port="0" />
|
|
<edge from-layer="509" from-port="4" to-layer="514" to-port="0" />
|
|
<edge from-layer="509" from-port="3" to-layer="625" to-port="0" />
|
|
<edge from-layer="509" from-port="4" to-layer="588" to-port="0" />
|
|
<edge from-layer="510" from-port="0" to-layer="511" to-port="0" />
|
|
<edge from-layer="511" from-port="1" to-layer="513" to-port="0" />
|
|
<edge from-layer="512" from-port="0" to-layer="513" to-port="1" />
|
|
<edge from-layer="513" from-port="2" to-layer="514" to-port="1" />
|
|
<edge from-layer="514" from-port="2" to-layer="516" to-port="0" />
|
|
<edge from-layer="515" from-port="0" to-layer="516" to-port="1" />
|
|
<edge from-layer="516" from-port="2" to-layer="521" to-port="0" />
|
|
<edge from-layer="517" from-port="0" to-layer="521" to-port="1" />
|
|
<edge from-layer="518" from-port="0" to-layer="521" to-port="2" />
|
|
<edge from-layer="519" from-port="0" to-layer="521" to-port="3" />
|
|
<edge from-layer="520" from-port="0" to-layer="521" to-port="4" />
|
|
<edge from-layer="521" from-port="5" to-layer="523" to-port="0" />
|
|
<edge from-layer="522" from-port="0" to-layer="523" to-port="1" />
|
|
<edge from-layer="523" from-port="2" to-layer="526" to-port="0" />
|
|
<edge from-layer="524" from-port="0" to-layer="526" to-port="1" />
|
|
<edge from-layer="525" from-port="0" to-layer="526" to-port="2" />
|
|
<edge from-layer="526" from-port="5" to-layer="557" to-port="0" />
|
|
<edge from-layer="526" from-port="4" to-layer="528" to-port="0" />
|
|
<edge from-layer="526" from-port="5" to-layer="541" to-port="0" />
|
|
<edge from-layer="526" from-port="3" to-layer="534" to-port="0" />
|
|
<edge from-layer="527" from-port="0" to-layer="528" to-port="1" />
|
|
<edge from-layer="528" from-port="2" to-layer="533" to-port="0" />
|
|
<edge from-layer="529" from-port="0" to-layer="533" to-port="1" />
|
|
<edge from-layer="530" from-port="0" to-layer="533" to-port="2" />
|
|
<edge from-layer="531" from-port="0" to-layer="533" to-port="3" />
|
|
<edge from-layer="532" from-port="0" to-layer="533" to-port="4" />
|
|
<edge from-layer="533" from-port="5" to-layer="534" to-port="1" />
|
|
<edge from-layer="534" from-port="2" to-layer="535" to-port="0" />
|
|
<edge from-layer="535" from-port="1" to-layer="540" to-port="0" />
|
|
<edge from-layer="536" from-port="0" to-layer="540" to-port="1" />
|
|
<edge from-layer="537" from-port="0" to-layer="540" to-port="2" />
|
|
<edge from-layer="538" from-port="0" to-layer="540" to-port="3" />
|
|
<edge from-layer="539" from-port="0" to-layer="540" to-port="4" />
|
|
<edge from-layer="540" from-port="5" to-layer="541" to-port="1" />
|
|
<edge from-layer="541" from-port="2" to-layer="546" to-port="0" />
|
|
<edge from-layer="542" from-port="0" to-layer="546" to-port="1" />
|
|
<edge from-layer="543" from-port="0" to-layer="546" to-port="2" />
|
|
<edge from-layer="544" from-port="0" to-layer="546" to-port="3" />
|
|
<edge from-layer="545" from-port="0" to-layer="546" to-port="4" />
|
|
<edge from-layer="546" from-port="5" to-layer="556" to-port="0" />
|
|
<edge from-layer="547" from-port="1" to-layer="550" to-port="0" />
|
|
<edge from-layer="548" from-port="0" to-layer="550" to-port="1" />
|
|
<edge from-layer="549" from-port="0" to-layer="550" to-port="2" />
|
|
<edge from-layer="550" from-port="3" to-layer="555" to-port="0" />
|
|
<edge from-layer="551" from-port="1" to-layer="554" to-port="0" />
|
|
<edge from-layer="552" from-port="0" to-layer="554" to-port="1" />
|
|
<edge from-layer="553" from-port="0" to-layer="554" to-port="2" />
|
|
<edge from-layer="554" from-port="3" to-layer="555" to-port="1" />
|
|
<edge from-layer="555" from-port="2" to-layer="557" to-port="1" />
|
|
<edge from-layer="555" from-port="2" to-layer="556" to-port="1" />
|
|
<edge from-layer="556" from-port="2" to-layer="570" to-port="0" />
|
|
<edge from-layer="557" from-port="2" to-layer="562" to-port="0" />
|
|
<edge from-layer="558" from-port="0" to-layer="559" to-port="0" />
|
|
<edge from-layer="559" from-port="1" to-layer="561" to-port="0" />
|
|
<edge from-layer="560" from-port="0" to-layer="561" to-port="1" />
|
|
<edge from-layer="561" from-port="2" to-layer="562" to-port="1" />
|
|
<edge from-layer="562" from-port="2" to-layer="564" to-port="0" />
|
|
<edge from-layer="563" from-port="0" to-layer="564" to-port="1" />
|
|
<edge from-layer="564" from-port="2" to-layer="569" to-port="0" />
|
|
<edge from-layer="565" from-port="0" to-layer="569" to-port="1" />
|
|
<edge from-layer="566" from-port="0" to-layer="569" to-port="2" />
|
|
<edge from-layer="567" from-port="0" to-layer="569" to-port="3" />
|
|
<edge from-layer="568" from-port="0" to-layer="569" to-port="4" />
|
|
<edge from-layer="569" from-port="5" to-layer="570" to-port="1" />
|
|
<edge from-layer="570" from-port="2" to-layer="575" to-port="0" />
|
|
<edge from-layer="571" from-port="0" to-layer="575" to-port="1" />
|
|
<edge from-layer="572" from-port="0" to-layer="575" to-port="2" />
|
|
<edge from-layer="573" from-port="0" to-layer="575" to-port="3" />
|
|
<edge from-layer="574" from-port="0" to-layer="575" to-port="4" />
|
|
<edge from-layer="575" from-port="5" to-layer="580" to-port="0" />
|
|
<edge from-layer="576" from-port="0" to-layer="577" to-port="0" />
|
|
<edge from-layer="577" from-port="1" to-layer="579" to-port="0" />
|
|
<edge from-layer="578" from-port="0" to-layer="579" to-port="1" />
|
|
<edge from-layer="579" from-port="2" to-layer="580" to-port="1" />
|
|
<edge from-layer="580" from-port="2" to-layer="582" to-port="0" />
|
|
<edge from-layer="581" from-port="0" to-layer="582" to-port="1" />
|
|
<edge from-layer="582" from-port="2" to-layer="587" to-port="0" />
|
|
<edge from-layer="583" from-port="0" to-layer="587" to-port="1" />
|
|
<edge from-layer="584" from-port="0" to-layer="587" to-port="2" />
|
|
<edge from-layer="585" from-port="0" to-layer="587" to-port="3" />
|
|
<edge from-layer="586" from-port="0" to-layer="587" to-port="4" />
|
|
<edge from-layer="587" from-port="5" to-layer="588" to-port="1" />
|
|
<edge from-layer="588" from-port="2" to-layer="593" to-port="0" />
|
|
<edge from-layer="589" from-port="0" to-layer="593" to-port="1" />
|
|
<edge from-layer="590" from-port="0" to-layer="593" to-port="2" />
|
|
<edge from-layer="591" from-port="0" to-layer="593" to-port="3" />
|
|
<edge from-layer="592" from-port="0" to-layer="593" to-port="4" />
|
|
<edge from-layer="593" from-port="5" to-layer="598" to-port="0" />
|
|
<edge from-layer="593" from-port="5" to-layer="619" to-port="0" />
|
|
<edge from-layer="594" from-port="0" to-layer="595" to-port="0" />
|
|
<edge from-layer="595" from-port="1" to-layer="597" to-port="0" />
|
|
<edge from-layer="596" from-port="0" to-layer="597" to-port="1" />
|
|
<edge from-layer="597" from-port="2" to-layer="598" to-port="1" />
|
|
<edge from-layer="598" from-port="2" to-layer="600" to-port="0" />
|
|
<edge from-layer="599" from-port="0" to-layer="600" to-port="1" />
|
|
<edge from-layer="600" from-port="2" to-layer="601" to-port="0" />
|
|
<edge from-layer="601" from-port="1" to-layer="606" to-port="0" />
|
|
<edge from-layer="602" from-port="0" to-layer="606" to-port="1" />
|
|
<edge from-layer="603" from-port="0" to-layer="606" to-port="2" />
|
|
<edge from-layer="604" from-port="0" to-layer="606" to-port="3" />
|
|
<edge from-layer="605" from-port="0" to-layer="606" to-port="4" />
|
|
<edge from-layer="606" from-port="5" to-layer="611" to-port="0" />
|
|
<edge from-layer="607" from-port="0" to-layer="608" to-port="0" />
|
|
<edge from-layer="608" from-port="1" to-layer="610" to-port="0" />
|
|
<edge from-layer="609" from-port="0" to-layer="610" to-port="1" />
|
|
<edge from-layer="610" from-port="2" to-layer="611" to-port="1" />
|
|
<edge from-layer="611" from-port="2" to-layer="613" to-port="0" />
|
|
<edge from-layer="612" from-port="0" to-layer="613" to-port="1" />
|
|
<edge from-layer="613" from-port="2" to-layer="618" to-port="0" />
|
|
<edge from-layer="614" from-port="0" to-layer="618" to-port="1" />
|
|
<edge from-layer="615" from-port="0" to-layer="618" to-port="2" />
|
|
<edge from-layer="616" from-port="0" to-layer="618" to-port="3" />
|
|
<edge from-layer="617" from-port="0" to-layer="618" to-port="4" />
|
|
<edge from-layer="618" from-port="5" to-layer="619" to-port="1" />
|
|
<edge from-layer="619" from-port="2" to-layer="624" to-port="0" />
|
|
<edge from-layer="620" from-port="0" to-layer="624" to-port="1" />
|
|
<edge from-layer="621" from-port="0" to-layer="624" to-port="2" />
|
|
<edge from-layer="622" from-port="0" to-layer="624" to-port="3" />
|
|
<edge from-layer="623" from-port="0" to-layer="624" to-port="4" />
|
|
<edge from-layer="624" from-port="5" to-layer="625" to-port="1" />
|
|
<edge from-layer="625" from-port="2" to-layer="630" to-port="0" />
|
|
<edge from-layer="626" from-port="0" to-layer="627" to-port="0" />
|
|
<edge from-layer="627" from-port="1" to-layer="629" to-port="0" />
|
|
<edge from-layer="628" from-port="0" to-layer="629" to-port="1" />
|
|
<edge from-layer="629" from-port="2" to-layer="630" to-port="1" />
|
|
<edge from-layer="630" from-port="2" to-layer="632" to-port="0" />
|
|
<edge from-layer="631" from-port="0" to-layer="632" to-port="1" />
|
|
<edge from-layer="632" from-port="2" to-layer="633" to-port="0" />
|
|
<edge from-layer="633" from-port="1" to-layer="638" to-port="0" />
|
|
<edge from-layer="634" from-port="0" to-layer="638" to-port="1" />
|
|
<edge from-layer="635" from-port="0" to-layer="638" to-port="2" />
|
|
<edge from-layer="636" from-port="0" to-layer="638" to-port="3" />
|
|
<edge from-layer="637" from-port="0" to-layer="638" to-port="4" />
|
|
<edge from-layer="638" from-port="5" to-layer="641" to-port="0" />
|
|
<edge from-layer="638" from-port="5" to-layer="1060" to-port="1" />
|
|
<edge from-layer="639" from-port="0" to-layer="641" to-port="1" />
|
|
<edge from-layer="640" from-port="0" to-layer="641" to-port="2" />
|
|
<edge from-layer="641" from-port="3" to-layer="646" to-port="0" />
|
|
<edge from-layer="642" from-port="0" to-layer="646" to-port="1" />
|
|
<edge from-layer="643" from-port="0" to-layer="646" to-port="2" />
|
|
<edge from-layer="644" from-port="0" to-layer="646" to-port="3" />
|
|
<edge from-layer="645" from-port="0" to-layer="646" to-port="4" />
|
|
<edge from-layer="646" from-port="5" to-layer="647" to-port="0" />
|
|
<edge from-layer="647" from-port="2" to-layer="652" to-port="0" />
|
|
<edge from-layer="648" from-port="0" to-layer="649" to-port="0" />
|
|
<edge from-layer="649" from-port="1" to-layer="651" to-port="0" />
|
|
<edge from-layer="650" from-port="0" to-layer="651" to-port="1" />
|
|
<edge from-layer="651" from-port="2" to-layer="652" to-port="1" />
|
|
<edge from-layer="652" from-port="2" to-layer="654" to-port="0" />
|
|
<edge from-layer="653" from-port="0" to-layer="654" to-port="1" />
|
|
<edge from-layer="654" from-port="2" to-layer="655" to-port="0" />
|
|
<edge from-layer="655" from-port="1" to-layer="660" to-port="0" />
|
|
<edge from-layer="656" from-port="0" to-layer="660" to-port="1" />
|
|
<edge from-layer="657" from-port="0" to-layer="660" to-port="2" />
|
|
<edge from-layer="658" from-port="0" to-layer="660" to-port="3" />
|
|
<edge from-layer="659" from-port="0" to-layer="660" to-port="4" />
|
|
<edge from-layer="660" from-port="5" to-layer="663" to-port="0" />
|
|
<edge from-layer="661" from-port="0" to-layer="663" to-port="1" />
|
|
<edge from-layer="662" from-port="0" to-layer="663" to-port="2" />
|
|
<edge from-layer="663" from-port="4" to-layer="668" to-port="0" />
|
|
<edge from-layer="663" from-port="3" to-layer="696" to-port="0" />
|
|
<edge from-layer="663" from-port="4" to-layer="696" to-port="1" />
|
|
<edge from-layer="663" from-port="4" to-layer="690" to-port="0" />
|
|
<edge from-layer="664" from-port="0" to-layer="665" to-port="0" />
|
|
<edge from-layer="665" from-port="1" to-layer="667" to-port="0" />
|
|
<edge from-layer="666" from-port="0" to-layer="667" to-port="1" />
|
|
<edge from-layer="667" from-port="2" to-layer="668" to-port="1" />
|
|
<edge from-layer="668" from-port="2" to-layer="670" to-port="0" />
|
|
<edge from-layer="669" from-port="0" to-layer="670" to-port="1" />
|
|
<edge from-layer="670" from-port="2" to-layer="671" to-port="0" />
|
|
<edge from-layer="671" from-port="1" to-layer="676" to-port="0" />
|
|
<edge from-layer="672" from-port="0" to-layer="676" to-port="1" />
|
|
<edge from-layer="673" from-port="0" to-layer="676" to-port="2" />
|
|
<edge from-layer="674" from-port="0" to-layer="676" to-port="3" />
|
|
<edge from-layer="675" from-port="0" to-layer="676" to-port="4" />
|
|
<edge from-layer="676" from-port="5" to-layer="681" to-port="0" />
|
|
<edge from-layer="677" from-port="0" to-layer="678" to-port="0" />
|
|
<edge from-layer="678" from-port="1" to-layer="680" to-port="0" />
|
|
<edge from-layer="679" from-port="0" to-layer="680" to-port="1" />
|
|
<edge from-layer="680" from-port="2" to-layer="681" to-port="1" />
|
|
<edge from-layer="681" from-port="2" to-layer="683" to-port="0" />
|
|
<edge from-layer="682" from-port="0" to-layer="683" to-port="1" />
|
|
<edge from-layer="683" from-port="2" to-layer="684" to-port="0" />
|
|
<edge from-layer="684" from-port="1" to-layer="689" to-port="0" />
|
|
<edge from-layer="685" from-port="0" to-layer="689" to-port="1" />
|
|
<edge from-layer="686" from-port="0" to-layer="689" to-port="2" />
|
|
<edge from-layer="687" from-port="0" to-layer="689" to-port="3" />
|
|
<edge from-layer="688" from-port="0" to-layer="689" to-port="4" />
|
|
<edge from-layer="689" from-port="5" to-layer="690" to-port="1" />
|
|
<edge from-layer="690" from-port="2" to-layer="695" to-port="0" />
|
|
<edge from-layer="691" from-port="0" to-layer="695" to-port="1" />
|
|
<edge from-layer="692" from-port="0" to-layer="695" to-port="2" />
|
|
<edge from-layer="693" from-port="0" to-layer="695" to-port="3" />
|
|
<edge from-layer="694" from-port="0" to-layer="695" to-port="4" />
|
|
<edge from-layer="695" from-port="5" to-layer="696" to-port="2" />
|
|
<edge from-layer="696" from-port="3" to-layer="701" to-port="0" />
|
|
<edge from-layer="697" from-port="0" to-layer="698" to-port="0" />
|
|
<edge from-layer="698" from-port="1" to-layer="700" to-port="0" />
|
|
<edge from-layer="699" from-port="0" to-layer="700" to-port="1" />
|
|
<edge from-layer="700" from-port="2" to-layer="701" to-port="1" />
|
|
<edge from-layer="701" from-port="2" to-layer="703" to-port="0" />
|
|
<edge from-layer="702" from-port="0" to-layer="703" to-port="1" />
|
|
<edge from-layer="703" from-port="2" to-layer="704" to-port="0" />
|
|
<edge from-layer="704" from-port="1" to-layer="709" to-port="0" />
|
|
<edge from-layer="705" from-port="0" to-layer="709" to-port="1" />
|
|
<edge from-layer="706" from-port="0" to-layer="709" to-port="2" />
|
|
<edge from-layer="707" from-port="0" to-layer="709" to-port="3" />
|
|
<edge from-layer="708" from-port="0" to-layer="709" to-port="4" />
|
|
<edge from-layer="709" from-port="5" to-layer="889" to-port="1" />
|
|
<edge from-layer="709" from-port="5" to-layer="712" to-port="0" />
|
|
<edge from-layer="710" from-port="0" to-layer="712" to-port="1" />
|
|
<edge from-layer="711" from-port="0" to-layer="712" to-port="2" />
|
|
<edge from-layer="712" from-port="3" to-layer="717" to-port="0" />
|
|
<edge from-layer="713" from-port="0" to-layer="717" to-port="1" />
|
|
<edge from-layer="714" from-port="0" to-layer="717" to-port="2" />
|
|
<edge from-layer="715" from-port="0" to-layer="717" to-port="3" />
|
|
<edge from-layer="716" from-port="0" to-layer="717" to-port="4" />
|
|
<edge from-layer="717" from-port="5" to-layer="718" to-port="0" />
|
|
<edge from-layer="718" from-port="2" to-layer="723" to-port="0" />
|
|
<edge from-layer="719" from-port="0" to-layer="720" to-port="0" />
|
|
<edge from-layer="720" from-port="1" to-layer="722" to-port="0" />
|
|
<edge from-layer="721" from-port="0" to-layer="722" to-port="1" />
|
|
<edge from-layer="722" from-port="2" to-layer="723" to-port="1" />
|
|
<edge from-layer="723" from-port="2" to-layer="725" to-port="0" />
|
|
<edge from-layer="724" from-port="0" to-layer="725" to-port="1" />
|
|
<edge from-layer="725" from-port="2" to-layer="726" to-port="0" />
|
|
<edge from-layer="726" from-port="1" to-layer="731" to-port="0" />
|
|
<edge from-layer="727" from-port="0" to-layer="731" to-port="1" />
|
|
<edge from-layer="728" from-port="0" to-layer="731" to-port="2" />
|
|
<edge from-layer="729" from-port="0" to-layer="731" to-port="3" />
|
|
<edge from-layer="730" from-port="0" to-layer="731" to-port="4" />
|
|
<edge from-layer="731" from-port="5" to-layer="734" to-port="0" />
|
|
<edge from-layer="732" from-port="0" to-layer="734" to-port="1" />
|
|
<edge from-layer="733" from-port="0" to-layer="734" to-port="2" />
|
|
<edge from-layer="734" from-port="4" to-layer="739" to-port="0" />
|
|
<edge from-layer="734" from-port="4" to-layer="767" to-port="1" />
|
|
<edge from-layer="734" from-port="3" to-layer="767" to-port="0" />
|
|
<edge from-layer="734" from-port="4" to-layer="761" to-port="0" />
|
|
<edge from-layer="735" from-port="0" to-layer="736" to-port="0" />
|
|
<edge from-layer="736" from-port="1" to-layer="738" to-port="0" />
|
|
<edge from-layer="737" from-port="0" to-layer="738" to-port="1" />
|
|
<edge from-layer="738" from-port="2" to-layer="739" to-port="1" />
|
|
<edge from-layer="739" from-port="2" to-layer="741" to-port="0" />
|
|
<edge from-layer="740" from-port="0" to-layer="741" to-port="1" />
|
|
<edge from-layer="741" from-port="2" to-layer="742" to-port="0" />
|
|
<edge from-layer="742" from-port="1" to-layer="747" to-port="0" />
|
|
<edge from-layer="743" from-port="0" to-layer="747" to-port="1" />
|
|
<edge from-layer="744" from-port="0" to-layer="747" to-port="2" />
|
|
<edge from-layer="745" from-port="0" to-layer="747" to-port="3" />
|
|
<edge from-layer="746" from-port="0" to-layer="747" to-port="4" />
|
|
<edge from-layer="747" from-port="5" to-layer="752" to-port="0" />
|
|
<edge from-layer="748" from-port="0" to-layer="749" to-port="0" />
|
|
<edge from-layer="749" from-port="1" to-layer="751" to-port="0" />
|
|
<edge from-layer="750" from-port="0" to-layer="751" to-port="1" />
|
|
<edge from-layer="751" from-port="2" to-layer="752" to-port="1" />
|
|
<edge from-layer="752" from-port="2" to-layer="754" to-port="0" />
|
|
<edge from-layer="753" from-port="0" to-layer="754" to-port="1" />
|
|
<edge from-layer="754" from-port="2" to-layer="755" to-port="0" />
|
|
<edge from-layer="755" from-port="1" to-layer="760" to-port="0" />
|
|
<edge from-layer="756" from-port="0" to-layer="760" to-port="1" />
|
|
<edge from-layer="757" from-port="0" to-layer="760" to-port="2" />
|
|
<edge from-layer="758" from-port="0" to-layer="760" to-port="3" />
|
|
<edge from-layer="759" from-port="0" to-layer="760" to-port="4" />
|
|
<edge from-layer="760" from-port="5" to-layer="761" to-port="1" />
|
|
<edge from-layer="761" from-port="2" to-layer="766" to-port="0" />
|
|
<edge from-layer="762" from-port="0" to-layer="766" to-port="1" />
|
|
<edge from-layer="763" from-port="0" to-layer="766" to-port="2" />
|
|
<edge from-layer="764" from-port="0" to-layer="766" to-port="3" />
|
|
<edge from-layer="765" from-port="0" to-layer="766" to-port="4" />
|
|
<edge from-layer="766" from-port="5" to-layer="767" to-port="2" />
|
|
<edge from-layer="767" from-port="3" to-layer="772" to-port="0" />
|
|
<edge from-layer="768" from-port="0" to-layer="769" to-port="0" />
|
|
<edge from-layer="769" from-port="1" to-layer="771" to-port="0" />
|
|
<edge from-layer="770" from-port="0" to-layer="771" to-port="1" />
|
|
<edge from-layer="771" from-port="2" to-layer="772" to-port="1" />
|
|
<edge from-layer="772" from-port="2" to-layer="774" to-port="0" />
|
|
<edge from-layer="773" from-port="0" to-layer="774" to-port="1" />
|
|
<edge from-layer="774" from-port="2" to-layer="775" to-port="0" />
|
|
<edge from-layer="775" from-port="1" to-layer="780" to-port="0" />
|
|
<edge from-layer="776" from-port="0" to-layer="780" to-port="1" />
|
|
<edge from-layer="777" from-port="0" to-layer="780" to-port="2" />
|
|
<edge from-layer="778" from-port="0" to-layer="780" to-port="3" />
|
|
<edge from-layer="779" from-port="0" to-layer="780" to-port="4" />
|
|
<edge from-layer="780" from-port="5" to-layer="818" to-port="0" />
|
|
<edge from-layer="780" from-port="5" to-layer="1329" to-port="0" />
|
|
<edge from-layer="780" from-port="5" to-layer="880" to-port="0" />
|
|
<edge from-layer="780" from-port="5" to-layer="785" to-port="0" />
|
|
<edge from-layer="781" from-port="0" to-layer="782" to-port="0" />
|
|
<edge from-layer="782" from-port="1" to-layer="784" to-port="0" />
|
|
<edge from-layer="783" from-port="0" to-layer="784" to-port="1" />
|
|
<edge from-layer="784" from-port="2" to-layer="785" to-port="1" />
|
|
<edge from-layer="785" from-port="2" to-layer="787" to-port="0" />
|
|
<edge from-layer="786" from-port="0" to-layer="787" to-port="1" />
|
|
<edge from-layer="787" from-port="2" to-layer="788" to-port="0" />
|
|
<edge from-layer="788" from-port="1" to-layer="793" to-port="0" />
|
|
<edge from-layer="789" from-port="0" to-layer="793" to-port="1" />
|
|
<edge from-layer="790" from-port="0" to-layer="793" to-port="2" />
|
|
<edge from-layer="791" from-port="0" to-layer="793" to-port="3" />
|
|
<edge from-layer="792" from-port="0" to-layer="793" to-port="4" />
|
|
<edge from-layer="793" from-port="5" to-layer="798" to-port="0" />
|
|
<edge from-layer="794" from-port="0" to-layer="795" to-port="0" />
|
|
<edge from-layer="795" from-port="1" to-layer="797" to-port="0" />
|
|
<edge from-layer="796" from-port="0" to-layer="797" to-port="1" />
|
|
<edge from-layer="797" from-port="2" to-layer="798" to-port="1" />
|
|
<edge from-layer="798" from-port="2" to-layer="800" to-port="0" />
|
|
<edge from-layer="799" from-port="0" to-layer="800" to-port="1" />
|
|
<edge from-layer="800" from-port="2" to-layer="801" to-port="0" />
|
|
<edge from-layer="801" from-port="1" to-layer="806" to-port="0" />
|
|
<edge from-layer="802" from-port="0" to-layer="806" to-port="1" />
|
|
<edge from-layer="803" from-port="0" to-layer="806" to-port="2" />
|
|
<edge from-layer="804" from-port="0" to-layer="806" to-port="3" />
|
|
<edge from-layer="805" from-port="0" to-layer="806" to-port="4" />
|
|
<edge from-layer="806" from-port="5" to-layer="811" to-port="0" />
|
|
<edge from-layer="807" from-port="0" to-layer="808" to-port="0" />
|
|
<edge from-layer="808" from-port="1" to-layer="810" to-port="0" />
|
|
<edge from-layer="809" from-port="0" to-layer="810" to-port="1" />
|
|
<edge from-layer="810" from-port="2" to-layer="811" to-port="1" />
|
|
<edge from-layer="811" from-port="2" to-layer="813" to-port="0" />
|
|
<edge from-layer="812" from-port="0" to-layer="813" to-port="1" />
|
|
<edge from-layer="813" from-port="2" to-layer="873" to-port="0" />
|
|
<edge from-layer="814" from-port="0" to-layer="815" to-port="0" />
|
|
<edge from-layer="815" from-port="1" to-layer="817" to-port="0" />
|
|
<edge from-layer="816" from-port="0" to-layer="817" to-port="1" />
|
|
<edge from-layer="817" from-port="2" to-layer="818" to-port="1" />
|
|
<edge from-layer="818" from-port="2" to-layer="820" to-port="0" />
|
|
<edge from-layer="819" from-port="0" to-layer="820" to-port="1" />
|
|
<edge from-layer="820" from-port="2" to-layer="821" to-port="0" />
|
|
<edge from-layer="821" from-port="1" to-layer="826" to-port="0" />
|
|
<edge from-layer="822" from-port="0" to-layer="826" to-port="1" />
|
|
<edge from-layer="823" from-port="0" to-layer="826" to-port="2" />
|
|
<edge from-layer="824" from-port="0" to-layer="826" to-port="3" />
|
|
<edge from-layer="825" from-port="0" to-layer="826" to-port="4" />
|
|
<edge from-layer="826" from-port="5" to-layer="831" to-port="0" />
|
|
<edge from-layer="827" from-port="0" to-layer="828" to-port="0" />
|
|
<edge from-layer="828" from-port="1" to-layer="830" to-port="0" />
|
|
<edge from-layer="829" from-port="0" to-layer="830" to-port="1" />
|
|
<edge from-layer="830" from-port="2" to-layer="831" to-port="1" />
|
|
<edge from-layer="831" from-port="2" to-layer="833" to-port="0" />
|
|
<edge from-layer="832" from-port="0" to-layer="833" to-port="1" />
|
|
<edge from-layer="833" from-port="2" to-layer="834" to-port="0" />
|
|
<edge from-layer="834" from-port="1" to-layer="839" to-port="0" />
|
|
<edge from-layer="835" from-port="0" to-layer="839" to-port="1" />
|
|
<edge from-layer="836" from-port="0" to-layer="839" to-port="2" />
|
|
<edge from-layer="837" from-port="0" to-layer="839" to-port="3" />
|
|
<edge from-layer="838" from-port="0" to-layer="839" to-port="4" />
|
|
<edge from-layer="839" from-port="5" to-layer="844" to-port="0" />
|
|
<edge from-layer="840" from-port="0" to-layer="841" to-port="0" />
|
|
<edge from-layer="841" from-port="1" to-layer="843" to-port="0" />
|
|
<edge from-layer="842" from-port="0" to-layer="843" to-port="1" />
|
|
<edge from-layer="843" from-port="2" to-layer="844" to-port="1" />
|
|
<edge from-layer="844" from-port="2" to-layer="846" to-port="0" />
|
|
<edge from-layer="845" from-port="0" to-layer="846" to-port="1" />
|
|
<edge from-layer="846" from-port="2" to-layer="847" to-port="0" />
|
|
<edge from-layer="847" from-port="1" to-layer="852" to-port="0" />
|
|
<edge from-layer="848" from-port="0" to-layer="852" to-port="1" />
|
|
<edge from-layer="849" from-port="0" to-layer="852" to-port="2" />
|
|
<edge from-layer="850" from-port="0" to-layer="852" to-port="3" />
|
|
<edge from-layer="851" from-port="0" to-layer="852" to-port="4" />
|
|
<edge from-layer="852" from-port="5" to-layer="857" to-port="0" />
|
|
<edge from-layer="853" from-port="0" to-layer="854" to-port="0" />
|
|
<edge from-layer="854" from-port="1" to-layer="856" to-port="0" />
|
|
<edge from-layer="855" from-port="0" to-layer="856" to-port="1" />
|
|
<edge from-layer="856" from-port="2" to-layer="857" to-port="1" />
|
|
<edge from-layer="857" from-port="2" to-layer="859" to-port="0" />
|
|
<edge from-layer="858" from-port="0" to-layer="859" to-port="1" />
|
|
<edge from-layer="859" from-port="2" to-layer="860" to-port="0" />
|
|
<edge from-layer="860" from-port="1" to-layer="865" to-port="0" />
|
|
<edge from-layer="861" from-port="0" to-layer="865" to-port="1" />
|
|
<edge from-layer="862" from-port="0" to-layer="865" to-port="2" />
|
|
<edge from-layer="863" from-port="0" to-layer="865" to-port="3" />
|
|
<edge from-layer="864" from-port="0" to-layer="865" to-port="4" />
|
|
<edge from-layer="865" from-port="5" to-layer="870" to-port="0" />
|
|
<edge from-layer="866" from-port="0" to-layer="867" to-port="0" />
|
|
<edge from-layer="867" from-port="1" to-layer="869" to-port="0" />
|
|
<edge from-layer="868" from-port="0" to-layer="869" to-port="1" />
|
|
<edge from-layer="869" from-port="2" to-layer="870" to-port="1" />
|
|
<edge from-layer="870" from-port="2" to-layer="872" to-port="0" />
|
|
<edge from-layer="871" from-port="0" to-layer="872" to-port="1" />
|
|
<edge from-layer="872" from-port="2" to-layer="873" to-port="1" />
|
|
<edge from-layer="873" from-port="2" to-layer="875" to-port="0" />
|
|
<edge from-layer="874" from-port="0" to-layer="875" to-port="1" />
|
|
<edge from-layer="875" from-port="2" to-layer="1290" to-port="0" />
|
|
<edge from-layer="876" from-port="0" to-layer="877" to-port="0" />
|
|
<edge from-layer="877" from-port="1" to-layer="879" to-port="0" />
|
|
<edge from-layer="878" from-port="0" to-layer="879" to-port="1" />
|
|
<edge from-layer="879" from-port="2" to-layer="880" to-port="1" />
|
|
<edge from-layer="880" from-port="2" to-layer="882" to-port="0" />
|
|
<edge from-layer="881" from-port="0" to-layer="882" to-port="1" />
|
|
<edge from-layer="882" from-port="2" to-layer="883" to-port="0" />
|
|
<edge from-layer="883" from-port="1" to-layer="888" to-port="0" />
|
|
<edge from-layer="884" from-port="0" to-layer="888" to-port="1" />
|
|
<edge from-layer="885" from-port="0" to-layer="888" to-port="2" />
|
|
<edge from-layer="886" from-port="0" to-layer="888" to-port="3" />
|
|
<edge from-layer="887" from-port="0" to-layer="888" to-port="4" />
|
|
<edge from-layer="888" from-port="5" to-layer="889" to-port="0" />
|
|
<edge from-layer="889" from-port="2" to-layer="894" to-port="0" />
|
|
<edge from-layer="890" from-port="0" to-layer="891" to-port="0" />
|
|
<edge from-layer="891" from-port="1" to-layer="893" to-port="0" />
|
|
<edge from-layer="892" from-port="0" to-layer="893" to-port="1" />
|
|
<edge from-layer="893" from-port="2" to-layer="894" to-port="1" />
|
|
<edge from-layer="894" from-port="2" to-layer="896" to-port="0" />
|
|
<edge from-layer="895" from-port="0" to-layer="896" to-port="1" />
|
|
<edge from-layer="896" from-port="2" to-layer="897" to-port="0" />
|
|
<edge from-layer="897" from-port="1" to-layer="902" to-port="0" />
|
|
<edge from-layer="898" from-port="0" to-layer="902" to-port="1" />
|
|
<edge from-layer="899" from-port="0" to-layer="902" to-port="2" />
|
|
<edge from-layer="900" from-port="0" to-layer="902" to-port="3" />
|
|
<edge from-layer="901" from-port="0" to-layer="902" to-port="4" />
|
|
<edge from-layer="902" from-port="5" to-layer="905" to-port="0" />
|
|
<edge from-layer="903" from-port="0" to-layer="905" to-port="1" />
|
|
<edge from-layer="904" from-port="0" to-layer="905" to-port="2" />
|
|
<edge from-layer="905" from-port="4" to-layer="938" to-port="1" />
|
|
<edge from-layer="905" from-port="3" to-layer="938" to-port="0" />
|
|
<edge from-layer="905" from-port="4" to-layer="932" to-port="0" />
|
|
<edge from-layer="905" from-port="4" to-layer="910" to-port="0" />
|
|
<edge from-layer="906" from-port="0" to-layer="907" to-port="0" />
|
|
<edge from-layer="907" from-port="1" to-layer="909" to-port="0" />
|
|
<edge from-layer="908" from-port="0" to-layer="909" to-port="1" />
|
|
<edge from-layer="909" from-port="2" to-layer="910" to-port="1" />
|
|
<edge from-layer="910" from-port="2" to-layer="912" to-port="0" />
|
|
<edge from-layer="911" from-port="0" to-layer="912" to-port="1" />
|
|
<edge from-layer="912" from-port="2" to-layer="913" to-port="0" />
|
|
<edge from-layer="913" from-port="1" to-layer="918" to-port="0" />
|
|
<edge from-layer="914" from-port="0" to-layer="918" to-port="1" />
|
|
<edge from-layer="915" from-port="0" to-layer="918" to-port="2" />
|
|
<edge from-layer="916" from-port="0" to-layer="918" to-port="3" />
|
|
<edge from-layer="917" from-port="0" to-layer="918" to-port="4" />
|
|
<edge from-layer="918" from-port="5" to-layer="923" to-port="0" />
|
|
<edge from-layer="919" from-port="0" to-layer="920" to-port="0" />
|
|
<edge from-layer="920" from-port="1" to-layer="922" to-port="0" />
|
|
<edge from-layer="921" from-port="0" to-layer="922" to-port="1" />
|
|
<edge from-layer="922" from-port="2" to-layer="923" to-port="1" />
|
|
<edge from-layer="923" from-port="2" to-layer="925" to-port="0" />
|
|
<edge from-layer="924" from-port="0" to-layer="925" to-port="1" />
|
|
<edge from-layer="925" from-port="2" to-layer="926" to-port="0" />
|
|
<edge from-layer="926" from-port="1" to-layer="931" to-port="0" />
|
|
<edge from-layer="927" from-port="0" to-layer="931" to-port="1" />
|
|
<edge from-layer="928" from-port="0" to-layer="931" to-port="2" />
|
|
<edge from-layer="929" from-port="0" to-layer="931" to-port="3" />
|
|
<edge from-layer="930" from-port="0" to-layer="931" to-port="4" />
|
|
<edge from-layer="931" from-port="5" to-layer="932" to-port="1" />
|
|
<edge from-layer="932" from-port="2" to-layer="937" to-port="0" />
|
|
<edge from-layer="933" from-port="0" to-layer="937" to-port="1" />
|
|
<edge from-layer="934" from-port="0" to-layer="937" to-port="2" />
|
|
<edge from-layer="935" from-port="0" to-layer="937" to-port="3" />
|
|
<edge from-layer="936" from-port="0" to-layer="937" to-port="4" />
|
|
<edge from-layer="937" from-port="5" to-layer="938" to-port="2" />
|
|
<edge from-layer="938" from-port="3" to-layer="943" to-port="0" />
|
|
<edge from-layer="939" from-port="0" to-layer="940" to-port="0" />
|
|
<edge from-layer="940" from-port="1" to-layer="942" to-port="0" />
|
|
<edge from-layer="941" from-port="0" to-layer="942" to-port="1" />
|
|
<edge from-layer="942" from-port="2" to-layer="943" to-port="1" />
|
|
<edge from-layer="943" from-port="2" to-layer="945" to-port="0" />
|
|
<edge from-layer="944" from-port="0" to-layer="945" to-port="1" />
|
|
<edge from-layer="945" from-port="2" to-layer="946" to-port="0" />
|
|
<edge from-layer="946" from-port="1" to-layer="951" to-port="0" />
|
|
<edge from-layer="947" from-port="0" to-layer="951" to-port="1" />
|
|
<edge from-layer="948" from-port="0" to-layer="951" to-port="2" />
|
|
<edge from-layer="949" from-port="0" to-layer="951" to-port="3" />
|
|
<edge from-layer="950" from-port="0" to-layer="951" to-port="4" />
|
|
<edge from-layer="951" from-port="5" to-layer="1051" to-port="0" />
|
|
<edge from-layer="951" from-port="5" to-layer="989" to-port="0" />
|
|
<edge from-layer="951" from-port="5" to-layer="1369" to-port="0" />
|
|
<edge from-layer="951" from-port="5" to-layer="956" to-port="0" />
|
|
<edge from-layer="952" from-port="0" to-layer="953" to-port="0" />
|
|
<edge from-layer="953" from-port="1" to-layer="955" to-port="0" />
|
|
<edge from-layer="954" from-port="0" to-layer="955" to-port="1" />
|
|
<edge from-layer="955" from-port="2" to-layer="956" to-port="1" />
|
|
<edge from-layer="956" from-port="2" to-layer="958" to-port="0" />
|
|
<edge from-layer="957" from-port="0" to-layer="958" to-port="1" />
|
|
<edge from-layer="958" from-port="2" to-layer="959" to-port="0" />
|
|
<edge from-layer="959" from-port="1" to-layer="964" to-port="0" />
|
|
<edge from-layer="960" from-port="0" to-layer="964" to-port="1" />
|
|
<edge from-layer="961" from-port="0" to-layer="964" to-port="2" />
|
|
<edge from-layer="962" from-port="0" to-layer="964" to-port="3" />
|
|
<edge from-layer="963" from-port="0" to-layer="964" to-port="4" />
|
|
<edge from-layer="964" from-port="5" to-layer="969" to-port="0" />
|
|
<edge from-layer="965" from-port="0" to-layer="966" to-port="0" />
|
|
<edge from-layer="966" from-port="1" to-layer="968" to-port="0" />
|
|
<edge from-layer="967" from-port="0" to-layer="968" to-port="1" />
|
|
<edge from-layer="968" from-port="2" to-layer="969" to-port="1" />
|
|
<edge from-layer="969" from-port="2" to-layer="971" to-port="0" />
|
|
<edge from-layer="970" from-port="0" to-layer="971" to-port="1" />
|
|
<edge from-layer="971" from-port="2" to-layer="972" to-port="0" />
|
|
<edge from-layer="972" from-port="1" to-layer="977" to-port="0" />
|
|
<edge from-layer="973" from-port="0" to-layer="977" to-port="1" />
|
|
<edge from-layer="974" from-port="0" to-layer="977" to-port="2" />
|
|
<edge from-layer="975" from-port="0" to-layer="977" to-port="3" />
|
|
<edge from-layer="976" from-port="0" to-layer="977" to-port="4" />
|
|
<edge from-layer="977" from-port="5" to-layer="982" to-port="0" />
|
|
<edge from-layer="978" from-port="0" to-layer="979" to-port="0" />
|
|
<edge from-layer="979" from-port="1" to-layer="981" to-port="0" />
|
|
<edge from-layer="980" from-port="0" to-layer="981" to-port="1" />
|
|
<edge from-layer="981" from-port="2" to-layer="982" to-port="1" />
|
|
<edge from-layer="982" from-port="2" to-layer="984" to-port="0" />
|
|
<edge from-layer="983" from-port="0" to-layer="984" to-port="1" />
|
|
<edge from-layer="984" from-port="2" to-layer="1044" to-port="0" />
|
|
<edge from-layer="985" from-port="0" to-layer="986" to-port="0" />
|
|
<edge from-layer="986" from-port="1" to-layer="988" to-port="0" />
|
|
<edge from-layer="987" from-port="0" to-layer="988" to-port="1" />
|
|
<edge from-layer="988" from-port="2" to-layer="989" to-port="1" />
|
|
<edge from-layer="989" from-port="2" to-layer="991" to-port="0" />
|
|
<edge from-layer="990" from-port="0" to-layer="991" to-port="1" />
|
|
<edge from-layer="991" from-port="2" to-layer="992" to-port="0" />
|
|
<edge from-layer="992" from-port="1" to-layer="997" to-port="0" />
|
|
<edge from-layer="993" from-port="0" to-layer="997" to-port="1" />
|
|
<edge from-layer="994" from-port="0" to-layer="997" to-port="2" />
|
|
<edge from-layer="995" from-port="0" to-layer="997" to-port="3" />
|
|
<edge from-layer="996" from-port="0" to-layer="997" to-port="4" />
|
|
<edge from-layer="997" from-port="5" to-layer="1002" to-port="0" />
|
|
<edge from-layer="998" from-port="0" to-layer="999" to-port="0" />
|
|
<edge from-layer="999" from-port="1" to-layer="1001" to-port="0" />
|
|
<edge from-layer="1000" from-port="0" to-layer="1001" to-port="1" />
|
|
<edge from-layer="1001" from-port="2" to-layer="1002" to-port="1" />
|
|
<edge from-layer="1002" from-port="2" to-layer="1004" to-port="0" />
|
|
<edge from-layer="1003" from-port="0" to-layer="1004" to-port="1" />
|
|
<edge from-layer="1004" from-port="2" to-layer="1005" to-port="0" />
|
|
<edge from-layer="1005" from-port="1" to-layer="1010" to-port="0" />
|
|
<edge from-layer="1006" from-port="0" to-layer="1010" to-port="1" />
|
|
<edge from-layer="1007" from-port="0" to-layer="1010" to-port="2" />
|
|
<edge from-layer="1008" from-port="0" to-layer="1010" to-port="3" />
|
|
<edge from-layer="1009" from-port="0" to-layer="1010" to-port="4" />
|
|
<edge from-layer="1010" from-port="5" to-layer="1015" to-port="0" />
|
|
<edge from-layer="1011" from-port="0" to-layer="1012" to-port="0" />
|
|
<edge from-layer="1012" from-port="1" to-layer="1014" to-port="0" />
|
|
<edge from-layer="1013" from-port="0" to-layer="1014" to-port="1" />
|
|
<edge from-layer="1014" from-port="2" to-layer="1015" to-port="1" />
|
|
<edge from-layer="1015" from-port="2" to-layer="1017" to-port="0" />
|
|
<edge from-layer="1016" from-port="0" to-layer="1017" to-port="1" />
|
|
<edge from-layer="1017" from-port="2" to-layer="1018" to-port="0" />
|
|
<edge from-layer="1018" from-port="1" to-layer="1023" to-port="0" />
|
|
<edge from-layer="1019" from-port="0" to-layer="1023" to-port="1" />
|
|
<edge from-layer="1020" from-port="0" to-layer="1023" to-port="2" />
|
|
<edge from-layer="1021" from-port="0" to-layer="1023" to-port="3" />
|
|
<edge from-layer="1022" from-port="0" to-layer="1023" to-port="4" />
|
|
<edge from-layer="1023" from-port="5" to-layer="1028" to-port="0" />
|
|
<edge from-layer="1024" from-port="0" to-layer="1025" to-port="0" />
|
|
<edge from-layer="1025" from-port="1" to-layer="1027" to-port="0" />
|
|
<edge from-layer="1026" from-port="0" to-layer="1027" to-port="1" />
|
|
<edge from-layer="1027" from-port="2" to-layer="1028" to-port="1" />
|
|
<edge from-layer="1028" from-port="2" to-layer="1030" to-port="0" />
|
|
<edge from-layer="1029" from-port="0" to-layer="1030" to-port="1" />
|
|
<edge from-layer="1030" from-port="2" to-layer="1031" to-port="0" />
|
|
<edge from-layer="1031" from-port="1" to-layer="1036" to-port="0" />
|
|
<edge from-layer="1032" from-port="0" to-layer="1036" to-port="1" />
|
|
<edge from-layer="1033" from-port="0" to-layer="1036" to-port="2" />
|
|
<edge from-layer="1034" from-port="0" to-layer="1036" to-port="3" />
|
|
<edge from-layer="1035" from-port="0" to-layer="1036" to-port="4" />
|
|
<edge from-layer="1036" from-port="5" to-layer="1041" to-port="0" />
|
|
<edge from-layer="1037" from-port="0" to-layer="1038" to-port="0" />
|
|
<edge from-layer="1038" from-port="1" to-layer="1040" to-port="0" />
|
|
<edge from-layer="1039" from-port="0" to-layer="1040" to-port="1" />
|
|
<edge from-layer="1040" from-port="2" to-layer="1041" to-port="1" />
|
|
<edge from-layer="1041" from-port="2" to-layer="1043" to-port="0" />
|
|
<edge from-layer="1042" from-port="0" to-layer="1043" to-port="1" />
|
|
<edge from-layer="1043" from-port="2" to-layer="1044" to-port="1" />
|
|
<edge from-layer="1044" from-port="2" to-layer="1046" to-port="0" />
|
|
<edge from-layer="1045" from-port="0" to-layer="1046" to-port="1" />
|
|
<edge from-layer="1046" from-port="2" to-layer="1290" to-port="1" />
|
|
<edge from-layer="1047" from-port="0" to-layer="1048" to-port="0" />
|
|
<edge from-layer="1048" from-port="1" to-layer="1050" to-port="0" />
|
|
<edge from-layer="1049" from-port="0" to-layer="1050" to-port="1" />
|
|
<edge from-layer="1050" from-port="2" to-layer="1051" to-port="1" />
|
|
<edge from-layer="1051" from-port="2" to-layer="1053" to-port="0" />
|
|
<edge from-layer="1052" from-port="0" to-layer="1053" to-port="1" />
|
|
<edge from-layer="1053" from-port="2" to-layer="1054" to-port="0" />
|
|
<edge from-layer="1054" from-port="1" to-layer="1059" to-port="0" />
|
|
<edge from-layer="1055" from-port="0" to-layer="1059" to-port="1" />
|
|
<edge from-layer="1056" from-port="0" to-layer="1059" to-port="2" />
|
|
<edge from-layer="1057" from-port="0" to-layer="1059" to-port="3" />
|
|
<edge from-layer="1058" from-port="0" to-layer="1059" to-port="4" />
|
|
<edge from-layer="1059" from-port="5" to-layer="1060" to-port="0" />
|
|
<edge from-layer="1060" from-port="2" to-layer="1065" to-port="0" />
|
|
<edge from-layer="1061" from-port="0" to-layer="1062" to-port="0" />
|
|
<edge from-layer="1062" from-port="1" to-layer="1064" to-port="0" />
|
|
<edge from-layer="1063" from-port="0" to-layer="1064" to-port="1" />
|
|
<edge from-layer="1064" from-port="2" to-layer="1065" to-port="1" />
|
|
<edge from-layer="1065" from-port="2" to-layer="1067" to-port="0" />
|
|
<edge from-layer="1066" from-port="0" to-layer="1067" to-port="1" />
|
|
<edge from-layer="1067" from-port="2" to-layer="1068" to-port="0" />
|
|
<edge from-layer="1068" from-port="1" to-layer="1073" to-port="0" />
|
|
<edge from-layer="1069" from-port="0" to-layer="1073" to-port="1" />
|
|
<edge from-layer="1070" from-port="0" to-layer="1073" to-port="2" />
|
|
<edge from-layer="1071" from-port="0" to-layer="1073" to-port="3" />
|
|
<edge from-layer="1072" from-port="0" to-layer="1073" to-port="4" />
|
|
<edge from-layer="1073" from-port="5" to-layer="1076" to-port="0" />
|
|
<edge from-layer="1074" from-port="0" to-layer="1076" to-port="1" />
|
|
<edge from-layer="1075" from-port="0" to-layer="1076" to-port="2" />
|
|
<edge from-layer="1076" from-port="3" to-layer="1181" to-port="0" />
|
|
<edge from-layer="1076" from-port="4" to-layer="1181" to-port="1" />
|
|
<edge from-layer="1076" from-port="4" to-layer="1158" to-port="0" />
|
|
<edge from-layer="1076" from-port="4" to-layer="1081" to-port="0" />
|
|
<edge from-layer="1077" from-port="0" to-layer="1078" to-port="0" />
|
|
<edge from-layer="1078" from-port="1" to-layer="1080" to-port="0" />
|
|
<edge from-layer="1079" from-port="0" to-layer="1080" to-port="1" />
|
|
<edge from-layer="1080" from-port="2" to-layer="1081" to-port="1" />
|
|
<edge from-layer="1081" from-port="2" to-layer="1083" to-port="0" />
|
|
<edge from-layer="1082" from-port="0" to-layer="1083" to-port="1" />
|
|
<edge from-layer="1083" from-port="2" to-layer="1084" to-port="0" />
|
|
<edge from-layer="1084" from-port="1" to-layer="1089" to-port="0" />
|
|
<edge from-layer="1085" from-port="0" to-layer="1089" to-port="1" />
|
|
<edge from-layer="1086" from-port="0" to-layer="1089" to-port="2" />
|
|
<edge from-layer="1087" from-port="0" to-layer="1089" to-port="3" />
|
|
<edge from-layer="1088" from-port="0" to-layer="1089" to-port="4" />
|
|
<edge from-layer="1089" from-port="5" to-layer="1116" to-port="0" />
|
|
<edge from-layer="1089" from-port="5" to-layer="1094" to-port="0" />
|
|
<edge from-layer="1090" from-port="0" to-layer="1091" to-port="0" />
|
|
<edge from-layer="1091" from-port="1" to-layer="1093" to-port="0" />
|
|
<edge from-layer="1092" from-port="0" to-layer="1093" to-port="1" />
|
|
<edge from-layer="1093" from-port="2" to-layer="1094" to-port="1" />
|
|
<edge from-layer="1094" from-port="2" to-layer="1096" to-port="0" />
|
|
<edge from-layer="1095" from-port="0" to-layer="1096" to-port="1" />
|
|
<edge from-layer="1096" from-port="2" to-layer="1097" to-port="0" />
|
|
<edge from-layer="1097" from-port="1" to-layer="1102" to-port="0" />
|
|
<edge from-layer="1098" from-port="0" to-layer="1102" to-port="1" />
|
|
<edge from-layer="1099" from-port="0" to-layer="1102" to-port="2" />
|
|
<edge from-layer="1100" from-port="0" to-layer="1102" to-port="3" />
|
|
<edge from-layer="1101" from-port="0" to-layer="1102" to-port="4" />
|
|
<edge from-layer="1102" from-port="5" to-layer="1107" to-port="0" />
|
|
<edge from-layer="1103" from-port="0" to-layer="1104" to-port="0" />
|
|
<edge from-layer="1104" from-port="1" to-layer="1106" to-port="0" />
|
|
<edge from-layer="1105" from-port="0" to-layer="1106" to-port="1" />
|
|
<edge from-layer="1106" from-port="2" to-layer="1107" to-port="1" />
|
|
<edge from-layer="1107" from-port="2" to-layer="1109" to-port="0" />
|
|
<edge from-layer="1108" from-port="0" to-layer="1109" to-port="1" />
|
|
<edge from-layer="1109" from-port="2" to-layer="1110" to-port="0" />
|
|
<edge from-layer="1110" from-port="1" to-layer="1115" to-port="0" />
|
|
<edge from-layer="1111" from-port="0" to-layer="1115" to-port="1" />
|
|
<edge from-layer="1112" from-port="0" to-layer="1115" to-port="2" />
|
|
<edge from-layer="1113" from-port="0" to-layer="1115" to-port="3" />
|
|
<edge from-layer="1114" from-port="0" to-layer="1115" to-port="4" />
|
|
<edge from-layer="1115" from-port="5" to-layer="1116" to-port="1" />
|
|
<edge from-layer="1116" from-port="2" to-layer="1121" to-port="0" />
|
|
<edge from-layer="1117" from-port="0" to-layer="1121" to-port="1" />
|
|
<edge from-layer="1118" from-port="0" to-layer="1121" to-port="2" />
|
|
<edge from-layer="1119" from-port="0" to-layer="1121" to-port="3" />
|
|
<edge from-layer="1120" from-port="0" to-layer="1121" to-port="4" />
|
|
<edge from-layer="1121" from-port="5" to-layer="1126" to-port="0" />
|
|
<edge from-layer="1121" from-port="5" to-layer="1148" to-port="0" />
|
|
<edge from-layer="1122" from-port="0" to-layer="1123" to-port="0" />
|
|
<edge from-layer="1123" from-port="1" to-layer="1125" to-port="0" />
|
|
<edge from-layer="1124" from-port="0" to-layer="1125" to-port="1" />
|
|
<edge from-layer="1125" from-port="2" to-layer="1126" to-port="1" />
|
|
<edge from-layer="1126" from-port="2" to-layer="1128" to-port="0" />
|
|
<edge from-layer="1127" from-port="0" to-layer="1128" to-port="1" />
|
|
<edge from-layer="1128" from-port="2" to-layer="1129" to-port="0" />
|
|
<edge from-layer="1129" from-port="1" to-layer="1134" to-port="0" />
|
|
<edge from-layer="1130" from-port="0" to-layer="1134" to-port="1" />
|
|
<edge from-layer="1131" from-port="0" to-layer="1134" to-port="2" />
|
|
<edge from-layer="1132" from-port="0" to-layer="1134" to-port="3" />
|
|
<edge from-layer="1133" from-port="0" to-layer="1134" to-port="4" />
|
|
<edge from-layer="1134" from-port="5" to-layer="1139" to-port="0" />
|
|
<edge from-layer="1135" from-port="0" to-layer="1136" to-port="0" />
|
|
<edge from-layer="1136" from-port="1" to-layer="1138" to-port="0" />
|
|
<edge from-layer="1137" from-port="0" to-layer="1138" to-port="1" />
|
|
<edge from-layer="1138" from-port="2" to-layer="1139" to-port="1" />
|
|
<edge from-layer="1139" from-port="2" to-layer="1141" to-port="0" />
|
|
<edge from-layer="1140" from-port="0" to-layer="1141" to-port="1" />
|
|
<edge from-layer="1141" from-port="2" to-layer="1142" to-port="0" />
|
|
<edge from-layer="1142" from-port="1" to-layer="1147" to-port="0" />
|
|
<edge from-layer="1143" from-port="0" to-layer="1147" to-port="1" />
|
|
<edge from-layer="1144" from-port="0" to-layer="1147" to-port="2" />
|
|
<edge from-layer="1145" from-port="0" to-layer="1147" to-port="3" />
|
|
<edge from-layer="1146" from-port="0" to-layer="1147" to-port="4" />
|
|
<edge from-layer="1147" from-port="5" to-layer="1148" to-port="1" />
|
|
<edge from-layer="1148" from-port="2" to-layer="1153" to-port="0" />
|
|
<edge from-layer="1149" from-port="0" to-layer="1153" to-port="1" />
|
|
<edge from-layer="1150" from-port="0" to-layer="1153" to-port="2" />
|
|
<edge from-layer="1151" from-port="0" to-layer="1153" to-port="3" />
|
|
<edge from-layer="1152" from-port="0" to-layer="1153" to-port="4" />
|
|
<edge from-layer="1153" from-port="5" to-layer="1167" to-port="0" />
|
|
<edge from-layer="1154" from-port="0" to-layer="1155" to-port="0" />
|
|
<edge from-layer="1155" from-port="1" to-layer="1157" to-port="0" />
|
|
<edge from-layer="1156" from-port="0" to-layer="1157" to-port="1" />
|
|
<edge from-layer="1157" from-port="2" to-layer="1158" to-port="1" />
|
|
<edge from-layer="1158" from-port="2" to-layer="1160" to-port="0" />
|
|
<edge from-layer="1159" from-port="0" to-layer="1160" to-port="1" />
|
|
<edge from-layer="1160" from-port="2" to-layer="1161" to-port="0" />
|
|
<edge from-layer="1161" from-port="1" to-layer="1166" to-port="0" />
|
|
<edge from-layer="1162" from-port="0" to-layer="1166" to-port="1" />
|
|
<edge from-layer="1163" from-port="0" to-layer="1166" to-port="2" />
|
|
<edge from-layer="1164" from-port="0" to-layer="1166" to-port="3" />
|
|
<edge from-layer="1165" from-port="0" to-layer="1166" to-port="4" />
|
|
<edge from-layer="1166" from-port="5" to-layer="1167" to-port="1" />
|
|
<edge from-layer="1167" from-port="2" to-layer="1172" to-port="0" />
|
|
<edge from-layer="1168" from-port="0" to-layer="1169" to-port="0" />
|
|
<edge from-layer="1169" from-port="1" to-layer="1171" to-port="0" />
|
|
<edge from-layer="1170" from-port="0" to-layer="1171" to-port="1" />
|
|
<edge from-layer="1171" from-port="2" to-layer="1172" to-port="1" />
|
|
<edge from-layer="1172" from-port="2" to-layer="1174" to-port="0" />
|
|
<edge from-layer="1173" from-port="0" to-layer="1174" to-port="1" />
|
|
<edge from-layer="1174" from-port="2" to-layer="1175" to-port="0" />
|
|
<edge from-layer="1175" from-port="1" to-layer="1180" to-port="0" />
|
|
<edge from-layer="1176" from-port="0" to-layer="1180" to-port="1" />
|
|
<edge from-layer="1177" from-port="0" to-layer="1180" to-port="2" />
|
|
<edge from-layer="1178" from-port="0" to-layer="1180" to-port="3" />
|
|
<edge from-layer="1179" from-port="0" to-layer="1180" to-port="4" />
|
|
<edge from-layer="1180" from-port="5" to-layer="1181" to-port="2" />
|
|
<edge from-layer="1181" from-port="3" to-layer="1186" to-port="0" />
|
|
<edge from-layer="1182" from-port="0" to-layer="1183" to-port="0" />
|
|
<edge from-layer="1183" from-port="1" to-layer="1185" to-port="0" />
|
|
<edge from-layer="1184" from-port="0" to-layer="1185" to-port="1" />
|
|
<edge from-layer="1185" from-port="2" to-layer="1186" to-port="1" />
|
|
<edge from-layer="1186" from-port="2" to-layer="1188" to-port="0" />
|
|
<edge from-layer="1187" from-port="0" to-layer="1188" to-port="1" />
|
|
<edge from-layer="1188" from-port="2" to-layer="1189" to-port="0" />
|
|
<edge from-layer="1189" from-port="1" to-layer="1194" to-port="0" />
|
|
<edge from-layer="1190" from-port="0" to-layer="1194" to-port="1" />
|
|
<edge from-layer="1191" from-port="0" to-layer="1194" to-port="2" />
|
|
<edge from-layer="1192" from-port="0" to-layer="1194" to-port="3" />
|
|
<edge from-layer="1193" from-port="0" to-layer="1194" to-port="4" />
|
|
<edge from-layer="1194" from-port="5" to-layer="1409" to-port="0" />
|
|
<edge from-layer="1194" from-port="5" to-layer="1232" to-port="0" />
|
|
<edge from-layer="1194" from-port="5" to-layer="1199" to-port="0" />
|
|
<edge from-layer="1195" from-port="0" to-layer="1196" to-port="0" />
|
|
<edge from-layer="1196" from-port="1" to-layer="1198" to-port="0" />
|
|
<edge from-layer="1197" from-port="0" to-layer="1198" to-port="1" />
|
|
<edge from-layer="1198" from-port="2" to-layer="1199" to-port="1" />
|
|
<edge from-layer="1199" from-port="2" to-layer="1201" to-port="0" />
|
|
<edge from-layer="1200" from-port="0" to-layer="1201" to-port="1" />
|
|
<edge from-layer="1201" from-port="2" to-layer="1202" to-port="0" />
|
|
<edge from-layer="1202" from-port="1" to-layer="1207" to-port="0" />
|
|
<edge from-layer="1203" from-port="0" to-layer="1207" to-port="1" />
|
|
<edge from-layer="1204" from-port="0" to-layer="1207" to-port="2" />
|
|
<edge from-layer="1205" from-port="0" to-layer="1207" to-port="3" />
|
|
<edge from-layer="1206" from-port="0" to-layer="1207" to-port="4" />
|
|
<edge from-layer="1207" from-port="5" to-layer="1212" to-port="0" />
|
|
<edge from-layer="1208" from-port="0" to-layer="1209" to-port="0" />
|
|
<edge from-layer="1209" from-port="1" to-layer="1211" to-port="0" />
|
|
<edge from-layer="1210" from-port="0" to-layer="1211" to-port="1" />
|
|
<edge from-layer="1211" from-port="2" to-layer="1212" to-port="1" />
|
|
<edge from-layer="1212" from-port="2" to-layer="1214" to-port="0" />
|
|
<edge from-layer="1213" from-port="0" to-layer="1214" to-port="1" />
|
|
<edge from-layer="1214" from-port="2" to-layer="1215" to-port="0" />
|
|
<edge from-layer="1215" from-port="1" to-layer="1220" to-port="0" />
|
|
<edge from-layer="1216" from-port="0" to-layer="1220" to-port="1" />
|
|
<edge from-layer="1217" from-port="0" to-layer="1220" to-port="2" />
|
|
<edge from-layer="1218" from-port="0" to-layer="1220" to-port="3" />
|
|
<edge from-layer="1219" from-port="0" to-layer="1220" to-port="4" />
|
|
<edge from-layer="1220" from-port="5" to-layer="1225" to-port="0" />
|
|
<edge from-layer="1221" from-port="0" to-layer="1222" to-port="0" />
|
|
<edge from-layer="1222" from-port="1" to-layer="1224" to-port="0" />
|
|
<edge from-layer="1223" from-port="0" to-layer="1224" to-port="1" />
|
|
<edge from-layer="1224" from-port="2" to-layer="1225" to-port="1" />
|
|
<edge from-layer="1225" from-port="2" to-layer="1227" to-port="0" />
|
|
<edge from-layer="1226" from-port="0" to-layer="1227" to-port="1" />
|
|
<edge from-layer="1227" from-port="2" to-layer="1287" to-port="0" />
|
|
<edge from-layer="1228" from-port="0" to-layer="1229" to-port="0" />
|
|
<edge from-layer="1229" from-port="1" to-layer="1231" to-port="0" />
|
|
<edge from-layer="1230" from-port="0" to-layer="1231" to-port="1" />
|
|
<edge from-layer="1231" from-port="2" to-layer="1232" to-port="1" />
|
|
<edge from-layer="1232" from-port="2" to-layer="1234" to-port="0" />
|
|
<edge from-layer="1233" from-port="0" to-layer="1234" to-port="1" />
|
|
<edge from-layer="1234" from-port="2" to-layer="1235" to-port="0" />
|
|
<edge from-layer="1235" from-port="1" to-layer="1240" to-port="0" />
|
|
<edge from-layer="1236" from-port="0" to-layer="1240" to-port="1" />
|
|
<edge from-layer="1237" from-port="0" to-layer="1240" to-port="2" />
|
|
<edge from-layer="1238" from-port="0" to-layer="1240" to-port="3" />
|
|
<edge from-layer="1239" from-port="0" to-layer="1240" to-port="4" />
|
|
<edge from-layer="1240" from-port="5" to-layer="1245" to-port="0" />
|
|
<edge from-layer="1241" from-port="0" to-layer="1242" to-port="0" />
|
|
<edge from-layer="1242" from-port="1" to-layer="1244" to-port="0" />
|
|
<edge from-layer="1243" from-port="0" to-layer="1244" to-port="1" />
|
|
<edge from-layer="1244" from-port="2" to-layer="1245" to-port="1" />
|
|
<edge from-layer="1245" from-port="2" to-layer="1247" to-port="0" />
|
|
<edge from-layer="1246" from-port="0" to-layer="1247" to-port="1" />
|
|
<edge from-layer="1247" from-port="2" to-layer="1248" to-port="0" />
|
|
<edge from-layer="1248" from-port="1" to-layer="1253" to-port="0" />
|
|
<edge from-layer="1249" from-port="0" to-layer="1253" to-port="1" />
|
|
<edge from-layer="1250" from-port="0" to-layer="1253" to-port="2" />
|
|
<edge from-layer="1251" from-port="0" to-layer="1253" to-port="3" />
|
|
<edge from-layer="1252" from-port="0" to-layer="1253" to-port="4" />
|
|
<edge from-layer="1253" from-port="5" to-layer="1258" to-port="0" />
|
|
<edge from-layer="1254" from-port="0" to-layer="1255" to-port="0" />
|
|
<edge from-layer="1255" from-port="1" to-layer="1257" to-port="0" />
|
|
<edge from-layer="1256" from-port="0" to-layer="1257" to-port="1" />
|
|
<edge from-layer="1257" from-port="2" to-layer="1258" to-port="1" />
|
|
<edge from-layer="1258" from-port="2" to-layer="1260" to-port="0" />
|
|
<edge from-layer="1259" from-port="0" to-layer="1260" to-port="1" />
|
|
<edge from-layer="1260" from-port="2" to-layer="1261" to-port="0" />
|
|
<edge from-layer="1261" from-port="1" to-layer="1266" to-port="0" />
|
|
<edge from-layer="1262" from-port="0" to-layer="1266" to-port="1" />
|
|
<edge from-layer="1263" from-port="0" to-layer="1266" to-port="2" />
|
|
<edge from-layer="1264" from-port="0" to-layer="1266" to-port="3" />
|
|
<edge from-layer="1265" from-port="0" to-layer="1266" to-port="4" />
|
|
<edge from-layer="1266" from-port="5" to-layer="1271" to-port="0" />
|
|
<edge from-layer="1267" from-port="0" to-layer="1268" to-port="0" />
|
|
<edge from-layer="1268" from-port="1" to-layer="1270" to-port="0" />
|
|
<edge from-layer="1269" from-port="0" to-layer="1270" to-port="1" />
|
|
<edge from-layer="1270" from-port="2" to-layer="1271" to-port="1" />
|
|
<edge from-layer="1271" from-port="2" to-layer="1273" to-port="0" />
|
|
<edge from-layer="1272" from-port="0" to-layer="1273" to-port="1" />
|
|
<edge from-layer="1273" from-port="2" to-layer="1274" to-port="0" />
|
|
<edge from-layer="1274" from-port="1" to-layer="1279" to-port="0" />
|
|
<edge from-layer="1275" from-port="0" to-layer="1279" to-port="1" />
|
|
<edge from-layer="1276" from-port="0" to-layer="1279" to-port="2" />
|
|
<edge from-layer="1277" from-port="0" to-layer="1279" to-port="3" />
|
|
<edge from-layer="1278" from-port="0" to-layer="1279" to-port="4" />
|
|
<edge from-layer="1279" from-port="5" to-layer="1284" to-port="0" />
|
|
<edge from-layer="1280" from-port="0" to-layer="1281" to-port="0" />
|
|
<edge from-layer="1281" from-port="1" to-layer="1283" to-port="0" />
|
|
<edge from-layer="1282" from-port="0" to-layer="1283" to-port="1" />
|
|
<edge from-layer="1283" from-port="2" to-layer="1284" to-port="1" />
|
|
<edge from-layer="1284" from-port="2" to-layer="1286" to-port="0" />
|
|
<edge from-layer="1285" from-port="0" to-layer="1286" to-port="1" />
|
|
<edge from-layer="1286" from-port="2" to-layer="1287" to-port="1" />
|
|
<edge from-layer="1287" from-port="2" to-layer="1289" to-port="0" />
|
|
<edge from-layer="1288" from-port="0" to-layer="1289" to-port="1" />
|
|
<edge from-layer="1289" from-port="2" to-layer="1290" to-port="2" />
|
|
<edge from-layer="1290" from-port="3" to-layer="1293" to-port="0" />
|
|
<edge from-layer="1291" from-port="0" to-layer="1293" to-port="1" />
|
|
<edge from-layer="1291" from-port="0" to-layer="1305" to-port="1" />
|
|
<edge from-layer="1291" from-port="0" to-layer="1314" to-port="1" />
|
|
<edge from-layer="1292" from-port="0" to-layer="1293" to-port="2" />
|
|
<edge from-layer="1293" from-port="3" to-layer="1295" to-port="0" />
|
|
<edge from-layer="1293" from-port="4" to-layer="1324" to-port="0" />
|
|
<edge from-layer="1294" from-port="0" to-layer="1295" to-port="1" />
|
|
<edge from-layer="1295" from-port="2" to-layer="1297" to-port="0" />
|
|
<edge from-layer="1296" from-port="0" to-layer="1297" to-port="1" />
|
|
<edge from-layer="1297" from-port="2" to-layer="1298" to-port="0" />
|
|
<edge from-layer="1298" from-port="1" to-layer="1300" to-port="0" />
|
|
<edge from-layer="1299" from-port="0" to-layer="1300" to-port="1" />
|
|
<edge from-layer="1300" from-port="2" to-layer="1302" to-port="0" />
|
|
<edge from-layer="1301" from-port="0" to-layer="1302" to-port="1" />
|
|
<edge from-layer="1302" from-port="2" to-layer="1303" to-port="0" />
|
|
<edge from-layer="1302" from-port="2" to-layer="1314" to-port="0" />
|
|
<edge from-layer="1303" from-port="1" to-layer="1305" to-port="0" />
|
|
<edge from-layer="1304" from-port="0" to-layer="1309" to-port="1" />
|
|
<edge from-layer="1304" from-port="0" to-layer="1305" to-port="2" />
|
|
<edge from-layer="1305" from-port="3" to-layer="1308" to-port="0" />
|
|
<edge from-layer="1305" from-port="3" to-layer="1307" to-port="0" />
|
|
<edge from-layer="1306" from-port="0" to-layer="1307" to-port="1" />
|
|
<edge from-layer="1306" from-port="0" to-layer="1308" to-port="1" />
|
|
<edge from-layer="1307" from-port="2" to-layer="1311" to-port="0" />
|
|
<edge from-layer="1308" from-port="2" to-layer="1309" to-port="0" />
|
|
<edge from-layer="1309" from-port="2" to-layer="1310" to-port="0" />
|
|
<edge from-layer="1310" from-port="1" to-layer="1311" to-port="1" />
|
|
<edge from-layer="1311" from-port="2" to-layer="1313" to-port="0" />
|
|
<edge from-layer="1312" from-port="0" to-layer="1313" to-port="1" />
|
|
<edge from-layer="1313" from-port="2" to-layer="1314" to-port="2" />
|
|
<edge from-layer="1314" from-port="4" to-layer="1316" to-port="1" />
|
|
<edge from-layer="1314" from-port="3" to-layer="1315" to-port="1" />
|
|
<edge from-layer="1315" from-port="2" to-layer="1317" to-port="0" />
|
|
<edge from-layer="1315" from-port="2" to-layer="1320" to-port="1" />
|
|
<edge from-layer="1316" from-port="2" to-layer="1320" to-port="0" />
|
|
<edge from-layer="1316" from-port="2" to-layer="1317" to-port="1" />
|
|
<edge from-layer="1317" from-port="2" to-layer="1319" to-port="0" />
|
|
<edge from-layer="1318" from-port="0" to-layer="1319" to-port="1" />
|
|
<edge from-layer="1319" from-port="2" to-layer="1321" to-port="0" />
|
|
<edge from-layer="1320" from-port="2" to-layer="1321" to-port="1" />
|
|
<edge from-layer="1321" from-port="2" to-layer="1323" to-port="0" />
|
|
<edge from-layer="1322" from-port="0" to-layer="1323" to-port="1" />
|
|
<edge from-layer="1323" from-port="2" to-layer="1464" to-port="0" />
|
|
<edge from-layer="1324" from-port="1" to-layer="1464" to-port="1" />
|
|
<edge from-layer="1325" from-port="0" to-layer="1326" to-port="0" />
|
|
<edge from-layer="1326" from-port="1" to-layer="1328" to-port="0" />
|
|
<edge from-layer="1327" from-port="0" to-layer="1328" to-port="1" />
|
|
<edge from-layer="1328" from-port="2" to-layer="1329" to-port="1" />
|
|
<edge from-layer="1329" from-port="2" to-layer="1331" to-port="0" />
|
|
<edge from-layer="1330" from-port="0" to-layer="1331" to-port="1" />
|
|
<edge from-layer="1331" from-port="2" to-layer="1332" to-port="0" />
|
|
<edge from-layer="1332" from-port="1" to-layer="1337" to-port="0" />
|
|
<edge from-layer="1333" from-port="0" to-layer="1337" to-port="1" />
|
|
<edge from-layer="1334" from-port="0" to-layer="1337" to-port="2" />
|
|
<edge from-layer="1335" from-port="0" to-layer="1337" to-port="3" />
|
|
<edge from-layer="1336" from-port="0" to-layer="1337" to-port="4" />
|
|
<edge from-layer="1337" from-port="5" to-layer="1342" to-port="0" />
|
|
<edge from-layer="1338" from-port="0" to-layer="1339" to-port="0" />
|
|
<edge from-layer="1339" from-port="1" to-layer="1341" to-port="0" />
|
|
<edge from-layer="1340" from-port="0" to-layer="1341" to-port="1" />
|
|
<edge from-layer="1341" from-port="2" to-layer="1342" to-port="1" />
|
|
<edge from-layer="1342" from-port="2" to-layer="1344" to-port="0" />
|
|
<edge from-layer="1343" from-port="0" to-layer="1344" to-port="1" />
|
|
<edge from-layer="1344" from-port="2" to-layer="1345" to-port="0" />
|
|
<edge from-layer="1345" from-port="1" to-layer="1350" to-port="0" />
|
|
<edge from-layer="1346" from-port="0" to-layer="1350" to-port="1" />
|
|
<edge from-layer="1347" from-port="0" to-layer="1350" to-port="2" />
|
|
<edge from-layer="1348" from-port="0" to-layer="1350" to-port="3" />
|
|
<edge from-layer="1349" from-port="0" to-layer="1350" to-port="4" />
|
|
<edge from-layer="1350" from-port="5" to-layer="1355" to-port="0" />
|
|
<edge from-layer="1351" from-port="0" to-layer="1352" to-port="0" />
|
|
<edge from-layer="1352" from-port="1" to-layer="1354" to-port="0" />
|
|
<edge from-layer="1353" from-port="0" to-layer="1354" to-port="1" />
|
|
<edge from-layer="1354" from-port="2" to-layer="1355" to-port="1" />
|
|
<edge from-layer="1355" from-port="2" to-layer="1357" to-port="0" />
|
|
<edge from-layer="1356" from-port="0" to-layer="1357" to-port="1" />
|
|
<edge from-layer="1357" from-port="2" to-layer="1362" to-port="0" />
|
|
<edge from-layer="1358" from-port="0" to-layer="1362" to-port="1" />
|
|
<edge from-layer="1359" from-port="0" to-layer="1362" to-port="2" />
|
|
<edge from-layer="1360" from-port="0" to-layer="1362" to-port="3" />
|
|
<edge from-layer="1361" from-port="0" to-layer="1362" to-port="4" />
|
|
<edge from-layer="1362" from-port="5" to-layer="1364" to-port="0" />
|
|
<edge from-layer="1363" from-port="0" to-layer="1364" to-port="1" />
|
|
<edge from-layer="1364" from-port="2" to-layer="1445" to-port="0" />
|
|
<edge from-layer="1365" from-port="0" to-layer="1366" to-port="0" />
|
|
<edge from-layer="1366" from-port="1" to-layer="1368" to-port="0" />
|
|
<edge from-layer="1367" from-port="0" to-layer="1368" to-port="1" />
|
|
<edge from-layer="1368" from-port="2" to-layer="1369" to-port="1" />
|
|
<edge from-layer="1369" from-port="2" to-layer="1371" to-port="0" />
|
|
<edge from-layer="1370" from-port="0" to-layer="1371" to-port="1" />
|
|
<edge from-layer="1371" from-port="2" to-layer="1372" to-port="0" />
|
|
<edge from-layer="1372" from-port="1" to-layer="1377" to-port="0" />
|
|
<edge from-layer="1373" from-port="0" to-layer="1377" to-port="1" />
|
|
<edge from-layer="1374" from-port="0" to-layer="1377" to-port="2" />
|
|
<edge from-layer="1375" from-port="0" to-layer="1377" to-port="3" />
|
|
<edge from-layer="1376" from-port="0" to-layer="1377" to-port="4" />
|
|
<edge from-layer="1377" from-port="5" to-layer="1382" to-port="0" />
|
|
<edge from-layer="1378" from-port="0" to-layer="1379" to-port="0" />
|
|
<edge from-layer="1379" from-port="1" to-layer="1381" to-port="0" />
|
|
<edge from-layer="1380" from-port="0" to-layer="1381" to-port="1" />
|
|
<edge from-layer="1381" from-port="2" to-layer="1382" to-port="1" />
|
|
<edge from-layer="1382" from-port="2" to-layer="1384" to-port="0" />
|
|
<edge from-layer="1383" from-port="0" to-layer="1384" to-port="1" />
|
|
<edge from-layer="1384" from-port="2" to-layer="1385" to-port="0" />
|
|
<edge from-layer="1385" from-port="1" to-layer="1390" to-port="0" />
|
|
<edge from-layer="1386" from-port="0" to-layer="1390" to-port="1" />
|
|
<edge from-layer="1387" from-port="0" to-layer="1390" to-port="2" />
|
|
<edge from-layer="1388" from-port="0" to-layer="1390" to-port="3" />
|
|
<edge from-layer="1389" from-port="0" to-layer="1390" to-port="4" />
|
|
<edge from-layer="1390" from-port="5" to-layer="1395" to-port="0" />
|
|
<edge from-layer="1391" from-port="0" to-layer="1392" to-port="0" />
|
|
<edge from-layer="1392" from-port="1" to-layer="1394" to-port="0" />
|
|
<edge from-layer="1393" from-port="0" to-layer="1394" to-port="1" />
|
|
<edge from-layer="1394" from-port="2" to-layer="1395" to-port="1" />
|
|
<edge from-layer="1395" from-port="2" to-layer="1397" to-port="0" />
|
|
<edge from-layer="1396" from-port="0" to-layer="1397" to-port="1" />
|
|
<edge from-layer="1397" from-port="2" to-layer="1402" to-port="0" />
|
|
<edge from-layer="1398" from-port="0" to-layer="1402" to-port="1" />
|
|
<edge from-layer="1399" from-port="0" to-layer="1402" to-port="2" />
|
|
<edge from-layer="1400" from-port="0" to-layer="1402" to-port="3" />
|
|
<edge from-layer="1401" from-port="0" to-layer="1402" to-port="4" />
|
|
<edge from-layer="1402" from-port="5" to-layer="1404" to-port="0" />
|
|
<edge from-layer="1403" from-port="0" to-layer="1404" to-port="1" />
|
|
<edge from-layer="1404" from-port="2" to-layer="1445" to-port="1" />
|
|
<edge from-layer="1405" from-port="0" to-layer="1406" to-port="0" />
|
|
<edge from-layer="1406" from-port="1" to-layer="1408" to-port="0" />
|
|
<edge from-layer="1407" from-port="0" to-layer="1408" to-port="1" />
|
|
<edge from-layer="1408" from-port="2" to-layer="1409" to-port="1" />
|
|
<edge from-layer="1409" from-port="2" to-layer="1411" to-port="0" />
|
|
<edge from-layer="1410" from-port="0" to-layer="1411" to-port="1" />
|
|
<edge from-layer="1411" from-port="2" to-layer="1412" to-port="0" />
|
|
<edge from-layer="1412" from-port="1" to-layer="1417" to-port="0" />
|
|
<edge from-layer="1413" from-port="0" to-layer="1417" to-port="1" />
|
|
<edge from-layer="1414" from-port="0" to-layer="1417" to-port="2" />
|
|
<edge from-layer="1415" from-port="0" to-layer="1417" to-port="3" />
|
|
<edge from-layer="1416" from-port="0" to-layer="1417" to-port="4" />
|
|
<edge from-layer="1417" from-port="5" to-layer="1422" to-port="0" />
|
|
<edge from-layer="1418" from-port="0" to-layer="1419" to-port="0" />
|
|
<edge from-layer="1419" from-port="1" to-layer="1421" to-port="0" />
|
|
<edge from-layer="1420" from-port="0" to-layer="1421" to-port="1" />
|
|
<edge from-layer="1421" from-port="2" to-layer="1422" to-port="1" />
|
|
<edge from-layer="1422" from-port="2" to-layer="1424" to-port="0" />
|
|
<edge from-layer="1423" from-port="0" to-layer="1424" to-port="1" />
|
|
<edge from-layer="1424" from-port="2" to-layer="1425" to-port="0" />
|
|
<edge from-layer="1425" from-port="1" to-layer="1430" to-port="0" />
|
|
<edge from-layer="1426" from-port="0" to-layer="1430" to-port="1" />
|
|
<edge from-layer="1427" from-port="0" to-layer="1430" to-port="2" />
|
|
<edge from-layer="1428" from-port="0" to-layer="1430" to-port="3" />
|
|
<edge from-layer="1429" from-port="0" to-layer="1430" to-port="4" />
|
|
<edge from-layer="1430" from-port="5" to-layer="1435" to-port="0" />
|
|
<edge from-layer="1431" from-port="0" to-layer="1432" to-port="0" />
|
|
<edge from-layer="1432" from-port="1" to-layer="1434" to-port="0" />
|
|
<edge from-layer="1433" from-port="0" to-layer="1434" to-port="1" />
|
|
<edge from-layer="1434" from-port="2" to-layer="1435" to-port="1" />
|
|
<edge from-layer="1435" from-port="2" to-layer="1437" to-port="0" />
|
|
<edge from-layer="1436" from-port="0" to-layer="1437" to-port="1" />
|
|
<edge from-layer="1437" from-port="2" to-layer="1442" to-port="0" />
|
|
<edge from-layer="1438" from-port="0" to-layer="1442" to-port="1" />
|
|
<edge from-layer="1439" from-port="0" to-layer="1442" to-port="2" />
|
|
<edge from-layer="1440" from-port="0" to-layer="1442" to-port="3" />
|
|
<edge from-layer="1441" from-port="0" to-layer="1442" to-port="4" />
|
|
<edge from-layer="1442" from-port="5" to-layer="1444" to-port="0" />
|
|
<edge from-layer="1443" from-port="0" to-layer="1444" to-port="1" />
|
|
<edge from-layer="1444" from-port="2" to-layer="1445" to-port="2" />
|
|
<edge from-layer="1445" from-port="3" to-layer="1447" to-port="0" />
|
|
<edge from-layer="1446" from-port="0" to-layer="1447" to-port="1" />
|
|
<edge from-layer="1447" from-port="2" to-layer="1451" to-port="0" />
|
|
<edge from-layer="1448" from-port="0" to-layer="1451" to-port="1" />
|
|
<edge from-layer="1449" from-port="0" to-layer="1451" to-port="2" />
|
|
<edge from-layer="1450" from-port="0" to-layer="1451" to-port="3" />
|
|
<edge from-layer="1451" from-port="4" to-layer="1453" to-port="0" />
|
|
<edge from-layer="1452" from-port="0" to-layer="1453" to-port="1" />
|
|
<edge from-layer="1453" from-port="2" to-layer="1455" to-port="0" />
|
|
<edge from-layer="1454" from-port="0" to-layer="1455" to-port="1" />
|
|
<edge from-layer="1455" from-port="2" to-layer="1463" to-port="0" />
|
|
<edge from-layer="1456" from-port="1" to-layer="1459" to-port="0" />
|
|
<edge from-layer="1457" from-port="0" to-layer="1459" to-port="1" />
|
|
<edge from-layer="1458" from-port="0" to-layer="1459" to-port="2" />
|
|
<edge from-layer="1459" from-port="3" to-layer="1462" to-port="0" />
|
|
<edge from-layer="1460" from-port="0" to-layer="1462" to-port="1" />
|
|
<edge from-layer="1461" from-port="0" to-layer="1462" to-port="2" />
|
|
<edge from-layer="1462" from-port="3" to-layer="1463" to-port="1" />
|
|
<edge from-layer="1463" from-port="2" to-layer="1464" to-port="2" />
|
|
<edge from-layer="1464" from-port="3" to-layer="1465" to-port="0" />
|
|
</edges>
|
|
<rt_info>
|
|
<Runtime_version value="2024.2.0-15519-5c0f38f83f6-releases/2024/2" />
|
|
<conversion_parameters>
|
|
<framework value="pytorch" />
|
|
<is_python_object value="True" />
|
|
</conversion_parameters>
|
|
<model_info>
|
|
<iou_threshold value="0.7" />
|
|
<labels value="Bsentry Rsentry Esentry Bone Rone Eone Btwo Rtwo Etwo Bthree Rthree Ethree Bfour Rfour Efour Bfive Rfive Efive Boutpost Routpost Eoutpost Bbase Rbase Ebase Pbase Bbasesmall Rbasesmall Ebasesmall Pbasesmall Bbalancethree Rbalancethree Ebalancethree Bbalancefour Rbalancefour Ebalancefour Bbalancefive Rbalancefive Ebalancefive" />
|
|
<model_type value="YOLO" />
|
|
<pad_value value="114" />
|
|
<resize_type value="fit_to_window_letterbox" />
|
|
<reverse_input_channels value="YES" />
|
|
<scale_values value="255" />
|
|
</model_info>
|
|
<nncf>
|
|
<friendly_names_were_updated value="True" />
|
|
<quantization>
|
|
<advanced_parameters value="{}" />
|
|
<fast_bias_correction value="True" />
|
|
<ignored_scope>
|
|
<patterns value="['.*model.23/.*/Add', '.*model.23/.*/Sub*', '.*model.23/.*/Mul*', '.*model.23/.*/Div*', '.*model.23\\.dfl.*']" />
|
|
<types value="['Sigmoid']" />
|
|
</ignored_scope>
|
|
<model_type value="None" />
|
|
<preset value="mixed" />
|
|
<subset_size value="300" />
|
|
<target_device value="ANY" />
|
|
</quantization>
|
|
</nncf>
|
|
</rt_info>
|
|
</net>
|