bloom-tiny-random / openvino_model.xml
yujiepan's picture
Upload folder using huggingface_hub
9a17f05
raw
history blame contribute delete
No virus
236 kB
<?xml version="1.0"?>
<net name="main_graph" version="11">
<layers>
<layer id="5" name="input_ids" type="Parameter" version="opset1">
<data shape="?,?" element_type="i64" />
<rt_info>
<attribute name="fused_names" version="0" value="input_ids" />
</rt_info>
<output>
<port id="0" precision="I64" names="input_ids">
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="4" name="past_key_values.0.key" type="Parameter" version="opset1">
<data shape="?,4,?" element_type="f32" />
<rt_info>
<attribute name="fused_names" version="0" value="past_key_values.0.key" />
<attribute name="old_api_map_element_type" version="0" value="f16" />
</rt_info>
<output>
<port id="0" precision="FP32" names="past_key_values.0.key">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="3" name="past_key_values.0.value" type="Parameter" version="opset1">
<data shape="?,?,4" element_type="f32" />
<rt_info>
<attribute name="fused_names" version="0" value="past_key_values.0.value" />
<attribute name="old_api_map_element_type" version="0" value="f16" />
</rt_info>
<output>
<port id="0" precision="FP32" names="past_key_values.0.value">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="2" name="past_key_values.1.key" type="Parameter" version="opset1">
<data shape="?,4,?" element_type="f32" />
<rt_info>
<attribute name="fused_names" version="0" value="past_key_values.1.key" />
<attribute name="old_api_map_element_type" version="0" value="f16" />
</rt_info>
<output>
<port id="0" precision="FP32" names="past_key_values.1.key">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="1" name="past_key_values.1.value" type="Parameter" version="opset1">
<data shape="?,?,4" element_type="f32" />
<rt_info>
<attribute name="fused_names" version="0" value="past_key_values.1.value" />
<attribute name="old_api_map_element_type" version="0" value="f16" />
</rt_info>
<output>
<port id="0" precision="FP32" names="past_key_values.1.value">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="0" name="attention_mask" type="Parameter" version="opset1">
<data shape="?,?" element_type="i64" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Cast, attention_mask" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Cast_output_0,attention_mask">
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="6" name="transformer.word_embeddings.weight_compressed" type="Const" version="opset1">
<data element_type="f16" shape="250880, 8" offset="0" size="4014080" />
<rt_info>
<attribute name="fused_names" version="0" value="transformer.word_embeddings.weight" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>250880</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="7" name="transformer.word_embeddings.weight" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="transformer.word_embeddings.weight" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>250880</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="1" precision="FP32" names="transformer.word_embeddings.weight">
<dim>250880</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="8" name="Constant_29852" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_29852" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="9" name="/transformer/word_embeddings/Gather" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/word_embeddings/Gather" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>250880</dim>
<dim>8</dim>
</port>
<port id="1" precision="I64">
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="FP32" names="/transformer/word_embeddings/Gather_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="10" name="Constant_29854" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014088" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_29854" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="11" name="/transformer/word_embeddings_layernorm/Div" type="MVN" version="opset6">
<data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/word_embeddings_layernorm/Add, /transformer/word_embeddings_layernorm/Add_1, /transformer/word_embeddings_layernorm/Div, /transformer/word_embeddings_layernorm/Mul, /transformer/word_embeddings_layernorm/Pow, /transformer/word_embeddings_layernorm/ReduceMean, /transformer/word_embeddings_layernorm/ReduceMean_1, /transformer/word_embeddings_layernorm/Sqrt, /transformer/word_embeddings_layernorm/Sub" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/word_embeddings_layernorm/Add_1_output_0,/transformer/word_embeddings_layernorm/Div_output_0,/transformer/word_embeddings_layernorm/Mul_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="12" name="Constant_30407" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30407" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="13" name="/transformer/Unsqueeze_16" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Unsqueeze_16" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Unsqueeze_16_output_0">
<dim>-1</dim>
<dim>1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="14" name="Constant_30409" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014104" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30409" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="15" name="/transformer/Unsqueeze_17" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Unsqueeze_17" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
<dim>1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Unsqueeze_17_output_0">
<dim>-1</dim>
<dim>1</dim>
<dim>1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="16" name="/transformer/Cast_6" type="Convert" version="opset1">
<data destination_type="boolean" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Cast_6" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
<dim>1</dim>
<dim>1</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="1" precision="BOOL" names="/transformer/Cast_6_output_0">
<dim>-1</dim>
<dim>1</dim>
<dim>1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="17" name="/transformer/Not" type="LogicalNot" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Not" />
</rt_info>
<input>
<port id="0" precision="BOOL">
<dim>-1</dim>
<dim>1</dim>
<dim>1</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="1" precision="BOOL" names="/transformer/Not_output_0">
<dim>-1</dim>
<dim>1</dim>
<dim>1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="18" name="/transformer/Shape_3" type="ShapeOf" version="opset3">
<data output_type="i64" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Shape_3, /transformer/Shape_4" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="1" precision="I64" names="/transformer/Shape_3_output_0,/transformer/Shape_4_output_0">
<dim>2</dim>
</port>
</output>
</layer>
<layer id="19" name="/transformer/Constant_6" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_6" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_6_output_0" />
</output>
</layer>
<layer id="20" name="Constant_29875" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_29875" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="21" name="/transformer/Gather_3" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_6, /transformer/Gather_3, Constant_29875" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>2</dim>
</port>
<port id="1" precision="I64" />
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/Gather_3_output_0" />
</output>
</layer>
<layer id="22" name="Constant_30413" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30413" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="23" name="/transformer/Unsqueeze_18" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Unsqueeze_18, Constant_30413" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Unsqueeze_18_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="24" name="/transformer/Constant_36" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_36" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_36_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="25" name="/transformer/Shape" type="ShapeOf" version="opset3">
<data output_type="i64" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Shape, /transformer/Shape_1" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="1" precision="I64" names="/transformer/Shape_1_output_0,/transformer/Shape_output_0">
<dim>2</dim>
</port>
</output>
</layer>
<layer id="26" name="/transformer/Constant_3" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_3" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_3_output_0" />
</output>
</layer>
<layer id="27" name="Constant_29850" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_29850" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="28" name="/transformer/Gather_1" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Cast_3, /transformer/Constant_3, /transformer/Gather_1, Constant_29850" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>2</dim>
</port>
<port id="1" precision="I64" />
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/Cast_3_output_0,/transformer/Gather_1_output_0" />
</output>
</layer>
<layer id="29" name="Constant_30416" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30416" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="30" name="/transformer/Unsqueeze_19" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Unsqueeze_14, /transformer/Unsqueeze_19, /transformer/Unsqueeze_3, Constant_29968, Constant_30281, Constant_30416" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Unsqueeze_14_output_0,/transformer/Unsqueeze_19_output_0,/transformer/Unsqueeze_3_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="31" name="Constant_52953" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_7, /transformer/Gather_4, /transformer/Unsqueeze_2, /transformer/Unsqueeze_20, Constant_29879, Constant_29896, Constant_30418" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="32" name="Constant_29879" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_29879" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="33" name="/transformer/Gather_4" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_7, /transformer/Gather_4, /transformer/Unsqueeze_2, /transformer/Unsqueeze_20, Constant_29879, Constant_29896, Constant_30418" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>2</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/Gather_4_output_0,/transformer/Unsqueeze_20_output_0,/transformer/Unsqueeze_2_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="34" name="/transformer/Concat_5" type="Concat" version="opset1">
<data axis="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Concat_5, /transformer/Constant_36" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64">
<dim>1</dim>
</port>
<port id="3" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="4" precision="I64" names="/transformer/Concat_5_output_0">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="35" name="Constant_34969" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014088" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Reshape_4" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="36" name="/transformer/Reshape_4" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Reshape_4" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Reshape_4_output_0">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="37" name="/transformer/Mul_6" type="Const" version="opset1">
<data element_type="i64" shape="4" offset="4014112" size="32" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/ConstantOfShape_4, /transformer/Constant_38, /transformer/Mul_6, /transformer/Shape_13, Constant_30427" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Mul_6_output_0">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="38" name="/transformer/Equal_3" type="Equal" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/ConstantOfShape_4, /transformer/Constant_38, /transformer/Equal_3, /transformer/Mul_6, /transformer/Shape_13, Constant_30427" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="BOOL" names="/transformer/Equal_3_output_0">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="39" name="/transformer/ConstantOfShape_4" type="Const" version="opset1">
<data element_type="i64" shape="4" offset="4014144" size="32" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/ConstantOfShape_4, /transformer/Shape_13, Constant_30427" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/ConstantOfShape_4_output_0">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="40" name="/transformer/Where_3" type="Select" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/ConstantOfShape_4, /transformer/Shape_13, /transformer/Where_3, Constant_30427" />
</rt_info>
<input>
<port id="0" precision="BOOL">
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>4</dim>
</port>
<port id="2" precision="I64">
<dim>4</dim>
</port>
</input>
<output>
<port id="3" precision="I64" names="/transformer/Where_3_output_0">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="41" name="/transformer/Expand_4" type="Broadcast" version="opset3">
<data mode="bidirectional" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Expand_4" />
</rt_info>
<input>
<port id="0" precision="BOOL">
<dim>-1</dim>
<dim>1</dim>
<dim>1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="BOOL" names="/transformer/Expand_4_output_0">
<dim>-1</dim>
<dim>1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="42" name="Constant_29974" type="Const" version="opset1">
<data element_type="boolean" shape="" offset="4014176" size="1" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_29974" />
</rt_info>
<output>
<port id="0" precision="BOOL" />
</output>
</layer>
<layer id="43" name="/transformer/Shape_2" type="ShapeOf" version="opset3">
<data output_type="i64" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Shape_2" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="1" precision="I64" names="/transformer/Shape_2_output_0">
<dim>3</dim>
</port>
</output>
</layer>
<layer id="44" name="/transformer/Constant_5" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014104" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_5" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_5_output_0" />
</output>
</layer>
<layer id="45" name="Constant_29870" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_29870" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="46" name="/transformer/Gather_2" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_5, /transformer/Gather_2, Constant_29870" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>3</dim>
</port>
<port id="1" precision="I64" />
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/Gather_2_output_0" />
</output>
</layer>
<layer id="47" name="/transformer/Add" type="Add" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Add" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64" />
</input>
<output>
<port id="2" precision="I64" names="/transformer/Add_output_0" />
</output>
</layer>
<layer id="48" name="Constant_30283" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30283" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="49" name="/transformer/Unsqueeze_15" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Unsqueeze_15, /transformer/Unsqueeze_4, Constant_29970, Constant_30283" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Unsqueeze_15_output_0,/transformer/Unsqueeze_4_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="50" name="/transformer/Concat_1" type="Concat" version="opset1">
<data axis="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Concat_1" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Concat_1_output_0">
<dim>2</dim>
</port>
</output>
</layer>
<layer id="51" name="/transformer/ConstantOfShape" type="Broadcast" version="opset3">
<data mode="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/ConstantOfShape" />
</rt_info>
<input>
<port id="0" precision="BOOL" />
<port id="1" precision="I64">
<dim>2</dim>
</port>
</input>
<output>
<port id="2" precision="BOOL" names="/transformer/ConstantOfShape_output_0">
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="52" name="/transformer/Constant_19" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_19" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_19_output_0" />
</output>
</layer>
<layer id="53" name="/transformer/Shape_7" type="ShapeOf" version="opset3">
<data output_type="i64" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Shape_6, /transformer/Shape_7" />
</rt_info>
<input>
<port id="0" precision="BOOL">
<dim>-1</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="1" precision="I64" names="/transformer/Shape_6_output_0,/transformer/Shape_7_output_0">
<dim>2</dim>
</port>
</output>
</layer>
<layer id="54" name="/transformer/Constant_18" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_18" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_18_output_0" />
</output>
</layer>
<layer id="55" name="Constant_30084" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30084" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="56" name="/transformer/Gather_5" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Cast_4, /transformer/Constant_18, /transformer/Gather_5, Constant_30084" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>2</dim>
</port>
<port id="1" precision="I64" />
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/Cast_4_output_0,/transformer/Gather_5_output_0" />
</output>
</layer>
<layer id="57" name="/transformer/Constant_20" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_20" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_20_output_0" />
</output>
</layer>
<layer id="58" name="/transformer/Range_1" type="Range" version="opset4">
<data output_type="i64" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Range_1" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64" />
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/Range_1_output_0">
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="59" name="/transformer/Constant_27" type="Const" version="opset1">
<data element_type="i64" shape="2" offset="4014177" size="16" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_27" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_27_output_0">
<dim>2</dim>
</port>
</output>
</layer>
<layer id="60" name="/transformer/Reshape_1" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Reshape_1" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>2</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Reshape_1_output_0">
<dim>-1</dim>
<dim>1</dim>
</port>
</output>
</layer>
<layer id="61" name="/transformer/Constant_22" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_22" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_22_output_0" />
</output>
</layer>
<layer id="62" name="/transformer/Constant_21" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_21" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_21_output_0" />
</output>
</layer>
<layer id="63" name="Constant_30105" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30105" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="64" name="/transformer/Gather_6" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Cast_5, /transformer/Constant_21, /transformer/Gather_6, Constant_30105" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>2</dim>
</port>
<port id="1" precision="I64" />
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/Cast_5_output_0,/transformer/Gather_6_output_0" />
</output>
</layer>
<layer id="65" name="/transformer/Constant_23" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_23" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_23_output_0" />
</output>
</layer>
<layer id="66" name="/transformer/Range_2" type="Range" version="opset4">
<data output_type="i64" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Range_2" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64" />
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/Range_2_output_0">
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="67" name="Constant_30124" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30124" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="68" name="/transformer/Unsqueeze_8" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Slice_1, Constant_30124" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Unsqueeze_7_output_0,/transformer/Unsqueeze_8_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="69" name="/transformer/Constant_25" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014193" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Slice_1" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_25_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="70" name="/transformer/Constant_26" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Slice_1" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_26_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="71" name="/transformer/Slice_1" type="StridedSlice" version="opset1">
<data begin_mask="0" end_mask="0" new_axis_mask="" shrink_axis_mask="" ellipsis_mask="" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Slice_1" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64">
<dim>1</dim>
</port>
<port id="3" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="4" precision="I64" names="/transformer/Slice_1_output_0">
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="72" name="/transformer/Add_1" type="Add" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Add_1" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>-1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Add_1_output_0">
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="73" name="/transformer/Shape_8" type="ShapeOf" version="opset3">
<data output_type="i64" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Concat_3, /transformer/Shape_8" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="1" precision="I64" names="/transformer/Concat_3_output_0,/transformer/Shape_8_output_0">
<dim>2</dim>
</port>
</output>
</layer>
<layer id="74" name="/transformer/Mul_3" type="Const" version="opset1">
<data element_type="i64" shape="2" offset="4014201" size="16" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/ConstantOfShape_1, /transformer/Constant_28, /transformer/Mul_3, /transformer/Shape_9, Constant_30141" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Mul_3_output_0">
<dim>2</dim>
</port>
</output>
</layer>
<layer id="75" name="/transformer/Equal" type="Equal" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/ConstantOfShape_1, /transformer/Constant_28, /transformer/Equal, /transformer/Mul_3, /transformer/Shape_9, Constant_30141" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>2</dim>
</port>
<port id="1" precision="I64">
<dim>2</dim>
</port>
</input>
<output>
<port id="2" precision="BOOL" names="/transformer/Equal_output_0">
<dim>2</dim>
</port>
</output>
</layer>
<layer id="76" name="/transformer/ConstantOfShape_1" type="Const" version="opset1">
<data element_type="i64" shape="2" offset="4014217" size="16" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/ConstantOfShape_1, /transformer/Shape_9, Constant_30141" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/ConstantOfShape_1_output_0">
<dim>2</dim>
</port>
</output>
</layer>
<layer id="77" name="/transformer/Where" type="Select" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/ConstantOfShape_1, /transformer/Shape_9, /transformer/Where, Constant_30141" />
</rt_info>
<input>
<port id="0" precision="BOOL">
<dim>2</dim>
</port>
<port id="1" precision="I64">
<dim>2</dim>
</port>
<port id="2" precision="I64">
<dim>2</dim>
</port>
</input>
<output>
<port id="3" precision="I64" names="/transformer/Where_output_0">
<dim>2</dim>
</port>
</output>
</layer>
<layer id="78" name="/transformer/Expand_1" type="Broadcast" version="opset3">
<data mode="bidirectional" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Expand_1" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>2</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Expand_1_output_0">
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="79" name="Constant_30192" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014088" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30192" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="80" name="/transformer/Unsqueeze_9" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Unsqueeze_9" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Unsqueeze_9_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>1</dim>
</port>
</output>
</layer>
<layer id="81" name="/transformer/Mul_4" type="Const" version="opset1">
<data element_type="i64" shape="2" offset="4014201" size="16" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/ConstantOfShape_2, /transformer/Constant_29, /transformer/Mul_4, /transformer/Shape_10, Constant_30196" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Mul_4_output_0">
<dim>2</dim>
</port>
</output>
</layer>
<layer id="82" name="/transformer/Equal_1" type="Equal" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/ConstantOfShape_2, /transformer/Constant_29, /transformer/Equal_1, /transformer/Mul_4, /transformer/Shape_10, Constant_30196" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>2</dim>
</port>
<port id="1" precision="I64">
<dim>2</dim>
</port>
</input>
<output>
<port id="2" precision="BOOL" names="/transformer/Equal_1_output_0">
<dim>2</dim>
</port>
</output>
</layer>
<layer id="83" name="/transformer/ConstantOfShape_2" type="Const" version="opset1">
<data element_type="i64" shape="2" offset="4014217" size="16" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/ConstantOfShape_2, /transformer/Shape_10, Constant_30196" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/ConstantOfShape_2_output_0">
<dim>2</dim>
</port>
</output>
</layer>
<layer id="84" name="/transformer/Where_1" type="Select" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/ConstantOfShape_2, /transformer/Shape_10, /transformer/Where_1, Constant_30196" />
</rt_info>
<input>
<port id="0" precision="BOOL">
<dim>2</dim>
</port>
<port id="1" precision="I64">
<dim>2</dim>
</port>
<port id="2" precision="I64">
<dim>2</dim>
</port>
</input>
<output>
<port id="3" precision="I64" names="/transformer/Where_1_output_0">
<dim>2</dim>
</port>
</output>
</layer>
<layer id="85" name="/transformer/Expand_2" type="Broadcast" version="opset3">
<data mode="bidirectional" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Expand_2" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>2</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Expand_2_output_0">
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="86" name="Constant_30247" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014088" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30247" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="87" name="/transformer/Unsqueeze_10" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Unsqueeze_10" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Unsqueeze_10_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>1</dim>
</port>
</output>
</layer>
<layer id="88" name="/transformer/Concat_2" type="Concat" version="opset1">
<data axis="-1" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Concat_2" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
<dim>-1</dim>
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>-1</dim>
<dim>-1</dim>
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Concat_2_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
</port>
</output>
</layer>
<layer id="89" name="/transformer/Constant_13" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_13" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_13_output_0" />
</output>
</layer>
<layer id="90" name="/transformer/Constant_14" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_14" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_14_output_0" />
</output>
</layer>
<layer id="91" name="/transformer/Range" type="Range" version="opset4">
<data output_type="i64" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Range" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64" />
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/Range_output_0">
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="92" name="Constant_30059" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30059" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="93" name="/transformer/Unsqueeze_5" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Unsqueeze_5" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Unsqueeze_5_output_0">
<dim>-1</dim>
<dim>1</dim>
</port>
</output>
</layer>
<layer id="94" name="Constant_30061" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30061" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="95" name="/transformer/Unsqueeze_6" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Unsqueeze_6" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Unsqueeze_6_output_0">
<dim>1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="96" name="/transformer/Less" type="Less" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Less" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="2" precision="BOOL" names="/transformer/Less_output_0">
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="97" name="Constant_55183" type="Const" version="opset1">
<data element_type="i64" shape="2" offset="4014233" size="16" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Slice" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>2</dim>
</port>
</output>
</layer>
<layer id="98" name="Constant_55182" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Slice" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="99" name="Constant_55181" type="Const" version="opset1">
<data element_type="i32" shape="1" offset="4014249" size="4" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Slice" />
</rt_info>
<output>
<port id="0" precision="I32">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="100" name="ScatterUpdate_55184" type="ScatterUpdate" version="opset3">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Slice" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>2</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64">
<dim>1</dim>
</port>
<port id="3" precision="I32">
<dim>1</dim>
</port>
</input>
<output>
<port id="4" precision="I64">
<dim>2</dim>
</port>
</output>
</layer>
<layer id="101" name="Constant_55187" type="Const" version="opset1">
<data element_type="i64" shape="2" offset="4014253" size="16" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Slice" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>2</dim>
</port>
</output>
</layer>
<layer id="102" name="Constant_55190" type="Const" version="opset1">
<data element_type="i64" shape="2" offset="4014217" size="16" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Slice" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>2</dim>
</port>
</output>
</layer>
<layer id="103" name="/transformer/Slice" type="StridedSlice" version="opset1">
<data begin_mask="1, 0" end_mask="1, 0" new_axis_mask="" shrink_axis_mask="" ellipsis_mask="" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Slice" />
</rt_info>
<input>
<port id="0" precision="BOOL">
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>2</dim>
</port>
<port id="2" precision="I64">
<dim>2</dim>
</port>
<port id="3" precision="I64">
<dim>2</dim>
</port>
</input>
<output>
<port id="4" precision="BOOL" names="/transformer/Slice_output_0">
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="104" name="/transformer/Shape_5" type="ShapeOf" version="opset3">
<data output_type="i64" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Shape_5" />
</rt_info>
<input>
<port id="0" precision="BOOL">
<dim>-1</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="1" precision="I64" names="/transformer/Shape_5_output_0">
<dim>2</dim>
</port>
</output>
</layer>
<layer id="105" name="/transformer/Expand" type="Broadcast" version="opset3">
<data mode="bidirectional" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Expand" />
</rt_info>
<input>
<port id="0" precision="BOOL">
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>2</dim>
</port>
</input>
<output>
<port id="2" precision="BOOL" names="/transformer/Expand_output_0">
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="106" name="/transformer/Reshape_2" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Reshape_2" />
</rt_info>
<input>
<port id="0" precision="BOOL">
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>2</dim>
</port>
</input>
<output>
<port id="2" precision="BOOL" names="/transformer/Reshape_2_output_0">
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="107" name="/transformer/ScatterND" type="ScatterNDUpdate" version="opset4">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/ScatterND" />
</rt_info>
<input>
<port id="0" precision="BOOL">
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
</port>
<port id="2" precision="BOOL">
<dim>-1</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="3" precision="BOOL" names="/transformer/ScatterND_output_0">
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="108" name="Constant_30274" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30274" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="109" name="/transformer/Unsqueeze_11" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Unsqueeze_11" />
</rt_info>
<input>
<port id="0" precision="BOOL">
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="BOOL" names="/transformer/Unsqueeze_11_output_0">
<dim>1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="110" name="Constant_30276" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30276" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="111" name="/transformer/Unsqueeze_12" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Unsqueeze_12" />
</rt_info>
<input>
<port id="0" precision="BOOL">
<dim>1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="BOOL" names="/transformer/Unsqueeze_12_output_0">
<dim>1</dim>
<dim>1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="112" name="Constant_52958" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_1, /transformer/Gather, /transformer/Unsqueeze_13, Constant_29845, Constant_30278" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="113" name="Constant_29845" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_29845" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="114" name="/transformer/Gather" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_1, /transformer/Gather, /transformer/Unsqueeze_13, Constant_29845, Constant_30278" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>2</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/Gather_output_0,/transformer/Unsqueeze_13_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="115" name="/transformer/Constant_33" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_33" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_33_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="116" name="/transformer/Concat_4" type="Concat" version="opset1">
<data axis="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Concat_4, /transformer/Constant_33" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64">
<dim>1</dim>
</port>
<port id="3" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="4" precision="I64" names="/transformer/Concat_4_output_0">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="117" name="Constant_34970" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014088" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Reshape_3" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="118" name="/transformer/Reshape_3" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Reshape_3" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Reshape_3_output_0">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="119" name="/transformer/Mul_5" type="Const" version="opset1">
<data element_type="i64" shape="4" offset="4014112" size="32" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/ConstantOfShape_3, /transformer/Constant_35, /transformer/Mul_5, /transformer/Shape_12, Constant_30292" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Mul_5_output_0">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="120" name="/transformer/Equal_2" type="Equal" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/ConstantOfShape_3, /transformer/Constant_35, /transformer/Equal_2, /transformer/Mul_5, /transformer/Shape_12, Constant_30292" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="BOOL" names="/transformer/Equal_2_output_0">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="121" name="/transformer/ConstantOfShape_3" type="Const" version="opset1">
<data element_type="i64" shape="4" offset="4014144" size="32" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/ConstantOfShape_3, /transformer/Shape_12, Constant_30292" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/ConstantOfShape_3_output_0">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="122" name="/transformer/Where_2" type="Select" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/ConstantOfShape_3, /transformer/Shape_12, /transformer/Where_2, Constant_30292" />
</rt_info>
<input>
<port id="0" precision="BOOL">
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>4</dim>
</port>
<port id="2" precision="I64">
<dim>4</dim>
</port>
</input>
<output>
<port id="3" precision="I64" names="/transformer/Where_2_output_0">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="123" name="/transformer/Expand_3" type="Broadcast" version="opset3">
<data mode="bidirectional" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Expand_3" />
</rt_info>
<input>
<port id="0" precision="BOOL">
<dim>1</dim>
<dim>1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="BOOL" names="/transformer/Expand_3_output_0">
<dim>-1</dim>
<dim>1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="124" name="/transformer/Or" type="LogicalOr" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Or, /transformer/h.0/self_attention/Cast, /transformer/h.1/self_attention/Cast" />
</rt_info>
<input>
<port id="0" precision="BOOL">
<dim>-1</dim>
<dim>1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="BOOL">
<dim>-1</dim>
<dim>1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="2" precision="BOOL" names="/transformer/Or_output_0,/transformer/h.0/self_attention/Cast_output_0,/transformer/h.1/self_attention/Cast_output_0">
<dim>-1</dim>
<dim>1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="125" name="/transformer/h.0/self_attention/Constant_15" type="Const" version="opset1">
<data element_type="f32" shape="" offset="4014269" size="4" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_15" />
</rt_info>
<output>
<port id="0" precision="FP32" names="/transformer/h.0/self_attention/Constant_15_output_0" />
</output>
</layer>
<layer id="126" name="Constant_30487" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014088" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30487" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="127" name="/transformer/h.0/input_layernorm/Div" type="MVN" version="opset6">
<data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/input_layernorm/Add, /transformer/h.0/input_layernorm/Add_1, /transformer/h.0/input_layernorm/Div, /transformer/h.0/input_layernorm/Mul, /transformer/h.0/input_layernorm/Pow, /transformer/h.0/input_layernorm/ReduceMean, /transformer/h.0/input_layernorm/ReduceMean_1, /transformer/h.0/input_layernorm/Sqrt, /transformer/h.0/input_layernorm/Sub" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/input_layernorm/Add_1_output_0,/transformer/h.0/input_layernorm/Div_output_0,/transformer/h.0/input_layernorm/Mul_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="128" name="Transpose_56360_compressed" type="Const" version="opset1">
<data element_type="f16" shape="24, 8" offset="4014273" size="384" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/query_key_value/Add, /transformer/h.0/self_attention/query_key_value/MatMul, onnx::MatMul_500" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>24</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="129" name="Transpose_56360" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/query_key_value/Add, /transformer/h.0/self_attention/query_key_value/MatMul, onnx::MatMul_500" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>24</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>24</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="130" name="/transformer/h.0/self_attention/query_key_value/MatMul" type="MatMul" version="opset1">
<data transpose_a="false" transpose_b="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/query_key_value/Add, /transformer/h.0/self_attention/query_key_value/MatMul" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="FP32">
<dim>24</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/self_attention/query_key_value/Add_output_0,/transformer/h.0/self_attention/query_key_value/MatMul_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>24</dim>
</port>
</output>
</layer>
<layer id="131" name="/transformer/h.0/self_attention/Concat" type="Const" version="opset1">
<data element_type="i64" shape="5" offset="4014657" size="40" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Concat, /transformer/h.0/self_attention/Constant_2, /transformer/h.0/self_attention/Constant_3, /transformer/h.0/self_attention/Constant_4" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>5</dim>
</port>
</output>
</layer>
<layer id="132" name="/transformer/h.0/self_attention/Reshape" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Reshape" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>24</dim>
</port>
<port id="1" precision="I64">
<dim>5</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/self_attention/Reshape_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>3</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="133" name="/transformer/Constant" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_output_0" />
</output>
</layer>
<layer id="134" name="Constant_30529" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014697" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30529" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="135" name="/transformer/h.0/self_attention/Gather_2" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Gather_2" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>3</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64" />
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="FP32" names="/transformer/h.0/self_attention/Gather_2_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="136" name="Constant_30543" type="Const" version="opset1">
<data element_type="i64" shape="4" offset="4014705" size="32" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30543" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="137" name="/transformer/h.0/self_attention/Transpose" type="Transpose" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Transpose" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/self_attention/Transpose_output_0">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="138" name="/transformer/h.0/self_attention/Shape_3" type="ShapeOf" version="opset3">
<data output_type="i64" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Shape_2, /transformer/h.0/self_attention/Shape_3" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
</port>
</input>
<output>
<port id="1" precision="I64" names="/transformer/h.0/self_attention/Shape_2_output_0,/transformer/h.0/self_attention/Shape_3_output_0">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="139" name="/transformer/h.0/self_attention/Constant_5" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_5" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.0/self_attention/Constant_5_output_0" />
</output>
</layer>
<layer id="140" name="Constant_30537" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30537" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="141" name="/transformer/h.0/self_attention/Gather_5" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_5, /transformer/h.0/self_attention/Gather_5, Constant_30537" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>4</dim>
</port>
<port id="1" precision="I64" />
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/h.0/self_attention/Gather_5_output_0" />
</output>
</layer>
<layer id="142" name="/transformer/h.0/self_attention/Constant_7" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014104" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_7" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.0/self_attention/Constant_7_output_0" />
</output>
</layer>
<layer id="143" name="/transformer/h.0/self_attention/Mul" type="Multiply" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_7, /transformer/h.0/self_attention/Mul" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64" />
</input>
<output>
<port id="2" precision="I64" names="/transformer/h.0/self_attention/Mul_output_0" />
</output>
</layer>
<layer id="144" name="Constant_30628" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30628" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="145" name="/transformer/h.0/self_attention/Unsqueeze_6" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Unsqueeze_11, /transformer/h.0/self_attention/Unsqueeze_2, /transformer/h.0/self_attention/Unsqueeze_4, /transformer/h.0/self_attention/Unsqueeze_6, Constant_30547, Constant_30553, Constant_30628, Constant_30811" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/h.0/self_attention/Unsqueeze_11_output_0,/transformer/h.0/self_attention/Unsqueeze_2_output_0,/transformer/h.0/self_attention/Unsqueeze_4_output_0,/transformer/h.0/self_attention/Unsqueeze_6_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="146" name="Constant_52973" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_6, /transformer/h.0/self_attention/Gather_6, /transformer/h.0/self_attention/Unsqueeze_12, /transformer/h.0/self_attention/Unsqueeze_3, /transformer/h.0/self_attention/Unsqueeze_5, /transformer/h.0/self_attention/Unsqueeze_7, /transformer/h.0/self_attention/Unsqueeze_9, Constant_30541, Constant_30549, Constant_30555, Constant_30631, Constant_30785, Constant_30813" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="147" name="Constant_30541" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30541" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="148" name="/transformer/h.0/self_attention/Gather_6" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_6, /transformer/h.0/self_attention/Gather_6, /transformer/h.0/self_attention/Unsqueeze_12, /transformer/h.0/self_attention/Unsqueeze_3, /transformer/h.0/self_attention/Unsqueeze_5, /transformer/h.0/self_attention/Unsqueeze_7, /transformer/h.0/self_attention/Unsqueeze_9, Constant_30541, Constant_30549, Constant_30555, Constant_30631, Constant_30785, Constant_30813" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/h.0/self_attention/Gather_6_output_0,/transformer/h.0/self_attention/Unsqueeze_12_output_0,/transformer/h.0/self_attention/Unsqueeze_3_output_0,/transformer/h.0/self_attention/Unsqueeze_5_output_0,/transformer/h.0/self_attention/Unsqueeze_7_output_0,/transformer/h.0/self_attention/Unsqueeze_9_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="149" name="/transformer/h.0/self_attention/Constant_8" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014737" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_8" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.0/self_attention/Constant_8_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="150" name="/transformer/h.0/self_attention/Concat_1" type="Concat" version="opset1">
<data axis="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Concat_1, /transformer/h.0/self_attention/Concat_2, /transformer/h.0/self_attention/Constant_8, /transformer/h.0/self_attention/Constant_9" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="3" precision="I64" names="/transformer/h.0/self_attention/Concat_1_output_0,/transformer/h.0/self_attention/Concat_2_output_0">
<dim>3</dim>
</port>
</output>
</layer>
<layer id="151" name="/transformer/h.0/self_attention/Reshape_1" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Reshape_1" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>3</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/self_attention/Reshape_1_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="152" name="/transformer/Constant_2" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_2" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_2_output_0" />
</output>
</layer>
<layer id="153" name="Constant_30531" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014697" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30531" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="154" name="/transformer/h.0/self_attention/Gather_3" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Gather_3" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>3</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64" />
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="FP32" names="/transformer/h.0/self_attention/Gather_3_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="155" name="Constant_30626" type="Const" version="opset1">
<data element_type="i64" shape="4" offset="4014745" size="32" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30626" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="156" name="/transformer/h.0/self_attention/Transpose_1" type="Transpose" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Transpose_1" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/self_attention/Transpose_1_output_0">
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="157" name="/transformer/h.0/self_attention/Constant_10" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014737" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_10" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.0/self_attention/Constant_10_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="158" name="/transformer/h.0/self_attention/Concat_3" type="Concat" version="opset1">
<data axis="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Concat_3, /transformer/h.0/self_attention/Constant_10" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="3" precision="I64" names="/transformer/h.0/self_attention/Concat_3_output_0">
<dim>3</dim>
</port>
</output>
</layer>
<layer id="159" name="/transformer/h.0/self_attention/Reshape_2" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Reshape_2" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>3</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/self_attention/Reshape_2_output_0">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="160" name="present.0.key" type="Concat" version="opset1">
<data axis="2" />
<rt_info>
<attribute name="fused_names" version="0" value="present.0.key" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="present.0.key">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="161" name="Constant_56554_compressed" type="Const" version="opset1">
<data element_type="f16" shape="1, 1, 1" offset="4014777" size="2" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_12" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</output>
</layer>
<layer id="162" name="Constant_56554" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_12" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</output>
</layer>
<layer id="163" name="Multiply_56394" type="Multiply" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/MatMul, /transformer/h.0/self_attention/Mul_1, present.0.key" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
<port id="1" precision="FP32">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="FP32">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="164" name="/transformer/h.0/self_attention/Mul_1" type="MatMul" version="opset1">
<data transpose_a="false" transpose_b="false" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/MatMul, /transformer/h.0/self_attention/Mul_1, present.0.key" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/self_attention/Mul_1_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="165" name="Constant_56556_compressed" type="Const" version="opset1">
<data element_type="f16" shape="1, 2, 1" offset="4014779" size="4" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_10" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>1</dim>
<dim>2</dim>
<dim>1</dim>
</port>
</output>
</layer>
<layer id="166" name="Constant_56556" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/Constant_10" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>1</dim>
<dim>2</dim>
<dim>1</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>1</dim>
<dim>2</dim>
<dim>1</dim>
</port>
</output>
</layer>
<layer id="167" name="/transformer/Constant_8" type="Const" version="opset1">
<data element_type="i32" shape="" offset="4014783" size="4" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_8" />
</rt_info>
<output>
<port id="0" precision="I32" names="/transformer/Constant_8_output_0" />
</output>
</layer>
<layer id="168" name="/transformer/CumSum" type="CumSum" version="opset3">
<data exclusive="false" reverse="false" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/CumSum" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I32" />
</input>
<output>
<port id="2" precision="I64" names="/transformer/CumSum_output_0">
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="169" name="Constant_56555" type="Const" version="opset1">
<data element_type="i64" shape="1, 1" offset="4014088" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_9, /transformer/Sub" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
<dim>1</dim>
</port>
</output>
</layer>
<layer id="170" name="/transformer/Sub" type="Add" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_9, /transformer/Sub" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Sub_output_0">
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="171" name="/transformer/Mul" type="Multiply" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Mul" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>-1</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Mul_output_0">
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="172" name="Constant_29886" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_29886" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="173" name="/transformer/Unsqueeze" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Unsqueeze" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Unsqueeze_output_0">
<dim>-1</dim>
<dim>1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="174" name="/transformer/Cast_1" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Cast_1" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>-1</dim>
<dim>1</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="1" precision="FP32" names="/transformer/Cast_1_output_0">
<dim>-1</dim>
<dim>1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="175" name="/transformer/Mul_1" type="Multiply" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Mul_1" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>1</dim>
<dim>2</dim>
<dim>1</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>1</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/Mul_1_output_0">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="176" name="/transformer/Constant_11" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014104" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_11" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_11_output_0" />
</output>
</layer>
<layer id="177" name="/transformer/Mul_2" type="Multiply" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_11, /transformer/Mul_2" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64" />
</input>
<output>
<port id="2" precision="I64" names="/transformer/Mul_2_output_0" />
</output>
</layer>
<layer id="178" name="Constant_29893" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_29893" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="179" name="/transformer/Unsqueeze_1" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Unsqueeze_1, Constant_29893" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/Unsqueeze_1_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="180" name="/transformer/Constant_12" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_12" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_12_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="181" name="/transformer/Concat" type="Concat" version="opset1">
<data axis="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Concat, /transformer/Constant_12" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="3" precision="I64" names="/transformer/Concat_output_0">
<dim>3</dim>
</port>
</output>
</layer>
<layer id="182" name="/transformer/Reshape" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Cast_2, /transformer/Reshape, /transformer/h.0/self_attention/Mul_2, /transformer/h.1/self_attention/Mul_2" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>3</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/Cast_2_output_0,/transformer/Reshape_output_0,/transformer/h.0/self_attention/Mul_2_output_0,/transformer/h.1/self_attention/Mul_2_output_0">
<dim>-1</dim>
<dim>1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="183" name="/transformer/h.0/self_attention/Add" type="Add" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Add" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>1</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/self_attention/Add_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="184" name="Constant_30782" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30782" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="185" name="/transformer/h.0/self_attention/Unsqueeze_8" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Unsqueeze_8, Constant_30782" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/h.0/self_attention/Unsqueeze_8_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="186" name="/transformer/h.0/self_attention/Constant_14" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014104" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_14" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.0/self_attention/Constant_14_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="187" name="/transformer/h.0/self_attention/Shape_4" type="ShapeOf" version="opset3">
<data output_type="i64" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Shape_4" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="1" precision="I64" names="/transformer/h.0/self_attention/Shape_4_output_0">
<dim>3</dim>
</port>
</output>
</layer>
<layer id="188" name="Constant_52978" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014104" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_11, /transformer/h.0/self_attention/Gather_7, /transformer/h.0/self_attention/Unsqueeze_10, /transformer/h.0/self_attention/Unsqueeze_13, Constant_30774, Constant_30787, Constant_30815" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="189" name="Constant_30774" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30774" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="190" name="/transformer/h.0/self_attention/Gather_7" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_11, /transformer/h.0/self_attention/Gather_7, /transformer/h.0/self_attention/Unsqueeze_10, /transformer/h.0/self_attention/Unsqueeze_13, Constant_30774, Constant_30787, Constant_30815" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>3</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/h.0/self_attention/Gather_7_output_0,/transformer/h.0/self_attention/Unsqueeze_10_output_0,/transformer/h.0/self_attention/Unsqueeze_13_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="191" name="/transformer/h.0/self_attention/Concat_6" type="Concat" version="opset1">
<data axis="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Concat_6, /transformer/h.0/self_attention/Constant_14" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64">
<dim>1</dim>
</port>
<port id="3" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="4" precision="I64" names="/transformer/h.0/self_attention/Concat_6_output_0">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="192" name="/transformer/h.0/self_attention/Reshape_4" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Reshape_4" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/self_attention/Reshape_4_output_0">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="193" name="/transformer/h.0/self_attention/Where" type="Select" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Where" />
</rt_info>
<input>
<port id="0" precision="BOOL">
<dim>-1</dim>
<dim>1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="FP32" />
<port id="2" precision="FP32">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="3" precision="FP32" names="/transformer/h.0/self_attention/Where_output_0">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="194" name="/transformer/h.0/self_attention/Softmax" type="SoftMax" version="opset8">
<data axis="3" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Cast_1, /transformer/h.0/self_attention/Cast_2, /transformer/h.0/self_attention/Softmax" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="1" precision="FP32" names="/transformer/h.0/self_attention/Cast_1_output_0,/transformer/h.0/self_attention/Cast_2_output_0,/transformer/h.0/self_attention/Softmax_output_0">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="195" name="/transformer/h.0/self_attention/Concat_7" type="Concat" version="opset1">
<data axis="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Concat_7" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="3" precision="I64" names="/transformer/h.0/self_attention/Concat_7_output_0">
<dim>3</dim>
</port>
</output>
</layer>
<layer id="196" name="/transformer/h.0/self_attention/Reshape_5" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Reshape_5" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>3</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/self_attention/Reshape_5_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="197" name="/transformer/Constant_4" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014104" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/Constant_4" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/Constant_4_output_0" />
</output>
</layer>
<layer id="198" name="Constant_30533" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014697" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30533" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="199" name="/transformer/h.0/self_attention/Gather_4" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Gather_4" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>3</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64" />
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="FP32" names="/transformer/h.0/self_attention/Gather_4_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="200" name="Constant_30701" type="Const" version="opset1">
<data element_type="i64" shape="4" offset="4014705" size="32" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30701" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="201" name="/transformer/h.0/self_attention/Transpose_2" type="Transpose" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Transpose_2" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/self_attention/Transpose_2_output_0">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="202" name="/transformer/h.0/self_attention/Reshape_3" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Reshape_3" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>3</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/self_attention/Reshape_3_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="203" name="present.0.value" type="Concat" version="opset1">
<data axis="1" />
<rt_info>
<attribute name="fused_names" version="0" value="present.0.value" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="present.0.value">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="204" name="/transformer/h.0/self_attention/MatMul_1" type="MatMul" version="opset1">
<data transpose_a="false" transpose_b="false" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/MatMul_1" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/self_attention/MatMul_1_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="205" name="/transformer/h.0/self_attention/Shape_5" type="ShapeOf" version="opset3">
<data output_type="i64" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Shape_5, /transformer/h.0/self_attention/Shape_6" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</input>
<output>
<port id="1" precision="I64" names="/transformer/h.0/self_attention/Shape_5_output_0,/transformer/h.0/self_attention/Shape_6_output_0">
<dim>3</dim>
</port>
</output>
</layer>
<layer id="206" name="/transformer/h.0/self_attention/Constant_16" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_16" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.0/self_attention/Constant_16_output_0" />
</output>
</layer>
<layer id="207" name="Constant_30892" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30892" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="208" name="/transformer/h.0/self_attention/Gather_8" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_16, /transformer/h.0/self_attention/Gather_8, Constant_30892" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>3</dim>
</port>
<port id="1" precision="I64" />
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/h.0/self_attention/Gather_8_output_0" />
</output>
</layer>
<layer id="209" name="/transformer/h.0/self_attention/Constant_18" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014104" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_18" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.0/self_attention/Constant_18_output_0" />
</output>
</layer>
<layer id="210" name="/transformer/h.0/self_attention/Div" type="Divide" version="opset1">
<data auto_broadcast="numpy" m_pythondiv="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Cast_3, /transformer/h.0/self_attention/Cast_4, /transformer/h.0/self_attention/Constant_18, /transformer/h.0/self_attention/Div" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64" />
</input>
<output>
<port id="2" precision="I64" names="/transformer/h.0/self_attention/Cast_3_output_0,/transformer/h.0/self_attention/Cast_4_output_0,/transformer/h.0/self_attention/Div_output_0" />
</output>
</layer>
<layer id="211" name="Constant_30976" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30976" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="212" name="/transformer/h.0/self_attention/Unsqueeze_16" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Unsqueeze_14, /transformer/h.0/self_attention/Unsqueeze_16, Constant_30902, Constant_30976" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/h.0/self_attention/Unsqueeze_14_output_0,/transformer/h.0/self_attention/Unsqueeze_16_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="213" name="/transformer/h.0/self_attention/Constant_19" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014104" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_19" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.0/self_attention/Constant_19_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="214" name="Constant_52983" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_17, /transformer/h.0/self_attention/Gather_9, /transformer/h.0/self_attention/Unsqueeze_15, /transformer/h.0/self_attention/Unsqueeze_17, Constant_30896, Constant_30905, Constant_30978" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="215" name="Constant_30896" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30896" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="216" name="/transformer/h.0/self_attention/Gather_9" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_17, /transformer/h.0/self_attention/Gather_9, /transformer/h.0/self_attention/Unsqueeze_15, /transformer/h.0/self_attention/Unsqueeze_17, Constant_30896, Constant_30905, Constant_30978" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>3</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/h.0/self_attention/Gather_9_output_0,/transformer/h.0/self_attention/Unsqueeze_15_output_0,/transformer/h.0/self_attention/Unsqueeze_17_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="217" name="/transformer/h.0/self_attention/Constant_20" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014737" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_20" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.0/self_attention/Constant_20_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="218" name="/transformer/h.0/self_attention/Concat_8" type="Concat" version="opset1">
<data axis="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Concat_8, /transformer/h.0/self_attention/Constant_19, /transformer/h.0/self_attention/Constant_20" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64">
<dim>1</dim>
</port>
<port id="3" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="4" precision="I64" names="/transformer/h.0/self_attention/Concat_8_output_0">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="219" name="/transformer/h.0/self_attention/Reshape_6" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Reshape_6" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/self_attention/Reshape_6_output_0">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="220" name="Constant_30974" type="Const" version="opset1">
<data element_type="i64" shape="4" offset="4014705" size="32" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_30974" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="221" name="/transformer/h.0/self_attention/Transpose_3" type="Transpose" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Transpose_3" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/self_attention/Transpose_3_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="222" name="/transformer/h.0/self_attention/Constant_21" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014787" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Constant_21" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.0/self_attention/Constant_21_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="223" name="/transformer/h.0/self_attention/Concat_9" type="Concat" version="opset1">
<data axis="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Concat_9, /transformer/h.0/self_attention/Constant_21" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="3" precision="I64" names="/transformer/h.0/self_attention/Concat_9_output_0">
<dim>3</dim>
</port>
</output>
</layer>
<layer id="224" name="/transformer/h.0/self_attention/Reshape_7" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Reshape_7" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>3</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/self_attention/Reshape_7_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="225" name="Transpose_56363_compressed" type="Const" version="opset1">
<data element_type="f16" shape="8, 8" offset="4014795" size="128" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/dense/Add, /transformer/h.0/self_attention/dense/MatMul, onnx::MatMul_513" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>8</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="226" name="Transpose_56363" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/dense/Add, /transformer/h.0/self_attention/dense/MatMul, onnx::MatMul_513" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>8</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>8</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="227" name="/transformer/h.0/self_attention/dense/MatMul" type="MatMul" version="opset1">
<data transpose_a="false" transpose_b="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/dense/Add, /transformer/h.0/self_attention/dense/MatMul" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="FP32">
<dim>8</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/self_attention/dense/Add_output_0,/transformer/h.0/self_attention/dense/MatMul_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="228" name="/transformer/h.0/self_attention/Add_1" type="Add" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/self_attention/Add_1" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/self_attention/Add_1_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="229" name="Constant_31050" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014088" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_31050" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="230" name="/transformer/h.0/post_attention_layernorm/Div" type="MVN" version="opset6">
<data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/post_attention_layernorm/Add, /transformer/h.0/post_attention_layernorm/Add_1, /transformer/h.0/post_attention_layernorm/Div, /transformer/h.0/post_attention_layernorm/Mul, /transformer/h.0/post_attention_layernorm/Pow, /transformer/h.0/post_attention_layernorm/ReduceMean, /transformer/h.0/post_attention_layernorm/ReduceMean_1, /transformer/h.0/post_attention_layernorm/Sqrt, /transformer/h.0/post_attention_layernorm/Sub" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/post_attention_layernorm/Add_1_output_0,/transformer/h.0/post_attention_layernorm/Div_output_0,/transformer/h.0/post_attention_layernorm/Mul_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="231" name="Multiply_56396_compressed" type="Const" version="opset1">
<data element_type="f16" shape="32, 8" offset="4014923" size="512" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/dense_h_to_4h/Add, /transformer/h.0/mlp/dense_h_to_4h/MatMul, /transformer/h.0/mlp/gelu_impl/Constant, /transformer/h.0/mlp/gelu_impl/Mul, onnx::MatMul_514" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>32</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="232" name="Multiply_56396" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/dense_h_to_4h/Add, /transformer/h.0/mlp/dense_h_to_4h/MatMul, /transformer/h.0/mlp/gelu_impl/Constant, /transformer/h.0/mlp/gelu_impl/Mul, onnx::MatMul_514" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>32</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>32</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="233" name="/transformer/h.0/mlp/gelu_impl/Mul" type="MatMul" version="opset1">
<data transpose_a="false" transpose_b="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/dense_h_to_4h/Add, /transformer/h.0/mlp/dense_h_to_4h/MatMul, /transformer/h.0/mlp/gelu_impl/Mul" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="FP32">
<dim>32</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/mlp/gelu_impl/Mul_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="234" name="Multiply_56398_compressed" type="Const" version="opset1">
<data element_type="f16" shape="32, 8" offset="4015435" size="512" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/dense_h_to_4h/Add, /transformer/h.0/mlp/dense_h_to_4h/MatMul, /transformer/h.0/mlp/gelu_impl/Constant_1, /transformer/h.0/mlp/gelu_impl/Mul_1, onnx::MatMul_514" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>32</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="235" name="Multiply_56398" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/dense_h_to_4h/Add, /transformer/h.0/mlp/dense_h_to_4h/MatMul, /transformer/h.0/mlp/gelu_impl/Constant_1, /transformer/h.0/mlp/gelu_impl/Mul_1, onnx::MatMul_514" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>32</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>32</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="236" name="/transformer/h.0/mlp/gelu_impl/Mul_1" type="MatMul" version="opset1">
<data transpose_a="false" transpose_b="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/dense_h_to_4h/Add, /transformer/h.0/mlp/dense_h_to_4h/MatMul, /transformer/h.0/mlp/gelu_impl/Mul_1" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="FP32">
<dim>32</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/mlp/gelu_impl/Mul_1_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="237" name="Multiply_56400_compressed" type="Const" version="opset1">
<data element_type="f16" shape="32, 8" offset="4015947" size="512" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/dense_h_to_4h/Add, /transformer/h.0/mlp/dense_h_to_4h/MatMul, /transformer/h.0/mlp/gelu_impl/Constant_2, /transformer/h.0/mlp/gelu_impl/Mul_2, onnx::MatMul_514" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>32</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="238" name="Multiply_56400" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/dense_h_to_4h/Add, /transformer/h.0/mlp/dense_h_to_4h/MatMul, /transformer/h.0/mlp/gelu_impl/Constant_2, /transformer/h.0/mlp/gelu_impl/Mul_2, onnx::MatMul_514" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>32</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>32</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="239" name="/transformer/h.0/mlp/gelu_impl/Mul_2" type="MatMul" version="opset1">
<data transpose_a="false" transpose_b="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/dense_h_to_4h/Add, /transformer/h.0/mlp/dense_h_to_4h/MatMul, /transformer/h.0/mlp/gelu_impl/Mul_2" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="FP32">
<dim>32</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/mlp/gelu_impl/Mul_2_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="240" name="Transpose_56367_compressed" type="Const" version="opset1">
<data element_type="f16" shape="32, 8" offset="4016459" size="512" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/dense_h_to_4h/Add, /transformer/h.0/mlp/dense_h_to_4h/MatMul, onnx::MatMul_514" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>32</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="241" name="Transpose_56367" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/dense_h_to_4h/Add, /transformer/h.0/mlp/dense_h_to_4h/MatMul, onnx::MatMul_514" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>32</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>32</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="242" name="/transformer/h.0/mlp/dense_h_to_4h/MatMul" type="MatMul" version="opset1">
<data transpose_a="false" transpose_b="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/dense_h_to_4h/Add, /transformer/h.0/mlp/dense_h_to_4h/MatMul" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="FP32">
<dim>32</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/mlp/dense_h_to_4h/Add_output_0,/transformer/h.0/mlp/dense_h_to_4h/MatMul_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="243" name="/transformer/h.0/mlp/gelu_impl/Mul_3" type="Multiply" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/gelu_impl/Mul_3" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/mlp/gelu_impl/Mul_3_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="244" name="Constant_56557_compressed" type="Const" version="opset1">
<data element_type="f16" shape="1, 1, 1" offset="4016971" size="2" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/gelu_impl/Constant_3" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</output>
</layer>
<layer id="245" name="Constant_56557" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/gelu_impl/Constant_3" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</output>
</layer>
<layer id="246" name="/transformer/h.0/mlp/gelu_impl/Add" type="Add" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/gelu_impl/Add" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
<port id="1" precision="FP32">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/mlp/gelu_impl/Add_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="247" name="/transformer/h.0/mlp/gelu_impl/Mul_4" type="Multiply" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/gelu_impl/Mul_4" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/mlp/gelu_impl/Mul_4_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="248" name="/transformer/h.0/mlp/gelu_impl/Tanh" type="Tanh" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/gelu_impl/Tanh" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</input>
<output>
<port id="1" precision="FP32" names="/transformer/h.0/mlp/gelu_impl/Tanh_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="249" name="Constant_56558_compressed" type="Const" version="opset1">
<data element_type="f16" shape="1, 1, 1" offset="4016971" size="2" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/gelu_impl/Constant_4" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</output>
</layer>
<layer id="250" name="Constant_56558" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/gelu_impl/Constant_4" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</output>
</layer>
<layer id="251" name="/transformer/h.0/mlp/gelu_impl/Add_1" type="Add" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/gelu_impl/Add_1" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
<port id="1" precision="FP32">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/mlp/gelu_impl/Add_1_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="252" name="/transformer/h.0/mlp/gelu_impl/Mul_5" type="Multiply" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/gelu_impl/Mul_5" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/mlp/gelu_impl/Mul_5_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="253" name="Transpose_56370_compressed" type="Const" version="opset1">
<data element_type="f16" shape="8, 32" offset="4016973" size="512" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/dense_4h_to_h/Add, /transformer/h.0/mlp/dense_4h_to_h/MatMul, onnx::MatMul_515" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>8</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="254" name="Transpose_56370" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/dense_4h_to_h/Add, /transformer/h.0/mlp/dense_4h_to_h/MatMul, onnx::MatMul_515" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>8</dim>
<dim>32</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>8</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="255" name="/transformer/h.0/mlp/dense_4h_to_h/MatMul" type="MatMul" version="opset1">
<data transpose_a="false" transpose_b="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/dense_4h_to_h/Add, /transformer/h.0/mlp/dense_4h_to_h/MatMul" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
<port id="1" precision="FP32">
<dim>8</dim>
<dim>32</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/mlp/dense_4h_to_h/Add_output_0,/transformer/h.0/mlp/dense_4h_to_h/MatMul_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="256" name="/transformer/h.0/mlp/Add" type="Add" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.0/mlp/Add" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.0/mlp/Add_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="257" name="Constant_31082" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014088" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_31082" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="258" name="/transformer/h.1/input_layernorm/Div" type="MVN" version="opset6">
<data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/input_layernorm/Add, /transformer/h.1/input_layernorm/Add_1, /transformer/h.1/input_layernorm/Div, /transformer/h.1/input_layernorm/Mul, /transformer/h.1/input_layernorm/Pow, /transformer/h.1/input_layernorm/ReduceMean, /transformer/h.1/input_layernorm/ReduceMean_1, /transformer/h.1/input_layernorm/Sqrt, /transformer/h.1/input_layernorm/Sub" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/input_layernorm/Add_1_output_0,/transformer/h.1/input_layernorm/Div_output_0,/transformer/h.1/input_layernorm/Mul_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="259" name="Transpose_56374_compressed" type="Const" version="opset1">
<data element_type="f16" shape="24, 8" offset="4017485" size="384" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/query_key_value/Add, /transformer/h.1/self_attention/query_key_value/MatMul, onnx::MatMul_516" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>24</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="260" name="Transpose_56374" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/query_key_value/Add, /transformer/h.1/self_attention/query_key_value/MatMul, onnx::MatMul_516" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>24</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>24</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="261" name="/transformer/h.1/self_attention/query_key_value/MatMul" type="MatMul" version="opset1">
<data transpose_a="false" transpose_b="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/query_key_value/Add, /transformer/h.1/self_attention/query_key_value/MatMul" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="FP32">
<dim>24</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/self_attention/query_key_value/Add_output_0,/transformer/h.1/self_attention/query_key_value/MatMul_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>24</dim>
</port>
</output>
</layer>
<layer id="262" name="/transformer/h.1/self_attention/Concat" type="Const" version="opset1">
<data element_type="i64" shape="5" offset="4014657" size="40" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Concat, /transformer/h.1/self_attention/Constant_2, /transformer/h.1/self_attention/Constant_3, /transformer/h.1/self_attention/Constant_4" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>5</dim>
</port>
</output>
</layer>
<layer id="263" name="/transformer/h.1/self_attention/Reshape" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Reshape" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>24</dim>
</port>
<port id="1" precision="I64">
<dim>5</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/self_attention/Reshape_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>3</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="264" name="Constant_31128" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014697" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_31128" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="265" name="/transformer/h.1/self_attention/Gather_4" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Gather_4" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>3</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64" />
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="FP32" names="/transformer/h.1/self_attention/Gather_4_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="266" name="Constant_31296" type="Const" version="opset1">
<data element_type="i64" shape="4" offset="4014705" size="32" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_31296" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="267" name="/transformer/h.1/self_attention/Transpose_2" type="Transpose" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Transpose_2" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/self_attention/Transpose_2_output_0">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="268" name="Constant_31124" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014697" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_31124" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="269" name="/transformer/h.1/self_attention/Gather_2" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Gather_2" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>3</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64" />
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="FP32" names="/transformer/h.1/self_attention/Gather_2_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="270" name="/transformer/h.1/self_attention/Shape_3" type="ShapeOf" version="opset3">
<data output_type="i64" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Shape_2, /transformer/h.1/self_attention/Shape_3" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
</port>
</input>
<output>
<port id="1" precision="I64" names="/transformer/h.1/self_attention/Shape_2_output_0,/transformer/h.1/self_attention/Shape_3_output_0">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="271" name="/transformer/h.1/self_attention/Constant_5" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_5" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.1/self_attention/Constant_5_output_0" />
</output>
</layer>
<layer id="272" name="Constant_31132" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_31132" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="273" name="/transformer/h.1/self_attention/Gather_5" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_5, /transformer/h.1/self_attention/Gather_5, Constant_31132" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>4</dim>
</port>
<port id="1" precision="I64" />
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/h.1/self_attention/Gather_5_output_0" />
</output>
</layer>
<layer id="274" name="/transformer/h.1/self_attention/Constant_7" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014104" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_7" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.1/self_attention/Constant_7_output_0" />
</output>
</layer>
<layer id="275" name="/transformer/h.1/self_attention/Mul" type="Multiply" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_7, /transformer/h.1/self_attention/Mul" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64" />
</input>
<output>
<port id="2" precision="I64" names="/transformer/h.1/self_attention/Mul_output_0" />
</output>
</layer>
<layer id="276" name="Constant_31142" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_31142" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="277" name="/transformer/h.1/self_attention/Unsqueeze_2" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Unsqueeze_11, /transformer/h.1/self_attention/Unsqueeze_2, /transformer/h.1/self_attention/Unsqueeze_4, /transformer/h.1/self_attention/Unsqueeze_6, Constant_31142, Constant_31148, Constant_31223, Constant_31406" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/h.1/self_attention/Unsqueeze_11_output_0,/transformer/h.1/self_attention/Unsqueeze_2_output_0,/transformer/h.1/self_attention/Unsqueeze_4_output_0,/transformer/h.1/self_attention/Unsqueeze_6_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="278" name="Constant_52998" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_6, /transformer/h.1/self_attention/Gather_6, /transformer/h.1/self_attention/Unsqueeze_12, /transformer/h.1/self_attention/Unsqueeze_3, /transformer/h.1/self_attention/Unsqueeze_5, /transformer/h.1/self_attention/Unsqueeze_7, /transformer/h.1/self_attention/Unsqueeze_9, Constant_31136, Constant_31144, Constant_31150, Constant_31226, Constant_31380, Constant_31408" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="279" name="Constant_31136" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_31136" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="280" name="/transformer/h.1/self_attention/Gather_6" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_6, /transformer/h.1/self_attention/Gather_6, /transformer/h.1/self_attention/Unsqueeze_12, /transformer/h.1/self_attention/Unsqueeze_3, /transformer/h.1/self_attention/Unsqueeze_5, /transformer/h.1/self_attention/Unsqueeze_7, /transformer/h.1/self_attention/Unsqueeze_9, Constant_31136, Constant_31144, Constant_31150, Constant_31226, Constant_31380, Constant_31408" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/h.1/self_attention/Gather_6_output_0,/transformer/h.1/self_attention/Unsqueeze_12_output_0,/transformer/h.1/self_attention/Unsqueeze_3_output_0,/transformer/h.1/self_attention/Unsqueeze_5_output_0,/transformer/h.1/self_attention/Unsqueeze_7_output_0,/transformer/h.1/self_attention/Unsqueeze_9_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="281" name="/transformer/h.1/self_attention/Constant_9" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014737" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_9" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.1/self_attention/Constant_9_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="282" name="/transformer/h.1/self_attention/Concat_2" type="Concat" version="opset1">
<data axis="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Concat_1, /transformer/h.1/self_attention/Concat_2, /transformer/h.1/self_attention/Constant_8, /transformer/h.1/self_attention/Constant_9" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="3" precision="I64" names="/transformer/h.1/self_attention/Concat_1_output_0,/transformer/h.1/self_attention/Concat_2_output_0">
<dim>3</dim>
</port>
</output>
</layer>
<layer id="283" name="/transformer/h.1/self_attention/Reshape_3" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Reshape_3" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>3</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/self_attention/Reshape_3_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="284" name="present.1.value" type="Concat" version="opset1">
<data axis="1" />
<rt_info>
<attribute name="fused_names" version="0" value="present.1.value" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="present.1.value">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="286" name="Constant_31126" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014697" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_31126" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="287" name="/transformer/h.1/self_attention/Gather_3" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Gather_3" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>3</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64" />
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="FP32" names="/transformer/h.1/self_attention/Gather_3_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="288" name="Constant_31221" type="Const" version="opset1">
<data element_type="i64" shape="4" offset="4014745" size="32" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_31221" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="289" name="/transformer/h.1/self_attention/Transpose_1" type="Transpose" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Transpose_1" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/self_attention/Transpose_1_output_0">
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="290" name="/transformer/h.1/self_attention/Constant_10" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014737" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_10" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.1/self_attention/Constant_10_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="291" name="/transformer/h.1/self_attention/Concat_3" type="Concat" version="opset1">
<data axis="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Concat_3, /transformer/h.1/self_attention/Constant_10" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="3" precision="I64" names="/transformer/h.1/self_attention/Concat_3_output_0">
<dim>3</dim>
</port>
</output>
</layer>
<layer id="292" name="/transformer/h.1/self_attention/Reshape_2" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Reshape_2" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>3</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/self_attention/Reshape_2_output_0">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="293" name="present.1.key" type="Concat" version="opset1">
<data axis="2" />
<rt_info>
<attribute name="fused_names" version="0" value="present.1.key" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="present.1.key">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="297" name="/transformer/h.1/self_attention/Constant_15" type="Const" version="opset1">
<data element_type="f32" shape="" offset="4014269" size="4" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_15" />
</rt_info>
<output>
<port id="0" precision="FP32" names="/transformer/h.1/self_attention/Constant_15_output_0" />
</output>
</layer>
<layer id="298" name="Constant_31138" type="Const" version="opset1">
<data element_type="i64" shape="4" offset="4014705" size="32" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_31138" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="299" name="/transformer/h.1/self_attention/Transpose" type="Transpose" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Transpose" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/self_attention/Transpose_output_0">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="300" name="/transformer/h.1/self_attention/Reshape_1" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Reshape_1" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>3</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/self_attention/Reshape_1_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="301" name="Constant_56559_compressed" type="Const" version="opset1">
<data element_type="f16" shape="1, 1, 1" offset="4014777" size="2" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_12" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</output>
</layer>
<layer id="302" name="Constant_56559" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_12" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</output>
</layer>
<layer id="303" name="Multiply_56402" type="Multiply" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/MatMul, /transformer/h.1/self_attention/Mul_1, present.1.key" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
<port id="1" precision="FP32">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="FP32">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="304" name="/transformer/h.1/self_attention/Mul_1" type="MatMul" version="opset1">
<data transpose_a="false" transpose_b="false" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/MatMul, /transformer/h.1/self_attention/Mul_1, present.1.key" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/self_attention/Mul_1_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="305" name="/transformer/h.1/self_attention/Add" type="Add" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Add" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>1</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/self_attention/Add_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="306" name="Constant_31377" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_31377" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="307" name="/transformer/h.1/self_attention/Unsqueeze_8" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Unsqueeze_8, Constant_31377" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/h.1/self_attention/Unsqueeze_8_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="308" name="/transformer/h.1/self_attention/Constant_14" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014104" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_14" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.1/self_attention/Constant_14_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="309" name="/transformer/h.1/self_attention/Shape_4" type="ShapeOf" version="opset3">
<data output_type="i64" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Shape_4" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="1" precision="I64" names="/transformer/h.1/self_attention/Shape_4_output_0">
<dim>3</dim>
</port>
</output>
</layer>
<layer id="310" name="Constant_53003" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014104" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_11, /transformer/h.1/self_attention/Gather_7, /transformer/h.1/self_attention/Unsqueeze_10, /transformer/h.1/self_attention/Unsqueeze_13, Constant_31369, Constant_31382, Constant_31410" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="311" name="Constant_31369" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_31369" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="312" name="/transformer/h.1/self_attention/Gather_7" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_11, /transformer/h.1/self_attention/Gather_7, /transformer/h.1/self_attention/Unsqueeze_10, /transformer/h.1/self_attention/Unsqueeze_13, Constant_31369, Constant_31382, Constant_31410" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>3</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/h.1/self_attention/Gather_7_output_0,/transformer/h.1/self_attention/Unsqueeze_10_output_0,/transformer/h.1/self_attention/Unsqueeze_13_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="313" name="/transformer/h.1/self_attention/Concat_6" type="Concat" version="opset1">
<data axis="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Concat_6, /transformer/h.1/self_attention/Constant_14" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64">
<dim>1</dim>
</port>
<port id="3" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="4" precision="I64" names="/transformer/h.1/self_attention/Concat_6_output_0">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="314" name="/transformer/h.1/self_attention/Reshape_4" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Reshape_4" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/self_attention/Reshape_4_output_0">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="315" name="/transformer/h.1/self_attention/Where" type="Select" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Where" />
</rt_info>
<input>
<port id="0" precision="BOOL">
<dim>-1</dim>
<dim>1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="FP32" />
<port id="2" precision="FP32">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="3" precision="FP32" names="/transformer/h.1/self_attention/Where_output_0">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="316" name="/transformer/h.1/self_attention/Softmax" type="SoftMax" version="opset8">
<data axis="3" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Cast_1, /transformer/h.1/self_attention/Cast_2, /transformer/h.1/self_attention/Softmax" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</input>
<output>
<port id="1" precision="FP32" names="/transformer/h.1/self_attention/Cast_1_output_0,/transformer/h.1/self_attention/Cast_2_output_0,/transformer/h.1/self_attention/Softmax_output_0">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="317" name="/transformer/h.1/self_attention/Concat_7" type="Concat" version="opset1">
<data axis="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Concat_7" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="3" precision="I64" names="/transformer/h.1/self_attention/Concat_7_output_0">
<dim>3</dim>
</port>
</output>
</layer>
<layer id="318" name="/transformer/h.1/self_attention/Reshape_5" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Reshape_5" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="I64">
<dim>3</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/self_attention/Reshape_5_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
</output>
</layer>
<layer id="319" name="/transformer/h.1/self_attention/MatMul_1" type="MatMul" version="opset1">
<data transpose_a="false" transpose_b="false" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/MatMul_1" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>-1</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/self_attention/MatMul_1_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="320" name="/transformer/h.1/self_attention/Shape_6" type="ShapeOf" version="opset3">
<data output_type="i64" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Shape_5, /transformer/h.1/self_attention/Shape_6" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</input>
<output>
<port id="1" precision="I64" names="/transformer/h.1/self_attention/Shape_5_output_0,/transformer/h.1/self_attention/Shape_6_output_0">
<dim>3</dim>
</port>
</output>
</layer>
<layer id="321" name="/transformer/h.1/self_attention/Constant_16" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_16" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.1/self_attention/Constant_16_output_0" />
</output>
</layer>
<layer id="322" name="Constant_31487" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_31487" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="323" name="/transformer/h.1/self_attention/Gather_8" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_16, /transformer/h.1/self_attention/Gather_8, Constant_31487" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>3</dim>
</port>
<port id="1" precision="I64" />
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/h.1/self_attention/Gather_8_output_0" />
</output>
</layer>
<layer id="324" name="/transformer/h.1/self_attention/Constant_18" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014104" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_18" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.1/self_attention/Constant_18_output_0" />
</output>
</layer>
<layer id="325" name="/transformer/h.1/self_attention/Div" type="Divide" version="opset1">
<data auto_broadcast="numpy" m_pythondiv="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Cast_3, /transformer/h.1/self_attention/Cast_4, /transformer/h.1/self_attention/Constant_18, /transformer/h.1/self_attention/Div" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64" />
</input>
<output>
<port id="2" precision="I64" names="/transformer/h.1/self_attention/Cast_3_output_0,/transformer/h.1/self_attention/Cast_4_output_0,/transformer/h.1/self_attention/Div_output_0" />
</output>
</layer>
<layer id="326" name="Constant_31571" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_31571" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="327" name="/transformer/h.1/self_attention/Unsqueeze_16" type="Unsqueeze" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Unsqueeze_14, /transformer/h.1/self_attention/Unsqueeze_16, Constant_31497, Constant_31571" />
</rt_info>
<input>
<port id="0" precision="I64" />
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="I64" names="/transformer/h.1/self_attention/Unsqueeze_14_output_0,/transformer/h.1/self_attention/Unsqueeze_16_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="328" name="/transformer/h.1/self_attention/Constant_19" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014104" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_19" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.1/self_attention/Constant_19_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="329" name="Constant_53008" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014096" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_17, /transformer/h.1/self_attention/Gather_9, /transformer/h.1/self_attention/Unsqueeze_15, /transformer/h.1/self_attention/Unsqueeze_17, Constant_31491, Constant_31500, Constant_31573" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="330" name="Constant_31491" type="Const" version="opset1">
<data element_type="i64" shape="" offset="4014080" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_31491" />
</rt_info>
<output>
<port id="0" precision="I64" />
</output>
</layer>
<layer id="331" name="/transformer/h.1/self_attention/Gather_9" type="Gather" version="opset8">
<data batch_dims="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_17, /transformer/h.1/self_attention/Gather_9, /transformer/h.1/self_attention/Unsqueeze_15, /transformer/h.1/self_attention/Unsqueeze_17, Constant_31491, Constant_31500, Constant_31573" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>3</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64" />
</input>
<output>
<port id="3" precision="I64" names="/transformer/h.1/self_attention/Gather_9_output_0,/transformer/h.1/self_attention/Unsqueeze_15_output_0,/transformer/h.1/self_attention/Unsqueeze_17_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="332" name="/transformer/h.1/self_attention/Constant_20" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014737" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_20" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.1/self_attention/Constant_20_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="333" name="/transformer/h.1/self_attention/Concat_8" type="Concat" version="opset1">
<data axis="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Concat_8, /transformer/h.1/self_attention/Constant_19, /transformer/h.1/self_attention/Constant_20" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64">
<dim>1</dim>
</port>
<port id="3" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="4" precision="I64" names="/transformer/h.1/self_attention/Concat_8_output_0">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="334" name="/transformer/h.1/self_attention/Reshape_6" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Reshape_6" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/self_attention/Reshape_6_output_0">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="335" name="Constant_31569" type="Const" version="opset1">
<data element_type="i64" shape="4" offset="4014705" size="32" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_31569" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>4</dim>
</port>
</output>
</layer>
<layer id="336" name="/transformer/h.1/self_attention/Transpose_3" type="Transpose" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Transpose_3" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>2</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>4</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/self_attention/Transpose_3_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
</port>
</output>
</layer>
<layer id="337" name="/transformer/h.1/self_attention/Constant_21" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014787" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Constant_21" />
</rt_info>
<output>
<port id="0" precision="I64" names="/transformer/h.1/self_attention/Constant_21_output_0">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="338" name="/transformer/h.1/self_attention/Concat_9" type="Concat" version="opset1">
<data axis="0" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Concat_9, /transformer/h.1/self_attention/Constant_21" />
</rt_info>
<input>
<port id="0" precision="I64">
<dim>1</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
<port id="2" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="3" precision="I64" names="/transformer/h.1/self_attention/Concat_9_output_0">
<dim>3</dim>
</port>
</output>
</layer>
<layer id="339" name="/transformer/h.1/self_attention/Reshape_7" type="Reshape" version="opset1">
<data special_zero="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Reshape_7" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>2</dim>
<dim>4</dim>
</port>
<port id="1" precision="I64">
<dim>3</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/self_attention/Reshape_7_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="340" name="Transpose_56377_compressed" type="Const" version="opset1">
<data element_type="f16" shape="8, 8" offset="4017869" size="128" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/dense/Add, /transformer/h.1/self_attention/dense/MatMul, onnx::MatMul_529" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>8</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="341" name="Transpose_56377" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/dense/Add, /transformer/h.1/self_attention/dense/MatMul, onnx::MatMul_529" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>8</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>8</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="342" name="/transformer/h.1/self_attention/dense/MatMul" type="MatMul" version="opset1">
<data transpose_a="false" transpose_b="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/dense/Add, /transformer/h.1/self_attention/dense/MatMul" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="FP32">
<dim>8</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/self_attention/dense/Add_output_0,/transformer/h.1/self_attention/dense/MatMul_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="343" name="/transformer/h.1/self_attention/Add_1" type="Add" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/self_attention/Add_1" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/self_attention/Add_1_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="344" name="Constant_31645" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014088" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_31645" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="345" name="/transformer/h.1/post_attention_layernorm/Div" type="MVN" version="opset6">
<data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/post_attention_layernorm/Add, /transformer/h.1/post_attention_layernorm/Add_1, /transformer/h.1/post_attention_layernorm/Div, /transformer/h.1/post_attention_layernorm/Mul, /transformer/h.1/post_attention_layernorm/Pow, /transformer/h.1/post_attention_layernorm/ReduceMean, /transformer/h.1/post_attention_layernorm/ReduceMean_1, /transformer/h.1/post_attention_layernorm/Sqrt, /transformer/h.1/post_attention_layernorm/Sub" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/post_attention_layernorm/Add_1_output_0,/transformer/h.1/post_attention_layernorm/Div_output_0,/transformer/h.1/post_attention_layernorm/Mul_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="346" name="Multiply_56404_compressed" type="Const" version="opset1">
<data element_type="f16" shape="32, 8" offset="4017997" size="512" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/dense_h_to_4h/Add, /transformer/h.1/mlp/dense_h_to_4h/MatMul, /transformer/h.1/mlp/gelu_impl/Constant, /transformer/h.1/mlp/gelu_impl/Mul, onnx::MatMul_530" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>32</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="347" name="Multiply_56404" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/dense_h_to_4h/Add, /transformer/h.1/mlp/dense_h_to_4h/MatMul, /transformer/h.1/mlp/gelu_impl/Constant, /transformer/h.1/mlp/gelu_impl/Mul, onnx::MatMul_530" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>32</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>32</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="348" name="/transformer/h.1/mlp/gelu_impl/Mul" type="MatMul" version="opset1">
<data transpose_a="false" transpose_b="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/dense_h_to_4h/Add, /transformer/h.1/mlp/dense_h_to_4h/MatMul, /transformer/h.1/mlp/gelu_impl/Mul" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="FP32">
<dim>32</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/mlp/gelu_impl/Mul_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="349" name="Multiply_56406_compressed" type="Const" version="opset1">
<data element_type="f16" shape="32, 8" offset="4018509" size="512" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/dense_h_to_4h/Add, /transformer/h.1/mlp/dense_h_to_4h/MatMul, /transformer/h.1/mlp/gelu_impl/Constant_1, /transformer/h.1/mlp/gelu_impl/Mul_1, onnx::MatMul_530" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>32</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="350" name="Multiply_56406" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/dense_h_to_4h/Add, /transformer/h.1/mlp/dense_h_to_4h/MatMul, /transformer/h.1/mlp/gelu_impl/Constant_1, /transformer/h.1/mlp/gelu_impl/Mul_1, onnx::MatMul_530" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>32</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>32</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="351" name="/transformer/h.1/mlp/gelu_impl/Mul_1" type="MatMul" version="opset1">
<data transpose_a="false" transpose_b="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/dense_h_to_4h/Add, /transformer/h.1/mlp/dense_h_to_4h/MatMul, /transformer/h.1/mlp/gelu_impl/Mul_1" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="FP32">
<dim>32</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/mlp/gelu_impl/Mul_1_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="352" name="Multiply_56408_compressed" type="Const" version="opset1">
<data element_type="f16" shape="32, 8" offset="4019021" size="512" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/dense_h_to_4h/Add, /transformer/h.1/mlp/dense_h_to_4h/MatMul, /transformer/h.1/mlp/gelu_impl/Constant_2, /transformer/h.1/mlp/gelu_impl/Mul_2, onnx::MatMul_530" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>32</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="353" name="Multiply_56408" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/dense_h_to_4h/Add, /transformer/h.1/mlp/dense_h_to_4h/MatMul, /transformer/h.1/mlp/gelu_impl/Constant_2, /transformer/h.1/mlp/gelu_impl/Mul_2, onnx::MatMul_530" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>32</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>32</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="354" name="/transformer/h.1/mlp/gelu_impl/Mul_2" type="MatMul" version="opset1">
<data transpose_a="false" transpose_b="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/dense_h_to_4h/Add, /transformer/h.1/mlp/dense_h_to_4h/MatMul, /transformer/h.1/mlp/gelu_impl/Mul_2" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="FP32">
<dim>32</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/mlp/gelu_impl/Mul_2_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="355" name="Transpose_56381_compressed" type="Const" version="opset1">
<data element_type="f16" shape="32, 8" offset="4019533" size="512" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/dense_h_to_4h/Add, /transformer/h.1/mlp/dense_h_to_4h/MatMul, onnx::MatMul_530" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>32</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="356" name="Transpose_56381" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/dense_h_to_4h/Add, /transformer/h.1/mlp/dense_h_to_4h/MatMul, onnx::MatMul_530" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>32</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>32</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="357" name="/transformer/h.1/mlp/dense_h_to_4h/MatMul" type="MatMul" version="opset1">
<data transpose_a="false" transpose_b="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/dense_h_to_4h/Add, /transformer/h.1/mlp/dense_h_to_4h/MatMul" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="FP32">
<dim>32</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/mlp/dense_h_to_4h/Add_output_0,/transformer/h.1/mlp/dense_h_to_4h/MatMul_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="358" name="/transformer/h.1/mlp/gelu_impl/Mul_3" type="Multiply" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/gelu_impl/Mul_3" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/mlp/gelu_impl/Mul_3_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="359" name="Constant_56560_compressed" type="Const" version="opset1">
<data element_type="f16" shape="1, 1, 1" offset="4016971" size="2" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/gelu_impl/Constant_3" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</output>
</layer>
<layer id="360" name="Constant_56560" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/gelu_impl/Constant_3" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</output>
</layer>
<layer id="361" name="/transformer/h.1/mlp/gelu_impl/Add" type="Add" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/gelu_impl/Add" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
<port id="1" precision="FP32">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/mlp/gelu_impl/Add_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="362" name="/transformer/h.1/mlp/gelu_impl/Mul_4" type="Multiply" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/gelu_impl/Mul_4" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/mlp/gelu_impl/Mul_4_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="363" name="/transformer/h.1/mlp/gelu_impl/Tanh" type="Tanh" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/gelu_impl/Tanh" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</input>
<output>
<port id="1" precision="FP32" names="/transformer/h.1/mlp/gelu_impl/Tanh_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="364" name="Constant_56561_compressed" type="Const" version="opset1">
<data element_type="f16" shape="1, 1, 1" offset="4016971" size="2" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/gelu_impl/Constant_4" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</output>
</layer>
<layer id="365" name="Constant_56561" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/gelu_impl/Constant_4" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</output>
</layer>
<layer id="366" name="/transformer/h.1/mlp/gelu_impl/Add_1" type="Add" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/gelu_impl/Add_1" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
<port id="1" precision="FP32">
<dim>1</dim>
<dim>1</dim>
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/mlp/gelu_impl/Add_1_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="367" name="/transformer/h.1/mlp/gelu_impl/Mul_5" type="Multiply" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/gelu_impl/Mul_5" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/mlp/gelu_impl/Mul_5_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="368" name="Transpose_56384_compressed" type="Const" version="opset1">
<data element_type="f16" shape="8, 32" offset="4020045" size="512" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/dense_4h_to_h/Add, /transformer/h.1/mlp/dense_4h_to_h/MatMul, onnx::MatMul_531" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>8</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="369" name="Transpose_56384" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/dense_4h_to_h/Add, /transformer/h.1/mlp/dense_4h_to_h/MatMul, onnx::MatMul_531" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>8</dim>
<dim>32</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>8</dim>
<dim>32</dim>
</port>
</output>
</layer>
<layer id="370" name="/transformer/h.1/mlp/dense_4h_to_h/MatMul" type="MatMul" version="opset1">
<data transpose_a="false" transpose_b="true" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/dense_4h_to_h/Add, /transformer/h.1/mlp/dense_4h_to_h/MatMul" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>32</dim>
</port>
<port id="1" precision="FP32">
<dim>8</dim>
<dim>32</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/mlp/dense_4h_to_h/Add_output_0,/transformer/h.1/mlp/dense_4h_to_h/MatMul_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="371" name="/transformer/h.1/mlp/Add" type="Add" version="opset1">
<data auto_broadcast="numpy" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/h.1/mlp/Add" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/h.1/mlp/Add_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="372" name="Constant_31677" type="Const" version="opset1">
<data element_type="i64" shape="1" offset="4014088" size="8" />
<rt_info>
<attribute name="fused_names" version="0" value="Constant_31677" />
</rt_info>
<output>
<port id="0" precision="I64">
<dim>1</dim>
</port>
</output>
</layer>
<layer id="373" name="/transformer/ln_f/Div" type="MVN" version="opset6">
<data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
<rt_info>
<attribute name="fused_names" version="0" value="/transformer/ln_f/Add, /transformer/ln_f/Add_1, /transformer/ln_f/Div, /transformer/ln_f/Mul, /transformer/ln_f/Pow, /transformer/ln_f/ReduceMean, /transformer/ln_f/ReduceMean_1, /transformer/ln_f/Sqrt, /transformer/ln_f/Sub" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="I64">
<dim>1</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="/transformer/ln_f/Add_1_output_0,/transformer/ln_f/Div_output_0,/transformer/ln_f/Mul_output_0">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="374" name="Transpose_56388_compressed" type="Const" version="opset1">
<data element_type="f16" shape="250880, 8" offset="0" size="4014080" />
<rt_info>
<attribute name="fused_names" version="0" value="logits, onnx::MatMul_532" />
</rt_info>
<output>
<port id="0" precision="FP16">
<dim>250880</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="375" name="Transpose_56388" type="Convert" version="opset1">
<data destination_type="f32" />
<rt_info>
<attribute name="decompression" version="0" />
<attribute name="fused_names" version="0" value="logits, onnx::MatMul_532" />
</rt_info>
<input>
<port id="0" precision="FP16">
<dim>250880</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="1" precision="FP32">
<dim>250880</dim>
<dim>8</dim>
</port>
</output>
</layer>
<layer id="376" name="logits" type="MatMul" version="opset1">
<data transpose_a="false" transpose_b="true" />
<rt_info>
<attribute name="fused_names" version="0" value="logits" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>8</dim>
</port>
<port id="1" precision="FP32">
<dim>250880</dim>
<dim>8</dim>
</port>
</input>
<output>
<port id="2" precision="FP32" names="logits">
<dim>-1</dim>
<dim>-1</dim>
<dim>250880</dim>
</port>
</output>
</layer>
<layer id="377" name="logits/sink_port_0" type="Result" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="logits/sink_port_0" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>250880</dim>
</port>
</input>
</layer>
<layer id="296" name="present.0.key/sink_port_0" type="Result" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="present.0.key/sink_port_0" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
</input>
</layer>
<layer id="295" name="present.0.value/sink_port_0" type="Result" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="present.0.value/sink_port_0" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</input>
</layer>
<layer id="294" name="present.1.key/sink_port_0" type="Result" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="present.1.key/sink_port_0" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>4</dim>
<dim>-1</dim>
</port>
</input>
</layer>
<layer id="285" name="present.1.value/sink_port_0" type="Result" version="opset1">
<rt_info>
<attribute name="fused_names" version="0" value="present.1.value/sink_port_0" />
</rt_info>
<input>
<port id="0" precision="FP32">
<dim>-1</dim>
<dim>-1</dim>
<dim>4</dim>
</port>
</input>
</layer>
</layers>
<edges>
<edge from-layer="0" from-port="0" to-layer="168" to-port="0" />
<edge from-layer="0" from-port="0" to-layer="171" to-port="1" />
<edge from-layer="0" from-port="0" to-layer="13" to-port="0" />
<edge from-layer="0" from-port="0" to-layer="18" to-port="0" />
<edge from-layer="1" from-port="0" to-layer="284" to-port="0" />
<edge from-layer="2" from-port="0" to-layer="293" to-port="0" />
<edge from-layer="3" from-port="0" to-layer="203" to-port="0" />
<edge from-layer="4" from-port="0" to-layer="160" to-port="0" />
<edge from-layer="4" from-port="0" to-layer="43" to-port="0" />
<edge from-layer="5" from-port="0" to-layer="9" to-port="1" />
<edge from-layer="5" from-port="0" to-layer="25" to-port="0" />
<edge from-layer="6" from-port="0" to-layer="7" to-port="0" />
<edge from-layer="7" from-port="1" to-layer="9" to-port="0" />
<edge from-layer="8" from-port="0" to-layer="9" to-port="2" />
<edge from-layer="9" from-port="3" to-layer="11" to-port="0" />
<edge from-layer="10" from-port="0" to-layer="11" to-port="1" />
<edge from-layer="11" from-port="2" to-layer="127" to-port="0" />
<edge from-layer="11" from-port="2" to-layer="228" to-port="0" />
<edge from-layer="12" from-port="0" to-layer="13" to-port="1" />
<edge from-layer="13" from-port="2" to-layer="15" to-port="0" />
<edge from-layer="14" from-port="0" to-layer="15" to-port="1" />
<edge from-layer="15" from-port="2" to-layer="16" to-port="0" />
<edge from-layer="16" from-port="1" to-layer="17" to-port="0" />
<edge from-layer="17" from-port="1" to-layer="41" to-port="0" />
<edge from-layer="18" from-port="1" to-layer="33" to-port="0" />
<edge from-layer="18" from-port="1" to-layer="21" to-port="0" />
<edge from-layer="19" from-port="0" to-layer="21" to-port="1" />
<edge from-layer="20" from-port="0" to-layer="21" to-port="2" />
<edge from-layer="21" from-port="3" to-layer="177" to-port="0" />
<edge from-layer="21" from-port="3" to-layer="23" to-port="0" />
<edge from-layer="22" from-port="0" to-layer="23" to-port="1" />
<edge from-layer="23" from-port="2" to-layer="34" to-port="0" />
<edge from-layer="24" from-port="0" to-layer="34" to-port="1" />
<edge from-layer="25" from-port="1" to-layer="28" to-port="0" />
<edge from-layer="25" from-port="1" to-layer="114" to-port="0" />
<edge from-layer="26" from-port="0" to-layer="28" to-port="1" />
<edge from-layer="27" from-port="0" to-layer="28" to-port="2" />
<edge from-layer="28" from-port="3" to-layer="47" to-port="0" />
<edge from-layer="28" from-port="3" to-layer="91" to-port="1" />
<edge from-layer="28" from-port="3" to-layer="30" to-port="0" />
<edge from-layer="29" from-port="0" to-layer="30" to-port="1" />
<edge from-layer="30" from-port="2" to-layer="116" to-port="2" />
<edge from-layer="30" from-port="2" to-layer="34" to-port="2" />
<edge from-layer="30" from-port="2" to-layer="50" to-port="0" />
<edge from-layer="31" from-port="0" to-layer="33" to-port="1" />
<edge from-layer="32" from-port="0" to-layer="33" to-port="2" />
<edge from-layer="33" from-port="3" to-layer="34" to-port="3" />
<edge from-layer="33" from-port="3" to-layer="181" to-port="2" />
<edge from-layer="34" from-port="4" to-layer="36" to-port="0" />
<edge from-layer="35" from-port="0" to-layer="36" to-port="1" />
<edge from-layer="36" from-port="2" to-layer="40" to-port="2" />
<edge from-layer="36" from-port="2" to-layer="38" to-port="0" />
<edge from-layer="37" from-port="0" to-layer="38" to-port="1" />
<edge from-layer="38" from-port="2" to-layer="40" to-port="0" />
<edge from-layer="39" from-port="0" to-layer="40" to-port="1" />
<edge from-layer="40" from-port="3" to-layer="41" to-port="1" />
<edge from-layer="41" from-port="2" to-layer="124" to-port="0" />
<edge from-layer="42" from-port="0" to-layer="51" to-port="0" />
<edge from-layer="43" from-port="1" to-layer="46" to-port="0" />
<edge from-layer="44" from-port="0" to-layer="46" to-port="1" />
<edge from-layer="45" from-port="0" to-layer="46" to-port="2" />
<edge from-layer="46" from-port="3" to-layer="68" to-port="0" />
<edge from-layer="46" from-port="3" to-layer="47" to-port="1" />
<edge from-layer="47" from-port="2" to-layer="49" to-port="0" />
<edge from-layer="48" from-port="0" to-layer="49" to-port="1" />
<edge from-layer="49" from-port="2" to-layer="50" to-port="1" />
<edge from-layer="49" from-port="2" to-layer="116" to-port="3" />
<edge from-layer="50" from-port="2" to-layer="51" to-port="1" />
<edge from-layer="51" from-port="2" to-layer="107" to-port="0" />
<edge from-layer="51" from-port="2" to-layer="53" to-port="0" />
<edge from-layer="51" from-port="2" to-layer="103" to-port="0" />
<edge from-layer="52" from-port="0" to-layer="58" to-port="0" />
<edge from-layer="53" from-port="1" to-layer="56" to-port="0" />
<edge from-layer="53" from-port="1" to-layer="64" to-port="0" />
<edge from-layer="54" from-port="0" to-layer="56" to-port="1" />
<edge from-layer="55" from-port="0" to-layer="56" to-port="2" />
<edge from-layer="56" from-port="3" to-layer="58" to-port="1" />
<edge from-layer="57" from-port="0" to-layer="58" to-port="2" />
<edge from-layer="58" from-port="3" to-layer="60" to-port="0" />
<edge from-layer="59" from-port="0" to-layer="60" to-port="1" />
<edge from-layer="60" from-port="2" to-layer="72" to-port="0" />
<edge from-layer="60" from-port="2" to-layer="78" to-port="0" />
<edge from-layer="61" from-port="0" to-layer="66" to-port="0" />
<edge from-layer="62" from-port="0" to-layer="64" to-port="1" />
<edge from-layer="63" from-port="0" to-layer="64" to-port="2" />
<edge from-layer="64" from-port="3" to-layer="66" to-port="1" />
<edge from-layer="65" from-port="0" to-layer="66" to-port="2" />
<edge from-layer="66" from-port="3" to-layer="71" to-port="0" />
<edge from-layer="67" from-port="0" to-layer="68" to-port="1" />
<edge from-layer="68" from-port="2" to-layer="71" to-port="1" />
<edge from-layer="68" from-port="2" to-layer="100" to-port="2" />
<edge from-layer="69" from-port="0" to-layer="71" to-port="2" />
<edge from-layer="70" from-port="0" to-layer="71" to-port="3" />
<edge from-layer="71" from-port="4" to-layer="85" to-port="0" />
<edge from-layer="71" from-port="4" to-layer="72" to-port="1" />
<edge from-layer="72" from-port="2" to-layer="73" to-port="0" />
<edge from-layer="73" from-port="1" to-layer="82" to-port="0" />
<edge from-layer="73" from-port="1" to-layer="84" to-port="2" />
<edge from-layer="73" from-port="1" to-layer="75" to-port="0" />
<edge from-layer="73" from-port="1" to-layer="77" to-port="2" />
<edge from-layer="73" from-port="1" to-layer="106" to-port="1" />
<edge from-layer="74" from-port="0" to-layer="75" to-port="1" />
<edge from-layer="75" from-port="2" to-layer="77" to-port="0" />
<edge from-layer="76" from-port="0" to-layer="77" to-port="1" />
<edge from-layer="77" from-port="3" to-layer="78" to-port="1" />
<edge from-layer="78" from-port="2" to-layer="80" to-port="0" />
<edge from-layer="79" from-port="0" to-layer="80" to-port="1" />
<edge from-layer="80" from-port="2" to-layer="88" to-port="0" />
<edge from-layer="81" from-port="0" to-layer="82" to-port="1" />
<edge from-layer="82" from-port="2" to-layer="84" to-port="0" />
<edge from-layer="83" from-port="0" to-layer="84" to-port="1" />
<edge from-layer="84" from-port="3" to-layer="85" to-port="1" />
<edge from-layer="85" from-port="2" to-layer="87" to-port="0" />
<edge from-layer="86" from-port="0" to-layer="87" to-port="1" />
<edge from-layer="87" from-port="2" to-layer="88" to-port="1" />
<edge from-layer="88" from-port="2" to-layer="107" to-port="1" />
<edge from-layer="89" from-port="0" to-layer="91" to-port="0" />
<edge from-layer="90" from-port="0" to-layer="91" to-port="2" />
<edge from-layer="91" from-port="3" to-layer="95" to-port="0" />
<edge from-layer="91" from-port="3" to-layer="93" to-port="0" />
<edge from-layer="92" from-port="0" to-layer="93" to-port="1" />
<edge from-layer="93" from-port="2" to-layer="96" to-port="0" />
<edge from-layer="94" from-port="0" to-layer="95" to-port="1" />
<edge from-layer="95" from-port="2" to-layer="96" to-port="1" />
<edge from-layer="96" from-port="2" to-layer="105" to-port="0" />
<edge from-layer="97" from-port="0" to-layer="100" to-port="0" />
<edge from-layer="98" from-port="0" to-layer="100" to-port="1" />
<edge from-layer="99" from-port="0" to-layer="100" to-port="3" />
<edge from-layer="100" from-port="4" to-layer="103" to-port="1" />
<edge from-layer="101" from-port="0" to-layer="103" to-port="2" />
<edge from-layer="102" from-port="0" to-layer="103" to-port="3" />
<edge from-layer="103" from-port="4" to-layer="104" to-port="0" />
<edge from-layer="104" from-port="1" to-layer="105" to-port="1" />
<edge from-layer="105" from-port="2" to-layer="106" to-port="0" />
<edge from-layer="106" from-port="2" to-layer="107" to-port="2" />
<edge from-layer="107" from-port="3" to-layer="109" to-port="0" />
<edge from-layer="108" from-port="0" to-layer="109" to-port="1" />
<edge from-layer="109" from-port="2" to-layer="111" to-port="0" />
<edge from-layer="110" from-port="0" to-layer="111" to-port="1" />
<edge from-layer="111" from-port="2" to-layer="123" to-port="0" />
<edge from-layer="112" from-port="0" to-layer="114" to-port="1" />
<edge from-layer="113" from-port="0" to-layer="114" to-port="2" />
<edge from-layer="114" from-port="3" to-layer="116" to-port="0" />
<edge from-layer="115" from-port="0" to-layer="116" to-port="1" />
<edge from-layer="116" from-port="4" to-layer="118" to-port="0" />
<edge from-layer="117" from-port="0" to-layer="118" to-port="1" />
<edge from-layer="118" from-port="2" to-layer="120" to-port="0" />
<edge from-layer="118" from-port="2" to-layer="122" to-port="2" />
<edge from-layer="119" from-port="0" to-layer="120" to-port="1" />
<edge from-layer="120" from-port="2" to-layer="122" to-port="0" />
<edge from-layer="121" from-port="0" to-layer="122" to-port="1" />
<edge from-layer="122" from-port="3" to-layer="123" to-port="1" />
<edge from-layer="123" from-port="2" to-layer="124" to-port="1" />
<edge from-layer="124" from-port="2" to-layer="315" to-port="0" />
<edge from-layer="124" from-port="2" to-layer="193" to-port="0" />
<edge from-layer="125" from-port="0" to-layer="193" to-port="1" />
<edge from-layer="126" from-port="0" to-layer="127" to-port="1" />
<edge from-layer="127" from-port="2" to-layer="130" to-port="0" />
<edge from-layer="128" from-port="0" to-layer="129" to-port="0" />
<edge from-layer="129" from-port="1" to-layer="130" to-port="1" />
<edge from-layer="130" from-port="2" to-layer="132" to-port="0" />
<edge from-layer="131" from-port="0" to-layer="132" to-port="1" />
<edge from-layer="132" from-port="2" to-layer="135" to-port="0" />
<edge from-layer="132" from-port="2" to-layer="199" to-port="0" />
<edge from-layer="132" from-port="2" to-layer="154" to-port="0" />
<edge from-layer="133" from-port="0" to-layer="135" to-port="1" />
<edge from-layer="133" from-port="0" to-layer="269" to-port="1" />
<edge from-layer="134" from-port="0" to-layer="135" to-port="2" />
<edge from-layer="135" from-port="3" to-layer="138" to-port="0" />
<edge from-layer="135" from-port="3" to-layer="137" to-port="0" />
<edge from-layer="136" from-port="0" to-layer="137" to-port="1" />
<edge from-layer="137" from-port="2" to-layer="151" to-port="0" />
<edge from-layer="138" from-port="1" to-layer="148" to-port="0" />
<edge from-layer="138" from-port="1" to-layer="141" to-port="0" />
<edge from-layer="139" from-port="0" to-layer="141" to-port="1" />
<edge from-layer="140" from-port="0" to-layer="141" to-port="2" />
<edge from-layer="141" from-port="3" to-layer="143" to-port="0" />
<edge from-layer="141" from-port="3" to-layer="185" to-port="0" />
<edge from-layer="142" from-port="0" to-layer="143" to-port="1" />
<edge from-layer="143" from-port="2" to-layer="145" to-port="0" />
<edge from-layer="144" from-port="0" to-layer="145" to-port="1" />
<edge from-layer="145" from-port="2" to-layer="195" to-port="0" />
<edge from-layer="145" from-port="2" to-layer="150" to-port="0" />
<edge from-layer="145" from-port="2" to-layer="158" to-port="0" />
<edge from-layer="146" from-port="0" to-layer="148" to-port="1" />
<edge from-layer="147" from-port="0" to-layer="148" to-port="2" />
<edge from-layer="148" from-port="3" to-layer="150" to-port="1" />
<edge from-layer="148" from-port="3" to-layer="195" to-port="1" />
<edge from-layer="148" from-port="3" to-layer="191" to-port="2" />
<edge from-layer="148" from-port="3" to-layer="158" to-port="2" />
<edge from-layer="149" from-port="0" to-layer="150" to-port="2" />
<edge from-layer="150" from-port="3" to-layer="151" to-port="1" />
<edge from-layer="150" from-port="3" to-layer="202" to-port="1" />
<edge from-layer="151" from-port="2" to-layer="164" to-port="0" />
<edge from-layer="152" from-port="0" to-layer="287" to-port="1" />
<edge from-layer="152" from-port="0" to-layer="154" to-port="1" />
<edge from-layer="153" from-port="0" to-layer="154" to-port="2" />
<edge from-layer="154" from-port="3" to-layer="156" to-port="0" />
<edge from-layer="155" from-port="0" to-layer="156" to-port="1" />
<edge from-layer="156" from-port="2" to-layer="159" to-port="0" />
<edge from-layer="157" from-port="0" to-layer="158" to-port="1" />
<edge from-layer="158" from-port="3" to-layer="159" to-port="1" />
<edge from-layer="159" from-port="2" to-layer="160" to-port="1" />
<edge from-layer="160" from-port="2" to-layer="296" to-port="0" />
<edge from-layer="160" from-port="2" to-layer="163" to-port="0" />
<edge from-layer="160" from-port="2" to-layer="187" to-port="0" />
<edge from-layer="161" from-port="0" to-layer="162" to-port="0" />
<edge from-layer="162" from-port="1" to-layer="163" to-port="1" />
<edge from-layer="163" from-port="2" to-layer="164" to-port="1" />
<edge from-layer="164" from-port="2" to-layer="183" to-port="0" />
<edge from-layer="165" from-port="0" to-layer="166" to-port="0" />
<edge from-layer="166" from-port="1" to-layer="175" to-port="0" />
<edge from-layer="167" from-port="0" to-layer="168" to-port="1" />
<edge from-layer="168" from-port="2" to-layer="170" to-port="0" />
<edge from-layer="169" from-port="0" to-layer="170" to-port="1" />
<edge from-layer="170" from-port="2" to-layer="171" to-port="0" />
<edge from-layer="171" from-port="2" to-layer="173" to-port="0" />
<edge from-layer="172" from-port="0" to-layer="173" to-port="1" />
<edge from-layer="173" from-port="2" to-layer="174" to-port="0" />
<edge from-layer="174" from-port="1" to-layer="175" to-port="1" />
<edge from-layer="175" from-port="2" to-layer="182" to-port="0" />
<edge from-layer="176" from-port="0" to-layer="177" to-port="1" />
<edge from-layer="177" from-port="2" to-layer="179" to-port="0" />
<edge from-layer="178" from-port="0" to-layer="179" to-port="1" />
<edge from-layer="179" from-port="2" to-layer="181" to-port="0" />
<edge from-layer="180" from-port="0" to-layer="181" to-port="1" />
<edge from-layer="181" from-port="3" to-layer="182" to-port="1" />
<edge from-layer="182" from-port="2" to-layer="183" to-port="1" />
<edge from-layer="182" from-port="2" to-layer="305" to-port="1" />
<edge from-layer="183" from-port="2" to-layer="192" to-port="0" />
<edge from-layer="184" from-port="0" to-layer="185" to-port="1" />
<edge from-layer="185" from-port="2" to-layer="191" to-port="0" />
<edge from-layer="186" from-port="0" to-layer="191" to-port="1" />
<edge from-layer="187" from-port="1" to-layer="190" to-port="0" />
<edge from-layer="188" from-port="0" to-layer="190" to-port="1" />
<edge from-layer="189" from-port="0" to-layer="190" to-port="2" />
<edge from-layer="190" from-port="3" to-layer="191" to-port="3" />
<edge from-layer="190" from-port="3" to-layer="195" to-port="2" />
<edge from-layer="191" from-port="4" to-layer="192" to-port="1" />
<edge from-layer="192" from-port="2" to-layer="193" to-port="2" />
<edge from-layer="193" from-port="3" to-layer="194" to-port="0" />
<edge from-layer="194" from-port="1" to-layer="196" to-port="0" />
<edge from-layer="195" from-port="3" to-layer="196" to-port="1" />
<edge from-layer="196" from-port="2" to-layer="204" to-port="0" />
<edge from-layer="197" from-port="0" to-layer="265" to-port="1" />
<edge from-layer="197" from-port="0" to-layer="199" to-port="1" />
<edge from-layer="198" from-port="0" to-layer="199" to-port="2" />
<edge from-layer="199" from-port="3" to-layer="201" to-port="0" />
<edge from-layer="200" from-port="0" to-layer="201" to-port="1" />
<edge from-layer="201" from-port="2" to-layer="202" to-port="0" />
<edge from-layer="202" from-port="2" to-layer="203" to-port="1" />
<edge from-layer="203" from-port="2" to-layer="295" to-port="0" />
<edge from-layer="203" from-port="2" to-layer="204" to-port="1" />
<edge from-layer="204" from-port="2" to-layer="219" to-port="0" />
<edge from-layer="204" from-port="2" to-layer="205" to-port="0" />
<edge from-layer="205" from-port="1" to-layer="216" to-port="0" />
<edge from-layer="205" from-port="1" to-layer="208" to-port="0" />
<edge from-layer="206" from-port="0" to-layer="208" to-port="1" />
<edge from-layer="207" from-port="0" to-layer="208" to-port="2" />
<edge from-layer="208" from-port="3" to-layer="210" to-port="0" />
<edge from-layer="209" from-port="0" to-layer="210" to-port="1" />
<edge from-layer="210" from-port="2" to-layer="212" to-port="0" />
<edge from-layer="211" from-port="0" to-layer="212" to-port="1" />
<edge from-layer="212" from-port="2" to-layer="218" to-port="0" />
<edge from-layer="212" from-port="2" to-layer="223" to-port="0" />
<edge from-layer="213" from-port="0" to-layer="218" to-port="1" />
<edge from-layer="214" from-port="0" to-layer="216" to-port="1" />
<edge from-layer="215" from-port="0" to-layer="216" to-port="2" />
<edge from-layer="216" from-port="3" to-layer="223" to-port="1" />
<edge from-layer="216" from-port="3" to-layer="218" to-port="2" />
<edge from-layer="217" from-port="0" to-layer="218" to-port="3" />
<edge from-layer="218" from-port="4" to-layer="219" to-port="1" />
<edge from-layer="219" from-port="2" to-layer="221" to-port="0" />
<edge from-layer="220" from-port="0" to-layer="221" to-port="1" />
<edge from-layer="221" from-port="2" to-layer="224" to-port="0" />
<edge from-layer="222" from-port="0" to-layer="223" to-port="2" />
<edge from-layer="223" from-port="3" to-layer="224" to-port="1" />
<edge from-layer="224" from-port="2" to-layer="227" to-port="0" />
<edge from-layer="225" from-port="0" to-layer="226" to-port="0" />
<edge from-layer="226" from-port="1" to-layer="227" to-port="1" />
<edge from-layer="227" from-port="2" to-layer="228" to-port="1" />
<edge from-layer="228" from-port="2" to-layer="230" to-port="0" />
<edge from-layer="228" from-port="2" to-layer="256" to-port="0" />
<edge from-layer="229" from-port="0" to-layer="230" to-port="1" />
<edge from-layer="230" from-port="2" to-layer="236" to-port="0" />
<edge from-layer="230" from-port="2" to-layer="239" to-port="0" />
<edge from-layer="230" from-port="2" to-layer="233" to-port="0" />
<edge from-layer="230" from-port="2" to-layer="242" to-port="0" />
<edge from-layer="231" from-port="0" to-layer="232" to-port="0" />
<edge from-layer="232" from-port="1" to-layer="233" to-port="1" />
<edge from-layer="233" from-port="2" to-layer="252" to-port="0" />
<edge from-layer="234" from-port="0" to-layer="235" to-port="0" />
<edge from-layer="235" from-port="1" to-layer="236" to-port="1" />
<edge from-layer="236" from-port="2" to-layer="247" to-port="0" />
<edge from-layer="237" from-port="0" to-layer="238" to-port="0" />
<edge from-layer="238" from-port="1" to-layer="239" to-port="1" />
<edge from-layer="239" from-port="2" to-layer="243" to-port="0" />
<edge from-layer="240" from-port="0" to-layer="241" to-port="0" />
<edge from-layer="241" from-port="1" to-layer="242" to-port="1" />
<edge from-layer="242" from-port="2" to-layer="243" to-port="1" />
<edge from-layer="243" from-port="2" to-layer="246" to-port="0" />
<edge from-layer="244" from-port="0" to-layer="245" to-port="0" />
<edge from-layer="245" from-port="1" to-layer="246" to-port="1" />
<edge from-layer="246" from-port="2" to-layer="247" to-port="1" />
<edge from-layer="247" from-port="2" to-layer="248" to-port="0" />
<edge from-layer="248" from-port="1" to-layer="251" to-port="0" />
<edge from-layer="249" from-port="0" to-layer="250" to-port="0" />
<edge from-layer="250" from-port="1" to-layer="251" to-port="1" />
<edge from-layer="251" from-port="2" to-layer="252" to-port="1" />
<edge from-layer="252" from-port="2" to-layer="255" to-port="0" />
<edge from-layer="253" from-port="0" to-layer="254" to-port="0" />
<edge from-layer="254" from-port="1" to-layer="255" to-port="1" />
<edge from-layer="255" from-port="2" to-layer="256" to-port="1" />
<edge from-layer="256" from-port="2" to-layer="258" to-port="0" />
<edge from-layer="256" from-port="2" to-layer="343" to-port="0" />
<edge from-layer="257" from-port="0" to-layer="258" to-port="1" />
<edge from-layer="258" from-port="2" to-layer="261" to-port="0" />
<edge from-layer="259" from-port="0" to-layer="260" to-port="0" />
<edge from-layer="260" from-port="1" to-layer="261" to-port="1" />
<edge from-layer="261" from-port="2" to-layer="263" to-port="0" />
<edge from-layer="262" from-port="0" to-layer="263" to-port="1" />
<edge from-layer="263" from-port="2" to-layer="269" to-port="0" />
<edge from-layer="263" from-port="2" to-layer="287" to-port="0" />
<edge from-layer="263" from-port="2" to-layer="265" to-port="0" />
<edge from-layer="264" from-port="0" to-layer="265" to-port="2" />
<edge from-layer="265" from-port="3" to-layer="267" to-port="0" />
<edge from-layer="266" from-port="0" to-layer="267" to-port="1" />
<edge from-layer="267" from-port="2" to-layer="283" to-port="0" />
<edge from-layer="268" from-port="0" to-layer="269" to-port="2" />
<edge from-layer="269" from-port="3" to-layer="270" to-port="0" />
<edge from-layer="269" from-port="3" to-layer="299" to-port="0" />
<edge from-layer="270" from-port="1" to-layer="273" to-port="0" />
<edge from-layer="270" from-port="1" to-layer="280" to-port="0" />
<edge from-layer="271" from-port="0" to-layer="273" to-port="1" />
<edge from-layer="272" from-port="0" to-layer="273" to-port="2" />
<edge from-layer="273" from-port="3" to-layer="307" to-port="0" />
<edge from-layer="273" from-port="3" to-layer="275" to-port="0" />
<edge from-layer="274" from-port="0" to-layer="275" to-port="1" />
<edge from-layer="275" from-port="2" to-layer="277" to-port="0" />
<edge from-layer="276" from-port="0" to-layer="277" to-port="1" />
<edge from-layer="277" from-port="2" to-layer="282" to-port="0" />
<edge from-layer="277" from-port="2" to-layer="317" to-port="0" />
<edge from-layer="277" from-port="2" to-layer="291" to-port="0" />
<edge from-layer="278" from-port="0" to-layer="280" to-port="1" />
<edge from-layer="279" from-port="0" to-layer="280" to-port="2" />
<edge from-layer="280" from-port="3" to-layer="291" to-port="2" />
<edge from-layer="280" from-port="3" to-layer="313" to-port="2" />
<edge from-layer="280" from-port="3" to-layer="282" to-port="1" />
<edge from-layer="280" from-port="3" to-layer="317" to-port="1" />
<edge from-layer="281" from-port="0" to-layer="282" to-port="2" />
<edge from-layer="282" from-port="3" to-layer="283" to-port="1" />
<edge from-layer="282" from-port="3" to-layer="300" to-port="1" />
<edge from-layer="283" from-port="2" to-layer="284" to-port="1" />
<edge from-layer="284" from-port="2" to-layer="285" to-port="0" />
<edge from-layer="284" from-port="2" to-layer="319" to-port="1" />
<edge from-layer="286" from-port="0" to-layer="287" to-port="2" />
<edge from-layer="287" from-port="3" to-layer="289" to-port="0" />
<edge from-layer="288" from-port="0" to-layer="289" to-port="1" />
<edge from-layer="289" from-port="2" to-layer="292" to-port="0" />
<edge from-layer="290" from-port="0" to-layer="291" to-port="1" />
<edge from-layer="291" from-port="3" to-layer="292" to-port="1" />
<edge from-layer="292" from-port="2" to-layer="293" to-port="1" />
<edge from-layer="293" from-port="2" to-layer="303" to-port="0" />
<edge from-layer="293" from-port="2" to-layer="309" to-port="0" />
<edge from-layer="293" from-port="2" to-layer="294" to-port="0" />
<edge from-layer="297" from-port="0" to-layer="315" to-port="1" />
<edge from-layer="298" from-port="0" to-layer="299" to-port="1" />
<edge from-layer="299" from-port="2" to-layer="300" to-port="0" />
<edge from-layer="300" from-port="2" to-layer="304" to-port="0" />
<edge from-layer="301" from-port="0" to-layer="302" to-port="0" />
<edge from-layer="302" from-port="1" to-layer="303" to-port="1" />
<edge from-layer="303" from-port="2" to-layer="304" to-port="1" />
<edge from-layer="304" from-port="2" to-layer="305" to-port="0" />
<edge from-layer="305" from-port="2" to-layer="314" to-port="0" />
<edge from-layer="306" from-port="0" to-layer="307" to-port="1" />
<edge from-layer="307" from-port="2" to-layer="313" to-port="0" />
<edge from-layer="308" from-port="0" to-layer="313" to-port="1" />
<edge from-layer="309" from-port="1" to-layer="312" to-port="0" />
<edge from-layer="310" from-port="0" to-layer="312" to-port="1" />
<edge from-layer="311" from-port="0" to-layer="312" to-port="2" />
<edge from-layer="312" from-port="3" to-layer="317" to-port="2" />
<edge from-layer="312" from-port="3" to-layer="313" to-port="3" />
<edge from-layer="313" from-port="4" to-layer="314" to-port="1" />
<edge from-layer="314" from-port="2" to-layer="315" to-port="2" />
<edge from-layer="315" from-port="3" to-layer="316" to-port="0" />
<edge from-layer="316" from-port="1" to-layer="318" to-port="0" />
<edge from-layer="317" from-port="3" to-layer="318" to-port="1" />
<edge from-layer="318" from-port="2" to-layer="319" to-port="0" />
<edge from-layer="319" from-port="2" to-layer="320" to-port="0" />
<edge from-layer="319" from-port="2" to-layer="334" to-port="0" />
<edge from-layer="320" from-port="1" to-layer="323" to-port="0" />
<edge from-layer="320" from-port="1" to-layer="331" to-port="0" />
<edge from-layer="321" from-port="0" to-layer="323" to-port="1" />
<edge from-layer="322" from-port="0" to-layer="323" to-port="2" />
<edge from-layer="323" from-port="3" to-layer="325" to-port="0" />
<edge from-layer="324" from-port="0" to-layer="325" to-port="1" />
<edge from-layer="325" from-port="2" to-layer="327" to-port="0" />
<edge from-layer="326" from-port="0" to-layer="327" to-port="1" />
<edge from-layer="327" from-port="2" to-layer="333" to-port="0" />
<edge from-layer="327" from-port="2" to-layer="338" to-port="0" />
<edge from-layer="328" from-port="0" to-layer="333" to-port="1" />
<edge from-layer="329" from-port="0" to-layer="331" to-port="1" />
<edge from-layer="330" from-port="0" to-layer="331" to-port="2" />
<edge from-layer="331" from-port="3" to-layer="333" to-port="2" />
<edge from-layer="331" from-port="3" to-layer="338" to-port="1" />
<edge from-layer="332" from-port="0" to-layer="333" to-port="3" />
<edge from-layer="333" from-port="4" to-layer="334" to-port="1" />
<edge from-layer="334" from-port="2" to-layer="336" to-port="0" />
<edge from-layer="335" from-port="0" to-layer="336" to-port="1" />
<edge from-layer="336" from-port="2" to-layer="339" to-port="0" />
<edge from-layer="337" from-port="0" to-layer="338" to-port="2" />
<edge from-layer="338" from-port="3" to-layer="339" to-port="1" />
<edge from-layer="339" from-port="2" to-layer="342" to-port="0" />
<edge from-layer="340" from-port="0" to-layer="341" to-port="0" />
<edge from-layer="341" from-port="1" to-layer="342" to-port="1" />
<edge from-layer="342" from-port="2" to-layer="343" to-port="1" />
<edge from-layer="343" from-port="2" to-layer="345" to-port="0" />
<edge from-layer="343" from-port="2" to-layer="371" to-port="0" />
<edge from-layer="344" from-port="0" to-layer="345" to-port="1" />
<edge from-layer="345" from-port="2" to-layer="348" to-port="0" />
<edge from-layer="345" from-port="2" to-layer="351" to-port="0" />
<edge from-layer="345" from-port="2" to-layer="354" to-port="0" />
<edge from-layer="345" from-port="2" to-layer="357" to-port="0" />
<edge from-layer="346" from-port="0" to-layer="347" to-port="0" />
<edge from-layer="347" from-port="1" to-layer="348" to-port="1" />
<edge from-layer="348" from-port="2" to-layer="367" to-port="0" />
<edge from-layer="349" from-port="0" to-layer="350" to-port="0" />
<edge from-layer="350" from-port="1" to-layer="351" to-port="1" />
<edge from-layer="351" from-port="2" to-layer="362" to-port="0" />
<edge from-layer="352" from-port="0" to-layer="353" to-port="0" />
<edge from-layer="353" from-port="1" to-layer="354" to-port="1" />
<edge from-layer="354" from-port="2" to-layer="358" to-port="0" />
<edge from-layer="355" from-port="0" to-layer="356" to-port="0" />
<edge from-layer="356" from-port="1" to-layer="357" to-port="1" />
<edge from-layer="357" from-port="2" to-layer="358" to-port="1" />
<edge from-layer="358" from-port="2" to-layer="361" to-port="0" />
<edge from-layer="359" from-port="0" to-layer="360" to-port="0" />
<edge from-layer="360" from-port="1" to-layer="361" to-port="1" />
<edge from-layer="361" from-port="2" to-layer="362" to-port="1" />
<edge from-layer="362" from-port="2" to-layer="363" to-port="0" />
<edge from-layer="363" from-port="1" to-layer="366" to-port="0" />
<edge from-layer="364" from-port="0" to-layer="365" to-port="0" />
<edge from-layer="365" from-port="1" to-layer="366" to-port="1" />
<edge from-layer="366" from-port="2" to-layer="367" to-port="1" />
<edge from-layer="367" from-port="2" to-layer="370" to-port="0" />
<edge from-layer="368" from-port="0" to-layer="369" to-port="0" />
<edge from-layer="369" from-port="1" to-layer="370" to-port="1" />
<edge from-layer="370" from-port="2" to-layer="371" to-port="1" />
<edge from-layer="371" from-port="2" to-layer="373" to-port="0" />
<edge from-layer="372" from-port="0" to-layer="373" to-port="1" />
<edge from-layer="373" from-port="2" to-layer="376" to-port="0" />
<edge from-layer="374" from-port="0" to-layer="375" to-port="0" />
<edge from-layer="375" from-port="1" to-layer="376" to-port="1" />
<edge from-layer="376" from-port="2" to-layer="377" to-port="0" />
</edges>
<rt_info />
</net>