diff --git "a/20x/modelpnnx20x.pnnx.param" "b/20x/modelpnnx20x.pnnx.param" new file mode 100644--- /dev/null +++ "b/20x/modelpnnx20x.pnnx.param" @@ -0,0 +1,3504 @@ +7767517 +3502 3717 +pnnx.Input pnnx_input_0 0 1 0 #0=(1,3,384,384)f32 +pnnx.Input pnnx_input_1 0 1 1 #1=(1,3,384,384)f32 +pnnx.Attribute pnnx_3 0 1 2 @pnnx_3=(1,3,1,1)f32 #2=(1,3,1,1)f32 +aten::type_as pnnx_61 2 1 2 0 3 #2=(1,3,1,1)f32 #0=(1,3,384,384)f32 #3=(1,3,1,1)f32 +pnnx.Expression pnnx_expr_17158 2 1 0 3 4 expr=sub(@0,@1) #0=(1,3,384,384)f32 #3=(1,3,1,1)f32 #4=(1,3,384,384)f32 +pnnx.Expression pnnx_expr_17156 2 1 1 3 5 expr=sub(@0,@1) #1=(1,3,384,384)f32 #3=(1,3,1,1)f32 #5=(1,3,384,384)f32 +nn.Conv2d conv_first 1 1 4 6 bias=True dilation=(1,1) groups=1 in_channels=3 kernel_size=(8,8) out_channels=192 padding=(0,0) padding_mode=zeros stride=(8,8) @bias=(192)f32 @weight=(192,3,8,8)f32 #4=(1,3,384,384)f32 #6=(1,192,48,48)f32 +nn.Conv2d pnnx_unique_0 1 1 5 7 bias=True dilation=(1,1) groups=1 in_channels=3 kernel_size=(8,8) out_channels=192 padding=(0,0) padding_mode=zeros stride=(8,8) @bias=(192)f32 @weight=(192,3,8,8)f32 #5=(1,3,384,384)f32 #7=(1,192,48,48)f32 +torch.flatten torch.flatten_2183 1 1 6 8 end_dim=-1 start_dim=2 $input=6 #6=(1,192,48,48)f32 #8=(1,192,2304)f32 +torch.transpose torch.transpose_2960 1 1 8 9 dim0=1 dim1=2 $input=8 #8=(1,192,2304)f32 #9=(1,2304,192)f32 +nn.LayerNorm patch_embed_dfe.norm 1 1 9 10 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #9=(1,2304,192)f32 #10=(1,2304,192)f32 +nn.LayerNorm layers_dfe.0.residual_group.blocks.0.norm1 1 1 10 11 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #10=(1,2304,192)f32 #11=(1,2304,192)f32 +Tensor.reshape Tensor.view_975 1 1 11 12 shape=(1,6,8,6,8,192) $input=11 #11=(1,2304,192)f32 #12=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2528 1 1 12 13 dims=(0,1,3,2,4,5) $input=12 #12=(1,6,8,6,8,192)f32 #13=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_977 1 1 13 14 shape=(36,64,192) $input=13 #13=(1,6,6,8,8,192)f32 #14=(36,64,192)f32 +nn.Linear layers_dfe.0.residual_group.blocks.0.attn.qkv 1 1 14 15 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #14=(36,64,192)f32 #15=(36,64,576)f32 +Tensor.reshape Tensor.reshape_434 1 1 15 16 shape=(36,64,3,6,32) $input=15 #15=(36,64,576)f32 #16=(36,64,3,6,32)f32 +torch.permute torch.permute_2529 1 1 16 17 dims=(2,0,3,1,4) $input=16 #16=(36,64,3,6,32)f32 #17=(3,36,6,64,32)f32 +torch.unbind Tensor.select_650 1 3 17 18 19 20 dim=0 #17=(3,36,6,64,32)f32 #18=(36,6,64,32)f32 #19=(36,6,64,32)f32 #20=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_16899 1 1 18 21 expr=mul(@0,1.767767e-01) #18=(36,6,64,32)f32 #21=(36,6,64,32)f32 +torch.transpose torch.transpose_2961 1 1 19 22 dim0=-2 dim1=-1 $input=19 #19=(36,6,64,32)f32 #22=(36,6,32,64)f32 +torch.matmul torch.matmul_2204 2 1 21 22 23 $input=21 $other=22 #21=(36,6,64,32)f32 #22=(36,6,32,64)f32 #23=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_2377 0 1 24 @pnnx_fold_2377=(1,6,64,64)f32 #24=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_16886 2 1 23 24 25 expr=add(@0,@1) #23=(36,6,64,64)f32 #24=(1,6,64,64)f32 #25=(36,6,64,64)f32 +nn.Softmax layers_dfe.0.residual_group.blocks.0.attn.softmax 1 1 25 26 dim=-1 #25=(36,6,64,64)f32 #26=(36,6,64,64)f32 +torch.matmul torch.matmul_2205 2 1 26 20 27 $input=26 $other=20 #26=(36,6,64,64)f32 #20=(36,6,64,32)f32 #27=(36,6,64,32)f32 +torch.transpose torch.transpose_2962 1 1 27 28 dim0=1 dim1=2 $input=27 #27=(36,6,64,32)f32 #28=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_435 1 1 28 29 shape=(36,64,192) $input=28 #28=(36,64,6,32)f32 #29=(36,64,192)f32 +nn.Linear layers_dfe.0.residual_group.blocks.0.attn.proj 1 1 29 30 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #29=(36,64,192)f32 #30=(36,64,192)f32 +Tensor.reshape Tensor.view_981 1 1 30 31 shape=(1,6,6,8,8,192) $input=30 #30=(36,64,192)f32 #31=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2531 1 1 31 32 dims=(0,1,3,2,4,5) $input=31 #31=(1,6,6,8,8,192)f32 #32=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_983 1 1 32 33 shape=(1,2304,192) $input=32 #32=(1,6,8,6,8,192)f32 #33=(1,2304,192)f32 +pnnx.Expression pnnx_expr_16853 2 1 10 33 34 expr=add(@0,@1) #10=(1,2304,192)f32 #33=(1,2304,192)f32 #34=(1,2304,192)f32 +nn.LayerNorm layers_dfe.0.residual_group.blocks.0.norm2 1 1 34 35 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #34=(1,2304,192)f32 #35=(1,2304,192)f32 +nn.Linear layers_dfe.0.residual_group.blocks.0.mlp.fc1 1 1 35 36 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #35=(1,2304,192)f32 #36=(1,2304,384)f32 +nn.GELU layers_dfe.0.residual_group.blocks.0.mlp.act 1 1 36 37 #36=(1,2304,384)f32 #37=(1,2304,384)f32 +nn.Linear layers_dfe.0.residual_group.blocks.0.mlp.fc2 1 1 37 38 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #37=(1,2304,384)f32 #38=(1,2304,192)f32 +pnnx.Expression pnnx_expr_16851 2 1 34 38 39 expr=add(@0,@1) #34=(1,2304,192)f32 #38=(1,2304,192)f32 #39=(1,2304,192)f32 +nn.LayerNorm layers_dfe.0.residual_group.blocks.1.norm1 1 1 39 40 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #39=(1,2304,192)f32 #40=(1,2304,192)f32 +Tensor.view Tensor.view_984 1 1 40 41 shape=(1,48,48,192) $input=40 #40=(1,2304,192)f32 #41=(1,48,48,192)f32 +torch.roll torch.roll_2420 1 1 41 42 dims=(1,2) shifts=(-4,-4) $input=41 #41=(1,48,48,192)f32 #42=(1,48,48,192)f32 +Tensor.view Tensor.view_985 1 1 42 43 shape=(1,6,8,6,8,192) $input=42 #42=(1,48,48,192)f32 #43=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2532 1 1 43 44 dims=(0,1,3,2,4,5) $input=43 #43=(1,6,8,6,8,192)f32 #44=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_987 1 1 44 45 shape=(36,64,192) $input=44 #44=(1,6,6,8,8,192)f32 #45=(36,64,192)f32 +nn.Linear layers_dfe.0.residual_group.blocks.1.attn.qkv 1 1 45 46 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #45=(36,64,192)f32 #46=(36,64,576)f32 +Tensor.reshape Tensor.reshape_436 1 1 46 47 shape=(36,64,3,6,32) $input=46 #46=(36,64,576)f32 #47=(36,64,3,6,32)f32 +torch.permute torch.permute_2533 1 1 47 48 dims=(2,0,3,1,4) $input=47 #47=(36,64,3,6,32)f32 #48=(3,36,6,64,32)f32 +torch.unbind Tensor.select_653 1 3 48 49 50 51 dim=0 #48=(3,36,6,64,32)f32 #49=(36,6,64,32)f32 #50=(36,6,64,32)f32 #51=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_16753 1 1 49 52 expr=mul(@0,1.767767e-01) #49=(36,6,64,32)f32 #52=(36,6,64,32)f32 +torch.transpose torch.transpose_2963 1 1 50 53 dim0=-2 dim1=-1 $input=50 #50=(36,6,64,32)f32 #53=(36,6,32,64)f32 +torch.matmul torch.matmul_2206 2 1 52 53 54 $input=52 $other=53 #52=(36,6,64,32)f32 #53=(36,6,32,64)f32 #54=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_2530 0 1 55 @pnnx_fold_2530=(1,6,64,64)f32 #55=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_16740 2 1 54 55 56 expr=add(@0,@1) #54=(36,6,64,64)f32 #55=(1,6,64,64)f32 #56=(36,6,64,64)f32 +Tensor.view Tensor.view_990 1 1 56 57 shape=(1,36,6,64,64) $input=56 #56=(36,6,64,64)f32 #57=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_2540 0 1 58 @pnnx_fold_2540=(1,36,1,64,64)f32 #58=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_16728 2 1 57 58 59 expr=add(@0,@1) #57=(1,36,6,64,64)f32 #58=(1,36,1,64,64)f32 #59=(1,36,6,64,64)f32 +Tensor.view Tensor.view_991 1 1 59 60 shape=(-1,6,64,64) $input=59 #59=(1,36,6,64,64)f32 #60=(36,6,64,64)f32 +nn.Softmax layers_dfe.0.residual_group.blocks.1.attn.softmax 1 1 60 61 dim=-1 #60=(36,6,64,64)f32 #61=(36,6,64,64)f32 +torch.matmul torch.matmul_2207 2 1 61 51 62 $input=61 $other=51 #61=(36,6,64,64)f32 #51=(36,6,64,32)f32 #62=(36,6,64,32)f32 +torch.transpose torch.transpose_2964 1 1 62 63 dim0=1 dim1=2 $input=62 #62=(36,6,64,32)f32 #63=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_437 1 1 63 64 shape=(36,64,192) $input=63 #63=(36,64,6,32)f32 #64=(36,64,192)f32 +nn.Linear layers_dfe.0.residual_group.blocks.1.attn.proj 1 1 64 65 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #64=(36,64,192)f32 #65=(36,64,192)f32 +Tensor.reshape Tensor.view_993 1 1 65 66 shape=(1,6,6,8,8,192) $input=65 #65=(36,64,192)f32 #66=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2535 1 1 66 67 dims=(0,1,3,2,4,5) $input=66 #66=(1,6,6,8,8,192)f32 #67=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_3 1 1 67 68 shape=(1,48,48,-1) $input=67 #67=(1,6,8,6,8,192)f32 #68=(1,48,48,192)f32 +torch.roll torch.roll_2421 1 1 68 69 dims=(1,2) shifts=(4,4) $input=68 #68=(1,48,48,192)f32 #69=(1,48,48,192)f32 +Tensor.view Tensor.view_995 1 1 69 70 shape=(1,2304,192) $input=69 #69=(1,48,48,192)f32 #70=(1,2304,192)f32 +pnnx.Expression pnnx_expr_16686 2 1 39 70 71 expr=add(@0,@1) #39=(1,2304,192)f32 #70=(1,2304,192)f32 #71=(1,2304,192)f32 +nn.LayerNorm layers_dfe.0.residual_group.blocks.1.norm2 1 1 71 72 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #71=(1,2304,192)f32 #72=(1,2304,192)f32 +nn.Linear layers_dfe.0.residual_group.blocks.1.mlp.fc1 1 1 72 73 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #72=(1,2304,192)f32 #73=(1,2304,384)f32 +nn.GELU layers_dfe.0.residual_group.blocks.1.mlp.act 1 1 73 74 #73=(1,2304,384)f32 #74=(1,2304,384)f32 +nn.Linear layers_dfe.0.residual_group.blocks.1.mlp.fc2 1 1 74 75 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #74=(1,2304,384)f32 #75=(1,2304,192)f32 +pnnx.Expression pnnx_expr_16684 2 1 71 75 76 expr=add(@0,@1) #71=(1,2304,192)f32 #75=(1,2304,192)f32 #76=(1,2304,192)f32 +nn.LayerNorm layers_dfe.0.residual_group.blocks.2.norm1 1 1 76 77 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #76=(1,2304,192)f32 #77=(1,2304,192)f32 +Tensor.reshape Tensor.view_997 1 1 77 78 shape=(1,6,8,6,8,192) $input=77 #77=(1,2304,192)f32 #78=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2536 1 1 78 79 dims=(0,1,3,2,4,5) $input=78 #78=(1,6,8,6,8,192)f32 #79=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_999 1 1 79 80 shape=(36,64,192) $input=79 #79=(1,6,6,8,8,192)f32 #80=(36,64,192)f32 +nn.Linear layers_dfe.0.residual_group.blocks.2.attn.qkv 1 1 80 81 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #80=(36,64,192)f32 #81=(36,64,576)f32 +Tensor.reshape Tensor.reshape_438 1 1 81 82 shape=(36,64,3,6,32) $input=81 #81=(36,64,576)f32 #82=(36,64,3,6,32)f32 +torch.permute torch.permute_2537 1 1 82 83 dims=(2,0,3,1,4) $input=82 #82=(36,64,3,6,32)f32 #83=(3,36,6,64,32)f32 +torch.unbind Tensor.select_656 1 3 83 84 85 86 dim=0 #83=(3,36,6,64,32)f32 #84=(36,6,64,32)f32 #85=(36,6,64,32)f32 #86=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_16596 1 1 84 87 expr=mul(@0,1.767767e-01) #84=(36,6,64,32)f32 #87=(36,6,64,32)f32 +torch.transpose torch.transpose_2965 1 1 85 88 dim0=-2 dim1=-1 $input=85 #85=(36,6,64,32)f32 #88=(36,6,32,64)f32 +torch.matmul torch.matmul_2208 2 1 87 88 89 $input=87 $other=88 #87=(36,6,64,32)f32 #88=(36,6,32,64)f32 #89=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_2689 0 1 90 @pnnx_fold_2689=(1,6,64,64)f32 #90=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_16583 2 1 89 90 91 expr=add(@0,@1) #89=(36,6,64,64)f32 #90=(1,6,64,64)f32 #91=(36,6,64,64)f32 +nn.Softmax layers_dfe.0.residual_group.blocks.2.attn.softmax 1 1 91 92 dim=-1 #91=(36,6,64,64)f32 #92=(36,6,64,64)f32 +torch.matmul torch.matmul_2209 2 1 92 86 93 $input=92 $other=86 #92=(36,6,64,64)f32 #86=(36,6,64,32)f32 #93=(36,6,64,32)f32 +torch.transpose torch.transpose_2966 1 1 93 94 dim0=1 dim1=2 $input=93 #93=(36,6,64,32)f32 #94=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_439 1 1 94 95 shape=(36,64,192) $input=94 #94=(36,64,6,32)f32 #95=(36,64,192)f32 +nn.Linear layers_dfe.0.residual_group.blocks.2.attn.proj 1 1 95 96 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #95=(36,64,192)f32 #96=(36,64,192)f32 +Tensor.reshape Tensor.view_1003 1 1 96 97 shape=(1,6,6,8,8,192) $input=96 #96=(36,64,192)f32 #97=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2539 1 1 97 98 dims=(0,1,3,2,4,5) $input=97 #97=(1,6,6,8,8,192)f32 #98=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1005 1 1 98 99 shape=(1,2304,192) $input=98 #98=(1,6,8,6,8,192)f32 #99=(1,2304,192)f32 +pnnx.Expression pnnx_expr_16550 2 1 76 99 100 expr=add(@0,@1) #76=(1,2304,192)f32 #99=(1,2304,192)f32 #100=(1,2304,192)f32 +nn.LayerNorm layers_dfe.0.residual_group.blocks.2.norm2 1 1 100 101 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #100=(1,2304,192)f32 #101=(1,2304,192)f32 +nn.Linear layers_dfe.0.residual_group.blocks.2.mlp.fc1 1 1 101 102 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #101=(1,2304,192)f32 #102=(1,2304,384)f32 +nn.GELU layers_dfe.0.residual_group.blocks.2.mlp.act 1 1 102 103 #102=(1,2304,384)f32 #103=(1,2304,384)f32 +nn.Linear layers_dfe.0.residual_group.blocks.2.mlp.fc2 1 1 103 104 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #103=(1,2304,384)f32 #104=(1,2304,192)f32 +pnnx.Expression pnnx_expr_16548 2 1 100 104 105 expr=add(@0,@1) #100=(1,2304,192)f32 #104=(1,2304,192)f32 #105=(1,2304,192)f32 +nn.LayerNorm layers_dfe.0.residual_group.blocks.3.norm1 1 1 105 106 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #105=(1,2304,192)f32 #106=(1,2304,192)f32 +Tensor.view Tensor.view_1006 1 1 106 107 shape=(1,48,48,192) $input=106 #106=(1,2304,192)f32 #107=(1,48,48,192)f32 +torch.roll torch.roll_2422 1 1 107 108 dims=(1,2) shifts=(-4,-4) $input=107 #107=(1,48,48,192)f32 #108=(1,48,48,192)f32 +Tensor.view Tensor.view_1007 1 1 108 109 shape=(1,6,8,6,8,192) $input=108 #108=(1,48,48,192)f32 #109=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2540 1 1 109 110 dims=(0,1,3,2,4,5) $input=109 #109=(1,6,8,6,8,192)f32 #110=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1009 1 1 110 111 shape=(36,64,192) $input=110 #110=(1,6,6,8,8,192)f32 #111=(36,64,192)f32 +nn.Linear layers_dfe.0.residual_group.blocks.3.attn.qkv 1 1 111 112 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #111=(36,64,192)f32 #112=(36,64,576)f32 +Tensor.reshape Tensor.reshape_440 1 1 112 113 shape=(36,64,3,6,32) $input=112 #112=(36,64,576)f32 #113=(36,64,3,6,32)f32 +torch.permute torch.permute_2541 1 1 113 114 dims=(2,0,3,1,4) $input=113 #113=(36,64,3,6,32)f32 #114=(3,36,6,64,32)f32 +torch.unbind Tensor.select_659 1 3 114 115 116 117 dim=0 #114=(3,36,6,64,32)f32 #115=(36,6,64,32)f32 #116=(36,6,64,32)f32 #117=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_16450 1 1 115 118 expr=mul(@0,1.767767e-01) #115=(36,6,64,32)f32 #118=(36,6,64,32)f32 +torch.transpose torch.transpose_2967 1 1 116 119 dim0=-2 dim1=-1 $input=116 #116=(36,6,64,32)f32 #119=(36,6,32,64)f32 +torch.matmul torch.matmul_2210 2 1 118 119 120 $input=118 $other=119 #118=(36,6,64,32)f32 #119=(36,6,32,64)f32 #120=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_2842 0 1 121 @pnnx_fold_2842=(1,6,64,64)f32 #121=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_16437 2 1 120 121 122 expr=add(@0,@1) #120=(36,6,64,64)f32 #121=(1,6,64,64)f32 #122=(36,6,64,64)f32 +Tensor.view Tensor.view_1012 1 1 122 123 shape=(1,36,6,64,64) $input=122 #122=(36,6,64,64)f32 #123=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_2852 0 1 124 @pnnx_fold_2852=(1,36,1,64,64)f32 #124=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_16425 2 1 123 124 125 expr=add(@0,@1) #123=(1,36,6,64,64)f32 #124=(1,36,1,64,64)f32 #125=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1013 1 1 125 126 shape=(-1,6,64,64) $input=125 #125=(1,36,6,64,64)f32 #126=(36,6,64,64)f32 +nn.Softmax layers_dfe.0.residual_group.blocks.3.attn.softmax 1 1 126 127 dim=-1 #126=(36,6,64,64)f32 #127=(36,6,64,64)f32 +torch.matmul torch.matmul_2211 2 1 127 117 128 $input=127 $other=117 #127=(36,6,64,64)f32 #117=(36,6,64,32)f32 #128=(36,6,64,32)f32 +torch.transpose torch.transpose_2968 1 1 128 129 dim0=1 dim1=2 $input=128 #128=(36,6,64,32)f32 #129=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_441 1 1 129 130 shape=(36,64,192) $input=129 #129=(36,64,6,32)f32 #130=(36,64,192)f32 +nn.Linear layers_dfe.0.residual_group.blocks.3.attn.proj 1 1 130 131 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #130=(36,64,192)f32 #131=(36,64,192)f32 +Tensor.reshape Tensor.view_1015 1 1 131 132 shape=(1,6,6,8,8,192) $input=131 #131=(36,64,192)f32 #132=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2543 1 1 132 133 dims=(0,1,3,2,4,5) $input=132 #132=(1,6,6,8,8,192)f32 #133=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_7 1 1 133 134 shape=(1,48,48,-1) $input=133 #133=(1,6,8,6,8,192)f32 #134=(1,48,48,192)f32 +torch.roll torch.roll_2423 1 1 134 135 dims=(1,2) shifts=(4,4) $input=134 #134=(1,48,48,192)f32 #135=(1,48,48,192)f32 +Tensor.view Tensor.view_1017 1 1 135 136 shape=(1,2304,192) $input=135 #135=(1,48,48,192)f32 #136=(1,2304,192)f32 +pnnx.Expression pnnx_expr_16383 2 1 105 136 137 expr=add(@0,@1) #105=(1,2304,192)f32 #136=(1,2304,192)f32 #137=(1,2304,192)f32 +nn.LayerNorm layers_dfe.0.residual_group.blocks.3.norm2 1 1 137 138 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #137=(1,2304,192)f32 #138=(1,2304,192)f32 +nn.Linear layers_dfe.0.residual_group.blocks.3.mlp.fc1 1 1 138 139 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #138=(1,2304,192)f32 #139=(1,2304,384)f32 +nn.GELU layers_dfe.0.residual_group.blocks.3.mlp.act 1 1 139 140 #139=(1,2304,384)f32 #140=(1,2304,384)f32 +nn.Linear layers_dfe.0.residual_group.blocks.3.mlp.fc2 1 1 140 141 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #140=(1,2304,384)f32 #141=(1,2304,192)f32 +pnnx.Expression pnnx_expr_16381 2 1 137 141 142 expr=add(@0,@1) #137=(1,2304,192)f32 #141=(1,2304,192)f32 #142=(1,2304,192)f32 +nn.LayerNorm layers_dfe.0.residual_group.blocks.4.norm1 1 1 142 143 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #142=(1,2304,192)f32 #143=(1,2304,192)f32 +Tensor.reshape Tensor.view_1019 1 1 143 144 shape=(1,6,8,6,8,192) $input=143 #143=(1,2304,192)f32 #144=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2544 1 1 144 145 dims=(0,1,3,2,4,5) $input=144 #144=(1,6,8,6,8,192)f32 #145=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1021 1 1 145 146 shape=(36,64,192) $input=145 #145=(1,6,6,8,8,192)f32 #146=(36,64,192)f32 +nn.Linear layers_dfe.0.residual_group.blocks.4.attn.qkv 1 1 146 147 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #146=(36,64,192)f32 #147=(36,64,576)f32 +Tensor.reshape Tensor.reshape_442 1 1 147 148 shape=(36,64,3,6,32) $input=147 #147=(36,64,576)f32 #148=(36,64,3,6,32)f32 +torch.permute torch.permute_2545 1 1 148 149 dims=(2,0,3,1,4) $input=148 #148=(36,64,3,6,32)f32 #149=(3,36,6,64,32)f32 +torch.unbind Tensor.select_662 1 3 149 150 151 152 dim=0 #149=(3,36,6,64,32)f32 #150=(36,6,64,32)f32 #151=(36,6,64,32)f32 #152=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_16293 1 1 150 153 expr=mul(@0,1.767767e-01) #150=(36,6,64,32)f32 #153=(36,6,64,32)f32 +torch.transpose torch.transpose_2969 1 1 151 154 dim0=-2 dim1=-1 $input=151 #151=(36,6,64,32)f32 #154=(36,6,32,64)f32 +torch.matmul torch.matmul_2212 2 1 153 154 155 $input=153 $other=154 #153=(36,6,64,32)f32 #154=(36,6,32,64)f32 #155=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_3001 0 1 156 @pnnx_fold_3001=(1,6,64,64)f32 #156=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_16280 2 1 155 156 157 expr=add(@0,@1) #155=(36,6,64,64)f32 #156=(1,6,64,64)f32 #157=(36,6,64,64)f32 +nn.Softmax layers_dfe.0.residual_group.blocks.4.attn.softmax 1 1 157 158 dim=-1 #157=(36,6,64,64)f32 #158=(36,6,64,64)f32 +torch.matmul torch.matmul_2213 2 1 158 152 159 $input=158 $other=152 #158=(36,6,64,64)f32 #152=(36,6,64,32)f32 #159=(36,6,64,32)f32 +torch.transpose torch.transpose_2970 1 1 159 160 dim0=1 dim1=2 $input=159 #159=(36,6,64,32)f32 #160=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_443 1 1 160 161 shape=(36,64,192) $input=160 #160=(36,64,6,32)f32 #161=(36,64,192)f32 +nn.Linear layers_dfe.0.residual_group.blocks.4.attn.proj 1 1 161 162 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #161=(36,64,192)f32 #162=(36,64,192)f32 +Tensor.reshape Tensor.view_1025 1 1 162 163 shape=(1,6,6,8,8,192) $input=162 #162=(36,64,192)f32 #163=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2547 1 1 163 164 dims=(0,1,3,2,4,5) $input=163 #163=(1,6,6,8,8,192)f32 #164=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1027 1 1 164 165 shape=(1,2304,192) $input=164 #164=(1,6,8,6,8,192)f32 #165=(1,2304,192)f32 +pnnx.Expression pnnx_expr_16247 2 1 142 165 166 expr=add(@0,@1) #142=(1,2304,192)f32 #165=(1,2304,192)f32 #166=(1,2304,192)f32 +nn.LayerNorm layers_dfe.0.residual_group.blocks.4.norm2 1 1 166 167 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #166=(1,2304,192)f32 #167=(1,2304,192)f32 +nn.Linear layers_dfe.0.residual_group.blocks.4.mlp.fc1 1 1 167 168 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #167=(1,2304,192)f32 #168=(1,2304,384)f32 +nn.GELU layers_dfe.0.residual_group.blocks.4.mlp.act 1 1 168 169 #168=(1,2304,384)f32 #169=(1,2304,384)f32 +nn.Linear layers_dfe.0.residual_group.blocks.4.mlp.fc2 1 1 169 170 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #169=(1,2304,384)f32 #170=(1,2304,192)f32 +pnnx.Expression pnnx_expr_16245 2 1 166 170 171 expr=add(@0,@1) #166=(1,2304,192)f32 #170=(1,2304,192)f32 #171=(1,2304,192)f32 +nn.LayerNorm layers_dfe.0.residual_group.blocks.5.norm1 1 1 171 172 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #171=(1,2304,192)f32 #172=(1,2304,192)f32 +Tensor.view Tensor.view_1028 1 1 172 173 shape=(1,48,48,192) $input=172 #172=(1,2304,192)f32 #173=(1,48,48,192)f32 +torch.roll torch.roll_2424 1 1 173 174 dims=(1,2) shifts=(-4,-4) $input=173 #173=(1,48,48,192)f32 #174=(1,48,48,192)f32 +Tensor.view Tensor.view_1029 1 1 174 175 shape=(1,6,8,6,8,192) $input=174 #174=(1,48,48,192)f32 #175=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2548 1 1 175 176 dims=(0,1,3,2,4,5) $input=175 #175=(1,6,8,6,8,192)f32 #176=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1031 1 1 176 177 shape=(36,64,192) $input=176 #176=(1,6,6,8,8,192)f32 #177=(36,64,192)f32 +nn.Linear layers_dfe.0.residual_group.blocks.5.attn.qkv 1 1 177 178 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #177=(36,64,192)f32 #178=(36,64,576)f32 +Tensor.reshape Tensor.reshape_444 1 1 178 179 shape=(36,64,3,6,32) $input=178 #178=(36,64,576)f32 #179=(36,64,3,6,32)f32 +torch.permute torch.permute_2549 1 1 179 180 dims=(2,0,3,1,4) $input=179 #179=(36,64,3,6,32)f32 #180=(3,36,6,64,32)f32 +torch.unbind Tensor.select_665 1 3 180 181 182 183 dim=0 #180=(3,36,6,64,32)f32 #181=(36,6,64,32)f32 #182=(36,6,64,32)f32 #183=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_16147 1 1 181 184 expr=mul(@0,1.767767e-01) #181=(36,6,64,32)f32 #184=(36,6,64,32)f32 +torch.transpose torch.transpose_2971 1 1 182 185 dim0=-2 dim1=-1 $input=182 #182=(36,6,64,32)f32 #185=(36,6,32,64)f32 +torch.matmul torch.matmul_2214 2 1 184 185 186 $input=184 $other=185 #184=(36,6,64,32)f32 #185=(36,6,32,64)f32 #186=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_3154 0 1 187 @pnnx_fold_3154=(1,6,64,64)f32 #187=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_16134 2 1 186 187 188 expr=add(@0,@1) #186=(36,6,64,64)f32 #187=(1,6,64,64)f32 #188=(36,6,64,64)f32 +Tensor.view Tensor.view_1034 1 1 188 189 shape=(1,36,6,64,64) $input=188 #188=(36,6,64,64)f32 #189=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_3164 0 1 190 @pnnx_fold_3164=(1,36,1,64,64)f32 #190=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_16122 2 1 189 190 191 expr=add(@0,@1) #189=(1,36,6,64,64)f32 #190=(1,36,1,64,64)f32 #191=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1035 1 1 191 192 shape=(-1,6,64,64) $input=191 #191=(1,36,6,64,64)f32 #192=(36,6,64,64)f32 +nn.Softmax layers_dfe.0.residual_group.blocks.5.attn.softmax 1 1 192 193 dim=-1 #192=(36,6,64,64)f32 #193=(36,6,64,64)f32 +torch.matmul torch.matmul_2215 2 1 193 183 194 $input=193 $other=183 #193=(36,6,64,64)f32 #183=(36,6,64,32)f32 #194=(36,6,64,32)f32 +torch.transpose torch.transpose_2972 1 1 194 195 dim0=1 dim1=2 $input=194 #194=(36,6,64,32)f32 #195=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_445 1 1 195 196 shape=(36,64,192) $input=195 #195=(36,64,6,32)f32 #196=(36,64,192)f32 +nn.Linear layers_dfe.0.residual_group.blocks.5.attn.proj 1 1 196 197 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #196=(36,64,192)f32 #197=(36,64,192)f32 +Tensor.reshape Tensor.view_1037 1 1 197 198 shape=(1,6,6,8,8,192) $input=197 #197=(36,64,192)f32 #198=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2551 1 1 198 199 dims=(0,1,3,2,4,5) $input=198 #198=(1,6,6,8,8,192)f32 #199=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_11 1 1 199 200 shape=(1,48,48,-1) $input=199 #199=(1,6,8,6,8,192)f32 #200=(1,48,48,192)f32 +torch.roll torch.roll_2425 1 1 200 201 dims=(1,2) shifts=(4,4) $input=200 #200=(1,48,48,192)f32 #201=(1,48,48,192)f32 +Tensor.view Tensor.view_1039 1 1 201 202 shape=(1,2304,192) $input=201 #201=(1,48,48,192)f32 #202=(1,2304,192)f32 +pnnx.Expression pnnx_expr_16080 2 1 171 202 203 expr=add(@0,@1) #171=(1,2304,192)f32 #202=(1,2304,192)f32 #203=(1,2304,192)f32 +nn.LayerNorm layers_dfe.0.residual_group.blocks.5.norm2 1 1 203 204 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #203=(1,2304,192)f32 #204=(1,2304,192)f32 +nn.Linear layers_dfe.0.residual_group.blocks.5.mlp.fc1 1 1 204 205 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #204=(1,2304,192)f32 #205=(1,2304,384)f32 +nn.GELU layers_dfe.0.residual_group.blocks.5.mlp.act 1 1 205 206 #205=(1,2304,384)f32 #206=(1,2304,384)f32 +nn.Linear layers_dfe.0.residual_group.blocks.5.mlp.fc2 1 1 206 207 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #206=(1,2304,384)f32 #207=(1,2304,192)f32 +pnnx.Expression pnnx_expr_16078 2 1 203 207 208 expr=add(@0,@1) #203=(1,2304,192)f32 #207=(1,2304,192)f32 #208=(1,2304,192)f32 +torch.transpose torch.transpose_2973 1 1 208 209 dim0=1 dim1=2 $input=208 #208=(1,2304,192)f32 #209=(1,192,2304)f32 +Tensor.view Tensor.view_1040 1 1 209 210 shape=(1,192,48,48) $input=209 #209=(1,192,2304)f32 #210=(1,192,48,48)f32 +nn.Conv2d layers_dfe.0.conv 1 1 210 211 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #210=(1,192,48,48)f32 #211=(1,192,48,48)f32 +torch.flatten torch.flatten_2184 1 1 211 212 end_dim=-1 start_dim=2 $input=211 #211=(1,192,48,48)f32 #212=(1,192,2304)f32 +torch.transpose torch.transpose_2974 1 1 212 213 dim0=1 dim1=2 $input=212 #212=(1,192,2304)f32 #213=(1,2304,192)f32 +pnnx.Expression pnnx_expr_16065 2 1 213 10 214 expr=add(@0,@1) #213=(1,2304,192)f32 #10=(1,2304,192)f32 #214=(1,2304,192)f32 +nn.LayerNorm layers_dfe.1.residual_group.blocks.0.norm1 1 1 214 215 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #214=(1,2304,192)f32 #215=(1,2304,192)f32 +Tensor.reshape Tensor.view_1042 1 1 215 216 shape=(1,6,8,6,8,192) $input=215 #215=(1,2304,192)f32 #216=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2552 1 1 216 217 dims=(0,1,3,2,4,5) $input=216 #216=(1,6,8,6,8,192)f32 #217=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1044 1 1 217 218 shape=(36,64,192) $input=217 #217=(1,6,6,8,8,192)f32 #218=(36,64,192)f32 +nn.Linear layers_dfe.1.residual_group.blocks.0.attn.qkv 1 1 218 219 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #218=(36,64,192)f32 #219=(36,64,576)f32 +Tensor.reshape Tensor.reshape_446 1 1 219 220 shape=(36,64,3,6,32) $input=219 #219=(36,64,576)f32 #220=(36,64,3,6,32)f32 +torch.permute torch.permute_2553 1 1 220 221 dims=(2,0,3,1,4) $input=220 #220=(36,64,3,6,32)f32 #221=(3,36,6,64,32)f32 +torch.unbind Tensor.select_668 1 3 221 222 223 224 dim=0 #221=(3,36,6,64,32)f32 #222=(36,6,64,32)f32 #223=(36,6,64,32)f32 #224=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_15976 1 1 222 225 expr=mul(@0,1.767767e-01) #222=(36,6,64,32)f32 #225=(36,6,64,32)f32 +torch.transpose torch.transpose_2975 1 1 223 226 dim0=-2 dim1=-1 $input=223 #223=(36,6,64,32)f32 #226=(36,6,32,64)f32 +torch.matmul torch.matmul_2216 2 1 225 226 227 $input=225 $other=226 #225=(36,6,64,32)f32 #226=(36,6,32,64)f32 #227=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_3347 0 1 228 @pnnx_fold_3347=(1,6,64,64)f32 #228=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_15963 2 1 227 228 229 expr=add(@0,@1) #227=(36,6,64,64)f32 #228=(1,6,64,64)f32 #229=(36,6,64,64)f32 +nn.Softmax layers_dfe.1.residual_group.blocks.0.attn.softmax 1 1 229 230 dim=-1 #229=(36,6,64,64)f32 #230=(36,6,64,64)f32 +torch.matmul torch.matmul_2217 2 1 230 224 231 $input=230 $other=224 #230=(36,6,64,64)f32 #224=(36,6,64,32)f32 #231=(36,6,64,32)f32 +torch.transpose torch.transpose_2976 1 1 231 232 dim0=1 dim1=2 $input=231 #231=(36,6,64,32)f32 #232=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_447 1 1 232 233 shape=(36,64,192) $input=232 #232=(36,64,6,32)f32 #233=(36,64,192)f32 +nn.Linear layers_dfe.1.residual_group.blocks.0.attn.proj 1 1 233 234 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #233=(36,64,192)f32 #234=(36,64,192)f32 +Tensor.reshape Tensor.view_1048 1 1 234 235 shape=(1,6,6,8,8,192) $input=234 #234=(36,64,192)f32 #235=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2555 1 1 235 236 dims=(0,1,3,2,4,5) $input=235 #235=(1,6,6,8,8,192)f32 #236=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1050 1 1 236 237 shape=(1,2304,192) $input=236 #236=(1,6,8,6,8,192)f32 #237=(1,2304,192)f32 +pnnx.Expression pnnx_expr_15930 2 1 214 237 238 expr=add(@0,@1) #214=(1,2304,192)f32 #237=(1,2304,192)f32 #238=(1,2304,192)f32 +nn.LayerNorm layers_dfe.1.residual_group.blocks.0.norm2 1 1 238 239 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #238=(1,2304,192)f32 #239=(1,2304,192)f32 +nn.Linear layers_dfe.1.residual_group.blocks.0.mlp.fc1 1 1 239 240 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #239=(1,2304,192)f32 #240=(1,2304,384)f32 +nn.GELU layers_dfe.1.residual_group.blocks.0.mlp.act 1 1 240 241 #240=(1,2304,384)f32 #241=(1,2304,384)f32 +nn.Linear layers_dfe.1.residual_group.blocks.0.mlp.fc2 1 1 241 242 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #241=(1,2304,384)f32 #242=(1,2304,192)f32 +pnnx.Expression pnnx_expr_15928 2 1 238 242 243 expr=add(@0,@1) #238=(1,2304,192)f32 #242=(1,2304,192)f32 #243=(1,2304,192)f32 +nn.LayerNorm layers_dfe.1.residual_group.blocks.1.norm1 1 1 243 244 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #243=(1,2304,192)f32 #244=(1,2304,192)f32 +Tensor.view Tensor.view_1051 1 1 244 245 shape=(1,48,48,192) $input=244 #244=(1,2304,192)f32 #245=(1,48,48,192)f32 +torch.roll torch.roll_2426 1 1 245 246 dims=(1,2) shifts=(-4,-4) $input=245 #245=(1,48,48,192)f32 #246=(1,48,48,192)f32 +Tensor.view Tensor.view_1052 1 1 246 247 shape=(1,6,8,6,8,192) $input=246 #246=(1,48,48,192)f32 #247=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2556 1 1 247 248 dims=(0,1,3,2,4,5) $input=247 #247=(1,6,8,6,8,192)f32 #248=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1054 1 1 248 249 shape=(36,64,192) $input=248 #248=(1,6,6,8,8,192)f32 #249=(36,64,192)f32 +nn.Linear layers_dfe.1.residual_group.blocks.1.attn.qkv 1 1 249 250 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #249=(36,64,192)f32 #250=(36,64,576)f32 +Tensor.reshape Tensor.reshape_448 1 1 250 251 shape=(36,64,3,6,32) $input=250 #250=(36,64,576)f32 #251=(36,64,3,6,32)f32 +torch.permute torch.permute_2557 1 1 251 252 dims=(2,0,3,1,4) $input=251 #251=(36,64,3,6,32)f32 #252=(3,36,6,64,32)f32 +torch.unbind Tensor.select_671 1 3 252 253 254 255 dim=0 #252=(3,36,6,64,32)f32 #253=(36,6,64,32)f32 #254=(36,6,64,32)f32 #255=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_15830 1 1 253 256 expr=mul(@0,1.767767e-01) #253=(36,6,64,32)f32 #256=(36,6,64,32)f32 +torch.transpose torch.transpose_2977 1 1 254 257 dim0=-2 dim1=-1 $input=254 #254=(36,6,64,32)f32 #257=(36,6,32,64)f32 +torch.matmul torch.matmul_2218 2 1 256 257 258 $input=256 $other=257 #256=(36,6,64,32)f32 #257=(36,6,32,64)f32 #258=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_3500 0 1 259 @pnnx_fold_3500=(1,6,64,64)f32 #259=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_15817 2 1 258 259 260 expr=add(@0,@1) #258=(36,6,64,64)f32 #259=(1,6,64,64)f32 #260=(36,6,64,64)f32 +Tensor.view Tensor.view_1057 1 1 260 261 shape=(1,36,6,64,64) $input=260 #260=(36,6,64,64)f32 #261=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_3510 0 1 262 @pnnx_fold_3510=(1,36,1,64,64)f32 #262=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_15805 2 1 261 262 263 expr=add(@0,@1) #261=(1,36,6,64,64)f32 #262=(1,36,1,64,64)f32 #263=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1058 1 1 263 264 shape=(-1,6,64,64) $input=263 #263=(1,36,6,64,64)f32 #264=(36,6,64,64)f32 +nn.Softmax layers_dfe.1.residual_group.blocks.1.attn.softmax 1 1 264 265 dim=-1 #264=(36,6,64,64)f32 #265=(36,6,64,64)f32 +torch.matmul torch.matmul_2219 2 1 265 255 266 $input=265 $other=255 #265=(36,6,64,64)f32 #255=(36,6,64,32)f32 #266=(36,6,64,32)f32 +torch.transpose torch.transpose_2978 1 1 266 267 dim0=1 dim1=2 $input=266 #266=(36,6,64,32)f32 #267=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_449 1 1 267 268 shape=(36,64,192) $input=267 #267=(36,64,6,32)f32 #268=(36,64,192)f32 +nn.Linear layers_dfe.1.residual_group.blocks.1.attn.proj 1 1 268 269 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #268=(36,64,192)f32 #269=(36,64,192)f32 +Tensor.reshape Tensor.view_1060 1 1 269 270 shape=(1,6,6,8,8,192) $input=269 #269=(36,64,192)f32 #270=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2559 1 1 270 271 dims=(0,1,3,2,4,5) $input=270 #270=(1,6,6,8,8,192)f32 #271=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_15 1 1 271 272 shape=(1,48,48,-1) $input=271 #271=(1,6,8,6,8,192)f32 #272=(1,48,48,192)f32 +torch.roll torch.roll_2427 1 1 272 273 dims=(1,2) shifts=(4,4) $input=272 #272=(1,48,48,192)f32 #273=(1,48,48,192)f32 +Tensor.view Tensor.view_1062 1 1 273 274 shape=(1,2304,192) $input=273 #273=(1,48,48,192)f32 #274=(1,2304,192)f32 +pnnx.Expression pnnx_expr_15763 2 1 243 274 275 expr=add(@0,@1) #243=(1,2304,192)f32 #274=(1,2304,192)f32 #275=(1,2304,192)f32 +nn.LayerNorm layers_dfe.1.residual_group.blocks.1.norm2 1 1 275 276 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #275=(1,2304,192)f32 #276=(1,2304,192)f32 +nn.Linear layers_dfe.1.residual_group.blocks.1.mlp.fc1 1 1 276 277 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #276=(1,2304,192)f32 #277=(1,2304,384)f32 +nn.GELU layers_dfe.1.residual_group.blocks.1.mlp.act 1 1 277 278 #277=(1,2304,384)f32 #278=(1,2304,384)f32 +nn.Linear layers_dfe.1.residual_group.blocks.1.mlp.fc2 1 1 278 279 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #278=(1,2304,384)f32 #279=(1,2304,192)f32 +pnnx.Expression pnnx_expr_15761 2 1 275 279 280 expr=add(@0,@1) #275=(1,2304,192)f32 #279=(1,2304,192)f32 #280=(1,2304,192)f32 +nn.LayerNorm layers_dfe.1.residual_group.blocks.2.norm1 1 1 280 281 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #280=(1,2304,192)f32 #281=(1,2304,192)f32 +Tensor.reshape Tensor.view_1064 1 1 281 282 shape=(1,6,8,6,8,192) $input=281 #281=(1,2304,192)f32 #282=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2560 1 1 282 283 dims=(0,1,3,2,4,5) $input=282 #282=(1,6,8,6,8,192)f32 #283=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1066 1 1 283 284 shape=(36,64,192) $input=283 #283=(1,6,6,8,8,192)f32 #284=(36,64,192)f32 +nn.Linear layers_dfe.1.residual_group.blocks.2.attn.qkv 1 1 284 285 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #284=(36,64,192)f32 #285=(36,64,576)f32 +Tensor.reshape Tensor.reshape_450 1 1 285 286 shape=(36,64,3,6,32) $input=285 #285=(36,64,576)f32 #286=(36,64,3,6,32)f32 +torch.permute torch.permute_2561 1 1 286 287 dims=(2,0,3,1,4) $input=286 #286=(36,64,3,6,32)f32 #287=(3,36,6,64,32)f32 +torch.unbind Tensor.select_674 1 3 287 288 289 290 dim=0 #287=(3,36,6,64,32)f32 #288=(36,6,64,32)f32 #289=(36,6,64,32)f32 #290=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_15673 1 1 288 291 expr=mul(@0,1.767767e-01) #288=(36,6,64,32)f32 #291=(36,6,64,32)f32 +torch.transpose torch.transpose_2979 1 1 289 292 dim0=-2 dim1=-1 $input=289 #289=(36,6,64,32)f32 #292=(36,6,32,64)f32 +torch.matmul torch.matmul_2220 2 1 291 292 293 $input=291 $other=292 #291=(36,6,64,32)f32 #292=(36,6,32,64)f32 #293=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_3659 0 1 294 @pnnx_fold_3659=(1,6,64,64)f32 #294=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_15660 2 1 293 294 295 expr=add(@0,@1) #293=(36,6,64,64)f32 #294=(1,6,64,64)f32 #295=(36,6,64,64)f32 +nn.Softmax layers_dfe.1.residual_group.blocks.2.attn.softmax 1 1 295 296 dim=-1 #295=(36,6,64,64)f32 #296=(36,6,64,64)f32 +torch.matmul torch.matmul_2221 2 1 296 290 297 $input=296 $other=290 #296=(36,6,64,64)f32 #290=(36,6,64,32)f32 #297=(36,6,64,32)f32 +torch.transpose torch.transpose_2980 1 1 297 298 dim0=1 dim1=2 $input=297 #297=(36,6,64,32)f32 #298=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_451 1 1 298 299 shape=(36,64,192) $input=298 #298=(36,64,6,32)f32 #299=(36,64,192)f32 +nn.Linear layers_dfe.1.residual_group.blocks.2.attn.proj 1 1 299 300 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #299=(36,64,192)f32 #300=(36,64,192)f32 +Tensor.reshape Tensor.view_1070 1 1 300 301 shape=(1,6,6,8,8,192) $input=300 #300=(36,64,192)f32 #301=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2563 1 1 301 302 dims=(0,1,3,2,4,5) $input=301 #301=(1,6,6,8,8,192)f32 #302=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1072 1 1 302 303 shape=(1,2304,192) $input=302 #302=(1,6,8,6,8,192)f32 #303=(1,2304,192)f32 +pnnx.Expression pnnx_expr_15627 2 1 280 303 304 expr=add(@0,@1) #280=(1,2304,192)f32 #303=(1,2304,192)f32 #304=(1,2304,192)f32 +nn.LayerNorm layers_dfe.1.residual_group.blocks.2.norm2 1 1 304 305 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #304=(1,2304,192)f32 #305=(1,2304,192)f32 +nn.Linear layers_dfe.1.residual_group.blocks.2.mlp.fc1 1 1 305 306 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #305=(1,2304,192)f32 #306=(1,2304,384)f32 +nn.GELU layers_dfe.1.residual_group.blocks.2.mlp.act 1 1 306 307 #306=(1,2304,384)f32 #307=(1,2304,384)f32 +nn.Linear layers_dfe.1.residual_group.blocks.2.mlp.fc2 1 1 307 308 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #307=(1,2304,384)f32 #308=(1,2304,192)f32 +pnnx.Expression pnnx_expr_15625 2 1 304 308 309 expr=add(@0,@1) #304=(1,2304,192)f32 #308=(1,2304,192)f32 #309=(1,2304,192)f32 +nn.LayerNorm layers_dfe.1.residual_group.blocks.3.norm1 1 1 309 310 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #309=(1,2304,192)f32 #310=(1,2304,192)f32 +Tensor.view Tensor.view_1073 1 1 310 311 shape=(1,48,48,192) $input=310 #310=(1,2304,192)f32 #311=(1,48,48,192)f32 +torch.roll torch.roll_2428 1 1 311 312 dims=(1,2) shifts=(-4,-4) $input=311 #311=(1,48,48,192)f32 #312=(1,48,48,192)f32 +Tensor.view Tensor.view_1074 1 1 312 313 shape=(1,6,8,6,8,192) $input=312 #312=(1,48,48,192)f32 #313=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2564 1 1 313 314 dims=(0,1,3,2,4,5) $input=313 #313=(1,6,8,6,8,192)f32 #314=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1076 1 1 314 315 shape=(36,64,192) $input=314 #314=(1,6,6,8,8,192)f32 #315=(36,64,192)f32 +nn.Linear layers_dfe.1.residual_group.blocks.3.attn.qkv 1 1 315 316 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #315=(36,64,192)f32 #316=(36,64,576)f32 +Tensor.reshape Tensor.reshape_452 1 1 316 317 shape=(36,64,3,6,32) $input=316 #316=(36,64,576)f32 #317=(36,64,3,6,32)f32 +torch.permute torch.permute_2565 1 1 317 318 dims=(2,0,3,1,4) $input=317 #317=(36,64,3,6,32)f32 #318=(3,36,6,64,32)f32 +torch.unbind Tensor.select_677 1 3 318 319 320 321 dim=0 #318=(3,36,6,64,32)f32 #319=(36,6,64,32)f32 #320=(36,6,64,32)f32 #321=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_15527 1 1 319 322 expr=mul(@0,1.767767e-01) #319=(36,6,64,32)f32 #322=(36,6,64,32)f32 +torch.transpose torch.transpose_2981 1 1 320 323 dim0=-2 dim1=-1 $input=320 #320=(36,6,64,32)f32 #323=(36,6,32,64)f32 +torch.matmul torch.matmul_2222 2 1 322 323 324 $input=322 $other=323 #322=(36,6,64,32)f32 #323=(36,6,32,64)f32 #324=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_3812 0 1 325 @pnnx_fold_3812=(1,6,64,64)f32 #325=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_15514 2 1 324 325 326 expr=add(@0,@1) #324=(36,6,64,64)f32 #325=(1,6,64,64)f32 #326=(36,6,64,64)f32 +Tensor.view Tensor.view_1079 1 1 326 327 shape=(1,36,6,64,64) $input=326 #326=(36,6,64,64)f32 #327=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_3822 0 1 328 @pnnx_fold_3822=(1,36,1,64,64)f32 #328=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_15502 2 1 327 328 329 expr=add(@0,@1) #327=(1,36,6,64,64)f32 #328=(1,36,1,64,64)f32 #329=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1080 1 1 329 330 shape=(-1,6,64,64) $input=329 #329=(1,36,6,64,64)f32 #330=(36,6,64,64)f32 +nn.Softmax layers_dfe.1.residual_group.blocks.3.attn.softmax 1 1 330 331 dim=-1 #330=(36,6,64,64)f32 #331=(36,6,64,64)f32 +torch.matmul torch.matmul_2223 2 1 331 321 332 $input=331 $other=321 #331=(36,6,64,64)f32 #321=(36,6,64,32)f32 #332=(36,6,64,32)f32 +torch.transpose torch.transpose_2982 1 1 332 333 dim0=1 dim1=2 $input=332 #332=(36,6,64,32)f32 #333=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_453 1 1 333 334 shape=(36,64,192) $input=333 #333=(36,64,6,32)f32 #334=(36,64,192)f32 +nn.Linear layers_dfe.1.residual_group.blocks.3.attn.proj 1 1 334 335 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #334=(36,64,192)f32 #335=(36,64,192)f32 +Tensor.reshape Tensor.view_1082 1 1 335 336 shape=(1,6,6,8,8,192) $input=335 #335=(36,64,192)f32 #336=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2567 1 1 336 337 dims=(0,1,3,2,4,5) $input=336 #336=(1,6,6,8,8,192)f32 #337=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_19 1 1 337 338 shape=(1,48,48,-1) $input=337 #337=(1,6,8,6,8,192)f32 #338=(1,48,48,192)f32 +torch.roll torch.roll_2429 1 1 338 339 dims=(1,2) shifts=(4,4) $input=338 #338=(1,48,48,192)f32 #339=(1,48,48,192)f32 +Tensor.view Tensor.view_1084 1 1 339 340 shape=(1,2304,192) $input=339 #339=(1,48,48,192)f32 #340=(1,2304,192)f32 +pnnx.Expression pnnx_expr_15460 2 1 309 340 341 expr=add(@0,@1) #309=(1,2304,192)f32 #340=(1,2304,192)f32 #341=(1,2304,192)f32 +nn.LayerNorm layers_dfe.1.residual_group.blocks.3.norm2 1 1 341 342 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #341=(1,2304,192)f32 #342=(1,2304,192)f32 +nn.Linear layers_dfe.1.residual_group.blocks.3.mlp.fc1 1 1 342 343 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #342=(1,2304,192)f32 #343=(1,2304,384)f32 +nn.GELU layers_dfe.1.residual_group.blocks.3.mlp.act 1 1 343 344 #343=(1,2304,384)f32 #344=(1,2304,384)f32 +nn.Linear layers_dfe.1.residual_group.blocks.3.mlp.fc2 1 1 344 345 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #344=(1,2304,384)f32 #345=(1,2304,192)f32 +pnnx.Expression pnnx_expr_15458 2 1 341 345 346 expr=add(@0,@1) #341=(1,2304,192)f32 #345=(1,2304,192)f32 #346=(1,2304,192)f32 +nn.LayerNorm layers_dfe.1.residual_group.blocks.4.norm1 1 1 346 347 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #346=(1,2304,192)f32 #347=(1,2304,192)f32 +Tensor.reshape Tensor.view_1086 1 1 347 348 shape=(1,6,8,6,8,192) $input=347 #347=(1,2304,192)f32 #348=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2568 1 1 348 349 dims=(0,1,3,2,4,5) $input=348 #348=(1,6,8,6,8,192)f32 #349=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1088 1 1 349 350 shape=(36,64,192) $input=349 #349=(1,6,6,8,8,192)f32 #350=(36,64,192)f32 +nn.Linear layers_dfe.1.residual_group.blocks.4.attn.qkv 1 1 350 351 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #350=(36,64,192)f32 #351=(36,64,576)f32 +Tensor.reshape Tensor.reshape_454 1 1 351 352 shape=(36,64,3,6,32) $input=351 #351=(36,64,576)f32 #352=(36,64,3,6,32)f32 +torch.permute torch.permute_2569 1 1 352 353 dims=(2,0,3,1,4) $input=352 #352=(36,64,3,6,32)f32 #353=(3,36,6,64,32)f32 +torch.unbind Tensor.select_680 1 3 353 354 355 356 dim=0 #353=(3,36,6,64,32)f32 #354=(36,6,64,32)f32 #355=(36,6,64,32)f32 #356=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_15370 1 1 354 357 expr=mul(@0,1.767767e-01) #354=(36,6,64,32)f32 #357=(36,6,64,32)f32 +torch.transpose torch.transpose_2983 1 1 355 358 dim0=-2 dim1=-1 $input=355 #355=(36,6,64,32)f32 #358=(36,6,32,64)f32 +torch.matmul torch.matmul_2224 2 1 357 358 359 $input=357 $other=358 #357=(36,6,64,32)f32 #358=(36,6,32,64)f32 #359=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_3971 0 1 360 @pnnx_fold_3971=(1,6,64,64)f32 #360=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_15357 2 1 359 360 361 expr=add(@0,@1) #359=(36,6,64,64)f32 #360=(1,6,64,64)f32 #361=(36,6,64,64)f32 +nn.Softmax layers_dfe.1.residual_group.blocks.4.attn.softmax 1 1 361 362 dim=-1 #361=(36,6,64,64)f32 #362=(36,6,64,64)f32 +torch.matmul torch.matmul_2225 2 1 362 356 363 $input=362 $other=356 #362=(36,6,64,64)f32 #356=(36,6,64,32)f32 #363=(36,6,64,32)f32 +torch.transpose torch.transpose_2984 1 1 363 364 dim0=1 dim1=2 $input=363 #363=(36,6,64,32)f32 #364=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_455 1 1 364 365 shape=(36,64,192) $input=364 #364=(36,64,6,32)f32 #365=(36,64,192)f32 +nn.Linear layers_dfe.1.residual_group.blocks.4.attn.proj 1 1 365 366 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #365=(36,64,192)f32 #366=(36,64,192)f32 +Tensor.reshape Tensor.view_1092 1 1 366 367 shape=(1,6,6,8,8,192) $input=366 #366=(36,64,192)f32 #367=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2571 1 1 367 368 dims=(0,1,3,2,4,5) $input=367 #367=(1,6,6,8,8,192)f32 #368=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1094 1 1 368 369 shape=(1,2304,192) $input=368 #368=(1,6,8,6,8,192)f32 #369=(1,2304,192)f32 +pnnx.Expression pnnx_expr_15324 2 1 346 369 370 expr=add(@0,@1) #346=(1,2304,192)f32 #369=(1,2304,192)f32 #370=(1,2304,192)f32 +nn.LayerNorm layers_dfe.1.residual_group.blocks.4.norm2 1 1 370 371 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #370=(1,2304,192)f32 #371=(1,2304,192)f32 +nn.Linear layers_dfe.1.residual_group.blocks.4.mlp.fc1 1 1 371 372 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #371=(1,2304,192)f32 #372=(1,2304,384)f32 +nn.GELU layers_dfe.1.residual_group.blocks.4.mlp.act 1 1 372 373 #372=(1,2304,384)f32 #373=(1,2304,384)f32 +nn.Linear layers_dfe.1.residual_group.blocks.4.mlp.fc2 1 1 373 374 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #373=(1,2304,384)f32 #374=(1,2304,192)f32 +pnnx.Expression pnnx_expr_15322 2 1 370 374 375 expr=add(@0,@1) #370=(1,2304,192)f32 #374=(1,2304,192)f32 #375=(1,2304,192)f32 +nn.LayerNorm layers_dfe.1.residual_group.blocks.5.norm1 1 1 375 376 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #375=(1,2304,192)f32 #376=(1,2304,192)f32 +Tensor.view Tensor.view_1095 1 1 376 377 shape=(1,48,48,192) $input=376 #376=(1,2304,192)f32 #377=(1,48,48,192)f32 +torch.roll torch.roll_2430 1 1 377 378 dims=(1,2) shifts=(-4,-4) $input=377 #377=(1,48,48,192)f32 #378=(1,48,48,192)f32 +Tensor.view Tensor.view_1096 1 1 378 379 shape=(1,6,8,6,8,192) $input=378 #378=(1,48,48,192)f32 #379=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2572 1 1 379 380 dims=(0,1,3,2,4,5) $input=379 #379=(1,6,8,6,8,192)f32 #380=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1098 1 1 380 381 shape=(36,64,192) $input=380 #380=(1,6,6,8,8,192)f32 #381=(36,64,192)f32 +nn.Linear layers_dfe.1.residual_group.blocks.5.attn.qkv 1 1 381 382 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #381=(36,64,192)f32 #382=(36,64,576)f32 +Tensor.reshape Tensor.reshape_456 1 1 382 383 shape=(36,64,3,6,32) $input=382 #382=(36,64,576)f32 #383=(36,64,3,6,32)f32 +torch.permute torch.permute_2573 1 1 383 384 dims=(2,0,3,1,4) $input=383 #383=(36,64,3,6,32)f32 #384=(3,36,6,64,32)f32 +torch.unbind Tensor.select_683 1 3 384 385 386 387 dim=0 #384=(3,36,6,64,32)f32 #385=(36,6,64,32)f32 #386=(36,6,64,32)f32 #387=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_15224 1 1 385 388 expr=mul(@0,1.767767e-01) #385=(36,6,64,32)f32 #388=(36,6,64,32)f32 +torch.transpose torch.transpose_2985 1 1 386 389 dim0=-2 dim1=-1 $input=386 #386=(36,6,64,32)f32 #389=(36,6,32,64)f32 +torch.matmul torch.matmul_2226 2 1 388 389 390 $input=388 $other=389 #388=(36,6,64,32)f32 #389=(36,6,32,64)f32 #390=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_4124 0 1 391 @pnnx_fold_4124=(1,6,64,64)f32 #391=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_15211 2 1 390 391 392 expr=add(@0,@1) #390=(36,6,64,64)f32 #391=(1,6,64,64)f32 #392=(36,6,64,64)f32 +Tensor.view Tensor.view_1101 1 1 392 393 shape=(1,36,6,64,64) $input=392 #392=(36,6,64,64)f32 #393=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_4134 0 1 394 @pnnx_fold_4134=(1,36,1,64,64)f32 #394=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_15199 2 1 393 394 395 expr=add(@0,@1) #393=(1,36,6,64,64)f32 #394=(1,36,1,64,64)f32 #395=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1102 1 1 395 396 shape=(-1,6,64,64) $input=395 #395=(1,36,6,64,64)f32 #396=(36,6,64,64)f32 +nn.Softmax layers_dfe.1.residual_group.blocks.5.attn.softmax 1 1 396 397 dim=-1 #396=(36,6,64,64)f32 #397=(36,6,64,64)f32 +torch.matmul torch.matmul_2227 2 1 397 387 398 $input=397 $other=387 #397=(36,6,64,64)f32 #387=(36,6,64,32)f32 #398=(36,6,64,32)f32 +torch.transpose torch.transpose_2986 1 1 398 399 dim0=1 dim1=2 $input=398 #398=(36,6,64,32)f32 #399=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_457 1 1 399 400 shape=(36,64,192) $input=399 #399=(36,64,6,32)f32 #400=(36,64,192)f32 +nn.Linear layers_dfe.1.residual_group.blocks.5.attn.proj 1 1 400 401 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #400=(36,64,192)f32 #401=(36,64,192)f32 +Tensor.reshape Tensor.view_1104 1 1 401 402 shape=(1,6,6,8,8,192) $input=401 #401=(36,64,192)f32 #402=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2575 1 1 402 403 dims=(0,1,3,2,4,5) $input=402 #402=(1,6,6,8,8,192)f32 #403=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_23 1 1 403 404 shape=(1,48,48,-1) $input=403 #403=(1,6,8,6,8,192)f32 #404=(1,48,48,192)f32 +torch.roll torch.roll_2431 1 1 404 405 dims=(1,2) shifts=(4,4) $input=404 #404=(1,48,48,192)f32 #405=(1,48,48,192)f32 +Tensor.view Tensor.view_1106 1 1 405 406 shape=(1,2304,192) $input=405 #405=(1,48,48,192)f32 #406=(1,2304,192)f32 +pnnx.Expression pnnx_expr_15157 2 1 375 406 407 expr=add(@0,@1) #375=(1,2304,192)f32 #406=(1,2304,192)f32 #407=(1,2304,192)f32 +nn.LayerNorm layers_dfe.1.residual_group.blocks.5.norm2 1 1 407 408 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #407=(1,2304,192)f32 #408=(1,2304,192)f32 +nn.Linear layers_dfe.1.residual_group.blocks.5.mlp.fc1 1 1 408 409 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #408=(1,2304,192)f32 #409=(1,2304,384)f32 +nn.GELU layers_dfe.1.residual_group.blocks.5.mlp.act 1 1 409 410 #409=(1,2304,384)f32 #410=(1,2304,384)f32 +nn.Linear layers_dfe.1.residual_group.blocks.5.mlp.fc2 1 1 410 411 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #410=(1,2304,384)f32 #411=(1,2304,192)f32 +pnnx.Expression pnnx_expr_15155 2 1 407 411 412 expr=add(@0,@1) #407=(1,2304,192)f32 #411=(1,2304,192)f32 #412=(1,2304,192)f32 +torch.transpose torch.transpose_2987 1 1 412 413 dim0=1 dim1=2 $input=412 #412=(1,2304,192)f32 #413=(1,192,2304)f32 +Tensor.view Tensor.view_1107 1 1 413 414 shape=(1,192,48,48) $input=413 #413=(1,192,2304)f32 #414=(1,192,48,48)f32 +nn.Conv2d layers_dfe.1.conv 1 1 414 415 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #414=(1,192,48,48)f32 #415=(1,192,48,48)f32 +torch.flatten torch.flatten_2185 1 1 415 416 end_dim=-1 start_dim=2 $input=415 #415=(1,192,48,48)f32 #416=(1,192,2304)f32 +torch.transpose torch.transpose_2988 1 1 416 417 dim0=1 dim1=2 $input=416 #416=(1,192,2304)f32 #417=(1,2304,192)f32 +pnnx.Expression pnnx_expr_15142 2 1 417 214 418 expr=add(@0,@1) #417=(1,2304,192)f32 #214=(1,2304,192)f32 #418=(1,2304,192)f32 +nn.LayerNorm layers_dfe.2.residual_group.blocks.0.norm1 1 1 418 419 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #418=(1,2304,192)f32 #419=(1,2304,192)f32 +Tensor.reshape Tensor.view_1109 1 1 419 420 shape=(1,6,8,6,8,192) $input=419 #419=(1,2304,192)f32 #420=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2576 1 1 420 421 dims=(0,1,3,2,4,5) $input=420 #420=(1,6,8,6,8,192)f32 #421=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1111 1 1 421 422 shape=(36,64,192) $input=421 #421=(1,6,6,8,8,192)f32 #422=(36,64,192)f32 +nn.Linear layers_dfe.2.residual_group.blocks.0.attn.qkv 1 1 422 423 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #422=(36,64,192)f32 #423=(36,64,576)f32 +Tensor.reshape Tensor.reshape_458 1 1 423 424 shape=(36,64,3,6,32) $input=423 #423=(36,64,576)f32 #424=(36,64,3,6,32)f32 +torch.permute torch.permute_2577 1 1 424 425 dims=(2,0,3,1,4) $input=424 #424=(36,64,3,6,32)f32 #425=(3,36,6,64,32)f32 +torch.unbind Tensor.select_686 1 3 425 426 427 428 dim=0 #425=(3,36,6,64,32)f32 #426=(36,6,64,32)f32 #427=(36,6,64,32)f32 #428=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_15053 1 1 426 429 expr=mul(@0,1.767767e-01) #426=(36,6,64,32)f32 #429=(36,6,64,32)f32 +torch.transpose torch.transpose_2989 1 1 427 430 dim0=-2 dim1=-1 $input=427 #427=(36,6,64,32)f32 #430=(36,6,32,64)f32 +torch.matmul torch.matmul_2228 2 1 429 430 431 $input=429 $other=430 #429=(36,6,64,32)f32 #430=(36,6,32,64)f32 #431=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_4317 0 1 432 @pnnx_fold_4317=(1,6,64,64)f32 #432=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_15040 2 1 431 432 433 expr=add(@0,@1) #431=(36,6,64,64)f32 #432=(1,6,64,64)f32 #433=(36,6,64,64)f32 +nn.Softmax layers_dfe.2.residual_group.blocks.0.attn.softmax 1 1 433 434 dim=-1 #433=(36,6,64,64)f32 #434=(36,6,64,64)f32 +torch.matmul torch.matmul_2229 2 1 434 428 435 $input=434 $other=428 #434=(36,6,64,64)f32 #428=(36,6,64,32)f32 #435=(36,6,64,32)f32 +torch.transpose torch.transpose_2990 1 1 435 436 dim0=1 dim1=2 $input=435 #435=(36,6,64,32)f32 #436=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_459 1 1 436 437 shape=(36,64,192) $input=436 #436=(36,64,6,32)f32 #437=(36,64,192)f32 +nn.Linear layers_dfe.2.residual_group.blocks.0.attn.proj 1 1 437 438 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #437=(36,64,192)f32 #438=(36,64,192)f32 +Tensor.reshape Tensor.view_1115 1 1 438 439 shape=(1,6,6,8,8,192) $input=438 #438=(36,64,192)f32 #439=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2579 1 1 439 440 dims=(0,1,3,2,4,5) $input=439 #439=(1,6,6,8,8,192)f32 #440=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1117 1 1 440 441 shape=(1,2304,192) $input=440 #440=(1,6,8,6,8,192)f32 #441=(1,2304,192)f32 +pnnx.Expression pnnx_expr_15007 2 1 418 441 442 expr=add(@0,@1) #418=(1,2304,192)f32 #441=(1,2304,192)f32 #442=(1,2304,192)f32 +nn.LayerNorm layers_dfe.2.residual_group.blocks.0.norm2 1 1 442 443 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #442=(1,2304,192)f32 #443=(1,2304,192)f32 +nn.Linear layers_dfe.2.residual_group.blocks.0.mlp.fc1 1 1 443 444 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #443=(1,2304,192)f32 #444=(1,2304,384)f32 +nn.GELU layers_dfe.2.residual_group.blocks.0.mlp.act 1 1 444 445 #444=(1,2304,384)f32 #445=(1,2304,384)f32 +nn.Linear layers_dfe.2.residual_group.blocks.0.mlp.fc2 1 1 445 446 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #445=(1,2304,384)f32 #446=(1,2304,192)f32 +pnnx.Expression pnnx_expr_15005 2 1 442 446 447 expr=add(@0,@1) #442=(1,2304,192)f32 #446=(1,2304,192)f32 #447=(1,2304,192)f32 +nn.LayerNorm layers_dfe.2.residual_group.blocks.1.norm1 1 1 447 448 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #447=(1,2304,192)f32 #448=(1,2304,192)f32 +Tensor.view Tensor.view_1118 1 1 448 449 shape=(1,48,48,192) $input=448 #448=(1,2304,192)f32 #449=(1,48,48,192)f32 +torch.roll torch.roll_2432 1 1 449 450 dims=(1,2) shifts=(-4,-4) $input=449 #449=(1,48,48,192)f32 #450=(1,48,48,192)f32 +Tensor.view Tensor.view_1119 1 1 450 451 shape=(1,6,8,6,8,192) $input=450 #450=(1,48,48,192)f32 #451=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2580 1 1 451 452 dims=(0,1,3,2,4,5) $input=451 #451=(1,6,8,6,8,192)f32 #452=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1121 1 1 452 453 shape=(36,64,192) $input=452 #452=(1,6,6,8,8,192)f32 #453=(36,64,192)f32 +nn.Linear layers_dfe.2.residual_group.blocks.1.attn.qkv 1 1 453 454 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #453=(36,64,192)f32 #454=(36,64,576)f32 +Tensor.reshape Tensor.reshape_460 1 1 454 455 shape=(36,64,3,6,32) $input=454 #454=(36,64,576)f32 #455=(36,64,3,6,32)f32 +torch.permute torch.permute_2581 1 1 455 456 dims=(2,0,3,1,4) $input=455 #455=(36,64,3,6,32)f32 #456=(3,36,6,64,32)f32 +torch.unbind Tensor.select_689 1 3 456 457 458 459 dim=0 #456=(3,36,6,64,32)f32 #457=(36,6,64,32)f32 #458=(36,6,64,32)f32 #459=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_14907 1 1 457 460 expr=mul(@0,1.767767e-01) #457=(36,6,64,32)f32 #460=(36,6,64,32)f32 +torch.transpose torch.transpose_2991 1 1 458 461 dim0=-2 dim1=-1 $input=458 #458=(36,6,64,32)f32 #461=(36,6,32,64)f32 +torch.matmul torch.matmul_2230 2 1 460 461 462 $input=460 $other=461 #460=(36,6,64,32)f32 #461=(36,6,32,64)f32 #462=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_4470 0 1 463 @pnnx_fold_4470=(1,6,64,64)f32 #463=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_14894 2 1 462 463 464 expr=add(@0,@1) #462=(36,6,64,64)f32 #463=(1,6,64,64)f32 #464=(36,6,64,64)f32 +Tensor.view Tensor.view_1124 1 1 464 465 shape=(1,36,6,64,64) $input=464 #464=(36,6,64,64)f32 #465=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_4480 0 1 466 @pnnx_fold_4480=(1,36,1,64,64)f32 #466=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_14882 2 1 465 466 467 expr=add(@0,@1) #465=(1,36,6,64,64)f32 #466=(1,36,1,64,64)f32 #467=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1125 1 1 467 468 shape=(-1,6,64,64) $input=467 #467=(1,36,6,64,64)f32 #468=(36,6,64,64)f32 +nn.Softmax layers_dfe.2.residual_group.blocks.1.attn.softmax 1 1 468 469 dim=-1 #468=(36,6,64,64)f32 #469=(36,6,64,64)f32 +torch.matmul torch.matmul_2231 2 1 469 459 470 $input=469 $other=459 #469=(36,6,64,64)f32 #459=(36,6,64,32)f32 #470=(36,6,64,32)f32 +torch.transpose torch.transpose_2992 1 1 470 471 dim0=1 dim1=2 $input=470 #470=(36,6,64,32)f32 #471=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_461 1 1 471 472 shape=(36,64,192) $input=471 #471=(36,64,6,32)f32 #472=(36,64,192)f32 +nn.Linear layers_dfe.2.residual_group.blocks.1.attn.proj 1 1 472 473 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #472=(36,64,192)f32 #473=(36,64,192)f32 +Tensor.reshape Tensor.view_1127 1 1 473 474 shape=(1,6,6,8,8,192) $input=473 #473=(36,64,192)f32 #474=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2583 1 1 474 475 dims=(0,1,3,2,4,5) $input=474 #474=(1,6,6,8,8,192)f32 #475=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_27 1 1 475 476 shape=(1,48,48,-1) $input=475 #475=(1,6,8,6,8,192)f32 #476=(1,48,48,192)f32 +torch.roll torch.roll_2433 1 1 476 477 dims=(1,2) shifts=(4,4) $input=476 #476=(1,48,48,192)f32 #477=(1,48,48,192)f32 +Tensor.view Tensor.view_1129 1 1 477 478 shape=(1,2304,192) $input=477 #477=(1,48,48,192)f32 #478=(1,2304,192)f32 +pnnx.Expression pnnx_expr_14840 2 1 447 478 479 expr=add(@0,@1) #447=(1,2304,192)f32 #478=(1,2304,192)f32 #479=(1,2304,192)f32 +nn.LayerNorm layers_dfe.2.residual_group.blocks.1.norm2 1 1 479 480 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #479=(1,2304,192)f32 #480=(1,2304,192)f32 +nn.Linear layers_dfe.2.residual_group.blocks.1.mlp.fc1 1 1 480 481 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #480=(1,2304,192)f32 #481=(1,2304,384)f32 +nn.GELU layers_dfe.2.residual_group.blocks.1.mlp.act 1 1 481 482 #481=(1,2304,384)f32 #482=(1,2304,384)f32 +nn.Linear layers_dfe.2.residual_group.blocks.1.mlp.fc2 1 1 482 483 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #482=(1,2304,384)f32 #483=(1,2304,192)f32 +pnnx.Expression pnnx_expr_14838 2 1 479 483 484 expr=add(@0,@1) #479=(1,2304,192)f32 #483=(1,2304,192)f32 #484=(1,2304,192)f32 +nn.LayerNorm layers_dfe.2.residual_group.blocks.2.norm1 1 1 484 485 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #484=(1,2304,192)f32 #485=(1,2304,192)f32 +Tensor.reshape Tensor.view_1131 1 1 485 486 shape=(1,6,8,6,8,192) $input=485 #485=(1,2304,192)f32 #486=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2584 1 1 486 487 dims=(0,1,3,2,4,5) $input=486 #486=(1,6,8,6,8,192)f32 #487=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1133 1 1 487 488 shape=(36,64,192) $input=487 #487=(1,6,6,8,8,192)f32 #488=(36,64,192)f32 +nn.Linear layers_dfe.2.residual_group.blocks.2.attn.qkv 1 1 488 489 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #488=(36,64,192)f32 #489=(36,64,576)f32 +Tensor.reshape Tensor.reshape_462 1 1 489 490 shape=(36,64,3,6,32) $input=489 #489=(36,64,576)f32 #490=(36,64,3,6,32)f32 +torch.permute torch.permute_2585 1 1 490 491 dims=(2,0,3,1,4) $input=490 #490=(36,64,3,6,32)f32 #491=(3,36,6,64,32)f32 +torch.unbind Tensor.select_692 1 3 491 492 493 494 dim=0 #491=(3,36,6,64,32)f32 #492=(36,6,64,32)f32 #493=(36,6,64,32)f32 #494=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_14750 1 1 492 495 expr=mul(@0,1.767767e-01) #492=(36,6,64,32)f32 #495=(36,6,64,32)f32 +torch.transpose torch.transpose_2993 1 1 493 496 dim0=-2 dim1=-1 $input=493 #493=(36,6,64,32)f32 #496=(36,6,32,64)f32 +torch.matmul torch.matmul_2232 2 1 495 496 497 $input=495 $other=496 #495=(36,6,64,32)f32 #496=(36,6,32,64)f32 #497=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_4629 0 1 498 @pnnx_fold_4629=(1,6,64,64)f32 #498=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_14737 2 1 497 498 499 expr=add(@0,@1) #497=(36,6,64,64)f32 #498=(1,6,64,64)f32 #499=(36,6,64,64)f32 +nn.Softmax layers_dfe.2.residual_group.blocks.2.attn.softmax 1 1 499 500 dim=-1 #499=(36,6,64,64)f32 #500=(36,6,64,64)f32 +torch.matmul torch.matmul_2233 2 1 500 494 501 $input=500 $other=494 #500=(36,6,64,64)f32 #494=(36,6,64,32)f32 #501=(36,6,64,32)f32 +torch.transpose torch.transpose_2994 1 1 501 502 dim0=1 dim1=2 $input=501 #501=(36,6,64,32)f32 #502=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_463 1 1 502 503 shape=(36,64,192) $input=502 #502=(36,64,6,32)f32 #503=(36,64,192)f32 +nn.Linear layers_dfe.2.residual_group.blocks.2.attn.proj 1 1 503 504 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #503=(36,64,192)f32 #504=(36,64,192)f32 +Tensor.reshape Tensor.view_1137 1 1 504 505 shape=(1,6,6,8,8,192) $input=504 #504=(36,64,192)f32 #505=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2587 1 1 505 506 dims=(0,1,3,2,4,5) $input=505 #505=(1,6,6,8,8,192)f32 #506=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1139 1 1 506 507 shape=(1,2304,192) $input=506 #506=(1,6,8,6,8,192)f32 #507=(1,2304,192)f32 +pnnx.Expression pnnx_expr_14704 2 1 484 507 508 expr=add(@0,@1) #484=(1,2304,192)f32 #507=(1,2304,192)f32 #508=(1,2304,192)f32 +nn.LayerNorm layers_dfe.2.residual_group.blocks.2.norm2 1 1 508 509 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #508=(1,2304,192)f32 #509=(1,2304,192)f32 +nn.Linear layers_dfe.2.residual_group.blocks.2.mlp.fc1 1 1 509 510 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #509=(1,2304,192)f32 #510=(1,2304,384)f32 +nn.GELU layers_dfe.2.residual_group.blocks.2.mlp.act 1 1 510 511 #510=(1,2304,384)f32 #511=(1,2304,384)f32 +nn.Linear layers_dfe.2.residual_group.blocks.2.mlp.fc2 1 1 511 512 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #511=(1,2304,384)f32 #512=(1,2304,192)f32 +pnnx.Expression pnnx_expr_14702 2 1 508 512 513 expr=add(@0,@1) #508=(1,2304,192)f32 #512=(1,2304,192)f32 #513=(1,2304,192)f32 +nn.LayerNorm layers_dfe.2.residual_group.blocks.3.norm1 1 1 513 514 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #513=(1,2304,192)f32 #514=(1,2304,192)f32 +Tensor.view Tensor.view_1140 1 1 514 515 shape=(1,48,48,192) $input=514 #514=(1,2304,192)f32 #515=(1,48,48,192)f32 +torch.roll torch.roll_2434 1 1 515 516 dims=(1,2) shifts=(-4,-4) $input=515 #515=(1,48,48,192)f32 #516=(1,48,48,192)f32 +Tensor.view Tensor.view_1141 1 1 516 517 shape=(1,6,8,6,8,192) $input=516 #516=(1,48,48,192)f32 #517=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2588 1 1 517 518 dims=(0,1,3,2,4,5) $input=517 #517=(1,6,8,6,8,192)f32 #518=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1143 1 1 518 519 shape=(36,64,192) $input=518 #518=(1,6,6,8,8,192)f32 #519=(36,64,192)f32 +nn.Linear layers_dfe.2.residual_group.blocks.3.attn.qkv 1 1 519 520 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #519=(36,64,192)f32 #520=(36,64,576)f32 +Tensor.reshape Tensor.reshape_464 1 1 520 521 shape=(36,64,3,6,32) $input=520 #520=(36,64,576)f32 #521=(36,64,3,6,32)f32 +torch.permute torch.permute_2589 1 1 521 522 dims=(2,0,3,1,4) $input=521 #521=(36,64,3,6,32)f32 #522=(3,36,6,64,32)f32 +torch.unbind Tensor.select_695 1 3 522 523 524 525 dim=0 #522=(3,36,6,64,32)f32 #523=(36,6,64,32)f32 #524=(36,6,64,32)f32 #525=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_14604 1 1 523 526 expr=mul(@0,1.767767e-01) #523=(36,6,64,32)f32 #526=(36,6,64,32)f32 +torch.transpose torch.transpose_2995 1 1 524 527 dim0=-2 dim1=-1 $input=524 #524=(36,6,64,32)f32 #527=(36,6,32,64)f32 +torch.matmul torch.matmul_2234 2 1 526 527 528 $input=526 $other=527 #526=(36,6,64,32)f32 #527=(36,6,32,64)f32 #528=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_4782 0 1 529 @pnnx_fold_4782=(1,6,64,64)f32 #529=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_14591 2 1 528 529 530 expr=add(@0,@1) #528=(36,6,64,64)f32 #529=(1,6,64,64)f32 #530=(36,6,64,64)f32 +Tensor.view Tensor.view_1146 1 1 530 531 shape=(1,36,6,64,64) $input=530 #530=(36,6,64,64)f32 #531=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_4792 0 1 532 @pnnx_fold_4792=(1,36,1,64,64)f32 #532=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_14579 2 1 531 532 533 expr=add(@0,@1) #531=(1,36,6,64,64)f32 #532=(1,36,1,64,64)f32 #533=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1147 1 1 533 534 shape=(-1,6,64,64) $input=533 #533=(1,36,6,64,64)f32 #534=(36,6,64,64)f32 +nn.Softmax layers_dfe.2.residual_group.blocks.3.attn.softmax 1 1 534 535 dim=-1 #534=(36,6,64,64)f32 #535=(36,6,64,64)f32 +torch.matmul torch.matmul_2235 2 1 535 525 536 $input=535 $other=525 #535=(36,6,64,64)f32 #525=(36,6,64,32)f32 #536=(36,6,64,32)f32 +torch.transpose torch.transpose_2996 1 1 536 537 dim0=1 dim1=2 $input=536 #536=(36,6,64,32)f32 #537=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_465 1 1 537 538 shape=(36,64,192) $input=537 #537=(36,64,6,32)f32 #538=(36,64,192)f32 +nn.Linear layers_dfe.2.residual_group.blocks.3.attn.proj 1 1 538 539 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #538=(36,64,192)f32 #539=(36,64,192)f32 +Tensor.reshape Tensor.view_1149 1 1 539 540 shape=(1,6,6,8,8,192) $input=539 #539=(36,64,192)f32 #540=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2591 1 1 540 541 dims=(0,1,3,2,4,5) $input=540 #540=(1,6,6,8,8,192)f32 #541=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_31 1 1 541 542 shape=(1,48,48,-1) $input=541 #541=(1,6,8,6,8,192)f32 #542=(1,48,48,192)f32 +torch.roll torch.roll_2435 1 1 542 543 dims=(1,2) shifts=(4,4) $input=542 #542=(1,48,48,192)f32 #543=(1,48,48,192)f32 +Tensor.view Tensor.view_1151 1 1 543 544 shape=(1,2304,192) $input=543 #543=(1,48,48,192)f32 #544=(1,2304,192)f32 +pnnx.Expression pnnx_expr_14537 2 1 513 544 545 expr=add(@0,@1) #513=(1,2304,192)f32 #544=(1,2304,192)f32 #545=(1,2304,192)f32 +nn.LayerNorm layers_dfe.2.residual_group.blocks.3.norm2 1 1 545 546 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #545=(1,2304,192)f32 #546=(1,2304,192)f32 +nn.Linear layers_dfe.2.residual_group.blocks.3.mlp.fc1 1 1 546 547 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #546=(1,2304,192)f32 #547=(1,2304,384)f32 +nn.GELU layers_dfe.2.residual_group.blocks.3.mlp.act 1 1 547 548 #547=(1,2304,384)f32 #548=(1,2304,384)f32 +nn.Linear layers_dfe.2.residual_group.blocks.3.mlp.fc2 1 1 548 549 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #548=(1,2304,384)f32 #549=(1,2304,192)f32 +pnnx.Expression pnnx_expr_14535 2 1 545 549 550 expr=add(@0,@1) #545=(1,2304,192)f32 #549=(1,2304,192)f32 #550=(1,2304,192)f32 +nn.LayerNorm layers_dfe.2.residual_group.blocks.4.norm1 1 1 550 551 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #550=(1,2304,192)f32 #551=(1,2304,192)f32 +Tensor.reshape Tensor.view_1153 1 1 551 552 shape=(1,6,8,6,8,192) $input=551 #551=(1,2304,192)f32 #552=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2592 1 1 552 553 dims=(0,1,3,2,4,5) $input=552 #552=(1,6,8,6,8,192)f32 #553=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1155 1 1 553 554 shape=(36,64,192) $input=553 #553=(1,6,6,8,8,192)f32 #554=(36,64,192)f32 +nn.Linear layers_dfe.2.residual_group.blocks.4.attn.qkv 1 1 554 555 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #554=(36,64,192)f32 #555=(36,64,576)f32 +Tensor.reshape Tensor.reshape_466 1 1 555 556 shape=(36,64,3,6,32) $input=555 #555=(36,64,576)f32 #556=(36,64,3,6,32)f32 +torch.permute torch.permute_2593 1 1 556 557 dims=(2,0,3,1,4) $input=556 #556=(36,64,3,6,32)f32 #557=(3,36,6,64,32)f32 +torch.unbind Tensor.select_698 1 3 557 558 559 560 dim=0 #557=(3,36,6,64,32)f32 #558=(36,6,64,32)f32 #559=(36,6,64,32)f32 #560=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_14447 1 1 558 561 expr=mul(@0,1.767767e-01) #558=(36,6,64,32)f32 #561=(36,6,64,32)f32 +torch.transpose torch.transpose_2997 1 1 559 562 dim0=-2 dim1=-1 $input=559 #559=(36,6,64,32)f32 #562=(36,6,32,64)f32 +torch.matmul torch.matmul_2236 2 1 561 562 563 $input=561 $other=562 #561=(36,6,64,32)f32 #562=(36,6,32,64)f32 #563=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_4941 0 1 564 @pnnx_fold_4941=(1,6,64,64)f32 #564=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_14434 2 1 563 564 565 expr=add(@0,@1) #563=(36,6,64,64)f32 #564=(1,6,64,64)f32 #565=(36,6,64,64)f32 +nn.Softmax layers_dfe.2.residual_group.blocks.4.attn.softmax 1 1 565 566 dim=-1 #565=(36,6,64,64)f32 #566=(36,6,64,64)f32 +torch.matmul torch.matmul_2237 2 1 566 560 567 $input=566 $other=560 #566=(36,6,64,64)f32 #560=(36,6,64,32)f32 #567=(36,6,64,32)f32 +torch.transpose torch.transpose_2998 1 1 567 568 dim0=1 dim1=2 $input=567 #567=(36,6,64,32)f32 #568=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_467 1 1 568 569 shape=(36,64,192) $input=568 #568=(36,64,6,32)f32 #569=(36,64,192)f32 +nn.Linear layers_dfe.2.residual_group.blocks.4.attn.proj 1 1 569 570 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #569=(36,64,192)f32 #570=(36,64,192)f32 +Tensor.reshape Tensor.view_1159 1 1 570 571 shape=(1,6,6,8,8,192) $input=570 #570=(36,64,192)f32 #571=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2595 1 1 571 572 dims=(0,1,3,2,4,5) $input=571 #571=(1,6,6,8,8,192)f32 #572=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1161 1 1 572 573 shape=(1,2304,192) $input=572 #572=(1,6,8,6,8,192)f32 #573=(1,2304,192)f32 +pnnx.Expression pnnx_expr_14401 2 1 550 573 574 expr=add(@0,@1) #550=(1,2304,192)f32 #573=(1,2304,192)f32 #574=(1,2304,192)f32 +nn.LayerNorm layers_dfe.2.residual_group.blocks.4.norm2 1 1 574 575 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #574=(1,2304,192)f32 #575=(1,2304,192)f32 +nn.Linear layers_dfe.2.residual_group.blocks.4.mlp.fc1 1 1 575 576 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #575=(1,2304,192)f32 #576=(1,2304,384)f32 +nn.GELU layers_dfe.2.residual_group.blocks.4.mlp.act 1 1 576 577 #576=(1,2304,384)f32 #577=(1,2304,384)f32 +nn.Linear layers_dfe.2.residual_group.blocks.4.mlp.fc2 1 1 577 578 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #577=(1,2304,384)f32 #578=(1,2304,192)f32 +pnnx.Expression pnnx_expr_14399 2 1 574 578 579 expr=add(@0,@1) #574=(1,2304,192)f32 #578=(1,2304,192)f32 #579=(1,2304,192)f32 +nn.LayerNorm layers_dfe.2.residual_group.blocks.5.norm1 1 1 579 580 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #579=(1,2304,192)f32 #580=(1,2304,192)f32 +Tensor.view Tensor.view_1162 1 1 580 581 shape=(1,48,48,192) $input=580 #580=(1,2304,192)f32 #581=(1,48,48,192)f32 +torch.roll torch.roll_2436 1 1 581 582 dims=(1,2) shifts=(-4,-4) $input=581 #581=(1,48,48,192)f32 #582=(1,48,48,192)f32 +Tensor.view Tensor.view_1163 1 1 582 583 shape=(1,6,8,6,8,192) $input=582 #582=(1,48,48,192)f32 #583=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2596 1 1 583 584 dims=(0,1,3,2,4,5) $input=583 #583=(1,6,8,6,8,192)f32 #584=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1165 1 1 584 585 shape=(36,64,192) $input=584 #584=(1,6,6,8,8,192)f32 #585=(36,64,192)f32 +nn.Linear layers_dfe.2.residual_group.blocks.5.attn.qkv 1 1 585 586 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #585=(36,64,192)f32 #586=(36,64,576)f32 +Tensor.reshape Tensor.reshape_468 1 1 586 587 shape=(36,64,3,6,32) $input=586 #586=(36,64,576)f32 #587=(36,64,3,6,32)f32 +torch.permute torch.permute_2597 1 1 587 588 dims=(2,0,3,1,4) $input=587 #587=(36,64,3,6,32)f32 #588=(3,36,6,64,32)f32 +torch.unbind Tensor.select_701 1 3 588 589 590 591 dim=0 #588=(3,36,6,64,32)f32 #589=(36,6,64,32)f32 #590=(36,6,64,32)f32 #591=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_14301 1 1 589 592 expr=mul(@0,1.767767e-01) #589=(36,6,64,32)f32 #592=(36,6,64,32)f32 +torch.transpose torch.transpose_2999 1 1 590 593 dim0=-2 dim1=-1 $input=590 #590=(36,6,64,32)f32 #593=(36,6,32,64)f32 +torch.matmul torch.matmul_2238 2 1 592 593 594 $input=592 $other=593 #592=(36,6,64,32)f32 #593=(36,6,32,64)f32 #594=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_5094 0 1 595 @pnnx_fold_5094=(1,6,64,64)f32 #595=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_14288 2 1 594 595 596 expr=add(@0,@1) #594=(36,6,64,64)f32 #595=(1,6,64,64)f32 #596=(36,6,64,64)f32 +Tensor.view Tensor.view_1168 1 1 596 597 shape=(1,36,6,64,64) $input=596 #596=(36,6,64,64)f32 #597=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_5104 0 1 598 @pnnx_fold_5104=(1,36,1,64,64)f32 #598=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_14276 2 1 597 598 599 expr=add(@0,@1) #597=(1,36,6,64,64)f32 #598=(1,36,1,64,64)f32 #599=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1169 1 1 599 600 shape=(-1,6,64,64) $input=599 #599=(1,36,6,64,64)f32 #600=(36,6,64,64)f32 +nn.Softmax layers_dfe.2.residual_group.blocks.5.attn.softmax 1 1 600 601 dim=-1 #600=(36,6,64,64)f32 #601=(36,6,64,64)f32 +torch.matmul torch.matmul_2239 2 1 601 591 602 $input=601 $other=591 #601=(36,6,64,64)f32 #591=(36,6,64,32)f32 #602=(36,6,64,32)f32 +torch.transpose torch.transpose_3000 1 1 602 603 dim0=1 dim1=2 $input=602 #602=(36,6,64,32)f32 #603=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_469 1 1 603 604 shape=(36,64,192) $input=603 #603=(36,64,6,32)f32 #604=(36,64,192)f32 +nn.Linear layers_dfe.2.residual_group.blocks.5.attn.proj 1 1 604 605 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #604=(36,64,192)f32 #605=(36,64,192)f32 +Tensor.reshape Tensor.view_1171 1 1 605 606 shape=(1,6,6,8,8,192) $input=605 #605=(36,64,192)f32 #606=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2599 1 1 606 607 dims=(0,1,3,2,4,5) $input=606 #606=(1,6,6,8,8,192)f32 #607=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_35 1 1 607 608 shape=(1,48,48,-1) $input=607 #607=(1,6,8,6,8,192)f32 #608=(1,48,48,192)f32 +torch.roll torch.roll_2437 1 1 608 609 dims=(1,2) shifts=(4,4) $input=608 #608=(1,48,48,192)f32 #609=(1,48,48,192)f32 +Tensor.view Tensor.view_1173 1 1 609 610 shape=(1,2304,192) $input=609 #609=(1,48,48,192)f32 #610=(1,2304,192)f32 +pnnx.Expression pnnx_expr_14234 2 1 579 610 611 expr=add(@0,@1) #579=(1,2304,192)f32 #610=(1,2304,192)f32 #611=(1,2304,192)f32 +nn.LayerNorm layers_dfe.2.residual_group.blocks.5.norm2 1 1 611 612 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #611=(1,2304,192)f32 #612=(1,2304,192)f32 +nn.Linear layers_dfe.2.residual_group.blocks.5.mlp.fc1 1 1 612 613 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #612=(1,2304,192)f32 #613=(1,2304,384)f32 +nn.GELU layers_dfe.2.residual_group.blocks.5.mlp.act 1 1 613 614 #613=(1,2304,384)f32 #614=(1,2304,384)f32 +nn.Linear layers_dfe.2.residual_group.blocks.5.mlp.fc2 1 1 614 615 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #614=(1,2304,384)f32 #615=(1,2304,192)f32 +pnnx.Expression pnnx_expr_14232 2 1 611 615 616 expr=add(@0,@1) #611=(1,2304,192)f32 #615=(1,2304,192)f32 #616=(1,2304,192)f32 +torch.transpose torch.transpose_3001 1 1 616 617 dim0=1 dim1=2 $input=616 #616=(1,2304,192)f32 #617=(1,192,2304)f32 +Tensor.view Tensor.view_1174 1 1 617 618 shape=(1,192,48,48) $input=617 #617=(1,192,2304)f32 #618=(1,192,48,48)f32 +nn.Conv2d layers_dfe.2.conv 1 1 618 619 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #618=(1,192,48,48)f32 #619=(1,192,48,48)f32 +torch.flatten torch.flatten_2186 1 1 619 620 end_dim=-1 start_dim=2 $input=619 #619=(1,192,48,48)f32 #620=(1,192,2304)f32 +torch.transpose torch.transpose_3002 1 1 620 621 dim0=1 dim1=2 $input=620 #620=(1,192,2304)f32 #621=(1,2304,192)f32 +pnnx.Expression pnnx_expr_14219 2 1 621 418 622 expr=add(@0,@1) #621=(1,2304,192)f32 #418=(1,2304,192)f32 #622=(1,2304,192)f32 +nn.LayerNorm layers_dfe.3.residual_group.blocks.0.norm1 1 1 622 623 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #622=(1,2304,192)f32 #623=(1,2304,192)f32 +Tensor.reshape Tensor.view_1176 1 1 623 624 shape=(1,6,8,6,8,192) $input=623 #623=(1,2304,192)f32 #624=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2600 1 1 624 625 dims=(0,1,3,2,4,5) $input=624 #624=(1,6,8,6,8,192)f32 #625=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1178 1 1 625 626 shape=(36,64,192) $input=625 #625=(1,6,6,8,8,192)f32 #626=(36,64,192)f32 +nn.Linear layers_dfe.3.residual_group.blocks.0.attn.qkv 1 1 626 627 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #626=(36,64,192)f32 #627=(36,64,576)f32 +Tensor.reshape Tensor.reshape_470 1 1 627 628 shape=(36,64,3,6,32) $input=627 #627=(36,64,576)f32 #628=(36,64,3,6,32)f32 +torch.permute torch.permute_2601 1 1 628 629 dims=(2,0,3,1,4) $input=628 #628=(36,64,3,6,32)f32 #629=(3,36,6,64,32)f32 +torch.unbind Tensor.select_704 1 3 629 630 631 632 dim=0 #629=(3,36,6,64,32)f32 #630=(36,6,64,32)f32 #631=(36,6,64,32)f32 #632=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_14130 1 1 630 633 expr=mul(@0,1.767767e-01) #630=(36,6,64,32)f32 #633=(36,6,64,32)f32 +torch.transpose torch.transpose_3003 1 1 631 634 dim0=-2 dim1=-1 $input=631 #631=(36,6,64,32)f32 #634=(36,6,32,64)f32 +torch.matmul torch.matmul_2240 2 1 633 634 635 $input=633 $other=634 #633=(36,6,64,32)f32 #634=(36,6,32,64)f32 #635=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_5287 0 1 636 @pnnx_fold_5287=(1,6,64,64)f32 #636=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_14117 2 1 635 636 637 expr=add(@0,@1) #635=(36,6,64,64)f32 #636=(1,6,64,64)f32 #637=(36,6,64,64)f32 +nn.Softmax layers_dfe.3.residual_group.blocks.0.attn.softmax 1 1 637 638 dim=-1 #637=(36,6,64,64)f32 #638=(36,6,64,64)f32 +torch.matmul torch.matmul_2241 2 1 638 632 639 $input=638 $other=632 #638=(36,6,64,64)f32 #632=(36,6,64,32)f32 #639=(36,6,64,32)f32 +torch.transpose torch.transpose_3004 1 1 639 640 dim0=1 dim1=2 $input=639 #639=(36,6,64,32)f32 #640=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_471 1 1 640 641 shape=(36,64,192) $input=640 #640=(36,64,6,32)f32 #641=(36,64,192)f32 +nn.Linear layers_dfe.3.residual_group.blocks.0.attn.proj 1 1 641 642 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #641=(36,64,192)f32 #642=(36,64,192)f32 +Tensor.reshape Tensor.view_1182 1 1 642 643 shape=(1,6,6,8,8,192) $input=642 #642=(36,64,192)f32 #643=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2603 1 1 643 644 dims=(0,1,3,2,4,5) $input=643 #643=(1,6,6,8,8,192)f32 #644=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1184 1 1 644 645 shape=(1,2304,192) $input=644 #644=(1,6,8,6,8,192)f32 #645=(1,2304,192)f32 +pnnx.Expression pnnx_expr_14084 2 1 622 645 646 expr=add(@0,@1) #622=(1,2304,192)f32 #645=(1,2304,192)f32 #646=(1,2304,192)f32 +nn.LayerNorm layers_dfe.3.residual_group.blocks.0.norm2 1 1 646 647 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #646=(1,2304,192)f32 #647=(1,2304,192)f32 +nn.Linear layers_dfe.3.residual_group.blocks.0.mlp.fc1 1 1 647 648 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #647=(1,2304,192)f32 #648=(1,2304,384)f32 +nn.GELU layers_dfe.3.residual_group.blocks.0.mlp.act 1 1 648 649 #648=(1,2304,384)f32 #649=(1,2304,384)f32 +nn.Linear layers_dfe.3.residual_group.blocks.0.mlp.fc2 1 1 649 650 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #649=(1,2304,384)f32 #650=(1,2304,192)f32 +pnnx.Expression pnnx_expr_14082 2 1 646 650 651 expr=add(@0,@1) #646=(1,2304,192)f32 #650=(1,2304,192)f32 #651=(1,2304,192)f32 +nn.LayerNorm layers_dfe.3.residual_group.blocks.1.norm1 1 1 651 652 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #651=(1,2304,192)f32 #652=(1,2304,192)f32 +Tensor.view Tensor.view_1185 1 1 652 653 shape=(1,48,48,192) $input=652 #652=(1,2304,192)f32 #653=(1,48,48,192)f32 +torch.roll torch.roll_2438 1 1 653 654 dims=(1,2) shifts=(-4,-4) $input=653 #653=(1,48,48,192)f32 #654=(1,48,48,192)f32 +Tensor.view Tensor.view_1186 1 1 654 655 shape=(1,6,8,6,8,192) $input=654 #654=(1,48,48,192)f32 #655=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2604 1 1 655 656 dims=(0,1,3,2,4,5) $input=655 #655=(1,6,8,6,8,192)f32 #656=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1188 1 1 656 657 shape=(36,64,192) $input=656 #656=(1,6,6,8,8,192)f32 #657=(36,64,192)f32 +nn.Linear layers_dfe.3.residual_group.blocks.1.attn.qkv 1 1 657 658 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #657=(36,64,192)f32 #658=(36,64,576)f32 +Tensor.reshape Tensor.reshape_472 1 1 658 659 shape=(36,64,3,6,32) $input=658 #658=(36,64,576)f32 #659=(36,64,3,6,32)f32 +torch.permute torch.permute_2605 1 1 659 660 dims=(2,0,3,1,4) $input=659 #659=(36,64,3,6,32)f32 #660=(3,36,6,64,32)f32 +torch.unbind Tensor.select_707 1 3 660 661 662 663 dim=0 #660=(3,36,6,64,32)f32 #661=(36,6,64,32)f32 #662=(36,6,64,32)f32 #663=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_13984 1 1 661 664 expr=mul(@0,1.767767e-01) #661=(36,6,64,32)f32 #664=(36,6,64,32)f32 +torch.transpose torch.transpose_3005 1 1 662 665 dim0=-2 dim1=-1 $input=662 #662=(36,6,64,32)f32 #665=(36,6,32,64)f32 +torch.matmul torch.matmul_2242 2 1 664 665 666 $input=664 $other=665 #664=(36,6,64,32)f32 #665=(36,6,32,64)f32 #666=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_5440 0 1 667 @pnnx_fold_5440=(1,6,64,64)f32 #667=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_13971 2 1 666 667 668 expr=add(@0,@1) #666=(36,6,64,64)f32 #667=(1,6,64,64)f32 #668=(36,6,64,64)f32 +Tensor.view Tensor.view_1191 1 1 668 669 shape=(1,36,6,64,64) $input=668 #668=(36,6,64,64)f32 #669=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_5450 0 1 670 @pnnx_fold_5450=(1,36,1,64,64)f32 #670=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_13959 2 1 669 670 671 expr=add(@0,@1) #669=(1,36,6,64,64)f32 #670=(1,36,1,64,64)f32 #671=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1192 1 1 671 672 shape=(-1,6,64,64) $input=671 #671=(1,36,6,64,64)f32 #672=(36,6,64,64)f32 +nn.Softmax layers_dfe.3.residual_group.blocks.1.attn.softmax 1 1 672 673 dim=-1 #672=(36,6,64,64)f32 #673=(36,6,64,64)f32 +torch.matmul torch.matmul_2243 2 1 673 663 674 $input=673 $other=663 #673=(36,6,64,64)f32 #663=(36,6,64,32)f32 #674=(36,6,64,32)f32 +torch.transpose torch.transpose_3006 1 1 674 675 dim0=1 dim1=2 $input=674 #674=(36,6,64,32)f32 #675=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_473 1 1 675 676 shape=(36,64,192) $input=675 #675=(36,64,6,32)f32 #676=(36,64,192)f32 +nn.Linear layers_dfe.3.residual_group.blocks.1.attn.proj 1 1 676 677 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #676=(36,64,192)f32 #677=(36,64,192)f32 +Tensor.reshape Tensor.view_1194 1 1 677 678 shape=(1,6,6,8,8,192) $input=677 #677=(36,64,192)f32 #678=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2607 1 1 678 679 dims=(0,1,3,2,4,5) $input=678 #678=(1,6,6,8,8,192)f32 #679=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_39 1 1 679 680 shape=(1,48,48,-1) $input=679 #679=(1,6,8,6,8,192)f32 #680=(1,48,48,192)f32 +torch.roll torch.roll_2439 1 1 680 681 dims=(1,2) shifts=(4,4) $input=680 #680=(1,48,48,192)f32 #681=(1,48,48,192)f32 +Tensor.view Tensor.view_1196 1 1 681 682 shape=(1,2304,192) $input=681 #681=(1,48,48,192)f32 #682=(1,2304,192)f32 +pnnx.Expression pnnx_expr_13917 2 1 651 682 683 expr=add(@0,@1) #651=(1,2304,192)f32 #682=(1,2304,192)f32 #683=(1,2304,192)f32 +nn.LayerNorm layers_dfe.3.residual_group.blocks.1.norm2 1 1 683 684 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #683=(1,2304,192)f32 #684=(1,2304,192)f32 +nn.Linear layers_dfe.3.residual_group.blocks.1.mlp.fc1 1 1 684 685 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #684=(1,2304,192)f32 #685=(1,2304,384)f32 +nn.GELU layers_dfe.3.residual_group.blocks.1.mlp.act 1 1 685 686 #685=(1,2304,384)f32 #686=(1,2304,384)f32 +nn.Linear layers_dfe.3.residual_group.blocks.1.mlp.fc2 1 1 686 687 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #686=(1,2304,384)f32 #687=(1,2304,192)f32 +pnnx.Expression pnnx_expr_13915 2 1 683 687 688 expr=add(@0,@1) #683=(1,2304,192)f32 #687=(1,2304,192)f32 #688=(1,2304,192)f32 +nn.LayerNorm layers_dfe.3.residual_group.blocks.2.norm1 1 1 688 689 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #688=(1,2304,192)f32 #689=(1,2304,192)f32 +Tensor.reshape Tensor.view_1198 1 1 689 690 shape=(1,6,8,6,8,192) $input=689 #689=(1,2304,192)f32 #690=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2608 1 1 690 691 dims=(0,1,3,2,4,5) $input=690 #690=(1,6,8,6,8,192)f32 #691=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1200 1 1 691 692 shape=(36,64,192) $input=691 #691=(1,6,6,8,8,192)f32 #692=(36,64,192)f32 +nn.Linear layers_dfe.3.residual_group.blocks.2.attn.qkv 1 1 692 693 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #692=(36,64,192)f32 #693=(36,64,576)f32 +Tensor.reshape Tensor.reshape_474 1 1 693 694 shape=(36,64,3,6,32) $input=693 #693=(36,64,576)f32 #694=(36,64,3,6,32)f32 +torch.permute torch.permute_2609 1 1 694 695 dims=(2,0,3,1,4) $input=694 #694=(36,64,3,6,32)f32 #695=(3,36,6,64,32)f32 +torch.unbind Tensor.select_710 1 3 695 696 697 698 dim=0 #695=(3,36,6,64,32)f32 #696=(36,6,64,32)f32 #697=(36,6,64,32)f32 #698=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_13827 1 1 696 699 expr=mul(@0,1.767767e-01) #696=(36,6,64,32)f32 #699=(36,6,64,32)f32 +torch.transpose torch.transpose_3007 1 1 697 700 dim0=-2 dim1=-1 $input=697 #697=(36,6,64,32)f32 #700=(36,6,32,64)f32 +torch.matmul torch.matmul_2244 2 1 699 700 701 $input=699 $other=700 #699=(36,6,64,32)f32 #700=(36,6,32,64)f32 #701=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_5599 0 1 702 @pnnx_fold_5599=(1,6,64,64)f32 #702=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_13814 2 1 701 702 703 expr=add(@0,@1) #701=(36,6,64,64)f32 #702=(1,6,64,64)f32 #703=(36,6,64,64)f32 +nn.Softmax layers_dfe.3.residual_group.blocks.2.attn.softmax 1 1 703 704 dim=-1 #703=(36,6,64,64)f32 #704=(36,6,64,64)f32 +torch.matmul torch.matmul_2245 2 1 704 698 705 $input=704 $other=698 #704=(36,6,64,64)f32 #698=(36,6,64,32)f32 #705=(36,6,64,32)f32 +torch.transpose torch.transpose_3008 1 1 705 706 dim0=1 dim1=2 $input=705 #705=(36,6,64,32)f32 #706=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_475 1 1 706 707 shape=(36,64,192) $input=706 #706=(36,64,6,32)f32 #707=(36,64,192)f32 +nn.Linear layers_dfe.3.residual_group.blocks.2.attn.proj 1 1 707 708 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #707=(36,64,192)f32 #708=(36,64,192)f32 +Tensor.reshape Tensor.view_1204 1 1 708 709 shape=(1,6,6,8,8,192) $input=708 #708=(36,64,192)f32 #709=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2611 1 1 709 710 dims=(0,1,3,2,4,5) $input=709 #709=(1,6,6,8,8,192)f32 #710=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1206 1 1 710 711 shape=(1,2304,192) $input=710 #710=(1,6,8,6,8,192)f32 #711=(1,2304,192)f32 +pnnx.Expression pnnx_expr_13781 2 1 688 711 712 expr=add(@0,@1) #688=(1,2304,192)f32 #711=(1,2304,192)f32 #712=(1,2304,192)f32 +nn.LayerNorm layers_dfe.3.residual_group.blocks.2.norm2 1 1 712 713 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #712=(1,2304,192)f32 #713=(1,2304,192)f32 +nn.Linear layers_dfe.3.residual_group.blocks.2.mlp.fc1 1 1 713 714 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #713=(1,2304,192)f32 #714=(1,2304,384)f32 +nn.GELU layers_dfe.3.residual_group.blocks.2.mlp.act 1 1 714 715 #714=(1,2304,384)f32 #715=(1,2304,384)f32 +nn.Linear layers_dfe.3.residual_group.blocks.2.mlp.fc2 1 1 715 716 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #715=(1,2304,384)f32 #716=(1,2304,192)f32 +pnnx.Expression pnnx_expr_13779 2 1 712 716 717 expr=add(@0,@1) #712=(1,2304,192)f32 #716=(1,2304,192)f32 #717=(1,2304,192)f32 +nn.LayerNorm layers_dfe.3.residual_group.blocks.3.norm1 1 1 717 718 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #717=(1,2304,192)f32 #718=(1,2304,192)f32 +Tensor.view Tensor.view_1207 1 1 718 719 shape=(1,48,48,192) $input=718 #718=(1,2304,192)f32 #719=(1,48,48,192)f32 +torch.roll torch.roll_2440 1 1 719 720 dims=(1,2) shifts=(-4,-4) $input=719 #719=(1,48,48,192)f32 #720=(1,48,48,192)f32 +Tensor.view Tensor.view_1208 1 1 720 721 shape=(1,6,8,6,8,192) $input=720 #720=(1,48,48,192)f32 #721=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2612 1 1 721 722 dims=(0,1,3,2,4,5) $input=721 #721=(1,6,8,6,8,192)f32 #722=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1210 1 1 722 723 shape=(36,64,192) $input=722 #722=(1,6,6,8,8,192)f32 #723=(36,64,192)f32 +nn.Linear layers_dfe.3.residual_group.blocks.3.attn.qkv 1 1 723 724 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #723=(36,64,192)f32 #724=(36,64,576)f32 +Tensor.reshape Tensor.reshape_476 1 1 724 725 shape=(36,64,3,6,32) $input=724 #724=(36,64,576)f32 #725=(36,64,3,6,32)f32 +torch.permute torch.permute_2613 1 1 725 726 dims=(2,0,3,1,4) $input=725 #725=(36,64,3,6,32)f32 #726=(3,36,6,64,32)f32 +torch.unbind Tensor.select_713 1 3 726 727 728 729 dim=0 #726=(3,36,6,64,32)f32 #727=(36,6,64,32)f32 #728=(36,6,64,32)f32 #729=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_13681 1 1 727 730 expr=mul(@0,1.767767e-01) #727=(36,6,64,32)f32 #730=(36,6,64,32)f32 +torch.transpose torch.transpose_3009 1 1 728 731 dim0=-2 dim1=-1 $input=728 #728=(36,6,64,32)f32 #731=(36,6,32,64)f32 +torch.matmul torch.matmul_2246 2 1 730 731 732 $input=730 $other=731 #730=(36,6,64,32)f32 #731=(36,6,32,64)f32 #732=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_5752 0 1 733 @pnnx_fold_5752=(1,6,64,64)f32 #733=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_13668 2 1 732 733 734 expr=add(@0,@1) #732=(36,6,64,64)f32 #733=(1,6,64,64)f32 #734=(36,6,64,64)f32 +Tensor.view Tensor.view_1213 1 1 734 735 shape=(1,36,6,64,64) $input=734 #734=(36,6,64,64)f32 #735=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_5762 0 1 736 @pnnx_fold_5762=(1,36,1,64,64)f32 #736=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_13656 2 1 735 736 737 expr=add(@0,@1) #735=(1,36,6,64,64)f32 #736=(1,36,1,64,64)f32 #737=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1214 1 1 737 738 shape=(-1,6,64,64) $input=737 #737=(1,36,6,64,64)f32 #738=(36,6,64,64)f32 +nn.Softmax layers_dfe.3.residual_group.blocks.3.attn.softmax 1 1 738 739 dim=-1 #738=(36,6,64,64)f32 #739=(36,6,64,64)f32 +torch.matmul torch.matmul_2247 2 1 739 729 740 $input=739 $other=729 #739=(36,6,64,64)f32 #729=(36,6,64,32)f32 #740=(36,6,64,32)f32 +torch.transpose torch.transpose_3010 1 1 740 741 dim0=1 dim1=2 $input=740 #740=(36,6,64,32)f32 #741=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_477 1 1 741 742 shape=(36,64,192) $input=741 #741=(36,64,6,32)f32 #742=(36,64,192)f32 +nn.Linear layers_dfe.3.residual_group.blocks.3.attn.proj 1 1 742 743 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #742=(36,64,192)f32 #743=(36,64,192)f32 +Tensor.reshape Tensor.view_1216 1 1 743 744 shape=(1,6,6,8,8,192) $input=743 #743=(36,64,192)f32 #744=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2615 1 1 744 745 dims=(0,1,3,2,4,5) $input=744 #744=(1,6,6,8,8,192)f32 #745=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_43 1 1 745 746 shape=(1,48,48,-1) $input=745 #745=(1,6,8,6,8,192)f32 #746=(1,48,48,192)f32 +torch.roll torch.roll_2441 1 1 746 747 dims=(1,2) shifts=(4,4) $input=746 #746=(1,48,48,192)f32 #747=(1,48,48,192)f32 +Tensor.view Tensor.view_1218 1 1 747 748 shape=(1,2304,192) $input=747 #747=(1,48,48,192)f32 #748=(1,2304,192)f32 +pnnx.Expression pnnx_expr_13614 2 1 717 748 749 expr=add(@0,@1) #717=(1,2304,192)f32 #748=(1,2304,192)f32 #749=(1,2304,192)f32 +nn.LayerNorm layers_dfe.3.residual_group.blocks.3.norm2 1 1 749 750 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #749=(1,2304,192)f32 #750=(1,2304,192)f32 +nn.Linear layers_dfe.3.residual_group.blocks.3.mlp.fc1 1 1 750 751 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #750=(1,2304,192)f32 #751=(1,2304,384)f32 +nn.GELU layers_dfe.3.residual_group.blocks.3.mlp.act 1 1 751 752 #751=(1,2304,384)f32 #752=(1,2304,384)f32 +nn.Linear layers_dfe.3.residual_group.blocks.3.mlp.fc2 1 1 752 753 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #752=(1,2304,384)f32 #753=(1,2304,192)f32 +pnnx.Expression pnnx_expr_13612 2 1 749 753 754 expr=add(@0,@1) #749=(1,2304,192)f32 #753=(1,2304,192)f32 #754=(1,2304,192)f32 +nn.LayerNorm layers_dfe.3.residual_group.blocks.4.norm1 1 1 754 755 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #754=(1,2304,192)f32 #755=(1,2304,192)f32 +Tensor.reshape Tensor.view_1220 1 1 755 756 shape=(1,6,8,6,8,192) $input=755 #755=(1,2304,192)f32 #756=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2616 1 1 756 757 dims=(0,1,3,2,4,5) $input=756 #756=(1,6,8,6,8,192)f32 #757=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1222 1 1 757 758 shape=(36,64,192) $input=757 #757=(1,6,6,8,8,192)f32 #758=(36,64,192)f32 +nn.Linear layers_dfe.3.residual_group.blocks.4.attn.qkv 1 1 758 759 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #758=(36,64,192)f32 #759=(36,64,576)f32 +Tensor.reshape Tensor.reshape_478 1 1 759 760 shape=(36,64,3,6,32) $input=759 #759=(36,64,576)f32 #760=(36,64,3,6,32)f32 +torch.permute torch.permute_2617 1 1 760 761 dims=(2,0,3,1,4) $input=760 #760=(36,64,3,6,32)f32 #761=(3,36,6,64,32)f32 +torch.unbind Tensor.select_716 1 3 761 762 763 764 dim=0 #761=(3,36,6,64,32)f32 #762=(36,6,64,32)f32 #763=(36,6,64,32)f32 #764=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_13524 1 1 762 765 expr=mul(@0,1.767767e-01) #762=(36,6,64,32)f32 #765=(36,6,64,32)f32 +torch.transpose torch.transpose_3011 1 1 763 766 dim0=-2 dim1=-1 $input=763 #763=(36,6,64,32)f32 #766=(36,6,32,64)f32 +torch.matmul torch.matmul_2248 2 1 765 766 767 $input=765 $other=766 #765=(36,6,64,32)f32 #766=(36,6,32,64)f32 #767=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_5911 0 1 768 @pnnx_fold_5911=(1,6,64,64)f32 #768=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_13511 2 1 767 768 769 expr=add(@0,@1) #767=(36,6,64,64)f32 #768=(1,6,64,64)f32 #769=(36,6,64,64)f32 +nn.Softmax layers_dfe.3.residual_group.blocks.4.attn.softmax 1 1 769 770 dim=-1 #769=(36,6,64,64)f32 #770=(36,6,64,64)f32 +torch.matmul torch.matmul_2249 2 1 770 764 771 $input=770 $other=764 #770=(36,6,64,64)f32 #764=(36,6,64,32)f32 #771=(36,6,64,32)f32 +torch.transpose torch.transpose_3012 1 1 771 772 dim0=1 dim1=2 $input=771 #771=(36,6,64,32)f32 #772=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_479 1 1 772 773 shape=(36,64,192) $input=772 #772=(36,64,6,32)f32 #773=(36,64,192)f32 +nn.Linear layers_dfe.3.residual_group.blocks.4.attn.proj 1 1 773 774 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #773=(36,64,192)f32 #774=(36,64,192)f32 +Tensor.reshape Tensor.view_1226 1 1 774 775 shape=(1,6,6,8,8,192) $input=774 #774=(36,64,192)f32 #775=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2619 1 1 775 776 dims=(0,1,3,2,4,5) $input=775 #775=(1,6,6,8,8,192)f32 #776=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1228 1 1 776 777 shape=(1,2304,192) $input=776 #776=(1,6,8,6,8,192)f32 #777=(1,2304,192)f32 +pnnx.Expression pnnx_expr_13478 2 1 754 777 778 expr=add(@0,@1) #754=(1,2304,192)f32 #777=(1,2304,192)f32 #778=(1,2304,192)f32 +nn.LayerNorm layers_dfe.3.residual_group.blocks.4.norm2 1 1 778 779 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #778=(1,2304,192)f32 #779=(1,2304,192)f32 +nn.Linear layers_dfe.3.residual_group.blocks.4.mlp.fc1 1 1 779 780 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #779=(1,2304,192)f32 #780=(1,2304,384)f32 +nn.GELU layers_dfe.3.residual_group.blocks.4.mlp.act 1 1 780 781 #780=(1,2304,384)f32 #781=(1,2304,384)f32 +nn.Linear layers_dfe.3.residual_group.blocks.4.mlp.fc2 1 1 781 782 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #781=(1,2304,384)f32 #782=(1,2304,192)f32 +pnnx.Expression pnnx_expr_13476 2 1 778 782 783 expr=add(@0,@1) #778=(1,2304,192)f32 #782=(1,2304,192)f32 #783=(1,2304,192)f32 +nn.LayerNorm layers_dfe.3.residual_group.blocks.5.norm1 1 1 783 784 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #783=(1,2304,192)f32 #784=(1,2304,192)f32 +Tensor.view Tensor.view_1229 1 1 784 785 shape=(1,48,48,192) $input=784 #784=(1,2304,192)f32 #785=(1,48,48,192)f32 +torch.roll torch.roll_2442 1 1 785 786 dims=(1,2) shifts=(-4,-4) $input=785 #785=(1,48,48,192)f32 #786=(1,48,48,192)f32 +Tensor.view Tensor.view_1230 1 1 786 787 shape=(1,6,8,6,8,192) $input=786 #786=(1,48,48,192)f32 #787=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2620 1 1 787 788 dims=(0,1,3,2,4,5) $input=787 #787=(1,6,8,6,8,192)f32 #788=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1232 1 1 788 789 shape=(36,64,192) $input=788 #788=(1,6,6,8,8,192)f32 #789=(36,64,192)f32 +nn.Linear layers_dfe.3.residual_group.blocks.5.attn.qkv 1 1 789 790 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #789=(36,64,192)f32 #790=(36,64,576)f32 +Tensor.reshape Tensor.reshape_480 1 1 790 791 shape=(36,64,3,6,32) $input=790 #790=(36,64,576)f32 #791=(36,64,3,6,32)f32 +torch.permute torch.permute_2621 1 1 791 792 dims=(2,0,3,1,4) $input=791 #791=(36,64,3,6,32)f32 #792=(3,36,6,64,32)f32 +torch.unbind Tensor.select_719 1 3 792 793 794 795 dim=0 #792=(3,36,6,64,32)f32 #793=(36,6,64,32)f32 #794=(36,6,64,32)f32 #795=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_13378 1 1 793 796 expr=mul(@0,1.767767e-01) #793=(36,6,64,32)f32 #796=(36,6,64,32)f32 +torch.transpose torch.transpose_3013 1 1 794 797 dim0=-2 dim1=-1 $input=794 #794=(36,6,64,32)f32 #797=(36,6,32,64)f32 +torch.matmul torch.matmul_2250 2 1 796 797 798 $input=796 $other=797 #796=(36,6,64,32)f32 #797=(36,6,32,64)f32 #798=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_6064 0 1 799 @pnnx_fold_6064=(1,6,64,64)f32 #799=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_13365 2 1 798 799 800 expr=add(@0,@1) #798=(36,6,64,64)f32 #799=(1,6,64,64)f32 #800=(36,6,64,64)f32 +Tensor.view Tensor.view_1235 1 1 800 801 shape=(1,36,6,64,64) $input=800 #800=(36,6,64,64)f32 #801=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_6074 0 1 802 @pnnx_fold_6074=(1,36,1,64,64)f32 #802=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_13353 2 1 801 802 803 expr=add(@0,@1) #801=(1,36,6,64,64)f32 #802=(1,36,1,64,64)f32 #803=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1236 1 1 803 804 shape=(-1,6,64,64) $input=803 #803=(1,36,6,64,64)f32 #804=(36,6,64,64)f32 +nn.Softmax layers_dfe.3.residual_group.blocks.5.attn.softmax 1 1 804 805 dim=-1 #804=(36,6,64,64)f32 #805=(36,6,64,64)f32 +torch.matmul torch.matmul_2251 2 1 805 795 806 $input=805 $other=795 #805=(36,6,64,64)f32 #795=(36,6,64,32)f32 #806=(36,6,64,32)f32 +torch.transpose torch.transpose_3014 1 1 806 807 dim0=1 dim1=2 $input=806 #806=(36,6,64,32)f32 #807=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_481 1 1 807 808 shape=(36,64,192) $input=807 #807=(36,64,6,32)f32 #808=(36,64,192)f32 +nn.Linear layers_dfe.3.residual_group.blocks.5.attn.proj 1 1 808 809 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #808=(36,64,192)f32 #809=(36,64,192)f32 +Tensor.reshape Tensor.view_1238 1 1 809 810 shape=(1,6,6,8,8,192) $input=809 #809=(36,64,192)f32 #810=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2623 1 1 810 811 dims=(0,1,3,2,4,5) $input=810 #810=(1,6,6,8,8,192)f32 #811=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_47 1 1 811 812 shape=(1,48,48,-1) $input=811 #811=(1,6,8,6,8,192)f32 #812=(1,48,48,192)f32 +torch.roll torch.roll_2443 1 1 812 813 dims=(1,2) shifts=(4,4) $input=812 #812=(1,48,48,192)f32 #813=(1,48,48,192)f32 +Tensor.view Tensor.view_1240 1 1 813 814 shape=(1,2304,192) $input=813 #813=(1,48,48,192)f32 #814=(1,2304,192)f32 +pnnx.Expression pnnx_expr_13311 2 1 783 814 815 expr=add(@0,@1) #783=(1,2304,192)f32 #814=(1,2304,192)f32 #815=(1,2304,192)f32 +nn.LayerNorm layers_dfe.3.residual_group.blocks.5.norm2 1 1 815 816 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #815=(1,2304,192)f32 #816=(1,2304,192)f32 +nn.Linear layers_dfe.3.residual_group.blocks.5.mlp.fc1 1 1 816 817 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #816=(1,2304,192)f32 #817=(1,2304,384)f32 +nn.GELU layers_dfe.3.residual_group.blocks.5.mlp.act 1 1 817 818 #817=(1,2304,384)f32 #818=(1,2304,384)f32 +nn.Linear layers_dfe.3.residual_group.blocks.5.mlp.fc2 1 1 818 819 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #818=(1,2304,384)f32 #819=(1,2304,192)f32 +pnnx.Expression pnnx_expr_13309 2 1 815 819 820 expr=add(@0,@1) #815=(1,2304,192)f32 #819=(1,2304,192)f32 #820=(1,2304,192)f32 +torch.transpose torch.transpose_3015 1 1 820 821 dim0=1 dim1=2 $input=820 #820=(1,2304,192)f32 #821=(1,192,2304)f32 +Tensor.view Tensor.view_1241 1 1 821 822 shape=(1,192,48,48) $input=821 #821=(1,192,2304)f32 #822=(1,192,48,48)f32 +nn.Conv2d layers_dfe.3.conv 1 1 822 823 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #822=(1,192,48,48)f32 #823=(1,192,48,48)f32 +torch.flatten torch.flatten_2187 1 1 823 824 end_dim=-1 start_dim=2 $input=823 #823=(1,192,48,48)f32 #824=(1,192,2304)f32 +torch.transpose torch.transpose_3016 1 1 824 825 dim0=1 dim1=2 $input=824 #824=(1,192,2304)f32 #825=(1,2304,192)f32 +pnnx.Expression pnnx_expr_13296 2 1 825 622 826 expr=add(@0,@1) #825=(1,2304,192)f32 #622=(1,2304,192)f32 #826=(1,2304,192)f32 +nn.LayerNorm layers_dfe.4.residual_group.blocks.0.norm1 1 1 826 827 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #826=(1,2304,192)f32 #827=(1,2304,192)f32 +Tensor.reshape Tensor.view_1243 1 1 827 828 shape=(1,6,8,6,8,192) $input=827 #827=(1,2304,192)f32 #828=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2624 1 1 828 829 dims=(0,1,3,2,4,5) $input=828 #828=(1,6,8,6,8,192)f32 #829=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1245 1 1 829 830 shape=(36,64,192) $input=829 #829=(1,6,6,8,8,192)f32 #830=(36,64,192)f32 +nn.Linear layers_dfe.4.residual_group.blocks.0.attn.qkv 1 1 830 831 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #830=(36,64,192)f32 #831=(36,64,576)f32 +Tensor.reshape Tensor.reshape_482 1 1 831 832 shape=(36,64,3,6,32) $input=831 #831=(36,64,576)f32 #832=(36,64,3,6,32)f32 +torch.permute torch.permute_2625 1 1 832 833 dims=(2,0,3,1,4) $input=832 #832=(36,64,3,6,32)f32 #833=(3,36,6,64,32)f32 +torch.unbind Tensor.select_722 1 3 833 834 835 836 dim=0 #833=(3,36,6,64,32)f32 #834=(36,6,64,32)f32 #835=(36,6,64,32)f32 #836=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_13207 1 1 834 837 expr=mul(@0,1.767767e-01) #834=(36,6,64,32)f32 #837=(36,6,64,32)f32 +torch.transpose torch.transpose_3017 1 1 835 838 dim0=-2 dim1=-1 $input=835 #835=(36,6,64,32)f32 #838=(36,6,32,64)f32 +torch.matmul torch.matmul_2252 2 1 837 838 839 $input=837 $other=838 #837=(36,6,64,32)f32 #838=(36,6,32,64)f32 #839=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_6257 0 1 840 @pnnx_fold_6257=(1,6,64,64)f32 #840=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_13194 2 1 839 840 841 expr=add(@0,@1) #839=(36,6,64,64)f32 #840=(1,6,64,64)f32 #841=(36,6,64,64)f32 +nn.Softmax layers_dfe.4.residual_group.blocks.0.attn.softmax 1 1 841 842 dim=-1 #841=(36,6,64,64)f32 #842=(36,6,64,64)f32 +torch.matmul torch.matmul_2253 2 1 842 836 843 $input=842 $other=836 #842=(36,6,64,64)f32 #836=(36,6,64,32)f32 #843=(36,6,64,32)f32 +torch.transpose torch.transpose_3018 1 1 843 844 dim0=1 dim1=2 $input=843 #843=(36,6,64,32)f32 #844=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_483 1 1 844 845 shape=(36,64,192) $input=844 #844=(36,64,6,32)f32 #845=(36,64,192)f32 +nn.Linear layers_dfe.4.residual_group.blocks.0.attn.proj 1 1 845 846 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #845=(36,64,192)f32 #846=(36,64,192)f32 +Tensor.reshape Tensor.view_1249 1 1 846 847 shape=(1,6,6,8,8,192) $input=846 #846=(36,64,192)f32 #847=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2627 1 1 847 848 dims=(0,1,3,2,4,5) $input=847 #847=(1,6,6,8,8,192)f32 #848=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1251 1 1 848 849 shape=(1,2304,192) $input=848 #848=(1,6,8,6,8,192)f32 #849=(1,2304,192)f32 +pnnx.Expression pnnx_expr_13161 2 1 826 849 850 expr=add(@0,@1) #826=(1,2304,192)f32 #849=(1,2304,192)f32 #850=(1,2304,192)f32 +nn.LayerNorm layers_dfe.4.residual_group.blocks.0.norm2 1 1 850 851 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #850=(1,2304,192)f32 #851=(1,2304,192)f32 +nn.Linear layers_dfe.4.residual_group.blocks.0.mlp.fc1 1 1 851 852 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #851=(1,2304,192)f32 #852=(1,2304,384)f32 +nn.GELU layers_dfe.4.residual_group.blocks.0.mlp.act 1 1 852 853 #852=(1,2304,384)f32 #853=(1,2304,384)f32 +nn.Linear layers_dfe.4.residual_group.blocks.0.mlp.fc2 1 1 853 854 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #853=(1,2304,384)f32 #854=(1,2304,192)f32 +pnnx.Expression pnnx_expr_13159 2 1 850 854 855 expr=add(@0,@1) #850=(1,2304,192)f32 #854=(1,2304,192)f32 #855=(1,2304,192)f32 +nn.LayerNorm layers_dfe.4.residual_group.blocks.1.norm1 1 1 855 856 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #855=(1,2304,192)f32 #856=(1,2304,192)f32 +Tensor.view Tensor.view_1252 1 1 856 857 shape=(1,48,48,192) $input=856 #856=(1,2304,192)f32 #857=(1,48,48,192)f32 +torch.roll torch.roll_2444 1 1 857 858 dims=(1,2) shifts=(-4,-4) $input=857 #857=(1,48,48,192)f32 #858=(1,48,48,192)f32 +Tensor.view Tensor.view_1253 1 1 858 859 shape=(1,6,8,6,8,192) $input=858 #858=(1,48,48,192)f32 #859=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2628 1 1 859 860 dims=(0,1,3,2,4,5) $input=859 #859=(1,6,8,6,8,192)f32 #860=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1255 1 1 860 861 shape=(36,64,192) $input=860 #860=(1,6,6,8,8,192)f32 #861=(36,64,192)f32 +nn.Linear layers_dfe.4.residual_group.blocks.1.attn.qkv 1 1 861 862 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #861=(36,64,192)f32 #862=(36,64,576)f32 +Tensor.reshape Tensor.reshape_484 1 1 862 863 shape=(36,64,3,6,32) $input=862 #862=(36,64,576)f32 #863=(36,64,3,6,32)f32 +torch.permute torch.permute_2629 1 1 863 864 dims=(2,0,3,1,4) $input=863 #863=(36,64,3,6,32)f32 #864=(3,36,6,64,32)f32 +torch.unbind Tensor.select_725 1 3 864 865 866 867 dim=0 #864=(3,36,6,64,32)f32 #865=(36,6,64,32)f32 #866=(36,6,64,32)f32 #867=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_13061 1 1 865 868 expr=mul(@0,1.767767e-01) #865=(36,6,64,32)f32 #868=(36,6,64,32)f32 +torch.transpose torch.transpose_3019 1 1 866 869 dim0=-2 dim1=-1 $input=866 #866=(36,6,64,32)f32 #869=(36,6,32,64)f32 +torch.matmul torch.matmul_2254 2 1 868 869 870 $input=868 $other=869 #868=(36,6,64,32)f32 #869=(36,6,32,64)f32 #870=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_6410 0 1 871 @pnnx_fold_6410=(1,6,64,64)f32 #871=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_13048 2 1 870 871 872 expr=add(@0,@1) #870=(36,6,64,64)f32 #871=(1,6,64,64)f32 #872=(36,6,64,64)f32 +Tensor.view Tensor.view_1258 1 1 872 873 shape=(1,36,6,64,64) $input=872 #872=(36,6,64,64)f32 #873=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_6420 0 1 874 @pnnx_fold_6420=(1,36,1,64,64)f32 #874=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_13036 2 1 873 874 875 expr=add(@0,@1) #873=(1,36,6,64,64)f32 #874=(1,36,1,64,64)f32 #875=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1259 1 1 875 876 shape=(-1,6,64,64) $input=875 #875=(1,36,6,64,64)f32 #876=(36,6,64,64)f32 +nn.Softmax layers_dfe.4.residual_group.blocks.1.attn.softmax 1 1 876 877 dim=-1 #876=(36,6,64,64)f32 #877=(36,6,64,64)f32 +torch.matmul torch.matmul_2255 2 1 877 867 878 $input=877 $other=867 #877=(36,6,64,64)f32 #867=(36,6,64,32)f32 #878=(36,6,64,32)f32 +torch.transpose torch.transpose_3020 1 1 878 879 dim0=1 dim1=2 $input=878 #878=(36,6,64,32)f32 #879=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_485 1 1 879 880 shape=(36,64,192) $input=879 #879=(36,64,6,32)f32 #880=(36,64,192)f32 +nn.Linear layers_dfe.4.residual_group.blocks.1.attn.proj 1 1 880 881 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #880=(36,64,192)f32 #881=(36,64,192)f32 +Tensor.reshape Tensor.view_1261 1 1 881 882 shape=(1,6,6,8,8,192) $input=881 #881=(36,64,192)f32 #882=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2631 1 1 882 883 dims=(0,1,3,2,4,5) $input=882 #882=(1,6,6,8,8,192)f32 #883=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_51 1 1 883 884 shape=(1,48,48,-1) $input=883 #883=(1,6,8,6,8,192)f32 #884=(1,48,48,192)f32 +torch.roll torch.roll_2445 1 1 884 885 dims=(1,2) shifts=(4,4) $input=884 #884=(1,48,48,192)f32 #885=(1,48,48,192)f32 +Tensor.view Tensor.view_1263 1 1 885 886 shape=(1,2304,192) $input=885 #885=(1,48,48,192)f32 #886=(1,2304,192)f32 +pnnx.Expression pnnx_expr_12994 2 1 855 886 887 expr=add(@0,@1) #855=(1,2304,192)f32 #886=(1,2304,192)f32 #887=(1,2304,192)f32 +nn.LayerNorm layers_dfe.4.residual_group.blocks.1.norm2 1 1 887 888 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #887=(1,2304,192)f32 #888=(1,2304,192)f32 +nn.Linear layers_dfe.4.residual_group.blocks.1.mlp.fc1 1 1 888 889 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #888=(1,2304,192)f32 #889=(1,2304,384)f32 +nn.GELU layers_dfe.4.residual_group.blocks.1.mlp.act 1 1 889 890 #889=(1,2304,384)f32 #890=(1,2304,384)f32 +nn.Linear layers_dfe.4.residual_group.blocks.1.mlp.fc2 1 1 890 891 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #890=(1,2304,384)f32 #891=(1,2304,192)f32 +pnnx.Expression pnnx_expr_12992 2 1 887 891 892 expr=add(@0,@1) #887=(1,2304,192)f32 #891=(1,2304,192)f32 #892=(1,2304,192)f32 +nn.LayerNorm layers_dfe.4.residual_group.blocks.2.norm1 1 1 892 893 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #892=(1,2304,192)f32 #893=(1,2304,192)f32 +Tensor.reshape Tensor.view_1265 1 1 893 894 shape=(1,6,8,6,8,192) $input=893 #893=(1,2304,192)f32 #894=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2632 1 1 894 895 dims=(0,1,3,2,4,5) $input=894 #894=(1,6,8,6,8,192)f32 #895=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1267 1 1 895 896 shape=(36,64,192) $input=895 #895=(1,6,6,8,8,192)f32 #896=(36,64,192)f32 +nn.Linear layers_dfe.4.residual_group.blocks.2.attn.qkv 1 1 896 897 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #896=(36,64,192)f32 #897=(36,64,576)f32 +Tensor.reshape Tensor.reshape_486 1 1 897 898 shape=(36,64,3,6,32) $input=897 #897=(36,64,576)f32 #898=(36,64,3,6,32)f32 +torch.permute torch.permute_2633 1 1 898 899 dims=(2,0,3,1,4) $input=898 #898=(36,64,3,6,32)f32 #899=(3,36,6,64,32)f32 +torch.unbind Tensor.select_728 1 3 899 900 901 902 dim=0 #899=(3,36,6,64,32)f32 #900=(36,6,64,32)f32 #901=(36,6,64,32)f32 #902=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_12904 1 1 900 903 expr=mul(@0,1.767767e-01) #900=(36,6,64,32)f32 #903=(36,6,64,32)f32 +torch.transpose torch.transpose_3021 1 1 901 904 dim0=-2 dim1=-1 $input=901 #901=(36,6,64,32)f32 #904=(36,6,32,64)f32 +torch.matmul torch.matmul_2256 2 1 903 904 905 $input=903 $other=904 #903=(36,6,64,32)f32 #904=(36,6,32,64)f32 #905=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_6569 0 1 906 @pnnx_fold_6569=(1,6,64,64)f32 #906=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_12891 2 1 905 906 907 expr=add(@0,@1) #905=(36,6,64,64)f32 #906=(1,6,64,64)f32 #907=(36,6,64,64)f32 +nn.Softmax layers_dfe.4.residual_group.blocks.2.attn.softmax 1 1 907 908 dim=-1 #907=(36,6,64,64)f32 #908=(36,6,64,64)f32 +torch.matmul torch.matmul_2257 2 1 908 902 909 $input=908 $other=902 #908=(36,6,64,64)f32 #902=(36,6,64,32)f32 #909=(36,6,64,32)f32 +torch.transpose torch.transpose_3022 1 1 909 910 dim0=1 dim1=2 $input=909 #909=(36,6,64,32)f32 #910=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_487 1 1 910 911 shape=(36,64,192) $input=910 #910=(36,64,6,32)f32 #911=(36,64,192)f32 +nn.Linear layers_dfe.4.residual_group.blocks.2.attn.proj 1 1 911 912 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #911=(36,64,192)f32 #912=(36,64,192)f32 +Tensor.reshape Tensor.view_1271 1 1 912 913 shape=(1,6,6,8,8,192) $input=912 #912=(36,64,192)f32 #913=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2635 1 1 913 914 dims=(0,1,3,2,4,5) $input=913 #913=(1,6,6,8,8,192)f32 #914=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1273 1 1 914 915 shape=(1,2304,192) $input=914 #914=(1,6,8,6,8,192)f32 #915=(1,2304,192)f32 +pnnx.Expression pnnx_expr_12858 2 1 892 915 916 expr=add(@0,@1) #892=(1,2304,192)f32 #915=(1,2304,192)f32 #916=(1,2304,192)f32 +nn.LayerNorm layers_dfe.4.residual_group.blocks.2.norm2 1 1 916 917 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #916=(1,2304,192)f32 #917=(1,2304,192)f32 +nn.Linear layers_dfe.4.residual_group.blocks.2.mlp.fc1 1 1 917 918 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #917=(1,2304,192)f32 #918=(1,2304,384)f32 +nn.GELU layers_dfe.4.residual_group.blocks.2.mlp.act 1 1 918 919 #918=(1,2304,384)f32 #919=(1,2304,384)f32 +nn.Linear layers_dfe.4.residual_group.blocks.2.mlp.fc2 1 1 919 920 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #919=(1,2304,384)f32 #920=(1,2304,192)f32 +pnnx.Expression pnnx_expr_12856 2 1 916 920 921 expr=add(@0,@1) #916=(1,2304,192)f32 #920=(1,2304,192)f32 #921=(1,2304,192)f32 +nn.LayerNorm layers_dfe.4.residual_group.blocks.3.norm1 1 1 921 922 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #921=(1,2304,192)f32 #922=(1,2304,192)f32 +Tensor.view Tensor.view_1274 1 1 922 923 shape=(1,48,48,192) $input=922 #922=(1,2304,192)f32 #923=(1,48,48,192)f32 +torch.roll torch.roll_2446 1 1 923 924 dims=(1,2) shifts=(-4,-4) $input=923 #923=(1,48,48,192)f32 #924=(1,48,48,192)f32 +Tensor.view Tensor.view_1275 1 1 924 925 shape=(1,6,8,6,8,192) $input=924 #924=(1,48,48,192)f32 #925=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2636 1 1 925 926 dims=(0,1,3,2,4,5) $input=925 #925=(1,6,8,6,8,192)f32 #926=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1277 1 1 926 927 shape=(36,64,192) $input=926 #926=(1,6,6,8,8,192)f32 #927=(36,64,192)f32 +nn.Linear layers_dfe.4.residual_group.blocks.3.attn.qkv 1 1 927 928 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #927=(36,64,192)f32 #928=(36,64,576)f32 +Tensor.reshape Tensor.reshape_488 1 1 928 929 shape=(36,64,3,6,32) $input=928 #928=(36,64,576)f32 #929=(36,64,3,6,32)f32 +torch.permute torch.permute_2637 1 1 929 930 dims=(2,0,3,1,4) $input=929 #929=(36,64,3,6,32)f32 #930=(3,36,6,64,32)f32 +torch.unbind Tensor.select_731 1 3 930 931 932 933 dim=0 #930=(3,36,6,64,32)f32 #931=(36,6,64,32)f32 #932=(36,6,64,32)f32 #933=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_12758 1 1 931 934 expr=mul(@0,1.767767e-01) #931=(36,6,64,32)f32 #934=(36,6,64,32)f32 +torch.transpose torch.transpose_3023 1 1 932 935 dim0=-2 dim1=-1 $input=932 #932=(36,6,64,32)f32 #935=(36,6,32,64)f32 +torch.matmul torch.matmul_2258 2 1 934 935 936 $input=934 $other=935 #934=(36,6,64,32)f32 #935=(36,6,32,64)f32 #936=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_6722 0 1 937 @pnnx_fold_6722=(1,6,64,64)f32 #937=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_12745 2 1 936 937 938 expr=add(@0,@1) #936=(36,6,64,64)f32 #937=(1,6,64,64)f32 #938=(36,6,64,64)f32 +Tensor.view Tensor.view_1280 1 1 938 939 shape=(1,36,6,64,64) $input=938 #938=(36,6,64,64)f32 #939=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_6732 0 1 940 @pnnx_fold_6732=(1,36,1,64,64)f32 #940=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_12733 2 1 939 940 941 expr=add(@0,@1) #939=(1,36,6,64,64)f32 #940=(1,36,1,64,64)f32 #941=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1281 1 1 941 942 shape=(-1,6,64,64) $input=941 #941=(1,36,6,64,64)f32 #942=(36,6,64,64)f32 +nn.Softmax layers_dfe.4.residual_group.blocks.3.attn.softmax 1 1 942 943 dim=-1 #942=(36,6,64,64)f32 #943=(36,6,64,64)f32 +torch.matmul torch.matmul_2259 2 1 943 933 944 $input=943 $other=933 #943=(36,6,64,64)f32 #933=(36,6,64,32)f32 #944=(36,6,64,32)f32 +torch.transpose torch.transpose_3024 1 1 944 945 dim0=1 dim1=2 $input=944 #944=(36,6,64,32)f32 #945=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_489 1 1 945 946 shape=(36,64,192) $input=945 #945=(36,64,6,32)f32 #946=(36,64,192)f32 +nn.Linear layers_dfe.4.residual_group.blocks.3.attn.proj 1 1 946 947 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #946=(36,64,192)f32 #947=(36,64,192)f32 +Tensor.reshape Tensor.view_1283 1 1 947 948 shape=(1,6,6,8,8,192) $input=947 #947=(36,64,192)f32 #948=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2639 1 1 948 949 dims=(0,1,3,2,4,5) $input=948 #948=(1,6,6,8,8,192)f32 #949=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_55 1 1 949 950 shape=(1,48,48,-1) $input=949 #949=(1,6,8,6,8,192)f32 #950=(1,48,48,192)f32 +torch.roll torch.roll_2447 1 1 950 951 dims=(1,2) shifts=(4,4) $input=950 #950=(1,48,48,192)f32 #951=(1,48,48,192)f32 +Tensor.view Tensor.view_1285 1 1 951 952 shape=(1,2304,192) $input=951 #951=(1,48,48,192)f32 #952=(1,2304,192)f32 +pnnx.Expression pnnx_expr_12691 2 1 921 952 953 expr=add(@0,@1) #921=(1,2304,192)f32 #952=(1,2304,192)f32 #953=(1,2304,192)f32 +nn.LayerNorm layers_dfe.4.residual_group.blocks.3.norm2 1 1 953 954 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #953=(1,2304,192)f32 #954=(1,2304,192)f32 +nn.Linear layers_dfe.4.residual_group.blocks.3.mlp.fc1 1 1 954 955 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #954=(1,2304,192)f32 #955=(1,2304,384)f32 +nn.GELU layers_dfe.4.residual_group.blocks.3.mlp.act 1 1 955 956 #955=(1,2304,384)f32 #956=(1,2304,384)f32 +nn.Linear layers_dfe.4.residual_group.blocks.3.mlp.fc2 1 1 956 957 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #956=(1,2304,384)f32 #957=(1,2304,192)f32 +pnnx.Expression pnnx_expr_12689 2 1 953 957 958 expr=add(@0,@1) #953=(1,2304,192)f32 #957=(1,2304,192)f32 #958=(1,2304,192)f32 +nn.LayerNorm layers_dfe.4.residual_group.blocks.4.norm1 1 1 958 959 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #958=(1,2304,192)f32 #959=(1,2304,192)f32 +Tensor.reshape Tensor.view_1287 1 1 959 960 shape=(1,6,8,6,8,192) $input=959 #959=(1,2304,192)f32 #960=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2640 1 1 960 961 dims=(0,1,3,2,4,5) $input=960 #960=(1,6,8,6,8,192)f32 #961=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1289 1 1 961 962 shape=(36,64,192) $input=961 #961=(1,6,6,8,8,192)f32 #962=(36,64,192)f32 +nn.Linear layers_dfe.4.residual_group.blocks.4.attn.qkv 1 1 962 963 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #962=(36,64,192)f32 #963=(36,64,576)f32 +Tensor.reshape Tensor.reshape_490 1 1 963 964 shape=(36,64,3,6,32) $input=963 #963=(36,64,576)f32 #964=(36,64,3,6,32)f32 +torch.permute torch.permute_2641 1 1 964 965 dims=(2,0,3,1,4) $input=964 #964=(36,64,3,6,32)f32 #965=(3,36,6,64,32)f32 +torch.unbind Tensor.select_734 1 3 965 966 967 968 dim=0 #965=(3,36,6,64,32)f32 #966=(36,6,64,32)f32 #967=(36,6,64,32)f32 #968=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_12601 1 1 966 969 expr=mul(@0,1.767767e-01) #966=(36,6,64,32)f32 #969=(36,6,64,32)f32 +torch.transpose torch.transpose_3025 1 1 967 970 dim0=-2 dim1=-1 $input=967 #967=(36,6,64,32)f32 #970=(36,6,32,64)f32 +torch.matmul torch.matmul_2260 2 1 969 970 971 $input=969 $other=970 #969=(36,6,64,32)f32 #970=(36,6,32,64)f32 #971=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_6881 0 1 972 @pnnx_fold_6881=(1,6,64,64)f32 #972=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_12588 2 1 971 972 973 expr=add(@0,@1) #971=(36,6,64,64)f32 #972=(1,6,64,64)f32 #973=(36,6,64,64)f32 +nn.Softmax layers_dfe.4.residual_group.blocks.4.attn.softmax 1 1 973 974 dim=-1 #973=(36,6,64,64)f32 #974=(36,6,64,64)f32 +torch.matmul torch.matmul_2261 2 1 974 968 975 $input=974 $other=968 #974=(36,6,64,64)f32 #968=(36,6,64,32)f32 #975=(36,6,64,32)f32 +torch.transpose torch.transpose_3026 1 1 975 976 dim0=1 dim1=2 $input=975 #975=(36,6,64,32)f32 #976=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_491 1 1 976 977 shape=(36,64,192) $input=976 #976=(36,64,6,32)f32 #977=(36,64,192)f32 +nn.Linear layers_dfe.4.residual_group.blocks.4.attn.proj 1 1 977 978 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #977=(36,64,192)f32 #978=(36,64,192)f32 +Tensor.reshape Tensor.view_1293 1 1 978 979 shape=(1,6,6,8,8,192) $input=978 #978=(36,64,192)f32 #979=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2643 1 1 979 980 dims=(0,1,3,2,4,5) $input=979 #979=(1,6,6,8,8,192)f32 #980=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1295 1 1 980 981 shape=(1,2304,192) $input=980 #980=(1,6,8,6,8,192)f32 #981=(1,2304,192)f32 +pnnx.Expression pnnx_expr_12555 2 1 958 981 982 expr=add(@0,@1) #958=(1,2304,192)f32 #981=(1,2304,192)f32 #982=(1,2304,192)f32 +nn.LayerNorm layers_dfe.4.residual_group.blocks.4.norm2 1 1 982 983 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #982=(1,2304,192)f32 #983=(1,2304,192)f32 +nn.Linear layers_dfe.4.residual_group.blocks.4.mlp.fc1 1 1 983 984 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #983=(1,2304,192)f32 #984=(1,2304,384)f32 +nn.GELU layers_dfe.4.residual_group.blocks.4.mlp.act 1 1 984 985 #984=(1,2304,384)f32 #985=(1,2304,384)f32 +nn.Linear layers_dfe.4.residual_group.blocks.4.mlp.fc2 1 1 985 986 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #985=(1,2304,384)f32 #986=(1,2304,192)f32 +pnnx.Expression pnnx_expr_12553 2 1 982 986 987 expr=add(@0,@1) #982=(1,2304,192)f32 #986=(1,2304,192)f32 #987=(1,2304,192)f32 +nn.LayerNorm layers_dfe.4.residual_group.blocks.5.norm1 1 1 987 988 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #987=(1,2304,192)f32 #988=(1,2304,192)f32 +Tensor.view Tensor.view_1296 1 1 988 989 shape=(1,48,48,192) $input=988 #988=(1,2304,192)f32 #989=(1,48,48,192)f32 +torch.roll torch.roll_2448 1 1 989 990 dims=(1,2) shifts=(-4,-4) $input=989 #989=(1,48,48,192)f32 #990=(1,48,48,192)f32 +Tensor.view Tensor.view_1297 1 1 990 991 shape=(1,6,8,6,8,192) $input=990 #990=(1,48,48,192)f32 #991=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2644 1 1 991 992 dims=(0,1,3,2,4,5) $input=991 #991=(1,6,8,6,8,192)f32 #992=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1299 1 1 992 993 shape=(36,64,192) $input=992 #992=(1,6,6,8,8,192)f32 #993=(36,64,192)f32 +nn.Linear layers_dfe.4.residual_group.blocks.5.attn.qkv 1 1 993 994 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #993=(36,64,192)f32 #994=(36,64,576)f32 +Tensor.reshape Tensor.reshape_492 1 1 994 995 shape=(36,64,3,6,32) $input=994 #994=(36,64,576)f32 #995=(36,64,3,6,32)f32 +torch.permute torch.permute_2645 1 1 995 996 dims=(2,0,3,1,4) $input=995 #995=(36,64,3,6,32)f32 #996=(3,36,6,64,32)f32 +torch.unbind Tensor.select_737 1 3 996 997 998 999 dim=0 #996=(3,36,6,64,32)f32 #997=(36,6,64,32)f32 #998=(36,6,64,32)f32 #999=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_12455 1 1 997 1000 expr=mul(@0,1.767767e-01) #997=(36,6,64,32)f32 #1000=(36,6,64,32)f32 +torch.transpose torch.transpose_3027 1 1 998 1001 dim0=-2 dim1=-1 $input=998 #998=(36,6,64,32)f32 #1001=(36,6,32,64)f32 +torch.matmul torch.matmul_2262 2 1 1000 1001 1002 $input=1000 $other=1001 #1000=(36,6,64,32)f32 #1001=(36,6,32,64)f32 #1002=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_7034 0 1 1003 @pnnx_fold_7034=(1,6,64,64)f32 #1003=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_12442 2 1 1002 1003 1004 expr=add(@0,@1) #1002=(36,6,64,64)f32 #1003=(1,6,64,64)f32 #1004=(36,6,64,64)f32 +Tensor.view Tensor.view_1302 1 1 1004 1005 shape=(1,36,6,64,64) $input=1004 #1004=(36,6,64,64)f32 #1005=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_7044 0 1 1006 @pnnx_fold_7044=(1,36,1,64,64)f32 #1006=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_12430 2 1 1005 1006 1007 expr=add(@0,@1) #1005=(1,36,6,64,64)f32 #1006=(1,36,1,64,64)f32 #1007=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1303 1 1 1007 1008 shape=(-1,6,64,64) $input=1007 #1007=(1,36,6,64,64)f32 #1008=(36,6,64,64)f32 +nn.Softmax layers_dfe.4.residual_group.blocks.5.attn.softmax 1 1 1008 1009 dim=-1 #1008=(36,6,64,64)f32 #1009=(36,6,64,64)f32 +torch.matmul torch.matmul_2263 2 1 1009 999 1010 $input=1009 $other=999 #1009=(36,6,64,64)f32 #999=(36,6,64,32)f32 #1010=(36,6,64,32)f32 +torch.transpose torch.transpose_3028 1 1 1010 1011 dim0=1 dim1=2 $input=1010 #1010=(36,6,64,32)f32 #1011=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_493 1 1 1011 1012 shape=(36,64,192) $input=1011 #1011=(36,64,6,32)f32 #1012=(36,64,192)f32 +nn.Linear layers_dfe.4.residual_group.blocks.5.attn.proj 1 1 1012 1013 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1012=(36,64,192)f32 #1013=(36,64,192)f32 +Tensor.reshape Tensor.view_1305 1 1 1013 1014 shape=(1,6,6,8,8,192) $input=1013 #1013=(36,64,192)f32 #1014=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2647 1 1 1014 1015 dims=(0,1,3,2,4,5) $input=1014 #1014=(1,6,6,8,8,192)f32 #1015=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_59 1 1 1015 1016 shape=(1,48,48,-1) $input=1015 #1015=(1,6,8,6,8,192)f32 #1016=(1,48,48,192)f32 +torch.roll torch.roll_2449 1 1 1016 1017 dims=(1,2) shifts=(4,4) $input=1016 #1016=(1,48,48,192)f32 #1017=(1,48,48,192)f32 +Tensor.view Tensor.view_1307 1 1 1017 1018 shape=(1,2304,192) $input=1017 #1017=(1,48,48,192)f32 #1018=(1,2304,192)f32 +pnnx.Expression pnnx_expr_12388 2 1 987 1018 1019 expr=add(@0,@1) #987=(1,2304,192)f32 #1018=(1,2304,192)f32 #1019=(1,2304,192)f32 +nn.LayerNorm layers_dfe.4.residual_group.blocks.5.norm2 1 1 1019 1020 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1019=(1,2304,192)f32 #1020=(1,2304,192)f32 +nn.Linear layers_dfe.4.residual_group.blocks.5.mlp.fc1 1 1 1020 1021 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1020=(1,2304,192)f32 #1021=(1,2304,384)f32 +nn.GELU layers_dfe.4.residual_group.blocks.5.mlp.act 1 1 1021 1022 #1021=(1,2304,384)f32 #1022=(1,2304,384)f32 +nn.Linear layers_dfe.4.residual_group.blocks.5.mlp.fc2 1 1 1022 1023 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1022=(1,2304,384)f32 #1023=(1,2304,192)f32 +pnnx.Expression pnnx_expr_12386 2 1 1019 1023 1024 expr=add(@0,@1) #1019=(1,2304,192)f32 #1023=(1,2304,192)f32 #1024=(1,2304,192)f32 +torch.transpose torch.transpose_3029 1 1 1024 1025 dim0=1 dim1=2 $input=1024 #1024=(1,2304,192)f32 #1025=(1,192,2304)f32 +Tensor.view Tensor.view_1308 1 1 1025 1026 shape=(1,192,48,48) $input=1025 #1025=(1,192,2304)f32 #1026=(1,192,48,48)f32 +nn.Conv2d layers_dfe.4.conv 1 1 1026 1027 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #1026=(1,192,48,48)f32 #1027=(1,192,48,48)f32 +torch.flatten torch.flatten_2188 1 1 1027 1028 end_dim=-1 start_dim=2 $input=1027 #1027=(1,192,48,48)f32 #1028=(1,192,2304)f32 +torch.transpose torch.transpose_3030 1 1 1028 1029 dim0=1 dim1=2 $input=1028 #1028=(1,192,2304)f32 #1029=(1,2304,192)f32 +pnnx.Expression pnnx_expr_12373 2 1 1029 826 1030 expr=add(@0,@1) #1029=(1,2304,192)f32 #826=(1,2304,192)f32 #1030=(1,2304,192)f32 +nn.LayerNorm layers_dfe.5.residual_group.blocks.0.norm1 1 1 1030 1031 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1030=(1,2304,192)f32 #1031=(1,2304,192)f32 +Tensor.reshape Tensor.view_1310 1 1 1031 1032 shape=(1,6,8,6,8,192) $input=1031 #1031=(1,2304,192)f32 #1032=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2648 1 1 1032 1033 dims=(0,1,3,2,4,5) $input=1032 #1032=(1,6,8,6,8,192)f32 #1033=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1312 1 1 1033 1034 shape=(36,64,192) $input=1033 #1033=(1,6,6,8,8,192)f32 #1034=(36,64,192)f32 +nn.Linear layers_dfe.5.residual_group.blocks.0.attn.qkv 1 1 1034 1035 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1034=(36,64,192)f32 #1035=(36,64,576)f32 +Tensor.reshape Tensor.reshape_494 1 1 1035 1036 shape=(36,64,3,6,32) $input=1035 #1035=(36,64,576)f32 #1036=(36,64,3,6,32)f32 +torch.permute torch.permute_2649 1 1 1036 1037 dims=(2,0,3,1,4) $input=1036 #1036=(36,64,3,6,32)f32 #1037=(3,36,6,64,32)f32 +torch.unbind Tensor.select_740 1 3 1037 1038 1039 1040 dim=0 #1037=(3,36,6,64,32)f32 #1038=(36,6,64,32)f32 #1039=(36,6,64,32)f32 #1040=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_12284 1 1 1038 1041 expr=mul(@0,1.767767e-01) #1038=(36,6,64,32)f32 #1041=(36,6,64,32)f32 +torch.transpose torch.transpose_3031 1 1 1039 1042 dim0=-2 dim1=-1 $input=1039 #1039=(36,6,64,32)f32 #1042=(36,6,32,64)f32 +torch.matmul torch.matmul_2264 2 1 1041 1042 1043 $input=1041 $other=1042 #1041=(36,6,64,32)f32 #1042=(36,6,32,64)f32 #1043=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_7227 0 1 1044 @pnnx_fold_7227=(1,6,64,64)f32 #1044=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_12271 2 1 1043 1044 1045 expr=add(@0,@1) #1043=(36,6,64,64)f32 #1044=(1,6,64,64)f32 #1045=(36,6,64,64)f32 +nn.Softmax layers_dfe.5.residual_group.blocks.0.attn.softmax 1 1 1045 1046 dim=-1 #1045=(36,6,64,64)f32 #1046=(36,6,64,64)f32 +torch.matmul torch.matmul_2265 2 1 1046 1040 1047 $input=1046 $other=1040 #1046=(36,6,64,64)f32 #1040=(36,6,64,32)f32 #1047=(36,6,64,32)f32 +torch.transpose torch.transpose_3032 1 1 1047 1048 dim0=1 dim1=2 $input=1047 #1047=(36,6,64,32)f32 #1048=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_495 1 1 1048 1049 shape=(36,64,192) $input=1048 #1048=(36,64,6,32)f32 #1049=(36,64,192)f32 +nn.Linear layers_dfe.5.residual_group.blocks.0.attn.proj 1 1 1049 1050 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1049=(36,64,192)f32 #1050=(36,64,192)f32 +Tensor.reshape Tensor.view_1316 1 1 1050 1051 shape=(1,6,6,8,8,192) $input=1050 #1050=(36,64,192)f32 #1051=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2651 1 1 1051 1052 dims=(0,1,3,2,4,5) $input=1051 #1051=(1,6,6,8,8,192)f32 #1052=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1318 1 1 1052 1053 shape=(1,2304,192) $input=1052 #1052=(1,6,8,6,8,192)f32 #1053=(1,2304,192)f32 +pnnx.Expression pnnx_expr_12238 2 1 1030 1053 1054 expr=add(@0,@1) #1030=(1,2304,192)f32 #1053=(1,2304,192)f32 #1054=(1,2304,192)f32 +nn.LayerNorm layers_dfe.5.residual_group.blocks.0.norm2 1 1 1054 1055 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1054=(1,2304,192)f32 #1055=(1,2304,192)f32 +nn.Linear layers_dfe.5.residual_group.blocks.0.mlp.fc1 1 1 1055 1056 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1055=(1,2304,192)f32 #1056=(1,2304,384)f32 +nn.GELU layers_dfe.5.residual_group.blocks.0.mlp.act 1 1 1056 1057 #1056=(1,2304,384)f32 #1057=(1,2304,384)f32 +nn.Linear layers_dfe.5.residual_group.blocks.0.mlp.fc2 1 1 1057 1058 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1057=(1,2304,384)f32 #1058=(1,2304,192)f32 +pnnx.Expression pnnx_expr_12236 2 1 1054 1058 1059 expr=add(@0,@1) #1054=(1,2304,192)f32 #1058=(1,2304,192)f32 #1059=(1,2304,192)f32 +nn.LayerNorm layers_dfe.5.residual_group.blocks.1.norm1 1 1 1059 1060 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1059=(1,2304,192)f32 #1060=(1,2304,192)f32 +Tensor.view Tensor.view_1319 1 1 1060 1061 shape=(1,48,48,192) $input=1060 #1060=(1,2304,192)f32 #1061=(1,48,48,192)f32 +torch.roll torch.roll_2450 1 1 1061 1062 dims=(1,2) shifts=(-4,-4) $input=1061 #1061=(1,48,48,192)f32 #1062=(1,48,48,192)f32 +Tensor.view Tensor.view_1320 1 1 1062 1063 shape=(1,6,8,6,8,192) $input=1062 #1062=(1,48,48,192)f32 #1063=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2652 1 1 1063 1064 dims=(0,1,3,2,4,5) $input=1063 #1063=(1,6,8,6,8,192)f32 #1064=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1322 1 1 1064 1065 shape=(36,64,192) $input=1064 #1064=(1,6,6,8,8,192)f32 #1065=(36,64,192)f32 +nn.Linear layers_dfe.5.residual_group.blocks.1.attn.qkv 1 1 1065 1066 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1065=(36,64,192)f32 #1066=(36,64,576)f32 +Tensor.reshape Tensor.reshape_496 1 1 1066 1067 shape=(36,64,3,6,32) $input=1066 #1066=(36,64,576)f32 #1067=(36,64,3,6,32)f32 +torch.permute torch.permute_2653 1 1 1067 1068 dims=(2,0,3,1,4) $input=1067 #1067=(36,64,3,6,32)f32 #1068=(3,36,6,64,32)f32 +torch.unbind Tensor.select_743 1 3 1068 1069 1070 1071 dim=0 #1068=(3,36,6,64,32)f32 #1069=(36,6,64,32)f32 #1070=(36,6,64,32)f32 #1071=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_12138 1 1 1069 1072 expr=mul(@0,1.767767e-01) #1069=(36,6,64,32)f32 #1072=(36,6,64,32)f32 +torch.transpose torch.transpose_3033 1 1 1070 1073 dim0=-2 dim1=-1 $input=1070 #1070=(36,6,64,32)f32 #1073=(36,6,32,64)f32 +torch.matmul torch.matmul_2266 2 1 1072 1073 1074 $input=1072 $other=1073 #1072=(36,6,64,32)f32 #1073=(36,6,32,64)f32 #1074=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_7380 0 1 1075 @pnnx_fold_7380=(1,6,64,64)f32 #1075=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_12125 2 1 1074 1075 1076 expr=add(@0,@1) #1074=(36,6,64,64)f32 #1075=(1,6,64,64)f32 #1076=(36,6,64,64)f32 +Tensor.view Tensor.view_1325 1 1 1076 1077 shape=(1,36,6,64,64) $input=1076 #1076=(36,6,64,64)f32 #1077=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_7390 0 1 1078 @pnnx_fold_7390=(1,36,1,64,64)f32 #1078=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_12113 2 1 1077 1078 1079 expr=add(@0,@1) #1077=(1,36,6,64,64)f32 #1078=(1,36,1,64,64)f32 #1079=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1326 1 1 1079 1080 shape=(-1,6,64,64) $input=1079 #1079=(1,36,6,64,64)f32 #1080=(36,6,64,64)f32 +nn.Softmax layers_dfe.5.residual_group.blocks.1.attn.softmax 1 1 1080 1081 dim=-1 #1080=(36,6,64,64)f32 #1081=(36,6,64,64)f32 +torch.matmul torch.matmul_2267 2 1 1081 1071 1082 $input=1081 $other=1071 #1081=(36,6,64,64)f32 #1071=(36,6,64,32)f32 #1082=(36,6,64,32)f32 +torch.transpose torch.transpose_3034 1 1 1082 1083 dim0=1 dim1=2 $input=1082 #1082=(36,6,64,32)f32 #1083=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_497 1 1 1083 1084 shape=(36,64,192) $input=1083 #1083=(36,64,6,32)f32 #1084=(36,64,192)f32 +nn.Linear layers_dfe.5.residual_group.blocks.1.attn.proj 1 1 1084 1085 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1084=(36,64,192)f32 #1085=(36,64,192)f32 +Tensor.reshape Tensor.view_1328 1 1 1085 1086 shape=(1,6,6,8,8,192) $input=1085 #1085=(36,64,192)f32 #1086=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2655 1 1 1086 1087 dims=(0,1,3,2,4,5) $input=1086 #1086=(1,6,6,8,8,192)f32 #1087=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_63 1 1 1087 1088 shape=(1,48,48,-1) $input=1087 #1087=(1,6,8,6,8,192)f32 #1088=(1,48,48,192)f32 +torch.roll torch.roll_2451 1 1 1088 1089 dims=(1,2) shifts=(4,4) $input=1088 #1088=(1,48,48,192)f32 #1089=(1,48,48,192)f32 +Tensor.view Tensor.view_1330 1 1 1089 1090 shape=(1,2304,192) $input=1089 #1089=(1,48,48,192)f32 #1090=(1,2304,192)f32 +pnnx.Expression pnnx_expr_12071 2 1 1059 1090 1091 expr=add(@0,@1) #1059=(1,2304,192)f32 #1090=(1,2304,192)f32 #1091=(1,2304,192)f32 +nn.LayerNorm layers_dfe.5.residual_group.blocks.1.norm2 1 1 1091 1092 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1091=(1,2304,192)f32 #1092=(1,2304,192)f32 +nn.Linear layers_dfe.5.residual_group.blocks.1.mlp.fc1 1 1 1092 1093 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1092=(1,2304,192)f32 #1093=(1,2304,384)f32 +nn.GELU layers_dfe.5.residual_group.blocks.1.mlp.act 1 1 1093 1094 #1093=(1,2304,384)f32 #1094=(1,2304,384)f32 +nn.Linear layers_dfe.5.residual_group.blocks.1.mlp.fc2 1 1 1094 1095 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1094=(1,2304,384)f32 #1095=(1,2304,192)f32 +pnnx.Expression pnnx_expr_12069 2 1 1091 1095 1096 expr=add(@0,@1) #1091=(1,2304,192)f32 #1095=(1,2304,192)f32 #1096=(1,2304,192)f32 +nn.LayerNorm layers_dfe.5.residual_group.blocks.2.norm1 1 1 1096 1097 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1096=(1,2304,192)f32 #1097=(1,2304,192)f32 +Tensor.reshape Tensor.view_1332 1 1 1097 1098 shape=(1,6,8,6,8,192) $input=1097 #1097=(1,2304,192)f32 #1098=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2656 1 1 1098 1099 dims=(0,1,3,2,4,5) $input=1098 #1098=(1,6,8,6,8,192)f32 #1099=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1334 1 1 1099 1100 shape=(36,64,192) $input=1099 #1099=(1,6,6,8,8,192)f32 #1100=(36,64,192)f32 +nn.Linear layers_dfe.5.residual_group.blocks.2.attn.qkv 1 1 1100 1101 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1100=(36,64,192)f32 #1101=(36,64,576)f32 +Tensor.reshape Tensor.reshape_498 1 1 1101 1102 shape=(36,64,3,6,32) $input=1101 #1101=(36,64,576)f32 #1102=(36,64,3,6,32)f32 +torch.permute torch.permute_2657 1 1 1102 1103 dims=(2,0,3,1,4) $input=1102 #1102=(36,64,3,6,32)f32 #1103=(3,36,6,64,32)f32 +torch.unbind Tensor.select_746 1 3 1103 1104 1105 1106 dim=0 #1103=(3,36,6,64,32)f32 #1104=(36,6,64,32)f32 #1105=(36,6,64,32)f32 #1106=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_11981 1 1 1104 1107 expr=mul(@0,1.767767e-01) #1104=(36,6,64,32)f32 #1107=(36,6,64,32)f32 +torch.transpose torch.transpose_3035 1 1 1105 1108 dim0=-2 dim1=-1 $input=1105 #1105=(36,6,64,32)f32 #1108=(36,6,32,64)f32 +torch.matmul torch.matmul_2268 2 1 1107 1108 1109 $input=1107 $other=1108 #1107=(36,6,64,32)f32 #1108=(36,6,32,64)f32 #1109=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_7539 0 1 1110 @pnnx_fold_7539=(1,6,64,64)f32 #1110=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_11968 2 1 1109 1110 1111 expr=add(@0,@1) #1109=(36,6,64,64)f32 #1110=(1,6,64,64)f32 #1111=(36,6,64,64)f32 +nn.Softmax layers_dfe.5.residual_group.blocks.2.attn.softmax 1 1 1111 1112 dim=-1 #1111=(36,6,64,64)f32 #1112=(36,6,64,64)f32 +torch.matmul torch.matmul_2269 2 1 1112 1106 1113 $input=1112 $other=1106 #1112=(36,6,64,64)f32 #1106=(36,6,64,32)f32 #1113=(36,6,64,32)f32 +torch.transpose torch.transpose_3036 1 1 1113 1114 dim0=1 dim1=2 $input=1113 #1113=(36,6,64,32)f32 #1114=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_499 1 1 1114 1115 shape=(36,64,192) $input=1114 #1114=(36,64,6,32)f32 #1115=(36,64,192)f32 +nn.Linear layers_dfe.5.residual_group.blocks.2.attn.proj 1 1 1115 1116 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1115=(36,64,192)f32 #1116=(36,64,192)f32 +Tensor.reshape Tensor.view_1338 1 1 1116 1117 shape=(1,6,6,8,8,192) $input=1116 #1116=(36,64,192)f32 #1117=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2659 1 1 1117 1118 dims=(0,1,3,2,4,5) $input=1117 #1117=(1,6,6,8,8,192)f32 #1118=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1340 1 1 1118 1119 shape=(1,2304,192) $input=1118 #1118=(1,6,8,6,8,192)f32 #1119=(1,2304,192)f32 +pnnx.Expression pnnx_expr_11935 2 1 1096 1119 1120 expr=add(@0,@1) #1096=(1,2304,192)f32 #1119=(1,2304,192)f32 #1120=(1,2304,192)f32 +nn.LayerNorm layers_dfe.5.residual_group.blocks.2.norm2 1 1 1120 1121 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1120=(1,2304,192)f32 #1121=(1,2304,192)f32 +nn.Linear layers_dfe.5.residual_group.blocks.2.mlp.fc1 1 1 1121 1122 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1121=(1,2304,192)f32 #1122=(1,2304,384)f32 +nn.GELU layers_dfe.5.residual_group.blocks.2.mlp.act 1 1 1122 1123 #1122=(1,2304,384)f32 #1123=(1,2304,384)f32 +nn.Linear layers_dfe.5.residual_group.blocks.2.mlp.fc2 1 1 1123 1124 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1123=(1,2304,384)f32 #1124=(1,2304,192)f32 +pnnx.Expression pnnx_expr_11933 2 1 1120 1124 1125 expr=add(@0,@1) #1120=(1,2304,192)f32 #1124=(1,2304,192)f32 #1125=(1,2304,192)f32 +nn.LayerNorm layers_dfe.5.residual_group.blocks.3.norm1 1 1 1125 1126 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1125=(1,2304,192)f32 #1126=(1,2304,192)f32 +Tensor.view Tensor.view_1341 1 1 1126 1127 shape=(1,48,48,192) $input=1126 #1126=(1,2304,192)f32 #1127=(1,48,48,192)f32 +torch.roll torch.roll_2452 1 1 1127 1128 dims=(1,2) shifts=(-4,-4) $input=1127 #1127=(1,48,48,192)f32 #1128=(1,48,48,192)f32 +Tensor.view Tensor.view_1342 1 1 1128 1129 shape=(1,6,8,6,8,192) $input=1128 #1128=(1,48,48,192)f32 #1129=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2660 1 1 1129 1130 dims=(0,1,3,2,4,5) $input=1129 #1129=(1,6,8,6,8,192)f32 #1130=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1344 1 1 1130 1131 shape=(36,64,192) $input=1130 #1130=(1,6,6,8,8,192)f32 #1131=(36,64,192)f32 +nn.Linear layers_dfe.5.residual_group.blocks.3.attn.qkv 1 1 1131 1132 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1131=(36,64,192)f32 #1132=(36,64,576)f32 +Tensor.reshape Tensor.reshape_500 1 1 1132 1133 shape=(36,64,3,6,32) $input=1132 #1132=(36,64,576)f32 #1133=(36,64,3,6,32)f32 +torch.permute torch.permute_2661 1 1 1133 1134 dims=(2,0,3,1,4) $input=1133 #1133=(36,64,3,6,32)f32 #1134=(3,36,6,64,32)f32 +torch.unbind Tensor.select_749 1 3 1134 1135 1136 1137 dim=0 #1134=(3,36,6,64,32)f32 #1135=(36,6,64,32)f32 #1136=(36,6,64,32)f32 #1137=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_11835 1 1 1135 1138 expr=mul(@0,1.767767e-01) #1135=(36,6,64,32)f32 #1138=(36,6,64,32)f32 +torch.transpose torch.transpose_3037 1 1 1136 1139 dim0=-2 dim1=-1 $input=1136 #1136=(36,6,64,32)f32 #1139=(36,6,32,64)f32 +torch.matmul torch.matmul_2270 2 1 1138 1139 1140 $input=1138 $other=1139 #1138=(36,6,64,32)f32 #1139=(36,6,32,64)f32 #1140=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_7692 0 1 1141 @pnnx_fold_7692=(1,6,64,64)f32 #1141=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_11822 2 1 1140 1141 1142 expr=add(@0,@1) #1140=(36,6,64,64)f32 #1141=(1,6,64,64)f32 #1142=(36,6,64,64)f32 +Tensor.view Tensor.view_1347 1 1 1142 1143 shape=(1,36,6,64,64) $input=1142 #1142=(36,6,64,64)f32 #1143=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_7702 0 1 1144 @pnnx_fold_7702=(1,36,1,64,64)f32 #1144=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_11810 2 1 1143 1144 1145 expr=add(@0,@1) #1143=(1,36,6,64,64)f32 #1144=(1,36,1,64,64)f32 #1145=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1348 1 1 1145 1146 shape=(-1,6,64,64) $input=1145 #1145=(1,36,6,64,64)f32 #1146=(36,6,64,64)f32 +nn.Softmax layers_dfe.5.residual_group.blocks.3.attn.softmax 1 1 1146 1147 dim=-1 #1146=(36,6,64,64)f32 #1147=(36,6,64,64)f32 +torch.matmul torch.matmul_2271 2 1 1147 1137 1148 $input=1147 $other=1137 #1147=(36,6,64,64)f32 #1137=(36,6,64,32)f32 #1148=(36,6,64,32)f32 +torch.transpose torch.transpose_3038 1 1 1148 1149 dim0=1 dim1=2 $input=1148 #1148=(36,6,64,32)f32 #1149=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_501 1 1 1149 1150 shape=(36,64,192) $input=1149 #1149=(36,64,6,32)f32 #1150=(36,64,192)f32 +nn.Linear layers_dfe.5.residual_group.blocks.3.attn.proj 1 1 1150 1151 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1150=(36,64,192)f32 #1151=(36,64,192)f32 +Tensor.reshape Tensor.view_1350 1 1 1151 1152 shape=(1,6,6,8,8,192) $input=1151 #1151=(36,64,192)f32 #1152=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2663 1 1 1152 1153 dims=(0,1,3,2,4,5) $input=1152 #1152=(1,6,6,8,8,192)f32 #1153=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_67 1 1 1153 1154 shape=(1,48,48,-1) $input=1153 #1153=(1,6,8,6,8,192)f32 #1154=(1,48,48,192)f32 +torch.roll torch.roll_2453 1 1 1154 1155 dims=(1,2) shifts=(4,4) $input=1154 #1154=(1,48,48,192)f32 #1155=(1,48,48,192)f32 +Tensor.view Tensor.view_1352 1 1 1155 1156 shape=(1,2304,192) $input=1155 #1155=(1,48,48,192)f32 #1156=(1,2304,192)f32 +pnnx.Expression pnnx_expr_11768 2 1 1125 1156 1157 expr=add(@0,@1) #1125=(1,2304,192)f32 #1156=(1,2304,192)f32 #1157=(1,2304,192)f32 +nn.LayerNorm layers_dfe.5.residual_group.blocks.3.norm2 1 1 1157 1158 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1157=(1,2304,192)f32 #1158=(1,2304,192)f32 +nn.Linear layers_dfe.5.residual_group.blocks.3.mlp.fc1 1 1 1158 1159 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1158=(1,2304,192)f32 #1159=(1,2304,384)f32 +nn.GELU layers_dfe.5.residual_group.blocks.3.mlp.act 1 1 1159 1160 #1159=(1,2304,384)f32 #1160=(1,2304,384)f32 +nn.Linear layers_dfe.5.residual_group.blocks.3.mlp.fc2 1 1 1160 1161 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1160=(1,2304,384)f32 #1161=(1,2304,192)f32 +pnnx.Expression pnnx_expr_11766 2 1 1157 1161 1162 expr=add(@0,@1) #1157=(1,2304,192)f32 #1161=(1,2304,192)f32 #1162=(1,2304,192)f32 +nn.LayerNorm layers_dfe.5.residual_group.blocks.4.norm1 1 1 1162 1163 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1162=(1,2304,192)f32 #1163=(1,2304,192)f32 +Tensor.reshape Tensor.view_1354 1 1 1163 1164 shape=(1,6,8,6,8,192) $input=1163 #1163=(1,2304,192)f32 #1164=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2664 1 1 1164 1165 dims=(0,1,3,2,4,5) $input=1164 #1164=(1,6,8,6,8,192)f32 #1165=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1356 1 1 1165 1166 shape=(36,64,192) $input=1165 #1165=(1,6,6,8,8,192)f32 #1166=(36,64,192)f32 +nn.Linear layers_dfe.5.residual_group.blocks.4.attn.qkv 1 1 1166 1167 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1166=(36,64,192)f32 #1167=(36,64,576)f32 +Tensor.reshape Tensor.reshape_502 1 1 1167 1168 shape=(36,64,3,6,32) $input=1167 #1167=(36,64,576)f32 #1168=(36,64,3,6,32)f32 +torch.permute torch.permute_2665 1 1 1168 1169 dims=(2,0,3,1,4) $input=1168 #1168=(36,64,3,6,32)f32 #1169=(3,36,6,64,32)f32 +torch.unbind Tensor.select_752 1 3 1169 1170 1171 1172 dim=0 #1169=(3,36,6,64,32)f32 #1170=(36,6,64,32)f32 #1171=(36,6,64,32)f32 #1172=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_11678 1 1 1170 1173 expr=mul(@0,1.767767e-01) #1170=(36,6,64,32)f32 #1173=(36,6,64,32)f32 +torch.transpose torch.transpose_3039 1 1 1171 1174 dim0=-2 dim1=-1 $input=1171 #1171=(36,6,64,32)f32 #1174=(36,6,32,64)f32 +torch.matmul torch.matmul_2272 2 1 1173 1174 1175 $input=1173 $other=1174 #1173=(36,6,64,32)f32 #1174=(36,6,32,64)f32 #1175=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_7851 0 1 1176 @pnnx_fold_7851=(1,6,64,64)f32 #1176=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_11665 2 1 1175 1176 1177 expr=add(@0,@1) #1175=(36,6,64,64)f32 #1176=(1,6,64,64)f32 #1177=(36,6,64,64)f32 +nn.Softmax layers_dfe.5.residual_group.blocks.4.attn.softmax 1 1 1177 1178 dim=-1 #1177=(36,6,64,64)f32 #1178=(36,6,64,64)f32 +torch.matmul torch.matmul_2273 2 1 1178 1172 1179 $input=1178 $other=1172 #1178=(36,6,64,64)f32 #1172=(36,6,64,32)f32 #1179=(36,6,64,32)f32 +torch.transpose torch.transpose_3040 1 1 1179 1180 dim0=1 dim1=2 $input=1179 #1179=(36,6,64,32)f32 #1180=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_503 1 1 1180 1181 shape=(36,64,192) $input=1180 #1180=(36,64,6,32)f32 #1181=(36,64,192)f32 +nn.Linear layers_dfe.5.residual_group.blocks.4.attn.proj 1 1 1181 1182 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1181=(36,64,192)f32 #1182=(36,64,192)f32 +Tensor.reshape Tensor.view_1360 1 1 1182 1183 shape=(1,6,6,8,8,192) $input=1182 #1182=(36,64,192)f32 #1183=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2667 1 1 1183 1184 dims=(0,1,3,2,4,5) $input=1183 #1183=(1,6,6,8,8,192)f32 #1184=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1362 1 1 1184 1185 shape=(1,2304,192) $input=1184 #1184=(1,6,8,6,8,192)f32 #1185=(1,2304,192)f32 +pnnx.Expression pnnx_expr_11632 2 1 1162 1185 1186 expr=add(@0,@1) #1162=(1,2304,192)f32 #1185=(1,2304,192)f32 #1186=(1,2304,192)f32 +nn.LayerNorm layers_dfe.5.residual_group.blocks.4.norm2 1 1 1186 1187 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1186=(1,2304,192)f32 #1187=(1,2304,192)f32 +nn.Linear layers_dfe.5.residual_group.blocks.4.mlp.fc1 1 1 1187 1188 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1187=(1,2304,192)f32 #1188=(1,2304,384)f32 +nn.GELU layers_dfe.5.residual_group.blocks.4.mlp.act 1 1 1188 1189 #1188=(1,2304,384)f32 #1189=(1,2304,384)f32 +nn.Linear layers_dfe.5.residual_group.blocks.4.mlp.fc2 1 1 1189 1190 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1189=(1,2304,384)f32 #1190=(1,2304,192)f32 +pnnx.Expression pnnx_expr_11630 2 1 1186 1190 1191 expr=add(@0,@1) #1186=(1,2304,192)f32 #1190=(1,2304,192)f32 #1191=(1,2304,192)f32 +nn.LayerNorm layers_dfe.5.residual_group.blocks.5.norm1 1 1 1191 1192 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1191=(1,2304,192)f32 #1192=(1,2304,192)f32 +Tensor.view Tensor.view_1363 1 1 1192 1193 shape=(1,48,48,192) $input=1192 #1192=(1,2304,192)f32 #1193=(1,48,48,192)f32 +torch.roll torch.roll_2454 1 1 1193 1194 dims=(1,2) shifts=(-4,-4) $input=1193 #1193=(1,48,48,192)f32 #1194=(1,48,48,192)f32 +Tensor.view Tensor.view_1364 1 1 1194 1195 shape=(1,6,8,6,8,192) $input=1194 #1194=(1,48,48,192)f32 #1195=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2668 1 1 1195 1196 dims=(0,1,3,2,4,5) $input=1195 #1195=(1,6,8,6,8,192)f32 #1196=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1366 1 1 1196 1197 shape=(36,64,192) $input=1196 #1196=(1,6,6,8,8,192)f32 #1197=(36,64,192)f32 +nn.Linear layers_dfe.5.residual_group.blocks.5.attn.qkv 1 1 1197 1198 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1197=(36,64,192)f32 #1198=(36,64,576)f32 +Tensor.reshape Tensor.reshape_504 1 1 1198 1199 shape=(36,64,3,6,32) $input=1198 #1198=(36,64,576)f32 #1199=(36,64,3,6,32)f32 +torch.permute torch.permute_2669 1 1 1199 1200 dims=(2,0,3,1,4) $input=1199 #1199=(36,64,3,6,32)f32 #1200=(3,36,6,64,32)f32 +torch.unbind Tensor.select_755 1 3 1200 1201 1202 1203 dim=0 #1200=(3,36,6,64,32)f32 #1201=(36,6,64,32)f32 #1202=(36,6,64,32)f32 #1203=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_11532 1 1 1201 1204 expr=mul(@0,1.767767e-01) #1201=(36,6,64,32)f32 #1204=(36,6,64,32)f32 +torch.transpose torch.transpose_3041 1 1 1202 1205 dim0=-2 dim1=-1 $input=1202 #1202=(36,6,64,32)f32 #1205=(36,6,32,64)f32 +torch.matmul torch.matmul_2274 2 1 1204 1205 1206 $input=1204 $other=1205 #1204=(36,6,64,32)f32 #1205=(36,6,32,64)f32 #1206=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_8004 0 1 1207 @pnnx_fold_8004=(1,6,64,64)f32 #1207=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_11519 2 1 1206 1207 1208 expr=add(@0,@1) #1206=(36,6,64,64)f32 #1207=(1,6,64,64)f32 #1208=(36,6,64,64)f32 +Tensor.view Tensor.view_1369 1 1 1208 1209 shape=(1,36,6,64,64) $input=1208 #1208=(36,6,64,64)f32 #1209=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_8014 0 1 1210 @pnnx_fold_8014=(1,36,1,64,64)f32 #1210=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_11507 2 1 1209 1210 1211 expr=add(@0,@1) #1209=(1,36,6,64,64)f32 #1210=(1,36,1,64,64)f32 #1211=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1370 1 1 1211 1212 shape=(-1,6,64,64) $input=1211 #1211=(1,36,6,64,64)f32 #1212=(36,6,64,64)f32 +nn.Softmax layers_dfe.5.residual_group.blocks.5.attn.softmax 1 1 1212 1213 dim=-1 #1212=(36,6,64,64)f32 #1213=(36,6,64,64)f32 +torch.matmul torch.matmul_2275 2 1 1213 1203 1214 $input=1213 $other=1203 #1213=(36,6,64,64)f32 #1203=(36,6,64,32)f32 #1214=(36,6,64,32)f32 +torch.transpose torch.transpose_3042 1 1 1214 1215 dim0=1 dim1=2 $input=1214 #1214=(36,6,64,32)f32 #1215=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_505 1 1 1215 1216 shape=(36,64,192) $input=1215 #1215=(36,64,6,32)f32 #1216=(36,64,192)f32 +nn.Linear layers_dfe.5.residual_group.blocks.5.attn.proj 1 1 1216 1217 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1216=(36,64,192)f32 #1217=(36,64,192)f32 +Tensor.reshape Tensor.view_1372 1 1 1217 1218 shape=(1,6,6,8,8,192) $input=1217 #1217=(36,64,192)f32 #1218=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2671 1 1 1218 1219 dims=(0,1,3,2,4,5) $input=1218 #1218=(1,6,6,8,8,192)f32 #1219=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_71 1 1 1219 1220 shape=(1,48,48,-1) $input=1219 #1219=(1,6,8,6,8,192)f32 #1220=(1,48,48,192)f32 +torch.roll torch.roll_2455 1 1 1220 1221 dims=(1,2) shifts=(4,4) $input=1220 #1220=(1,48,48,192)f32 #1221=(1,48,48,192)f32 +Tensor.view Tensor.view_1374 1 1 1221 1222 shape=(1,2304,192) $input=1221 #1221=(1,48,48,192)f32 #1222=(1,2304,192)f32 +pnnx.Expression pnnx_expr_11465 2 1 1191 1222 1223 expr=add(@0,@1) #1191=(1,2304,192)f32 #1222=(1,2304,192)f32 #1223=(1,2304,192)f32 +nn.LayerNorm layers_dfe.5.residual_group.blocks.5.norm2 1 1 1223 1224 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1223=(1,2304,192)f32 #1224=(1,2304,192)f32 +nn.Linear layers_dfe.5.residual_group.blocks.5.mlp.fc1 1 1 1224 1225 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1224=(1,2304,192)f32 #1225=(1,2304,384)f32 +nn.GELU layers_dfe.5.residual_group.blocks.5.mlp.act 1 1 1225 1226 #1225=(1,2304,384)f32 #1226=(1,2304,384)f32 +nn.Linear layers_dfe.5.residual_group.blocks.5.mlp.fc2 1 1 1226 1227 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1226=(1,2304,384)f32 #1227=(1,2304,192)f32 +pnnx.Expression pnnx_expr_11463 2 1 1223 1227 1228 expr=add(@0,@1) #1223=(1,2304,192)f32 #1227=(1,2304,192)f32 #1228=(1,2304,192)f32 +torch.transpose torch.transpose_3043 1 1 1228 1229 dim0=1 dim1=2 $input=1228 #1228=(1,2304,192)f32 #1229=(1,192,2304)f32 +Tensor.view Tensor.view_1375 1 1 1229 1230 shape=(1,192,48,48) $input=1229 #1229=(1,192,2304)f32 #1230=(1,192,48,48)f32 +nn.Conv2d layers_dfe.5.conv 1 1 1230 1231 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #1230=(1,192,48,48)f32 #1231=(1,192,48,48)f32 +torch.flatten torch.flatten_2189 1 1 1231 1232 end_dim=-1 start_dim=2 $input=1231 #1231=(1,192,48,48)f32 #1232=(1,192,2304)f32 +torch.transpose torch.transpose_3044 1 1 1232 1233 dim0=1 dim1=2 $input=1232 #1232=(1,192,2304)f32 #1233=(1,2304,192)f32 +pnnx.Expression pnnx_expr_11450 2 1 1233 1030 1234 expr=add(@0,@1) #1233=(1,2304,192)f32 #1030=(1,2304,192)f32 #1234=(1,2304,192)f32 +nn.LayerNorm norm_dfe 1 1 1234 1235 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1234=(1,2304,192)f32 #1235=(1,2304,192)f32 +torch.transpose torch.transpose_3045 1 1 1235 1236 dim0=1 dim1=2 $input=1235 #1235=(1,2304,192)f32 #1236=(1,192,2304)f32 +Tensor.view Tensor.view_1376 1 1 1236 1237 shape=(1,192,48,48) $input=1236 #1236=(1,192,2304)f32 #1237=(1,192,48,48)f32 +nn.Conv2d conv_after_body_dfe 1 1 1237 1238 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #1237=(1,192,48,48)f32 #1238=(1,192,48,48)f32 +pnnx.Expression pnnx_expr_11440 2 1 1238 6 1239 expr=add(@0,@1) #1238=(1,192,48,48)f32 #6=(1,192,48,48)f32 #1239=(1,192,48,48)f32 +torch.flatten torch.flatten_2190 1 1 7 1240 end_dim=-1 start_dim=2 $input=7 #7=(1,192,48,48)f32 #1240=(1,192,2304)f32 +torch.transpose torch.transpose_3046 1 1 1240 1241 dim0=1 dim1=2 $input=1240 #1240=(1,192,2304)f32 #1241=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_73 1 1 1241 1242 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1241=(1,2304,192)f32 #1242=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_75 1 1 1242 1243 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1242=(1,2304,192)f32 #1243=(1,2304,192)f32 +Tensor.reshape Tensor.view_1378 1 1 1243 1244 shape=(1,6,8,6,8,192) $input=1243 #1243=(1,2304,192)f32 #1244=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2672 1 1 1244 1245 dims=(0,1,3,2,4,5) $input=1244 #1244=(1,6,8,6,8,192)f32 #1245=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1380 1 1 1245 1246 shape=(36,64,192) $input=1245 #1245=(1,6,6,8,8,192)f32 #1246=(36,64,192)f32 +nn.Linear pnnx_unique_78 1 1 1246 1247 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1246=(36,64,192)f32 #1247=(36,64,576)f32 +Tensor.reshape Tensor.reshape_506 1 1 1247 1248 shape=(36,64,3,6,32) $input=1247 #1247=(36,64,576)f32 #1248=(36,64,3,6,32)f32 +torch.permute torch.permute_2673 1 1 1248 1249 dims=(2,0,3,1,4) $input=1248 #1248=(36,64,3,6,32)f32 #1249=(3,36,6,64,32)f32 +torch.unbind Tensor.select_758 1 3 1249 1250 1251 1252 dim=0 #1249=(3,36,6,64,32)f32 #1250=(36,6,64,32)f32 #1251=(36,6,64,32)f32 #1252=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_11183 1 1 1250 1253 expr=mul(@0,1.767767e-01) #1250=(36,6,64,32)f32 #1253=(36,6,64,32)f32 +torch.transpose torch.transpose_3047 1 1 1251 1254 dim0=-2 dim1=-1 $input=1251 #1251=(36,6,64,32)f32 #1254=(36,6,32,64)f32 +torch.matmul torch.matmul_2276 2 1 1253 1254 1255 $input=1253 $other=1254 #1253=(36,6,64,32)f32 #1254=(36,6,32,64)f32 #1255=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_8214 0 1 1256 @pnnx_fold_8214=(1,6,64,64)f32 #1256=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_11170 2 1 1255 1256 1257 expr=add(@0,@1) #1255=(36,6,64,64)f32 #1256=(1,6,64,64)f32 #1257=(36,6,64,64)f32 +nn.Softmax pnnx_unique_79 1 1 1257 1258 dim=-1 #1257=(36,6,64,64)f32 #1258=(36,6,64,64)f32 +torch.matmul torch.matmul_2277 2 1 1258 1252 1259 $input=1258 $other=1252 #1258=(36,6,64,64)f32 #1252=(36,6,64,32)f32 #1259=(36,6,64,32)f32 +torch.transpose torch.transpose_3048 1 1 1259 1260 dim0=1 dim1=2 $input=1259 #1259=(36,6,64,32)f32 #1260=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_507 1 1 1260 1261 shape=(36,64,192) $input=1260 #1260=(36,64,6,32)f32 #1261=(36,64,192)f32 +nn.Linear pnnx_unique_81 1 1 1261 1262 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1261=(36,64,192)f32 #1262=(36,64,192)f32 +Tensor.reshape Tensor.view_1384 1 1 1262 1263 shape=(1,6,6,8,8,192) $input=1262 #1262=(36,64,192)f32 #1263=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2675 1 1 1263 1264 dims=(0,1,3,2,4,5) $input=1263 #1263=(1,6,6,8,8,192)f32 #1264=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1386 1 1 1264 1265 shape=(1,2304,192) $input=1264 #1264=(1,6,8,6,8,192)f32 #1265=(1,2304,192)f32 +pnnx.Expression pnnx_expr_11137 2 1 1242 1265 1266 expr=add(@0,@1) #1242=(1,2304,192)f32 #1265=(1,2304,192)f32 #1266=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_83 1 1 1266 1267 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1266=(1,2304,192)f32 #1267=(1,2304,192)f32 +nn.Linear pnnx_unique_84 1 1 1267 1268 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1267=(1,2304,192)f32 #1268=(1,2304,384)f32 +nn.GELU pnnx_unique_85 1 1 1268 1269 #1268=(1,2304,384)f32 #1269=(1,2304,384)f32 +nn.Linear pnnx_unique_87 1 1 1269 1270 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1269=(1,2304,384)f32 #1270=(1,2304,192)f32 +pnnx.Expression pnnx_expr_11135 2 1 1266 1270 1271 expr=add(@0,@1) #1266=(1,2304,192)f32 #1270=(1,2304,192)f32 #1271=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_90 1 1 1271 1272 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1271=(1,2304,192)f32 #1272=(1,2304,192)f32 +Tensor.view Tensor.view_1387 1 1 1272 1273 shape=(1,48,48,192) $input=1272 #1272=(1,2304,192)f32 #1273=(1,48,48,192)f32 +torch.roll torch.roll_2456 1 1 1273 1274 dims=(1,2) shifts=(-4,-4) $input=1273 #1273=(1,48,48,192)f32 #1274=(1,48,48,192)f32 +Tensor.view Tensor.view_1388 1 1 1274 1275 shape=(1,6,8,6,8,192) $input=1274 #1274=(1,48,48,192)f32 #1275=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2676 1 1 1275 1276 dims=(0,1,3,2,4,5) $input=1275 #1275=(1,6,8,6,8,192)f32 #1276=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1390 1 1 1276 1277 shape=(36,64,192) $input=1276 #1276=(1,6,6,8,8,192)f32 #1277=(36,64,192)f32 +nn.Linear pnnx_unique_93 1 1 1277 1278 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1277=(36,64,192)f32 #1278=(36,64,576)f32 +Tensor.reshape Tensor.reshape_508 1 1 1278 1279 shape=(36,64,3,6,32) $input=1278 #1278=(36,64,576)f32 #1279=(36,64,3,6,32)f32 +torch.permute torch.permute_2677 1 1 1279 1280 dims=(2,0,3,1,4) $input=1279 #1279=(36,64,3,6,32)f32 #1280=(3,36,6,64,32)f32 +torch.unbind Tensor.select_761 1 3 1280 1281 1282 1283 dim=0 #1280=(3,36,6,64,32)f32 #1281=(36,6,64,32)f32 #1282=(36,6,64,32)f32 #1283=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_11037 1 1 1281 1284 expr=mul(@0,1.767767e-01) #1281=(36,6,64,32)f32 #1284=(36,6,64,32)f32 +torch.transpose torch.transpose_3049 1 1 1282 1285 dim0=-2 dim1=-1 $input=1282 #1282=(36,6,64,32)f32 #1285=(36,6,32,64)f32 +torch.matmul torch.matmul_2278 2 1 1284 1285 1286 $input=1284 $other=1285 #1284=(36,6,64,32)f32 #1285=(36,6,32,64)f32 #1286=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_8367 0 1 1287 @pnnx_fold_8367=(1,6,64,64)f32 #1287=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_11024 2 1 1286 1287 1288 expr=add(@0,@1) #1286=(36,6,64,64)f32 #1287=(1,6,64,64)f32 #1288=(36,6,64,64)f32 +Tensor.view Tensor.view_1393 1 1 1288 1289 shape=(1,36,6,64,64) $input=1288 #1288=(36,6,64,64)f32 #1289=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_8377 0 1 1290 @pnnx_fold_8377=(1,36,1,64,64)f32 #1290=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_11012 2 1 1289 1290 1291 expr=add(@0,@1) #1289=(1,36,6,64,64)f32 #1290=(1,36,1,64,64)f32 #1291=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1394 1 1 1291 1292 shape=(-1,6,64,64) $input=1291 #1291=(1,36,6,64,64)f32 #1292=(36,6,64,64)f32 +nn.Softmax pnnx_unique_94 1 1 1292 1293 dim=-1 #1292=(36,6,64,64)f32 #1293=(36,6,64,64)f32 +torch.matmul torch.matmul_2279 2 1 1293 1283 1294 $input=1293 $other=1283 #1293=(36,6,64,64)f32 #1283=(36,6,64,32)f32 #1294=(36,6,64,32)f32 +torch.transpose torch.transpose_3050 1 1 1294 1295 dim0=1 dim1=2 $input=1294 #1294=(36,6,64,32)f32 #1295=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_509 1 1 1295 1296 shape=(36,64,192) $input=1295 #1295=(36,64,6,32)f32 #1296=(36,64,192)f32 +nn.Linear pnnx_unique_96 1 1 1296 1297 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1296=(36,64,192)f32 #1297=(36,64,192)f32 +Tensor.reshape Tensor.view_1396 1 1 1297 1298 shape=(1,6,6,8,8,192) $input=1297 #1297=(36,64,192)f32 #1298=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2679 1 1 1298 1299 dims=(0,1,3,2,4,5) $input=1298 #1298=(1,6,6,8,8,192)f32 #1299=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_75 1 1 1299 1300 shape=(1,48,48,-1) $input=1299 #1299=(1,6,8,6,8,192)f32 #1300=(1,48,48,192)f32 +torch.roll torch.roll_2457 1 1 1300 1301 dims=(1,2) shifts=(4,4) $input=1300 #1300=(1,48,48,192)f32 #1301=(1,48,48,192)f32 +Tensor.view Tensor.view_1398 1 1 1301 1302 shape=(1,2304,192) $input=1301 #1301=(1,48,48,192)f32 #1302=(1,2304,192)f32 +pnnx.Expression pnnx_expr_10970 2 1 1271 1302 1303 expr=add(@0,@1) #1271=(1,2304,192)f32 #1302=(1,2304,192)f32 #1303=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_98 1 1 1303 1304 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1303=(1,2304,192)f32 #1304=(1,2304,192)f32 +nn.Linear pnnx_unique_99 1 1 1304 1305 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1304=(1,2304,192)f32 #1305=(1,2304,384)f32 +nn.GELU pnnx_unique_100 1 1 1305 1306 #1305=(1,2304,384)f32 #1306=(1,2304,384)f32 +nn.Linear pnnx_unique_102 1 1 1306 1307 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1306=(1,2304,384)f32 #1307=(1,2304,192)f32 +pnnx.Expression pnnx_expr_10968 2 1 1303 1307 1308 expr=add(@0,@1) #1303=(1,2304,192)f32 #1307=(1,2304,192)f32 #1308=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_104 1 1 1308 1309 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1308=(1,2304,192)f32 #1309=(1,2304,192)f32 +Tensor.reshape Tensor.view_1400 1 1 1309 1310 shape=(1,6,8,6,8,192) $input=1309 #1309=(1,2304,192)f32 #1310=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2680 1 1 1310 1311 dims=(0,1,3,2,4,5) $input=1310 #1310=(1,6,8,6,8,192)f32 #1311=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1402 1 1 1311 1312 shape=(36,64,192) $input=1311 #1311=(1,6,6,8,8,192)f32 #1312=(36,64,192)f32 +nn.Linear pnnx_unique_107 1 1 1312 1313 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1312=(36,64,192)f32 #1313=(36,64,576)f32 +Tensor.reshape Tensor.reshape_510 1 1 1313 1314 shape=(36,64,3,6,32) $input=1313 #1313=(36,64,576)f32 #1314=(36,64,3,6,32)f32 +torch.permute torch.permute_2681 1 1 1314 1315 dims=(2,0,3,1,4) $input=1314 #1314=(36,64,3,6,32)f32 #1315=(3,36,6,64,32)f32 +torch.unbind Tensor.select_764 1 3 1315 1316 1317 1318 dim=0 #1315=(3,36,6,64,32)f32 #1316=(36,6,64,32)f32 #1317=(36,6,64,32)f32 #1318=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_10880 1 1 1316 1319 expr=mul(@0,1.767767e-01) #1316=(36,6,64,32)f32 #1319=(36,6,64,32)f32 +torch.transpose torch.transpose_3051 1 1 1317 1320 dim0=-2 dim1=-1 $input=1317 #1317=(36,6,64,32)f32 #1320=(36,6,32,64)f32 +torch.matmul torch.matmul_2280 2 1 1319 1320 1321 $input=1319 $other=1320 #1319=(36,6,64,32)f32 #1320=(36,6,32,64)f32 #1321=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_8526 0 1 1322 @pnnx_fold_8526=(1,6,64,64)f32 #1322=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_10867 2 1 1321 1322 1323 expr=add(@0,@1) #1321=(36,6,64,64)f32 #1322=(1,6,64,64)f32 #1323=(36,6,64,64)f32 +nn.Softmax pnnx_unique_108 1 1 1323 1324 dim=-1 #1323=(36,6,64,64)f32 #1324=(36,6,64,64)f32 +torch.matmul torch.matmul_2281 2 1 1324 1318 1325 $input=1324 $other=1318 #1324=(36,6,64,64)f32 #1318=(36,6,64,32)f32 #1325=(36,6,64,32)f32 +torch.transpose torch.transpose_3052 1 1 1325 1326 dim0=1 dim1=2 $input=1325 #1325=(36,6,64,32)f32 #1326=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_511 1 1 1326 1327 shape=(36,64,192) $input=1326 #1326=(36,64,6,32)f32 #1327=(36,64,192)f32 +nn.Linear pnnx_unique_110 1 1 1327 1328 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1327=(36,64,192)f32 #1328=(36,64,192)f32 +Tensor.reshape Tensor.view_1406 1 1 1328 1329 shape=(1,6,6,8,8,192) $input=1328 #1328=(36,64,192)f32 #1329=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2683 1 1 1329 1330 dims=(0,1,3,2,4,5) $input=1329 #1329=(1,6,6,8,8,192)f32 #1330=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1408 1 1 1330 1331 shape=(1,2304,192) $input=1330 #1330=(1,6,8,6,8,192)f32 #1331=(1,2304,192)f32 +pnnx.Expression pnnx_expr_10834 2 1 1308 1331 1332 expr=add(@0,@1) #1308=(1,2304,192)f32 #1331=(1,2304,192)f32 #1332=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_112 1 1 1332 1333 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1332=(1,2304,192)f32 #1333=(1,2304,192)f32 +nn.Linear pnnx_unique_113 1 1 1333 1334 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1333=(1,2304,192)f32 #1334=(1,2304,384)f32 +nn.GELU pnnx_unique_114 1 1 1334 1335 #1334=(1,2304,384)f32 #1335=(1,2304,384)f32 +nn.Linear pnnx_unique_116 1 1 1335 1336 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1335=(1,2304,384)f32 #1336=(1,2304,192)f32 +pnnx.Expression pnnx_expr_10832 2 1 1332 1336 1337 expr=add(@0,@1) #1332=(1,2304,192)f32 #1336=(1,2304,192)f32 #1337=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_119 1 1 1337 1338 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1337=(1,2304,192)f32 #1338=(1,2304,192)f32 +Tensor.view Tensor.view_1409 1 1 1338 1339 shape=(1,48,48,192) $input=1338 #1338=(1,2304,192)f32 #1339=(1,48,48,192)f32 +torch.roll torch.roll_2458 1 1 1339 1340 dims=(1,2) shifts=(-4,-4) $input=1339 #1339=(1,48,48,192)f32 #1340=(1,48,48,192)f32 +Tensor.view Tensor.view_1410 1 1 1340 1341 shape=(1,6,8,6,8,192) $input=1340 #1340=(1,48,48,192)f32 #1341=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2684 1 1 1341 1342 dims=(0,1,3,2,4,5) $input=1341 #1341=(1,6,8,6,8,192)f32 #1342=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1412 1 1 1342 1343 shape=(36,64,192) $input=1342 #1342=(1,6,6,8,8,192)f32 #1343=(36,64,192)f32 +nn.Linear pnnx_unique_122 1 1 1343 1344 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1343=(36,64,192)f32 #1344=(36,64,576)f32 +Tensor.reshape Tensor.reshape_512 1 1 1344 1345 shape=(36,64,3,6,32) $input=1344 #1344=(36,64,576)f32 #1345=(36,64,3,6,32)f32 +torch.permute torch.permute_2685 1 1 1345 1346 dims=(2,0,3,1,4) $input=1345 #1345=(36,64,3,6,32)f32 #1346=(3,36,6,64,32)f32 +torch.unbind Tensor.select_767 1 3 1346 1347 1348 1349 dim=0 #1346=(3,36,6,64,32)f32 #1347=(36,6,64,32)f32 #1348=(36,6,64,32)f32 #1349=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_10734 1 1 1347 1350 expr=mul(@0,1.767767e-01) #1347=(36,6,64,32)f32 #1350=(36,6,64,32)f32 +torch.transpose torch.transpose_3053 1 1 1348 1351 dim0=-2 dim1=-1 $input=1348 #1348=(36,6,64,32)f32 #1351=(36,6,32,64)f32 +torch.matmul torch.matmul_2282 2 1 1350 1351 1352 $input=1350 $other=1351 #1350=(36,6,64,32)f32 #1351=(36,6,32,64)f32 #1352=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_8679 0 1 1353 @pnnx_fold_8679=(1,6,64,64)f32 #1353=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_10721 2 1 1352 1353 1354 expr=add(@0,@1) #1352=(36,6,64,64)f32 #1353=(1,6,64,64)f32 #1354=(36,6,64,64)f32 +Tensor.view Tensor.view_1415 1 1 1354 1355 shape=(1,36,6,64,64) $input=1354 #1354=(36,6,64,64)f32 #1355=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_8689 0 1 1356 @pnnx_fold_8689=(1,36,1,64,64)f32 #1356=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_10709 2 1 1355 1356 1357 expr=add(@0,@1) #1355=(1,36,6,64,64)f32 #1356=(1,36,1,64,64)f32 #1357=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1416 1 1 1357 1358 shape=(-1,6,64,64) $input=1357 #1357=(1,36,6,64,64)f32 #1358=(36,6,64,64)f32 +nn.Softmax pnnx_unique_123 1 1 1358 1359 dim=-1 #1358=(36,6,64,64)f32 #1359=(36,6,64,64)f32 +torch.matmul torch.matmul_2283 2 1 1359 1349 1360 $input=1359 $other=1349 #1359=(36,6,64,64)f32 #1349=(36,6,64,32)f32 #1360=(36,6,64,32)f32 +torch.transpose torch.transpose_3054 1 1 1360 1361 dim0=1 dim1=2 $input=1360 #1360=(36,6,64,32)f32 #1361=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_513 1 1 1361 1362 shape=(36,64,192) $input=1361 #1361=(36,64,6,32)f32 #1362=(36,64,192)f32 +nn.Linear pnnx_unique_125 1 1 1362 1363 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1362=(36,64,192)f32 #1363=(36,64,192)f32 +Tensor.reshape Tensor.view_1418 1 1 1363 1364 shape=(1,6,6,8,8,192) $input=1363 #1363=(36,64,192)f32 #1364=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2687 1 1 1364 1365 dims=(0,1,3,2,4,5) $input=1364 #1364=(1,6,6,8,8,192)f32 #1365=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_79 1 1 1365 1366 shape=(1,48,48,-1) $input=1365 #1365=(1,6,8,6,8,192)f32 #1366=(1,48,48,192)f32 +torch.roll torch.roll_2459 1 1 1366 1367 dims=(1,2) shifts=(4,4) $input=1366 #1366=(1,48,48,192)f32 #1367=(1,48,48,192)f32 +Tensor.view Tensor.view_1420 1 1 1367 1368 shape=(1,2304,192) $input=1367 #1367=(1,48,48,192)f32 #1368=(1,2304,192)f32 +pnnx.Expression pnnx_expr_10667 2 1 1337 1368 1369 expr=add(@0,@1) #1337=(1,2304,192)f32 #1368=(1,2304,192)f32 #1369=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_127 1 1 1369 1370 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1369=(1,2304,192)f32 #1370=(1,2304,192)f32 +nn.Linear pnnx_unique_128 1 1 1370 1371 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1370=(1,2304,192)f32 #1371=(1,2304,384)f32 +nn.GELU pnnx_unique_129 1 1 1371 1372 #1371=(1,2304,384)f32 #1372=(1,2304,384)f32 +nn.Linear pnnx_unique_131 1 1 1372 1373 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1372=(1,2304,384)f32 #1373=(1,2304,192)f32 +pnnx.Expression pnnx_expr_10665 2 1 1369 1373 1374 expr=add(@0,@1) #1369=(1,2304,192)f32 #1373=(1,2304,192)f32 #1374=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_133 1 1 1374 1375 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1374=(1,2304,192)f32 #1375=(1,2304,192)f32 +Tensor.reshape Tensor.view_1422 1 1 1375 1376 shape=(1,6,8,6,8,192) $input=1375 #1375=(1,2304,192)f32 #1376=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2688 1 1 1376 1377 dims=(0,1,3,2,4,5) $input=1376 #1376=(1,6,8,6,8,192)f32 #1377=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1424 1 1 1377 1378 shape=(36,64,192) $input=1377 #1377=(1,6,6,8,8,192)f32 #1378=(36,64,192)f32 +nn.Linear pnnx_unique_136 1 1 1378 1379 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1378=(36,64,192)f32 #1379=(36,64,576)f32 +Tensor.reshape Tensor.reshape_514 1 1 1379 1380 shape=(36,64,3,6,32) $input=1379 #1379=(36,64,576)f32 #1380=(36,64,3,6,32)f32 +torch.permute torch.permute_2689 1 1 1380 1381 dims=(2,0,3,1,4) $input=1380 #1380=(36,64,3,6,32)f32 #1381=(3,36,6,64,32)f32 +torch.unbind Tensor.select_770 1 3 1381 1382 1383 1384 dim=0 #1381=(3,36,6,64,32)f32 #1382=(36,6,64,32)f32 #1383=(36,6,64,32)f32 #1384=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_10577 1 1 1382 1385 expr=mul(@0,1.767767e-01) #1382=(36,6,64,32)f32 #1385=(36,6,64,32)f32 +torch.transpose torch.transpose_3055 1 1 1383 1386 dim0=-2 dim1=-1 $input=1383 #1383=(36,6,64,32)f32 #1386=(36,6,32,64)f32 +torch.matmul torch.matmul_2284 2 1 1385 1386 1387 $input=1385 $other=1386 #1385=(36,6,64,32)f32 #1386=(36,6,32,64)f32 #1387=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_8838 0 1 1388 @pnnx_fold_8838=(1,6,64,64)f32 #1388=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_10564 2 1 1387 1388 1389 expr=add(@0,@1) #1387=(36,6,64,64)f32 #1388=(1,6,64,64)f32 #1389=(36,6,64,64)f32 +nn.Softmax pnnx_unique_137 1 1 1389 1390 dim=-1 #1389=(36,6,64,64)f32 #1390=(36,6,64,64)f32 +torch.matmul torch.matmul_2285 2 1 1390 1384 1391 $input=1390 $other=1384 #1390=(36,6,64,64)f32 #1384=(36,6,64,32)f32 #1391=(36,6,64,32)f32 +torch.transpose torch.transpose_3056 1 1 1391 1392 dim0=1 dim1=2 $input=1391 #1391=(36,6,64,32)f32 #1392=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_515 1 1 1392 1393 shape=(36,64,192) $input=1392 #1392=(36,64,6,32)f32 #1393=(36,64,192)f32 +nn.Linear pnnx_unique_139 1 1 1393 1394 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1393=(36,64,192)f32 #1394=(36,64,192)f32 +Tensor.reshape Tensor.view_1428 1 1 1394 1395 shape=(1,6,6,8,8,192) $input=1394 #1394=(36,64,192)f32 #1395=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2691 1 1 1395 1396 dims=(0,1,3,2,4,5) $input=1395 #1395=(1,6,6,8,8,192)f32 #1396=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1430 1 1 1396 1397 shape=(1,2304,192) $input=1396 #1396=(1,6,8,6,8,192)f32 #1397=(1,2304,192)f32 +pnnx.Expression pnnx_expr_10531 2 1 1374 1397 1398 expr=add(@0,@1) #1374=(1,2304,192)f32 #1397=(1,2304,192)f32 #1398=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_141 1 1 1398 1399 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1398=(1,2304,192)f32 #1399=(1,2304,192)f32 +nn.Linear pnnx_unique_142 1 1 1399 1400 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1399=(1,2304,192)f32 #1400=(1,2304,384)f32 +nn.GELU pnnx_unique_143 1 1 1400 1401 #1400=(1,2304,384)f32 #1401=(1,2304,384)f32 +nn.Linear pnnx_unique_145 1 1 1401 1402 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1401=(1,2304,384)f32 #1402=(1,2304,192)f32 +pnnx.Expression pnnx_expr_10529 2 1 1398 1402 1403 expr=add(@0,@1) #1398=(1,2304,192)f32 #1402=(1,2304,192)f32 #1403=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_148 1 1 1403 1404 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1403=(1,2304,192)f32 #1404=(1,2304,192)f32 +Tensor.view Tensor.view_1431 1 1 1404 1405 shape=(1,48,48,192) $input=1404 #1404=(1,2304,192)f32 #1405=(1,48,48,192)f32 +torch.roll torch.roll_2460 1 1 1405 1406 dims=(1,2) shifts=(-4,-4) $input=1405 #1405=(1,48,48,192)f32 #1406=(1,48,48,192)f32 +Tensor.view Tensor.view_1432 1 1 1406 1407 shape=(1,6,8,6,8,192) $input=1406 #1406=(1,48,48,192)f32 #1407=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2692 1 1 1407 1408 dims=(0,1,3,2,4,5) $input=1407 #1407=(1,6,8,6,8,192)f32 #1408=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1434 1 1 1408 1409 shape=(36,64,192) $input=1408 #1408=(1,6,6,8,8,192)f32 #1409=(36,64,192)f32 +nn.Linear pnnx_unique_151 1 1 1409 1410 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1409=(36,64,192)f32 #1410=(36,64,576)f32 +Tensor.reshape Tensor.reshape_516 1 1 1410 1411 shape=(36,64,3,6,32) $input=1410 #1410=(36,64,576)f32 #1411=(36,64,3,6,32)f32 +torch.permute torch.permute_2693 1 1 1411 1412 dims=(2,0,3,1,4) $input=1411 #1411=(36,64,3,6,32)f32 #1412=(3,36,6,64,32)f32 +torch.unbind Tensor.select_773 1 3 1412 1413 1414 1415 dim=0 #1412=(3,36,6,64,32)f32 #1413=(36,6,64,32)f32 #1414=(36,6,64,32)f32 #1415=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_10431 1 1 1413 1416 expr=mul(@0,1.767767e-01) #1413=(36,6,64,32)f32 #1416=(36,6,64,32)f32 +torch.transpose torch.transpose_3057 1 1 1414 1417 dim0=-2 dim1=-1 $input=1414 #1414=(36,6,64,32)f32 #1417=(36,6,32,64)f32 +torch.matmul torch.matmul_2286 2 1 1416 1417 1418 $input=1416 $other=1417 #1416=(36,6,64,32)f32 #1417=(36,6,32,64)f32 #1418=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_8991 0 1 1419 @pnnx_fold_8991=(1,6,64,64)f32 #1419=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_10418 2 1 1418 1419 1420 expr=add(@0,@1) #1418=(36,6,64,64)f32 #1419=(1,6,64,64)f32 #1420=(36,6,64,64)f32 +Tensor.view Tensor.view_1437 1 1 1420 1421 shape=(1,36,6,64,64) $input=1420 #1420=(36,6,64,64)f32 #1421=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_9001 0 1 1422 @pnnx_fold_9001=(1,36,1,64,64)f32 #1422=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_10406 2 1 1421 1422 1423 expr=add(@0,@1) #1421=(1,36,6,64,64)f32 #1422=(1,36,1,64,64)f32 #1423=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1438 1 1 1423 1424 shape=(-1,6,64,64) $input=1423 #1423=(1,36,6,64,64)f32 #1424=(36,6,64,64)f32 +nn.Softmax pnnx_unique_152 1 1 1424 1425 dim=-1 #1424=(36,6,64,64)f32 #1425=(36,6,64,64)f32 +torch.matmul torch.matmul_2287 2 1 1425 1415 1426 $input=1425 $other=1415 #1425=(36,6,64,64)f32 #1415=(36,6,64,32)f32 #1426=(36,6,64,32)f32 +torch.transpose torch.transpose_3058 1 1 1426 1427 dim0=1 dim1=2 $input=1426 #1426=(36,6,64,32)f32 #1427=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_517 1 1 1427 1428 shape=(36,64,192) $input=1427 #1427=(36,64,6,32)f32 #1428=(36,64,192)f32 +nn.Linear pnnx_unique_154 1 1 1428 1429 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1428=(36,64,192)f32 #1429=(36,64,192)f32 +Tensor.reshape Tensor.view_1440 1 1 1429 1430 shape=(1,6,6,8,8,192) $input=1429 #1429=(36,64,192)f32 #1430=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2695 1 1 1430 1431 dims=(0,1,3,2,4,5) $input=1430 #1430=(1,6,6,8,8,192)f32 #1431=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_83 1 1 1431 1432 shape=(1,48,48,-1) $input=1431 #1431=(1,6,8,6,8,192)f32 #1432=(1,48,48,192)f32 +torch.roll torch.roll_2461 1 1 1432 1433 dims=(1,2) shifts=(4,4) $input=1432 #1432=(1,48,48,192)f32 #1433=(1,48,48,192)f32 +Tensor.view Tensor.view_1442 1 1 1433 1434 shape=(1,2304,192) $input=1433 #1433=(1,48,48,192)f32 #1434=(1,2304,192)f32 +pnnx.Expression pnnx_expr_10364 2 1 1403 1434 1435 expr=add(@0,@1) #1403=(1,2304,192)f32 #1434=(1,2304,192)f32 #1435=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_156 1 1 1435 1436 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1435=(1,2304,192)f32 #1436=(1,2304,192)f32 +nn.Linear pnnx_unique_157 1 1 1436 1437 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1436=(1,2304,192)f32 #1437=(1,2304,384)f32 +nn.GELU pnnx_unique_158 1 1 1437 1438 #1437=(1,2304,384)f32 #1438=(1,2304,384)f32 +nn.Linear pnnx_unique_160 1 1 1438 1439 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1438=(1,2304,384)f32 #1439=(1,2304,192)f32 +pnnx.Expression pnnx_expr_10362 2 1 1435 1439 1440 expr=add(@0,@1) #1435=(1,2304,192)f32 #1439=(1,2304,192)f32 #1440=(1,2304,192)f32 +torch.transpose torch.transpose_3059 1 1 1440 1441 dim0=1 dim1=2 $input=1440 #1440=(1,2304,192)f32 #1441=(1,192,2304)f32 +Tensor.view Tensor.view_1443 1 1 1441 1442 shape=(1,192,48,48) $input=1441 #1441=(1,192,2304)f32 #1442=(1,192,48,48)f32 +nn.Conv2d pnnx_unique_162 1 1 1442 1443 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #1442=(1,192,48,48)f32 #1443=(1,192,48,48)f32 +torch.flatten torch.flatten_2191 1 1 1443 1444 end_dim=-1 start_dim=2 $input=1443 #1443=(1,192,48,48)f32 #1444=(1,192,2304)f32 +torch.transpose torch.transpose_3060 1 1 1444 1445 dim0=1 dim1=2 $input=1444 #1444=(1,192,2304)f32 #1445=(1,2304,192)f32 +pnnx.Expression pnnx_expr_10349 2 1 1445 1242 1446 expr=add(@0,@1) #1445=(1,2304,192)f32 #1242=(1,2304,192)f32 #1446=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_163 1 1 1446 1447 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1446=(1,2304,192)f32 #1447=(1,2304,192)f32 +Tensor.reshape Tensor.view_1445 1 1 1447 1448 shape=(1,6,8,6,8,192) $input=1447 #1447=(1,2304,192)f32 #1448=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2696 1 1 1448 1449 dims=(0,1,3,2,4,5) $input=1448 #1448=(1,6,8,6,8,192)f32 #1449=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1447 1 1 1449 1450 shape=(36,64,192) $input=1449 #1449=(1,6,6,8,8,192)f32 #1450=(36,64,192)f32 +nn.Linear pnnx_unique_166 1 1 1450 1451 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1450=(36,64,192)f32 #1451=(36,64,576)f32 +Tensor.reshape Tensor.reshape_518 1 1 1451 1452 shape=(36,64,3,6,32) $input=1451 #1451=(36,64,576)f32 #1452=(36,64,3,6,32)f32 +torch.permute torch.permute_2697 1 1 1452 1453 dims=(2,0,3,1,4) $input=1452 #1452=(36,64,3,6,32)f32 #1453=(3,36,6,64,32)f32 +torch.unbind Tensor.select_776 1 3 1453 1454 1455 1456 dim=0 #1453=(3,36,6,64,32)f32 #1454=(36,6,64,32)f32 #1455=(36,6,64,32)f32 #1456=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_10260 1 1 1454 1457 expr=mul(@0,1.767767e-01) #1454=(36,6,64,32)f32 #1457=(36,6,64,32)f32 +torch.transpose torch.transpose_3061 1 1 1455 1458 dim0=-2 dim1=-1 $input=1455 #1455=(36,6,64,32)f32 #1458=(36,6,32,64)f32 +torch.matmul torch.matmul_2288 2 1 1457 1458 1459 $input=1457 $other=1458 #1457=(36,6,64,32)f32 #1458=(36,6,32,64)f32 #1459=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_9184 0 1 1460 @pnnx_fold_9184=(1,6,64,64)f32 #1460=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_10247 2 1 1459 1460 1461 expr=add(@0,@1) #1459=(36,6,64,64)f32 #1460=(1,6,64,64)f32 #1461=(36,6,64,64)f32 +nn.Softmax pnnx_unique_167 1 1 1461 1462 dim=-1 #1461=(36,6,64,64)f32 #1462=(36,6,64,64)f32 +torch.matmul torch.matmul_2289 2 1 1462 1456 1463 $input=1462 $other=1456 #1462=(36,6,64,64)f32 #1456=(36,6,64,32)f32 #1463=(36,6,64,32)f32 +torch.transpose torch.transpose_3062 1 1 1463 1464 dim0=1 dim1=2 $input=1463 #1463=(36,6,64,32)f32 #1464=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_519 1 1 1464 1465 shape=(36,64,192) $input=1464 #1464=(36,64,6,32)f32 #1465=(36,64,192)f32 +nn.Linear pnnx_unique_169 1 1 1465 1466 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1465=(36,64,192)f32 #1466=(36,64,192)f32 +Tensor.reshape Tensor.view_1451 1 1 1466 1467 shape=(1,6,6,8,8,192) $input=1466 #1466=(36,64,192)f32 #1467=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2699 1 1 1467 1468 dims=(0,1,3,2,4,5) $input=1467 #1467=(1,6,6,8,8,192)f32 #1468=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1453 1 1 1468 1469 shape=(1,2304,192) $input=1468 #1468=(1,6,8,6,8,192)f32 #1469=(1,2304,192)f32 +pnnx.Expression pnnx_expr_10214 2 1 1446 1469 1470 expr=add(@0,@1) #1446=(1,2304,192)f32 #1469=(1,2304,192)f32 #1470=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_171 1 1 1470 1471 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1470=(1,2304,192)f32 #1471=(1,2304,192)f32 +nn.Linear pnnx_unique_172 1 1 1471 1472 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1471=(1,2304,192)f32 #1472=(1,2304,384)f32 +nn.GELU pnnx_unique_173 1 1 1472 1473 #1472=(1,2304,384)f32 #1473=(1,2304,384)f32 +nn.Linear pnnx_unique_175 1 1 1473 1474 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1473=(1,2304,384)f32 #1474=(1,2304,192)f32 +pnnx.Expression pnnx_expr_10212 2 1 1470 1474 1475 expr=add(@0,@1) #1470=(1,2304,192)f32 #1474=(1,2304,192)f32 #1475=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_178 1 1 1475 1476 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1475=(1,2304,192)f32 #1476=(1,2304,192)f32 +Tensor.view Tensor.view_1454 1 1 1476 1477 shape=(1,48,48,192) $input=1476 #1476=(1,2304,192)f32 #1477=(1,48,48,192)f32 +torch.roll torch.roll_2462 1 1 1477 1478 dims=(1,2) shifts=(-4,-4) $input=1477 #1477=(1,48,48,192)f32 #1478=(1,48,48,192)f32 +Tensor.view Tensor.view_1455 1 1 1478 1479 shape=(1,6,8,6,8,192) $input=1478 #1478=(1,48,48,192)f32 #1479=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2700 1 1 1479 1480 dims=(0,1,3,2,4,5) $input=1479 #1479=(1,6,8,6,8,192)f32 #1480=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1457 1 1 1480 1481 shape=(36,64,192) $input=1480 #1480=(1,6,6,8,8,192)f32 #1481=(36,64,192)f32 +nn.Linear pnnx_unique_181 1 1 1481 1482 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1481=(36,64,192)f32 #1482=(36,64,576)f32 +Tensor.reshape Tensor.reshape_520 1 1 1482 1483 shape=(36,64,3,6,32) $input=1482 #1482=(36,64,576)f32 #1483=(36,64,3,6,32)f32 +torch.permute torch.permute_2701 1 1 1483 1484 dims=(2,0,3,1,4) $input=1483 #1483=(36,64,3,6,32)f32 #1484=(3,36,6,64,32)f32 +torch.unbind Tensor.select_779 1 3 1484 1485 1486 1487 dim=0 #1484=(3,36,6,64,32)f32 #1485=(36,6,64,32)f32 #1486=(36,6,64,32)f32 #1487=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_10114 1 1 1485 1488 expr=mul(@0,1.767767e-01) #1485=(36,6,64,32)f32 #1488=(36,6,64,32)f32 +torch.transpose torch.transpose_3063 1 1 1486 1489 dim0=-2 dim1=-1 $input=1486 #1486=(36,6,64,32)f32 #1489=(36,6,32,64)f32 +torch.matmul torch.matmul_2290 2 1 1488 1489 1490 $input=1488 $other=1489 #1488=(36,6,64,32)f32 #1489=(36,6,32,64)f32 #1490=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_9337 0 1 1491 @pnnx_fold_9337=(1,6,64,64)f32 #1491=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_10101 2 1 1490 1491 1492 expr=add(@0,@1) #1490=(36,6,64,64)f32 #1491=(1,6,64,64)f32 #1492=(36,6,64,64)f32 +Tensor.view Tensor.view_1460 1 1 1492 1493 shape=(1,36,6,64,64) $input=1492 #1492=(36,6,64,64)f32 #1493=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_9347 0 1 1494 @pnnx_fold_9347=(1,36,1,64,64)f32 #1494=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_10089 2 1 1493 1494 1495 expr=add(@0,@1) #1493=(1,36,6,64,64)f32 #1494=(1,36,1,64,64)f32 #1495=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1461 1 1 1495 1496 shape=(-1,6,64,64) $input=1495 #1495=(1,36,6,64,64)f32 #1496=(36,6,64,64)f32 +nn.Softmax pnnx_unique_182 1 1 1496 1497 dim=-1 #1496=(36,6,64,64)f32 #1497=(36,6,64,64)f32 +torch.matmul torch.matmul_2291 2 1 1497 1487 1498 $input=1497 $other=1487 #1497=(36,6,64,64)f32 #1487=(36,6,64,32)f32 #1498=(36,6,64,32)f32 +torch.transpose torch.transpose_3064 1 1 1498 1499 dim0=1 dim1=2 $input=1498 #1498=(36,6,64,32)f32 #1499=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_521 1 1 1499 1500 shape=(36,64,192) $input=1499 #1499=(36,64,6,32)f32 #1500=(36,64,192)f32 +nn.Linear pnnx_unique_184 1 1 1500 1501 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1500=(36,64,192)f32 #1501=(36,64,192)f32 +Tensor.reshape Tensor.view_1463 1 1 1501 1502 shape=(1,6,6,8,8,192) $input=1501 #1501=(36,64,192)f32 #1502=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2703 1 1 1502 1503 dims=(0,1,3,2,4,5) $input=1502 #1502=(1,6,6,8,8,192)f32 #1503=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_87 1 1 1503 1504 shape=(1,48,48,-1) $input=1503 #1503=(1,6,8,6,8,192)f32 #1504=(1,48,48,192)f32 +torch.roll torch.roll_2463 1 1 1504 1505 dims=(1,2) shifts=(4,4) $input=1504 #1504=(1,48,48,192)f32 #1505=(1,48,48,192)f32 +Tensor.view Tensor.view_1465 1 1 1505 1506 shape=(1,2304,192) $input=1505 #1505=(1,48,48,192)f32 #1506=(1,2304,192)f32 +pnnx.Expression pnnx_expr_10047 2 1 1475 1506 1507 expr=add(@0,@1) #1475=(1,2304,192)f32 #1506=(1,2304,192)f32 #1507=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_186 1 1 1507 1508 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1507=(1,2304,192)f32 #1508=(1,2304,192)f32 +nn.Linear pnnx_unique_187 1 1 1508 1509 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1508=(1,2304,192)f32 #1509=(1,2304,384)f32 +nn.GELU pnnx_unique_188 1 1 1509 1510 #1509=(1,2304,384)f32 #1510=(1,2304,384)f32 +nn.Linear pnnx_unique_190 1 1 1510 1511 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1510=(1,2304,384)f32 #1511=(1,2304,192)f32 +pnnx.Expression pnnx_expr_10045 2 1 1507 1511 1512 expr=add(@0,@1) #1507=(1,2304,192)f32 #1511=(1,2304,192)f32 #1512=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_192 1 1 1512 1513 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1512=(1,2304,192)f32 #1513=(1,2304,192)f32 +Tensor.reshape Tensor.view_1467 1 1 1513 1514 shape=(1,6,8,6,8,192) $input=1513 #1513=(1,2304,192)f32 #1514=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2704 1 1 1514 1515 dims=(0,1,3,2,4,5) $input=1514 #1514=(1,6,8,6,8,192)f32 #1515=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1469 1 1 1515 1516 shape=(36,64,192) $input=1515 #1515=(1,6,6,8,8,192)f32 #1516=(36,64,192)f32 +nn.Linear pnnx_unique_195 1 1 1516 1517 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1516=(36,64,192)f32 #1517=(36,64,576)f32 +Tensor.reshape Tensor.reshape_522 1 1 1517 1518 shape=(36,64,3,6,32) $input=1517 #1517=(36,64,576)f32 #1518=(36,64,3,6,32)f32 +torch.permute torch.permute_2705 1 1 1518 1519 dims=(2,0,3,1,4) $input=1518 #1518=(36,64,3,6,32)f32 #1519=(3,36,6,64,32)f32 +torch.unbind Tensor.select_782 1 3 1519 1520 1521 1522 dim=0 #1519=(3,36,6,64,32)f32 #1520=(36,6,64,32)f32 #1521=(36,6,64,32)f32 #1522=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_9957 1 1 1520 1523 expr=mul(@0,1.767767e-01) #1520=(36,6,64,32)f32 #1523=(36,6,64,32)f32 +torch.transpose torch.transpose_3065 1 1 1521 1524 dim0=-2 dim1=-1 $input=1521 #1521=(36,6,64,32)f32 #1524=(36,6,32,64)f32 +torch.matmul torch.matmul_2292 2 1 1523 1524 1525 $input=1523 $other=1524 #1523=(36,6,64,32)f32 #1524=(36,6,32,64)f32 #1525=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_9496 0 1 1526 @pnnx_fold_9496=(1,6,64,64)f32 #1526=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_9944 2 1 1525 1526 1527 expr=add(@0,@1) #1525=(36,6,64,64)f32 #1526=(1,6,64,64)f32 #1527=(36,6,64,64)f32 +nn.Softmax pnnx_unique_196 1 1 1527 1528 dim=-1 #1527=(36,6,64,64)f32 #1528=(36,6,64,64)f32 +torch.matmul torch.matmul_2293 2 1 1528 1522 1529 $input=1528 $other=1522 #1528=(36,6,64,64)f32 #1522=(36,6,64,32)f32 #1529=(36,6,64,32)f32 +torch.transpose torch.transpose_3066 1 1 1529 1530 dim0=1 dim1=2 $input=1529 #1529=(36,6,64,32)f32 #1530=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_523 1 1 1530 1531 shape=(36,64,192) $input=1530 #1530=(36,64,6,32)f32 #1531=(36,64,192)f32 +nn.Linear pnnx_unique_198 1 1 1531 1532 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1531=(36,64,192)f32 #1532=(36,64,192)f32 +Tensor.reshape Tensor.view_1473 1 1 1532 1533 shape=(1,6,6,8,8,192) $input=1532 #1532=(36,64,192)f32 #1533=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2707 1 1 1533 1534 dims=(0,1,3,2,4,5) $input=1533 #1533=(1,6,6,8,8,192)f32 #1534=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1475 1 1 1534 1535 shape=(1,2304,192) $input=1534 #1534=(1,6,8,6,8,192)f32 #1535=(1,2304,192)f32 +pnnx.Expression pnnx_expr_9911 2 1 1512 1535 1536 expr=add(@0,@1) #1512=(1,2304,192)f32 #1535=(1,2304,192)f32 #1536=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_200 1 1 1536 1537 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1536=(1,2304,192)f32 #1537=(1,2304,192)f32 +nn.Linear pnnx_unique_201 1 1 1537 1538 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1537=(1,2304,192)f32 #1538=(1,2304,384)f32 +nn.GELU pnnx_unique_202 1 1 1538 1539 #1538=(1,2304,384)f32 #1539=(1,2304,384)f32 +nn.Linear pnnx_unique_204 1 1 1539 1540 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1539=(1,2304,384)f32 #1540=(1,2304,192)f32 +pnnx.Expression pnnx_expr_9909 2 1 1536 1540 1541 expr=add(@0,@1) #1536=(1,2304,192)f32 #1540=(1,2304,192)f32 #1541=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_207 1 1 1541 1542 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1541=(1,2304,192)f32 #1542=(1,2304,192)f32 +Tensor.view Tensor.view_1476 1 1 1542 1543 shape=(1,48,48,192) $input=1542 #1542=(1,2304,192)f32 #1543=(1,48,48,192)f32 +torch.roll torch.roll_2464 1 1 1543 1544 dims=(1,2) shifts=(-4,-4) $input=1543 #1543=(1,48,48,192)f32 #1544=(1,48,48,192)f32 +Tensor.view Tensor.view_1477 1 1 1544 1545 shape=(1,6,8,6,8,192) $input=1544 #1544=(1,48,48,192)f32 #1545=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2708 1 1 1545 1546 dims=(0,1,3,2,4,5) $input=1545 #1545=(1,6,8,6,8,192)f32 #1546=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1479 1 1 1546 1547 shape=(36,64,192) $input=1546 #1546=(1,6,6,8,8,192)f32 #1547=(36,64,192)f32 +nn.Linear pnnx_unique_210 1 1 1547 1548 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1547=(36,64,192)f32 #1548=(36,64,576)f32 +Tensor.reshape Tensor.reshape_524 1 1 1548 1549 shape=(36,64,3,6,32) $input=1548 #1548=(36,64,576)f32 #1549=(36,64,3,6,32)f32 +torch.permute torch.permute_2709 1 1 1549 1550 dims=(2,0,3,1,4) $input=1549 #1549=(36,64,3,6,32)f32 #1550=(3,36,6,64,32)f32 +torch.unbind Tensor.select_785 1 3 1550 1551 1552 1553 dim=0 #1550=(3,36,6,64,32)f32 #1551=(36,6,64,32)f32 #1552=(36,6,64,32)f32 #1553=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_9811 1 1 1551 1554 expr=mul(@0,1.767767e-01) #1551=(36,6,64,32)f32 #1554=(36,6,64,32)f32 +torch.transpose torch.transpose_3067 1 1 1552 1555 dim0=-2 dim1=-1 $input=1552 #1552=(36,6,64,32)f32 #1555=(36,6,32,64)f32 +torch.matmul torch.matmul_2294 2 1 1554 1555 1556 $input=1554 $other=1555 #1554=(36,6,64,32)f32 #1555=(36,6,32,64)f32 #1556=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_9649 0 1 1557 @pnnx_fold_9649=(1,6,64,64)f32 #1557=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_9798 2 1 1556 1557 1558 expr=add(@0,@1) #1556=(36,6,64,64)f32 #1557=(1,6,64,64)f32 #1558=(36,6,64,64)f32 +Tensor.view Tensor.view_1482 1 1 1558 1559 shape=(1,36,6,64,64) $input=1558 #1558=(36,6,64,64)f32 #1559=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_9659 0 1 1560 @pnnx_fold_9659=(1,36,1,64,64)f32 #1560=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_9786 2 1 1559 1560 1561 expr=add(@0,@1) #1559=(1,36,6,64,64)f32 #1560=(1,36,1,64,64)f32 #1561=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1483 1 1 1561 1562 shape=(-1,6,64,64) $input=1561 #1561=(1,36,6,64,64)f32 #1562=(36,6,64,64)f32 +nn.Softmax pnnx_unique_211 1 1 1562 1563 dim=-1 #1562=(36,6,64,64)f32 #1563=(36,6,64,64)f32 +torch.matmul torch.matmul_2295 2 1 1563 1553 1564 $input=1563 $other=1553 #1563=(36,6,64,64)f32 #1553=(36,6,64,32)f32 #1564=(36,6,64,32)f32 +torch.transpose torch.transpose_3068 1 1 1564 1565 dim0=1 dim1=2 $input=1564 #1564=(36,6,64,32)f32 #1565=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_525 1 1 1565 1566 shape=(36,64,192) $input=1565 #1565=(36,64,6,32)f32 #1566=(36,64,192)f32 +nn.Linear pnnx_unique_213 1 1 1566 1567 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1566=(36,64,192)f32 #1567=(36,64,192)f32 +Tensor.reshape Tensor.view_1485 1 1 1567 1568 shape=(1,6,6,8,8,192) $input=1567 #1567=(36,64,192)f32 #1568=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2711 1 1 1568 1569 dims=(0,1,3,2,4,5) $input=1568 #1568=(1,6,6,8,8,192)f32 #1569=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_91 1 1 1569 1570 shape=(1,48,48,-1) $input=1569 #1569=(1,6,8,6,8,192)f32 #1570=(1,48,48,192)f32 +torch.roll torch.roll_2465 1 1 1570 1571 dims=(1,2) shifts=(4,4) $input=1570 #1570=(1,48,48,192)f32 #1571=(1,48,48,192)f32 +Tensor.view Tensor.view_1487 1 1 1571 1572 shape=(1,2304,192) $input=1571 #1571=(1,48,48,192)f32 #1572=(1,2304,192)f32 +pnnx.Expression pnnx_expr_9744 2 1 1541 1572 1573 expr=add(@0,@1) #1541=(1,2304,192)f32 #1572=(1,2304,192)f32 #1573=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_215 1 1 1573 1574 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1573=(1,2304,192)f32 #1574=(1,2304,192)f32 +nn.Linear pnnx_unique_216 1 1 1574 1575 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1574=(1,2304,192)f32 #1575=(1,2304,384)f32 +nn.GELU pnnx_unique_217 1 1 1575 1576 #1575=(1,2304,384)f32 #1576=(1,2304,384)f32 +nn.Linear pnnx_unique_219 1 1 1576 1577 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1576=(1,2304,384)f32 #1577=(1,2304,192)f32 +pnnx.Expression pnnx_expr_9742 2 1 1573 1577 1578 expr=add(@0,@1) #1573=(1,2304,192)f32 #1577=(1,2304,192)f32 #1578=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_221 1 1 1578 1579 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1578=(1,2304,192)f32 #1579=(1,2304,192)f32 +Tensor.reshape Tensor.view_1489 1 1 1579 1580 shape=(1,6,8,6,8,192) $input=1579 #1579=(1,2304,192)f32 #1580=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2712 1 1 1580 1581 dims=(0,1,3,2,4,5) $input=1580 #1580=(1,6,8,6,8,192)f32 #1581=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1491 1 1 1581 1582 shape=(36,64,192) $input=1581 #1581=(1,6,6,8,8,192)f32 #1582=(36,64,192)f32 +nn.Linear pnnx_unique_224 1 1 1582 1583 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1582=(36,64,192)f32 #1583=(36,64,576)f32 +Tensor.reshape Tensor.reshape_526 1 1 1583 1584 shape=(36,64,3,6,32) $input=1583 #1583=(36,64,576)f32 #1584=(36,64,3,6,32)f32 +torch.permute torch.permute_2713 1 1 1584 1585 dims=(2,0,3,1,4) $input=1584 #1584=(36,64,3,6,32)f32 #1585=(3,36,6,64,32)f32 +torch.unbind Tensor.select_788 1 3 1585 1586 1587 1588 dim=0 #1585=(3,36,6,64,32)f32 #1586=(36,6,64,32)f32 #1587=(36,6,64,32)f32 #1588=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_9654 1 1 1586 1589 expr=mul(@0,1.767767e-01) #1586=(36,6,64,32)f32 #1589=(36,6,64,32)f32 +torch.transpose torch.transpose_3069 1 1 1587 1590 dim0=-2 dim1=-1 $input=1587 #1587=(36,6,64,32)f32 #1590=(36,6,32,64)f32 +torch.matmul torch.matmul_2296 2 1 1589 1590 1591 $input=1589 $other=1590 #1589=(36,6,64,32)f32 #1590=(36,6,32,64)f32 #1591=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_9808 0 1 1592 @pnnx_fold_9808=(1,6,64,64)f32 #1592=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_9641 2 1 1591 1592 1593 expr=add(@0,@1) #1591=(36,6,64,64)f32 #1592=(1,6,64,64)f32 #1593=(36,6,64,64)f32 +nn.Softmax pnnx_unique_225 1 1 1593 1594 dim=-1 #1593=(36,6,64,64)f32 #1594=(36,6,64,64)f32 +torch.matmul torch.matmul_2297 2 1 1594 1588 1595 $input=1594 $other=1588 #1594=(36,6,64,64)f32 #1588=(36,6,64,32)f32 #1595=(36,6,64,32)f32 +torch.transpose torch.transpose_3070 1 1 1595 1596 dim0=1 dim1=2 $input=1595 #1595=(36,6,64,32)f32 #1596=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_527 1 1 1596 1597 shape=(36,64,192) $input=1596 #1596=(36,64,6,32)f32 #1597=(36,64,192)f32 +nn.Linear pnnx_unique_227 1 1 1597 1598 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1597=(36,64,192)f32 #1598=(36,64,192)f32 +Tensor.reshape Tensor.view_1495 1 1 1598 1599 shape=(1,6,6,8,8,192) $input=1598 #1598=(36,64,192)f32 #1599=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2715 1 1 1599 1600 dims=(0,1,3,2,4,5) $input=1599 #1599=(1,6,6,8,8,192)f32 #1600=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1497 1 1 1600 1601 shape=(1,2304,192) $input=1600 #1600=(1,6,8,6,8,192)f32 #1601=(1,2304,192)f32 +pnnx.Expression pnnx_expr_9608 2 1 1578 1601 1602 expr=add(@0,@1) #1578=(1,2304,192)f32 #1601=(1,2304,192)f32 #1602=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_229 1 1 1602 1603 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1602=(1,2304,192)f32 #1603=(1,2304,192)f32 +nn.Linear pnnx_unique_230 1 1 1603 1604 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1603=(1,2304,192)f32 #1604=(1,2304,384)f32 +nn.GELU pnnx_unique_231 1 1 1604 1605 #1604=(1,2304,384)f32 #1605=(1,2304,384)f32 +nn.Linear pnnx_unique_233 1 1 1605 1606 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1605=(1,2304,384)f32 #1606=(1,2304,192)f32 +pnnx.Expression pnnx_expr_9606 2 1 1602 1606 1607 expr=add(@0,@1) #1602=(1,2304,192)f32 #1606=(1,2304,192)f32 #1607=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_236 1 1 1607 1608 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1607=(1,2304,192)f32 #1608=(1,2304,192)f32 +Tensor.view Tensor.view_1498 1 1 1608 1609 shape=(1,48,48,192) $input=1608 #1608=(1,2304,192)f32 #1609=(1,48,48,192)f32 +torch.roll torch.roll_2466 1 1 1609 1610 dims=(1,2) shifts=(-4,-4) $input=1609 #1609=(1,48,48,192)f32 #1610=(1,48,48,192)f32 +Tensor.view Tensor.view_1499 1 1 1610 1611 shape=(1,6,8,6,8,192) $input=1610 #1610=(1,48,48,192)f32 #1611=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2716 1 1 1611 1612 dims=(0,1,3,2,4,5) $input=1611 #1611=(1,6,8,6,8,192)f32 #1612=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1501 1 1 1612 1613 shape=(36,64,192) $input=1612 #1612=(1,6,6,8,8,192)f32 #1613=(36,64,192)f32 +nn.Linear pnnx_unique_239 1 1 1613 1614 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1613=(36,64,192)f32 #1614=(36,64,576)f32 +Tensor.reshape Tensor.reshape_528 1 1 1614 1615 shape=(36,64,3,6,32) $input=1614 #1614=(36,64,576)f32 #1615=(36,64,3,6,32)f32 +torch.permute torch.permute_2717 1 1 1615 1616 dims=(2,0,3,1,4) $input=1615 #1615=(36,64,3,6,32)f32 #1616=(3,36,6,64,32)f32 +torch.unbind Tensor.select_791 1 3 1616 1617 1618 1619 dim=0 #1616=(3,36,6,64,32)f32 #1617=(36,6,64,32)f32 #1618=(36,6,64,32)f32 #1619=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_9508 1 1 1617 1620 expr=mul(@0,1.767767e-01) #1617=(36,6,64,32)f32 #1620=(36,6,64,32)f32 +torch.transpose torch.transpose_3071 1 1 1618 1621 dim0=-2 dim1=-1 $input=1618 #1618=(36,6,64,32)f32 #1621=(36,6,32,64)f32 +torch.matmul torch.matmul_2298 2 1 1620 1621 1622 $input=1620 $other=1621 #1620=(36,6,64,32)f32 #1621=(36,6,32,64)f32 #1622=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_9961 0 1 1623 @pnnx_fold_9961=(1,6,64,64)f32 #1623=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_9495 2 1 1622 1623 1624 expr=add(@0,@1) #1622=(36,6,64,64)f32 #1623=(1,6,64,64)f32 #1624=(36,6,64,64)f32 +Tensor.view Tensor.view_1504 1 1 1624 1625 shape=(1,36,6,64,64) $input=1624 #1624=(36,6,64,64)f32 #1625=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_9971 0 1 1626 @pnnx_fold_9971=(1,36,1,64,64)f32 #1626=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_9483 2 1 1625 1626 1627 expr=add(@0,@1) #1625=(1,36,6,64,64)f32 #1626=(1,36,1,64,64)f32 #1627=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1505 1 1 1627 1628 shape=(-1,6,64,64) $input=1627 #1627=(1,36,6,64,64)f32 #1628=(36,6,64,64)f32 +nn.Softmax pnnx_unique_240 1 1 1628 1629 dim=-1 #1628=(36,6,64,64)f32 #1629=(36,6,64,64)f32 +torch.matmul torch.matmul_2299 2 1 1629 1619 1630 $input=1629 $other=1619 #1629=(36,6,64,64)f32 #1619=(36,6,64,32)f32 #1630=(36,6,64,32)f32 +torch.transpose torch.transpose_3072 1 1 1630 1631 dim0=1 dim1=2 $input=1630 #1630=(36,6,64,32)f32 #1631=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_529 1 1 1631 1632 shape=(36,64,192) $input=1631 #1631=(36,64,6,32)f32 #1632=(36,64,192)f32 +nn.Linear pnnx_unique_242 1 1 1632 1633 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1632=(36,64,192)f32 #1633=(36,64,192)f32 +Tensor.reshape Tensor.view_1507 1 1 1633 1634 shape=(1,6,6,8,8,192) $input=1633 #1633=(36,64,192)f32 #1634=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2719 1 1 1634 1635 dims=(0,1,3,2,4,5) $input=1634 #1634=(1,6,6,8,8,192)f32 #1635=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_95 1 1 1635 1636 shape=(1,48,48,-1) $input=1635 #1635=(1,6,8,6,8,192)f32 #1636=(1,48,48,192)f32 +torch.roll torch.roll_2467 1 1 1636 1637 dims=(1,2) shifts=(4,4) $input=1636 #1636=(1,48,48,192)f32 #1637=(1,48,48,192)f32 +Tensor.view Tensor.view_1509 1 1 1637 1638 shape=(1,2304,192) $input=1637 #1637=(1,48,48,192)f32 #1638=(1,2304,192)f32 +pnnx.Expression pnnx_expr_9441 2 1 1607 1638 1639 expr=add(@0,@1) #1607=(1,2304,192)f32 #1638=(1,2304,192)f32 #1639=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_244 1 1 1639 1640 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1639=(1,2304,192)f32 #1640=(1,2304,192)f32 +nn.Linear pnnx_unique_245 1 1 1640 1641 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1640=(1,2304,192)f32 #1641=(1,2304,384)f32 +nn.GELU pnnx_unique_246 1 1 1641 1642 #1641=(1,2304,384)f32 #1642=(1,2304,384)f32 +nn.Linear pnnx_unique_248 1 1 1642 1643 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1642=(1,2304,384)f32 #1643=(1,2304,192)f32 +pnnx.Expression pnnx_expr_9439 2 1 1639 1643 1644 expr=add(@0,@1) #1639=(1,2304,192)f32 #1643=(1,2304,192)f32 #1644=(1,2304,192)f32 +torch.transpose torch.transpose_3073 1 1 1644 1645 dim0=1 dim1=2 $input=1644 #1644=(1,2304,192)f32 #1645=(1,192,2304)f32 +Tensor.view Tensor.view_1510 1 1 1645 1646 shape=(1,192,48,48) $input=1645 #1645=(1,192,2304)f32 #1646=(1,192,48,48)f32 +nn.Conv2d pnnx_unique_250 1 1 1646 1647 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #1646=(1,192,48,48)f32 #1647=(1,192,48,48)f32 +torch.flatten torch.flatten_2192 1 1 1647 1648 end_dim=-1 start_dim=2 $input=1647 #1647=(1,192,48,48)f32 #1648=(1,192,2304)f32 +torch.transpose torch.transpose_3074 1 1 1648 1649 dim0=1 dim1=2 $input=1648 #1648=(1,192,2304)f32 #1649=(1,2304,192)f32 +pnnx.Expression pnnx_expr_9426 2 1 1649 1446 1650 expr=add(@0,@1) #1649=(1,2304,192)f32 #1446=(1,2304,192)f32 #1650=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_251 1 1 1650 1651 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1650=(1,2304,192)f32 #1651=(1,2304,192)f32 +Tensor.reshape Tensor.view_1512 1 1 1651 1652 shape=(1,6,8,6,8,192) $input=1651 #1651=(1,2304,192)f32 #1652=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2720 1 1 1652 1653 dims=(0,1,3,2,4,5) $input=1652 #1652=(1,6,8,6,8,192)f32 #1653=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1514 1 1 1653 1654 shape=(36,64,192) $input=1653 #1653=(1,6,6,8,8,192)f32 #1654=(36,64,192)f32 +nn.Linear pnnx_unique_254 1 1 1654 1655 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1654=(36,64,192)f32 #1655=(36,64,576)f32 +Tensor.reshape Tensor.reshape_530 1 1 1655 1656 shape=(36,64,3,6,32) $input=1655 #1655=(36,64,576)f32 #1656=(36,64,3,6,32)f32 +torch.permute torch.permute_2721 1 1 1656 1657 dims=(2,0,3,1,4) $input=1656 #1656=(36,64,3,6,32)f32 #1657=(3,36,6,64,32)f32 +torch.unbind Tensor.select_794 1 3 1657 1658 1659 1660 dim=0 #1657=(3,36,6,64,32)f32 #1658=(36,6,64,32)f32 #1659=(36,6,64,32)f32 #1660=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_9337 1 1 1658 1661 expr=mul(@0,1.767767e-01) #1658=(36,6,64,32)f32 #1661=(36,6,64,32)f32 +torch.transpose torch.transpose_3075 1 1 1659 1662 dim0=-2 dim1=-1 $input=1659 #1659=(36,6,64,32)f32 #1662=(36,6,32,64)f32 +torch.matmul torch.matmul_2300 2 1 1661 1662 1663 $input=1661 $other=1662 #1661=(36,6,64,32)f32 #1662=(36,6,32,64)f32 #1663=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_10154 0 1 1664 @pnnx_fold_10154=(1,6,64,64)f32 #1664=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_9324 2 1 1663 1664 1665 expr=add(@0,@1) #1663=(36,6,64,64)f32 #1664=(1,6,64,64)f32 #1665=(36,6,64,64)f32 +nn.Softmax pnnx_unique_255 1 1 1665 1666 dim=-1 #1665=(36,6,64,64)f32 #1666=(36,6,64,64)f32 +torch.matmul torch.matmul_2301 2 1 1666 1660 1667 $input=1666 $other=1660 #1666=(36,6,64,64)f32 #1660=(36,6,64,32)f32 #1667=(36,6,64,32)f32 +torch.transpose torch.transpose_3076 1 1 1667 1668 dim0=1 dim1=2 $input=1667 #1667=(36,6,64,32)f32 #1668=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_531 1 1 1668 1669 shape=(36,64,192) $input=1668 #1668=(36,64,6,32)f32 #1669=(36,64,192)f32 +nn.Linear pnnx_unique_257 1 1 1669 1670 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1669=(36,64,192)f32 #1670=(36,64,192)f32 +Tensor.reshape Tensor.view_1518 1 1 1670 1671 shape=(1,6,6,8,8,192) $input=1670 #1670=(36,64,192)f32 #1671=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2723 1 1 1671 1672 dims=(0,1,3,2,4,5) $input=1671 #1671=(1,6,6,8,8,192)f32 #1672=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1520 1 1 1672 1673 shape=(1,2304,192) $input=1672 #1672=(1,6,8,6,8,192)f32 #1673=(1,2304,192)f32 +pnnx.Expression pnnx_expr_9291 2 1 1650 1673 1674 expr=add(@0,@1) #1650=(1,2304,192)f32 #1673=(1,2304,192)f32 #1674=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_259 1 1 1674 1675 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1674=(1,2304,192)f32 #1675=(1,2304,192)f32 +nn.Linear pnnx_unique_260 1 1 1675 1676 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1675=(1,2304,192)f32 #1676=(1,2304,384)f32 +nn.GELU pnnx_unique_261 1 1 1676 1677 #1676=(1,2304,384)f32 #1677=(1,2304,384)f32 +nn.Linear pnnx_unique_263 1 1 1677 1678 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1677=(1,2304,384)f32 #1678=(1,2304,192)f32 +pnnx.Expression pnnx_expr_9289 2 1 1674 1678 1679 expr=add(@0,@1) #1674=(1,2304,192)f32 #1678=(1,2304,192)f32 #1679=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_266 1 1 1679 1680 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1679=(1,2304,192)f32 #1680=(1,2304,192)f32 +Tensor.view Tensor.view_1521 1 1 1680 1681 shape=(1,48,48,192) $input=1680 #1680=(1,2304,192)f32 #1681=(1,48,48,192)f32 +torch.roll torch.roll_2468 1 1 1681 1682 dims=(1,2) shifts=(-4,-4) $input=1681 #1681=(1,48,48,192)f32 #1682=(1,48,48,192)f32 +Tensor.view Tensor.view_1522 1 1 1682 1683 shape=(1,6,8,6,8,192) $input=1682 #1682=(1,48,48,192)f32 #1683=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2724 1 1 1683 1684 dims=(0,1,3,2,4,5) $input=1683 #1683=(1,6,8,6,8,192)f32 #1684=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1524 1 1 1684 1685 shape=(36,64,192) $input=1684 #1684=(1,6,6,8,8,192)f32 #1685=(36,64,192)f32 +nn.Linear pnnx_unique_269 1 1 1685 1686 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1685=(36,64,192)f32 #1686=(36,64,576)f32 +Tensor.reshape Tensor.reshape_532 1 1 1686 1687 shape=(36,64,3,6,32) $input=1686 #1686=(36,64,576)f32 #1687=(36,64,3,6,32)f32 +torch.permute torch.permute_2725 1 1 1687 1688 dims=(2,0,3,1,4) $input=1687 #1687=(36,64,3,6,32)f32 #1688=(3,36,6,64,32)f32 +torch.unbind Tensor.select_797 1 3 1688 1689 1690 1691 dim=0 #1688=(3,36,6,64,32)f32 #1689=(36,6,64,32)f32 #1690=(36,6,64,32)f32 #1691=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_9191 1 1 1689 1692 expr=mul(@0,1.767767e-01) #1689=(36,6,64,32)f32 #1692=(36,6,64,32)f32 +torch.transpose torch.transpose_3077 1 1 1690 1693 dim0=-2 dim1=-1 $input=1690 #1690=(36,6,64,32)f32 #1693=(36,6,32,64)f32 +torch.matmul torch.matmul_2302 2 1 1692 1693 1694 $input=1692 $other=1693 #1692=(36,6,64,32)f32 #1693=(36,6,32,64)f32 #1694=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_10307 0 1 1695 @pnnx_fold_10307=(1,6,64,64)f32 #1695=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_9178 2 1 1694 1695 1696 expr=add(@0,@1) #1694=(36,6,64,64)f32 #1695=(1,6,64,64)f32 #1696=(36,6,64,64)f32 +Tensor.view Tensor.view_1527 1 1 1696 1697 shape=(1,36,6,64,64) $input=1696 #1696=(36,6,64,64)f32 #1697=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_10317 0 1 1698 @pnnx_fold_10317=(1,36,1,64,64)f32 #1698=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_9166 2 1 1697 1698 1699 expr=add(@0,@1) #1697=(1,36,6,64,64)f32 #1698=(1,36,1,64,64)f32 #1699=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1528 1 1 1699 1700 shape=(-1,6,64,64) $input=1699 #1699=(1,36,6,64,64)f32 #1700=(36,6,64,64)f32 +nn.Softmax pnnx_unique_270 1 1 1700 1701 dim=-1 #1700=(36,6,64,64)f32 #1701=(36,6,64,64)f32 +torch.matmul torch.matmul_2303 2 1 1701 1691 1702 $input=1701 $other=1691 #1701=(36,6,64,64)f32 #1691=(36,6,64,32)f32 #1702=(36,6,64,32)f32 +torch.transpose torch.transpose_3078 1 1 1702 1703 dim0=1 dim1=2 $input=1702 #1702=(36,6,64,32)f32 #1703=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_533 1 1 1703 1704 shape=(36,64,192) $input=1703 #1703=(36,64,6,32)f32 #1704=(36,64,192)f32 +nn.Linear pnnx_unique_272 1 1 1704 1705 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1704=(36,64,192)f32 #1705=(36,64,192)f32 +Tensor.reshape Tensor.view_1530 1 1 1705 1706 shape=(1,6,6,8,8,192) $input=1705 #1705=(36,64,192)f32 #1706=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2727 1 1 1706 1707 dims=(0,1,3,2,4,5) $input=1706 #1706=(1,6,6,8,8,192)f32 #1707=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_99 1 1 1707 1708 shape=(1,48,48,-1) $input=1707 #1707=(1,6,8,6,8,192)f32 #1708=(1,48,48,192)f32 +torch.roll torch.roll_2469 1 1 1708 1709 dims=(1,2) shifts=(4,4) $input=1708 #1708=(1,48,48,192)f32 #1709=(1,48,48,192)f32 +Tensor.view Tensor.view_1532 1 1 1709 1710 shape=(1,2304,192) $input=1709 #1709=(1,48,48,192)f32 #1710=(1,2304,192)f32 +pnnx.Expression pnnx_expr_9124 2 1 1679 1710 1711 expr=add(@0,@1) #1679=(1,2304,192)f32 #1710=(1,2304,192)f32 #1711=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_274 1 1 1711 1712 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1711=(1,2304,192)f32 #1712=(1,2304,192)f32 +nn.Linear pnnx_unique_275 1 1 1712 1713 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1712=(1,2304,192)f32 #1713=(1,2304,384)f32 +nn.GELU pnnx_unique_276 1 1 1713 1714 #1713=(1,2304,384)f32 #1714=(1,2304,384)f32 +nn.Linear pnnx_unique_278 1 1 1714 1715 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1714=(1,2304,384)f32 #1715=(1,2304,192)f32 +pnnx.Expression pnnx_expr_9122 2 1 1711 1715 1716 expr=add(@0,@1) #1711=(1,2304,192)f32 #1715=(1,2304,192)f32 #1716=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_280 1 1 1716 1717 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1716=(1,2304,192)f32 #1717=(1,2304,192)f32 +Tensor.reshape Tensor.view_1534 1 1 1717 1718 shape=(1,6,8,6,8,192) $input=1717 #1717=(1,2304,192)f32 #1718=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2728 1 1 1718 1719 dims=(0,1,3,2,4,5) $input=1718 #1718=(1,6,8,6,8,192)f32 #1719=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1536 1 1 1719 1720 shape=(36,64,192) $input=1719 #1719=(1,6,6,8,8,192)f32 #1720=(36,64,192)f32 +nn.Linear pnnx_unique_283 1 1 1720 1721 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1720=(36,64,192)f32 #1721=(36,64,576)f32 +Tensor.reshape Tensor.reshape_534 1 1 1721 1722 shape=(36,64,3,6,32) $input=1721 #1721=(36,64,576)f32 #1722=(36,64,3,6,32)f32 +torch.permute torch.permute_2729 1 1 1722 1723 dims=(2,0,3,1,4) $input=1722 #1722=(36,64,3,6,32)f32 #1723=(3,36,6,64,32)f32 +torch.unbind Tensor.select_800 1 3 1723 1724 1725 1726 dim=0 #1723=(3,36,6,64,32)f32 #1724=(36,6,64,32)f32 #1725=(36,6,64,32)f32 #1726=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_9034 1 1 1724 1727 expr=mul(@0,1.767767e-01) #1724=(36,6,64,32)f32 #1727=(36,6,64,32)f32 +torch.transpose torch.transpose_3079 1 1 1725 1728 dim0=-2 dim1=-1 $input=1725 #1725=(36,6,64,32)f32 #1728=(36,6,32,64)f32 +torch.matmul torch.matmul_2304 2 1 1727 1728 1729 $input=1727 $other=1728 #1727=(36,6,64,32)f32 #1728=(36,6,32,64)f32 #1729=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_10466 0 1 1730 @pnnx_fold_10466=(1,6,64,64)f32 #1730=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_9021 2 1 1729 1730 1731 expr=add(@0,@1) #1729=(36,6,64,64)f32 #1730=(1,6,64,64)f32 #1731=(36,6,64,64)f32 +nn.Softmax pnnx_unique_284 1 1 1731 1732 dim=-1 #1731=(36,6,64,64)f32 #1732=(36,6,64,64)f32 +torch.matmul torch.matmul_2305 2 1 1732 1726 1733 $input=1732 $other=1726 #1732=(36,6,64,64)f32 #1726=(36,6,64,32)f32 #1733=(36,6,64,32)f32 +torch.transpose torch.transpose_3080 1 1 1733 1734 dim0=1 dim1=2 $input=1733 #1733=(36,6,64,32)f32 #1734=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_535 1 1 1734 1735 shape=(36,64,192) $input=1734 #1734=(36,64,6,32)f32 #1735=(36,64,192)f32 +nn.Linear pnnx_unique_286 1 1 1735 1736 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1735=(36,64,192)f32 #1736=(36,64,192)f32 +Tensor.reshape Tensor.view_1540 1 1 1736 1737 shape=(1,6,6,8,8,192) $input=1736 #1736=(36,64,192)f32 #1737=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2731 1 1 1737 1738 dims=(0,1,3,2,4,5) $input=1737 #1737=(1,6,6,8,8,192)f32 #1738=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1542 1 1 1738 1739 shape=(1,2304,192) $input=1738 #1738=(1,6,8,6,8,192)f32 #1739=(1,2304,192)f32 +pnnx.Expression pnnx_expr_8988 2 1 1716 1739 1740 expr=add(@0,@1) #1716=(1,2304,192)f32 #1739=(1,2304,192)f32 #1740=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_288 1 1 1740 1741 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1740=(1,2304,192)f32 #1741=(1,2304,192)f32 +nn.Linear pnnx_unique_289 1 1 1741 1742 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1741=(1,2304,192)f32 #1742=(1,2304,384)f32 +nn.GELU pnnx_unique_290 1 1 1742 1743 #1742=(1,2304,384)f32 #1743=(1,2304,384)f32 +nn.Linear pnnx_unique_292 1 1 1743 1744 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1743=(1,2304,384)f32 #1744=(1,2304,192)f32 +pnnx.Expression pnnx_expr_8986 2 1 1740 1744 1745 expr=add(@0,@1) #1740=(1,2304,192)f32 #1744=(1,2304,192)f32 #1745=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_295 1 1 1745 1746 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1745=(1,2304,192)f32 #1746=(1,2304,192)f32 +Tensor.view Tensor.view_1543 1 1 1746 1747 shape=(1,48,48,192) $input=1746 #1746=(1,2304,192)f32 #1747=(1,48,48,192)f32 +torch.roll torch.roll_2470 1 1 1747 1748 dims=(1,2) shifts=(-4,-4) $input=1747 #1747=(1,48,48,192)f32 #1748=(1,48,48,192)f32 +Tensor.view Tensor.view_1544 1 1 1748 1749 shape=(1,6,8,6,8,192) $input=1748 #1748=(1,48,48,192)f32 #1749=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2732 1 1 1749 1750 dims=(0,1,3,2,4,5) $input=1749 #1749=(1,6,8,6,8,192)f32 #1750=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1546 1 1 1750 1751 shape=(36,64,192) $input=1750 #1750=(1,6,6,8,8,192)f32 #1751=(36,64,192)f32 +nn.Linear pnnx_unique_298 1 1 1751 1752 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1751=(36,64,192)f32 #1752=(36,64,576)f32 +Tensor.reshape Tensor.reshape_536 1 1 1752 1753 shape=(36,64,3,6,32) $input=1752 #1752=(36,64,576)f32 #1753=(36,64,3,6,32)f32 +torch.permute torch.permute_2733 1 1 1753 1754 dims=(2,0,3,1,4) $input=1753 #1753=(36,64,3,6,32)f32 #1754=(3,36,6,64,32)f32 +torch.unbind Tensor.select_803 1 3 1754 1755 1756 1757 dim=0 #1754=(3,36,6,64,32)f32 #1755=(36,6,64,32)f32 #1756=(36,6,64,32)f32 #1757=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_8888 1 1 1755 1758 expr=mul(@0,1.767767e-01) #1755=(36,6,64,32)f32 #1758=(36,6,64,32)f32 +torch.transpose torch.transpose_3081 1 1 1756 1759 dim0=-2 dim1=-1 $input=1756 #1756=(36,6,64,32)f32 #1759=(36,6,32,64)f32 +torch.matmul torch.matmul_2306 2 1 1758 1759 1760 $input=1758 $other=1759 #1758=(36,6,64,32)f32 #1759=(36,6,32,64)f32 #1760=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_10619 0 1 1761 @pnnx_fold_10619=(1,6,64,64)f32 #1761=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_8875 2 1 1760 1761 1762 expr=add(@0,@1) #1760=(36,6,64,64)f32 #1761=(1,6,64,64)f32 #1762=(36,6,64,64)f32 +Tensor.view Tensor.view_1549 1 1 1762 1763 shape=(1,36,6,64,64) $input=1762 #1762=(36,6,64,64)f32 #1763=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_10629 0 1 1764 @pnnx_fold_10629=(1,36,1,64,64)f32 #1764=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_8863 2 1 1763 1764 1765 expr=add(@0,@1) #1763=(1,36,6,64,64)f32 #1764=(1,36,1,64,64)f32 #1765=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1550 1 1 1765 1766 shape=(-1,6,64,64) $input=1765 #1765=(1,36,6,64,64)f32 #1766=(36,6,64,64)f32 +nn.Softmax pnnx_unique_299 1 1 1766 1767 dim=-1 #1766=(36,6,64,64)f32 #1767=(36,6,64,64)f32 +torch.matmul torch.matmul_2307 2 1 1767 1757 1768 $input=1767 $other=1757 #1767=(36,6,64,64)f32 #1757=(36,6,64,32)f32 #1768=(36,6,64,32)f32 +torch.transpose torch.transpose_3082 1 1 1768 1769 dim0=1 dim1=2 $input=1768 #1768=(36,6,64,32)f32 #1769=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_537 1 1 1769 1770 shape=(36,64,192) $input=1769 #1769=(36,64,6,32)f32 #1770=(36,64,192)f32 +nn.Linear pnnx_unique_301 1 1 1770 1771 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1770=(36,64,192)f32 #1771=(36,64,192)f32 +Tensor.reshape Tensor.view_1552 1 1 1771 1772 shape=(1,6,6,8,8,192) $input=1771 #1771=(36,64,192)f32 #1772=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2735 1 1 1772 1773 dims=(0,1,3,2,4,5) $input=1772 #1772=(1,6,6,8,8,192)f32 #1773=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_103 1 1 1773 1774 shape=(1,48,48,-1) $input=1773 #1773=(1,6,8,6,8,192)f32 #1774=(1,48,48,192)f32 +torch.roll torch.roll_2471 1 1 1774 1775 dims=(1,2) shifts=(4,4) $input=1774 #1774=(1,48,48,192)f32 #1775=(1,48,48,192)f32 +Tensor.view Tensor.view_1554 1 1 1775 1776 shape=(1,2304,192) $input=1775 #1775=(1,48,48,192)f32 #1776=(1,2304,192)f32 +pnnx.Expression pnnx_expr_8821 2 1 1745 1776 1777 expr=add(@0,@1) #1745=(1,2304,192)f32 #1776=(1,2304,192)f32 #1777=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_303 1 1 1777 1778 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1777=(1,2304,192)f32 #1778=(1,2304,192)f32 +nn.Linear pnnx_unique_304 1 1 1778 1779 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1778=(1,2304,192)f32 #1779=(1,2304,384)f32 +nn.GELU pnnx_unique_305 1 1 1779 1780 #1779=(1,2304,384)f32 #1780=(1,2304,384)f32 +nn.Linear pnnx_unique_307 1 1 1780 1781 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1780=(1,2304,384)f32 #1781=(1,2304,192)f32 +pnnx.Expression pnnx_expr_8819 2 1 1777 1781 1782 expr=add(@0,@1) #1777=(1,2304,192)f32 #1781=(1,2304,192)f32 #1782=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_309 1 1 1782 1783 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1782=(1,2304,192)f32 #1783=(1,2304,192)f32 +Tensor.reshape Tensor.view_1556 1 1 1783 1784 shape=(1,6,8,6,8,192) $input=1783 #1783=(1,2304,192)f32 #1784=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2736 1 1 1784 1785 dims=(0,1,3,2,4,5) $input=1784 #1784=(1,6,8,6,8,192)f32 #1785=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1558 1 1 1785 1786 shape=(36,64,192) $input=1785 #1785=(1,6,6,8,8,192)f32 #1786=(36,64,192)f32 +nn.Linear pnnx_unique_312 1 1 1786 1787 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1786=(36,64,192)f32 #1787=(36,64,576)f32 +Tensor.reshape Tensor.reshape_538 1 1 1787 1788 shape=(36,64,3,6,32) $input=1787 #1787=(36,64,576)f32 #1788=(36,64,3,6,32)f32 +torch.permute torch.permute_2737 1 1 1788 1789 dims=(2,0,3,1,4) $input=1788 #1788=(36,64,3,6,32)f32 #1789=(3,36,6,64,32)f32 +torch.unbind Tensor.select_806 1 3 1789 1790 1791 1792 dim=0 #1789=(3,36,6,64,32)f32 #1790=(36,6,64,32)f32 #1791=(36,6,64,32)f32 #1792=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_8731 1 1 1790 1793 expr=mul(@0,1.767767e-01) #1790=(36,6,64,32)f32 #1793=(36,6,64,32)f32 +torch.transpose torch.transpose_3083 1 1 1791 1794 dim0=-2 dim1=-1 $input=1791 #1791=(36,6,64,32)f32 #1794=(36,6,32,64)f32 +torch.matmul torch.matmul_2308 2 1 1793 1794 1795 $input=1793 $other=1794 #1793=(36,6,64,32)f32 #1794=(36,6,32,64)f32 #1795=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_10778 0 1 1796 @pnnx_fold_10778=(1,6,64,64)f32 #1796=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_8718 2 1 1795 1796 1797 expr=add(@0,@1) #1795=(36,6,64,64)f32 #1796=(1,6,64,64)f32 #1797=(36,6,64,64)f32 +nn.Softmax pnnx_unique_313 1 1 1797 1798 dim=-1 #1797=(36,6,64,64)f32 #1798=(36,6,64,64)f32 +torch.matmul torch.matmul_2309 2 1 1798 1792 1799 $input=1798 $other=1792 #1798=(36,6,64,64)f32 #1792=(36,6,64,32)f32 #1799=(36,6,64,32)f32 +torch.transpose torch.transpose_3084 1 1 1799 1800 dim0=1 dim1=2 $input=1799 #1799=(36,6,64,32)f32 #1800=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_539 1 1 1800 1801 shape=(36,64,192) $input=1800 #1800=(36,64,6,32)f32 #1801=(36,64,192)f32 +nn.Linear pnnx_unique_315 1 1 1801 1802 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1801=(36,64,192)f32 #1802=(36,64,192)f32 +Tensor.reshape Tensor.view_1562 1 1 1802 1803 shape=(1,6,6,8,8,192) $input=1802 #1802=(36,64,192)f32 #1803=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2739 1 1 1803 1804 dims=(0,1,3,2,4,5) $input=1803 #1803=(1,6,6,8,8,192)f32 #1804=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1564 1 1 1804 1805 shape=(1,2304,192) $input=1804 #1804=(1,6,8,6,8,192)f32 #1805=(1,2304,192)f32 +pnnx.Expression pnnx_expr_8685 2 1 1782 1805 1806 expr=add(@0,@1) #1782=(1,2304,192)f32 #1805=(1,2304,192)f32 #1806=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_317 1 1 1806 1807 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1806=(1,2304,192)f32 #1807=(1,2304,192)f32 +nn.Linear pnnx_unique_318 1 1 1807 1808 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1807=(1,2304,192)f32 #1808=(1,2304,384)f32 +nn.GELU pnnx_unique_319 1 1 1808 1809 #1808=(1,2304,384)f32 #1809=(1,2304,384)f32 +nn.Linear pnnx_unique_321 1 1 1809 1810 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1809=(1,2304,384)f32 #1810=(1,2304,192)f32 +pnnx.Expression pnnx_expr_8683 2 1 1806 1810 1811 expr=add(@0,@1) #1806=(1,2304,192)f32 #1810=(1,2304,192)f32 #1811=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_324 1 1 1811 1812 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1811=(1,2304,192)f32 #1812=(1,2304,192)f32 +Tensor.view Tensor.view_1565 1 1 1812 1813 shape=(1,48,48,192) $input=1812 #1812=(1,2304,192)f32 #1813=(1,48,48,192)f32 +torch.roll torch.roll_2472 1 1 1813 1814 dims=(1,2) shifts=(-4,-4) $input=1813 #1813=(1,48,48,192)f32 #1814=(1,48,48,192)f32 +Tensor.view Tensor.view_1566 1 1 1814 1815 shape=(1,6,8,6,8,192) $input=1814 #1814=(1,48,48,192)f32 #1815=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2740 1 1 1815 1816 dims=(0,1,3,2,4,5) $input=1815 #1815=(1,6,8,6,8,192)f32 #1816=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1568 1 1 1816 1817 shape=(36,64,192) $input=1816 #1816=(1,6,6,8,8,192)f32 #1817=(36,64,192)f32 +nn.Linear pnnx_unique_327 1 1 1817 1818 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1817=(36,64,192)f32 #1818=(36,64,576)f32 +Tensor.reshape Tensor.reshape_540 1 1 1818 1819 shape=(36,64,3,6,32) $input=1818 #1818=(36,64,576)f32 #1819=(36,64,3,6,32)f32 +torch.permute torch.permute_2741 1 1 1819 1820 dims=(2,0,3,1,4) $input=1819 #1819=(36,64,3,6,32)f32 #1820=(3,36,6,64,32)f32 +torch.unbind Tensor.select_809 1 3 1820 1821 1822 1823 dim=0 #1820=(3,36,6,64,32)f32 #1821=(36,6,64,32)f32 #1822=(36,6,64,32)f32 #1823=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_8585 1 1 1821 1824 expr=mul(@0,1.767767e-01) #1821=(36,6,64,32)f32 #1824=(36,6,64,32)f32 +torch.transpose torch.transpose_3085 1 1 1822 1825 dim0=-2 dim1=-1 $input=1822 #1822=(36,6,64,32)f32 #1825=(36,6,32,64)f32 +torch.matmul torch.matmul_2310 2 1 1824 1825 1826 $input=1824 $other=1825 #1824=(36,6,64,32)f32 #1825=(36,6,32,64)f32 #1826=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_10931 0 1 1827 @pnnx_fold_10931=(1,6,64,64)f32 #1827=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_8572 2 1 1826 1827 1828 expr=add(@0,@1) #1826=(36,6,64,64)f32 #1827=(1,6,64,64)f32 #1828=(36,6,64,64)f32 +Tensor.view Tensor.view_1571 1 1 1828 1829 shape=(1,36,6,64,64) $input=1828 #1828=(36,6,64,64)f32 #1829=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_10941 0 1 1830 @pnnx_fold_10941=(1,36,1,64,64)f32 #1830=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_8560 2 1 1829 1830 1831 expr=add(@0,@1) #1829=(1,36,6,64,64)f32 #1830=(1,36,1,64,64)f32 #1831=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1572 1 1 1831 1832 shape=(-1,6,64,64) $input=1831 #1831=(1,36,6,64,64)f32 #1832=(36,6,64,64)f32 +nn.Softmax pnnx_unique_328 1 1 1832 1833 dim=-1 #1832=(36,6,64,64)f32 #1833=(36,6,64,64)f32 +torch.matmul torch.matmul_2311 2 1 1833 1823 1834 $input=1833 $other=1823 #1833=(36,6,64,64)f32 #1823=(36,6,64,32)f32 #1834=(36,6,64,32)f32 +torch.transpose torch.transpose_3086 1 1 1834 1835 dim0=1 dim1=2 $input=1834 #1834=(36,6,64,32)f32 #1835=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_541 1 1 1835 1836 shape=(36,64,192) $input=1835 #1835=(36,64,6,32)f32 #1836=(36,64,192)f32 +nn.Linear pnnx_unique_330 1 1 1836 1837 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1836=(36,64,192)f32 #1837=(36,64,192)f32 +Tensor.reshape Tensor.view_1574 1 1 1837 1838 shape=(1,6,6,8,8,192) $input=1837 #1837=(36,64,192)f32 #1838=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2743 1 1 1838 1839 dims=(0,1,3,2,4,5) $input=1838 #1838=(1,6,6,8,8,192)f32 #1839=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_107 1 1 1839 1840 shape=(1,48,48,-1) $input=1839 #1839=(1,6,8,6,8,192)f32 #1840=(1,48,48,192)f32 +torch.roll torch.roll_2473 1 1 1840 1841 dims=(1,2) shifts=(4,4) $input=1840 #1840=(1,48,48,192)f32 #1841=(1,48,48,192)f32 +Tensor.view Tensor.view_1576 1 1 1841 1842 shape=(1,2304,192) $input=1841 #1841=(1,48,48,192)f32 #1842=(1,2304,192)f32 +pnnx.Expression pnnx_expr_8518 2 1 1811 1842 1843 expr=add(@0,@1) #1811=(1,2304,192)f32 #1842=(1,2304,192)f32 #1843=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_332 1 1 1843 1844 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1843=(1,2304,192)f32 #1844=(1,2304,192)f32 +nn.Linear pnnx_unique_333 1 1 1844 1845 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1844=(1,2304,192)f32 #1845=(1,2304,384)f32 +nn.GELU pnnx_unique_334 1 1 1845 1846 #1845=(1,2304,384)f32 #1846=(1,2304,384)f32 +nn.Linear pnnx_unique_336 1 1 1846 1847 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1846=(1,2304,384)f32 #1847=(1,2304,192)f32 +pnnx.Expression pnnx_expr_8516 2 1 1843 1847 1848 expr=add(@0,@1) #1843=(1,2304,192)f32 #1847=(1,2304,192)f32 #1848=(1,2304,192)f32 +torch.transpose torch.transpose_3087 1 1 1848 1849 dim0=1 dim1=2 $input=1848 #1848=(1,2304,192)f32 #1849=(1,192,2304)f32 +Tensor.view Tensor.view_1577 1 1 1849 1850 shape=(1,192,48,48) $input=1849 #1849=(1,192,2304)f32 #1850=(1,192,48,48)f32 +nn.Conv2d pnnx_unique_338 1 1 1850 1851 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #1850=(1,192,48,48)f32 #1851=(1,192,48,48)f32 +torch.flatten torch.flatten_2193 1 1 1851 1852 end_dim=-1 start_dim=2 $input=1851 #1851=(1,192,48,48)f32 #1852=(1,192,2304)f32 +torch.transpose torch.transpose_3088 1 1 1852 1853 dim0=1 dim1=2 $input=1852 #1852=(1,192,2304)f32 #1853=(1,2304,192)f32 +pnnx.Expression pnnx_expr_8503 2 1 1853 1650 1854 expr=add(@0,@1) #1853=(1,2304,192)f32 #1650=(1,2304,192)f32 #1854=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_339 1 1 1854 1855 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1854=(1,2304,192)f32 #1855=(1,2304,192)f32 +Tensor.reshape Tensor.view_1579 1 1 1855 1856 shape=(1,6,8,6,8,192) $input=1855 #1855=(1,2304,192)f32 #1856=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2744 1 1 1856 1857 dims=(0,1,3,2,4,5) $input=1856 #1856=(1,6,8,6,8,192)f32 #1857=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1581 1 1 1857 1858 shape=(36,64,192) $input=1857 #1857=(1,6,6,8,8,192)f32 #1858=(36,64,192)f32 +nn.Linear pnnx_unique_342 1 1 1858 1859 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1858=(36,64,192)f32 #1859=(36,64,576)f32 +Tensor.reshape Tensor.reshape_542 1 1 1859 1860 shape=(36,64,3,6,32) $input=1859 #1859=(36,64,576)f32 #1860=(36,64,3,6,32)f32 +torch.permute torch.permute_2745 1 1 1860 1861 dims=(2,0,3,1,4) $input=1860 #1860=(36,64,3,6,32)f32 #1861=(3,36,6,64,32)f32 +torch.unbind Tensor.select_812 1 3 1861 1862 1863 1864 dim=0 #1861=(3,36,6,64,32)f32 #1862=(36,6,64,32)f32 #1863=(36,6,64,32)f32 #1864=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_8414 1 1 1862 1865 expr=mul(@0,1.767767e-01) #1862=(36,6,64,32)f32 #1865=(36,6,64,32)f32 +torch.transpose torch.transpose_3089 1 1 1863 1866 dim0=-2 dim1=-1 $input=1863 #1863=(36,6,64,32)f32 #1866=(36,6,32,64)f32 +torch.matmul torch.matmul_2312 2 1 1865 1866 1867 $input=1865 $other=1866 #1865=(36,6,64,32)f32 #1866=(36,6,32,64)f32 #1867=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_11124 0 1 1868 @pnnx_fold_11124=(1,6,64,64)f32 #1868=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_8401 2 1 1867 1868 1869 expr=add(@0,@1) #1867=(36,6,64,64)f32 #1868=(1,6,64,64)f32 #1869=(36,6,64,64)f32 +nn.Softmax pnnx_unique_343 1 1 1869 1870 dim=-1 #1869=(36,6,64,64)f32 #1870=(36,6,64,64)f32 +torch.matmul torch.matmul_2313 2 1 1870 1864 1871 $input=1870 $other=1864 #1870=(36,6,64,64)f32 #1864=(36,6,64,32)f32 #1871=(36,6,64,32)f32 +torch.transpose torch.transpose_3090 1 1 1871 1872 dim0=1 dim1=2 $input=1871 #1871=(36,6,64,32)f32 #1872=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_543 1 1 1872 1873 shape=(36,64,192) $input=1872 #1872=(36,64,6,32)f32 #1873=(36,64,192)f32 +nn.Linear pnnx_unique_345 1 1 1873 1874 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1873=(36,64,192)f32 #1874=(36,64,192)f32 +Tensor.reshape Tensor.view_1585 1 1 1874 1875 shape=(1,6,6,8,8,192) $input=1874 #1874=(36,64,192)f32 #1875=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2747 1 1 1875 1876 dims=(0,1,3,2,4,5) $input=1875 #1875=(1,6,6,8,8,192)f32 #1876=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1587 1 1 1876 1877 shape=(1,2304,192) $input=1876 #1876=(1,6,8,6,8,192)f32 #1877=(1,2304,192)f32 +pnnx.Expression pnnx_expr_8368 2 1 1854 1877 1878 expr=add(@0,@1) #1854=(1,2304,192)f32 #1877=(1,2304,192)f32 #1878=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_347 1 1 1878 1879 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1878=(1,2304,192)f32 #1879=(1,2304,192)f32 +nn.Linear pnnx_unique_348 1 1 1879 1880 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1879=(1,2304,192)f32 #1880=(1,2304,384)f32 +nn.GELU pnnx_unique_349 1 1 1880 1881 #1880=(1,2304,384)f32 #1881=(1,2304,384)f32 +nn.Linear pnnx_unique_351 1 1 1881 1882 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1881=(1,2304,384)f32 #1882=(1,2304,192)f32 +pnnx.Expression pnnx_expr_8366 2 1 1878 1882 1883 expr=add(@0,@1) #1878=(1,2304,192)f32 #1882=(1,2304,192)f32 #1883=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_354 1 1 1883 1884 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1883=(1,2304,192)f32 #1884=(1,2304,192)f32 +Tensor.view Tensor.view_1588 1 1 1884 1885 shape=(1,48,48,192) $input=1884 #1884=(1,2304,192)f32 #1885=(1,48,48,192)f32 +torch.roll torch.roll_2474 1 1 1885 1886 dims=(1,2) shifts=(-4,-4) $input=1885 #1885=(1,48,48,192)f32 #1886=(1,48,48,192)f32 +Tensor.view Tensor.view_1589 1 1 1886 1887 shape=(1,6,8,6,8,192) $input=1886 #1886=(1,48,48,192)f32 #1887=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2748 1 1 1887 1888 dims=(0,1,3,2,4,5) $input=1887 #1887=(1,6,8,6,8,192)f32 #1888=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1591 1 1 1888 1889 shape=(36,64,192) $input=1888 #1888=(1,6,6,8,8,192)f32 #1889=(36,64,192)f32 +nn.Linear pnnx_unique_357 1 1 1889 1890 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1889=(36,64,192)f32 #1890=(36,64,576)f32 +Tensor.reshape Tensor.reshape_544 1 1 1890 1891 shape=(36,64,3,6,32) $input=1890 #1890=(36,64,576)f32 #1891=(36,64,3,6,32)f32 +torch.permute torch.permute_2749 1 1 1891 1892 dims=(2,0,3,1,4) $input=1891 #1891=(36,64,3,6,32)f32 #1892=(3,36,6,64,32)f32 +torch.unbind Tensor.select_815 1 3 1892 1893 1894 1895 dim=0 #1892=(3,36,6,64,32)f32 #1893=(36,6,64,32)f32 #1894=(36,6,64,32)f32 #1895=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_8268 1 1 1893 1896 expr=mul(@0,1.767767e-01) #1893=(36,6,64,32)f32 #1896=(36,6,64,32)f32 +torch.transpose torch.transpose_3091 1 1 1894 1897 dim0=-2 dim1=-1 $input=1894 #1894=(36,6,64,32)f32 #1897=(36,6,32,64)f32 +torch.matmul torch.matmul_2314 2 1 1896 1897 1898 $input=1896 $other=1897 #1896=(36,6,64,32)f32 #1897=(36,6,32,64)f32 #1898=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_11277 0 1 1899 @pnnx_fold_11277=(1,6,64,64)f32 #1899=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_8255 2 1 1898 1899 1900 expr=add(@0,@1) #1898=(36,6,64,64)f32 #1899=(1,6,64,64)f32 #1900=(36,6,64,64)f32 +Tensor.view Tensor.view_1594 1 1 1900 1901 shape=(1,36,6,64,64) $input=1900 #1900=(36,6,64,64)f32 #1901=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_11287 0 1 1902 @pnnx_fold_11287=(1,36,1,64,64)f32 #1902=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_8243 2 1 1901 1902 1903 expr=add(@0,@1) #1901=(1,36,6,64,64)f32 #1902=(1,36,1,64,64)f32 #1903=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1595 1 1 1903 1904 shape=(-1,6,64,64) $input=1903 #1903=(1,36,6,64,64)f32 #1904=(36,6,64,64)f32 +nn.Softmax pnnx_unique_358 1 1 1904 1905 dim=-1 #1904=(36,6,64,64)f32 #1905=(36,6,64,64)f32 +torch.matmul torch.matmul_2315 2 1 1905 1895 1906 $input=1905 $other=1895 #1905=(36,6,64,64)f32 #1895=(36,6,64,32)f32 #1906=(36,6,64,32)f32 +torch.transpose torch.transpose_3092 1 1 1906 1907 dim0=1 dim1=2 $input=1906 #1906=(36,6,64,32)f32 #1907=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_545 1 1 1907 1908 shape=(36,64,192) $input=1907 #1907=(36,64,6,32)f32 #1908=(36,64,192)f32 +nn.Linear pnnx_unique_360 1 1 1908 1909 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1908=(36,64,192)f32 #1909=(36,64,192)f32 +Tensor.reshape Tensor.view_1597 1 1 1909 1910 shape=(1,6,6,8,8,192) $input=1909 #1909=(36,64,192)f32 #1910=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2751 1 1 1910 1911 dims=(0,1,3,2,4,5) $input=1910 #1910=(1,6,6,8,8,192)f32 #1911=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_111 1 1 1911 1912 shape=(1,48,48,-1) $input=1911 #1911=(1,6,8,6,8,192)f32 #1912=(1,48,48,192)f32 +torch.roll torch.roll_2475 1 1 1912 1913 dims=(1,2) shifts=(4,4) $input=1912 #1912=(1,48,48,192)f32 #1913=(1,48,48,192)f32 +Tensor.view Tensor.view_1599 1 1 1913 1914 shape=(1,2304,192) $input=1913 #1913=(1,48,48,192)f32 #1914=(1,2304,192)f32 +pnnx.Expression pnnx_expr_8201 2 1 1883 1914 1915 expr=add(@0,@1) #1883=(1,2304,192)f32 #1914=(1,2304,192)f32 #1915=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_362 1 1 1915 1916 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1915=(1,2304,192)f32 #1916=(1,2304,192)f32 +nn.Linear pnnx_unique_363 1 1 1916 1917 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1916=(1,2304,192)f32 #1917=(1,2304,384)f32 +nn.GELU pnnx_unique_364 1 1 1917 1918 #1917=(1,2304,384)f32 #1918=(1,2304,384)f32 +nn.Linear pnnx_unique_366 1 1 1918 1919 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1918=(1,2304,384)f32 #1919=(1,2304,192)f32 +pnnx.Expression pnnx_expr_8199 2 1 1915 1919 1920 expr=add(@0,@1) #1915=(1,2304,192)f32 #1919=(1,2304,192)f32 #1920=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_368 1 1 1920 1921 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1920=(1,2304,192)f32 #1921=(1,2304,192)f32 +Tensor.reshape Tensor.view_1601 1 1 1921 1922 shape=(1,6,8,6,8,192) $input=1921 #1921=(1,2304,192)f32 #1922=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2752 1 1 1922 1923 dims=(0,1,3,2,4,5) $input=1922 #1922=(1,6,8,6,8,192)f32 #1923=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1603 1 1 1923 1924 shape=(36,64,192) $input=1923 #1923=(1,6,6,8,8,192)f32 #1924=(36,64,192)f32 +nn.Linear pnnx_unique_371 1 1 1924 1925 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1924=(36,64,192)f32 #1925=(36,64,576)f32 +Tensor.reshape Tensor.reshape_546 1 1 1925 1926 shape=(36,64,3,6,32) $input=1925 #1925=(36,64,576)f32 #1926=(36,64,3,6,32)f32 +torch.permute torch.permute_2753 1 1 1926 1927 dims=(2,0,3,1,4) $input=1926 #1926=(36,64,3,6,32)f32 #1927=(3,36,6,64,32)f32 +torch.unbind Tensor.select_818 1 3 1927 1928 1929 1930 dim=0 #1927=(3,36,6,64,32)f32 #1928=(36,6,64,32)f32 #1929=(36,6,64,32)f32 #1930=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_8111 1 1 1928 1931 expr=mul(@0,1.767767e-01) #1928=(36,6,64,32)f32 #1931=(36,6,64,32)f32 +torch.transpose torch.transpose_3093 1 1 1929 1932 dim0=-2 dim1=-1 $input=1929 #1929=(36,6,64,32)f32 #1932=(36,6,32,64)f32 +torch.matmul torch.matmul_2316 2 1 1931 1932 1933 $input=1931 $other=1932 #1931=(36,6,64,32)f32 #1932=(36,6,32,64)f32 #1933=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_11436 0 1 1934 @pnnx_fold_11436=(1,6,64,64)f32 #1934=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_8098 2 1 1933 1934 1935 expr=add(@0,@1) #1933=(36,6,64,64)f32 #1934=(1,6,64,64)f32 #1935=(36,6,64,64)f32 +nn.Softmax pnnx_unique_372 1 1 1935 1936 dim=-1 #1935=(36,6,64,64)f32 #1936=(36,6,64,64)f32 +torch.matmul torch.matmul_2317 2 1 1936 1930 1937 $input=1936 $other=1930 #1936=(36,6,64,64)f32 #1930=(36,6,64,32)f32 #1937=(36,6,64,32)f32 +torch.transpose torch.transpose_3094 1 1 1937 1938 dim0=1 dim1=2 $input=1937 #1937=(36,6,64,32)f32 #1938=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_547 1 1 1938 1939 shape=(36,64,192) $input=1938 #1938=(36,64,6,32)f32 #1939=(36,64,192)f32 +nn.Linear pnnx_unique_374 1 1 1939 1940 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1939=(36,64,192)f32 #1940=(36,64,192)f32 +Tensor.reshape Tensor.view_1607 1 1 1940 1941 shape=(1,6,6,8,8,192) $input=1940 #1940=(36,64,192)f32 #1941=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2755 1 1 1941 1942 dims=(0,1,3,2,4,5) $input=1941 #1941=(1,6,6,8,8,192)f32 #1942=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1609 1 1 1942 1943 shape=(1,2304,192) $input=1942 #1942=(1,6,8,6,8,192)f32 #1943=(1,2304,192)f32 +pnnx.Expression pnnx_expr_8065 2 1 1920 1943 1944 expr=add(@0,@1) #1920=(1,2304,192)f32 #1943=(1,2304,192)f32 #1944=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_376 1 1 1944 1945 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1944=(1,2304,192)f32 #1945=(1,2304,192)f32 +nn.Linear pnnx_unique_377 1 1 1945 1946 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1945=(1,2304,192)f32 #1946=(1,2304,384)f32 +nn.GELU pnnx_unique_378 1 1 1946 1947 #1946=(1,2304,384)f32 #1947=(1,2304,384)f32 +nn.Linear pnnx_unique_380 1 1 1947 1948 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1947=(1,2304,384)f32 #1948=(1,2304,192)f32 +pnnx.Expression pnnx_expr_8063 2 1 1944 1948 1949 expr=add(@0,@1) #1944=(1,2304,192)f32 #1948=(1,2304,192)f32 #1949=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_383 1 1 1949 1950 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1949=(1,2304,192)f32 #1950=(1,2304,192)f32 +Tensor.view Tensor.view_1610 1 1 1950 1951 shape=(1,48,48,192) $input=1950 #1950=(1,2304,192)f32 #1951=(1,48,48,192)f32 +torch.roll torch.roll_2476 1 1 1951 1952 dims=(1,2) shifts=(-4,-4) $input=1951 #1951=(1,48,48,192)f32 #1952=(1,48,48,192)f32 +Tensor.view Tensor.view_1611 1 1 1952 1953 shape=(1,6,8,6,8,192) $input=1952 #1952=(1,48,48,192)f32 #1953=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2756 1 1 1953 1954 dims=(0,1,3,2,4,5) $input=1953 #1953=(1,6,8,6,8,192)f32 #1954=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1613 1 1 1954 1955 shape=(36,64,192) $input=1954 #1954=(1,6,6,8,8,192)f32 #1955=(36,64,192)f32 +nn.Linear pnnx_unique_386 1 1 1955 1956 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1955=(36,64,192)f32 #1956=(36,64,576)f32 +Tensor.reshape Tensor.reshape_548 1 1 1956 1957 shape=(36,64,3,6,32) $input=1956 #1956=(36,64,576)f32 #1957=(36,64,3,6,32)f32 +torch.permute torch.permute_2757 1 1 1957 1958 dims=(2,0,3,1,4) $input=1957 #1957=(36,64,3,6,32)f32 #1958=(3,36,6,64,32)f32 +torch.unbind Tensor.select_821 1 3 1958 1959 1960 1961 dim=0 #1958=(3,36,6,64,32)f32 #1959=(36,6,64,32)f32 #1960=(36,6,64,32)f32 #1961=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_7965 1 1 1959 1962 expr=mul(@0,1.767767e-01) #1959=(36,6,64,32)f32 #1962=(36,6,64,32)f32 +torch.transpose torch.transpose_3095 1 1 1960 1963 dim0=-2 dim1=-1 $input=1960 #1960=(36,6,64,32)f32 #1963=(36,6,32,64)f32 +torch.matmul torch.matmul_2318 2 1 1962 1963 1964 $input=1962 $other=1963 #1962=(36,6,64,32)f32 #1963=(36,6,32,64)f32 #1964=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_11589 0 1 1965 @pnnx_fold_11589=(1,6,64,64)f32 #1965=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_7952 2 1 1964 1965 1966 expr=add(@0,@1) #1964=(36,6,64,64)f32 #1965=(1,6,64,64)f32 #1966=(36,6,64,64)f32 +Tensor.view Tensor.view_1616 1 1 1966 1967 shape=(1,36,6,64,64) $input=1966 #1966=(36,6,64,64)f32 #1967=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_11599 0 1 1968 @pnnx_fold_11599=(1,36,1,64,64)f32 #1968=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_7940 2 1 1967 1968 1969 expr=add(@0,@1) #1967=(1,36,6,64,64)f32 #1968=(1,36,1,64,64)f32 #1969=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1617 1 1 1969 1970 shape=(-1,6,64,64) $input=1969 #1969=(1,36,6,64,64)f32 #1970=(36,6,64,64)f32 +nn.Softmax pnnx_unique_387 1 1 1970 1971 dim=-1 #1970=(36,6,64,64)f32 #1971=(36,6,64,64)f32 +torch.matmul torch.matmul_2319 2 1 1971 1961 1972 $input=1971 $other=1961 #1971=(36,6,64,64)f32 #1961=(36,6,64,32)f32 #1972=(36,6,64,32)f32 +torch.transpose torch.transpose_3096 1 1 1972 1973 dim0=1 dim1=2 $input=1972 #1972=(36,6,64,32)f32 #1973=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_549 1 1 1973 1974 shape=(36,64,192) $input=1973 #1973=(36,64,6,32)f32 #1974=(36,64,192)f32 +nn.Linear pnnx_unique_389 1 1 1974 1975 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #1974=(36,64,192)f32 #1975=(36,64,192)f32 +Tensor.reshape Tensor.view_1619 1 1 1975 1976 shape=(1,6,6,8,8,192) $input=1975 #1975=(36,64,192)f32 #1976=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2759 1 1 1976 1977 dims=(0,1,3,2,4,5) $input=1976 #1976=(1,6,6,8,8,192)f32 #1977=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_115 1 1 1977 1978 shape=(1,48,48,-1) $input=1977 #1977=(1,6,8,6,8,192)f32 #1978=(1,48,48,192)f32 +torch.roll torch.roll_2477 1 1 1978 1979 dims=(1,2) shifts=(4,4) $input=1978 #1978=(1,48,48,192)f32 #1979=(1,48,48,192)f32 +Tensor.view Tensor.view_1621 1 1 1979 1980 shape=(1,2304,192) $input=1979 #1979=(1,48,48,192)f32 #1980=(1,2304,192)f32 +pnnx.Expression pnnx_expr_7898 2 1 1949 1980 1981 expr=add(@0,@1) #1949=(1,2304,192)f32 #1980=(1,2304,192)f32 #1981=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_391 1 1 1981 1982 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1981=(1,2304,192)f32 #1982=(1,2304,192)f32 +nn.Linear pnnx_unique_392 1 1 1982 1983 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #1982=(1,2304,192)f32 #1983=(1,2304,384)f32 +nn.GELU pnnx_unique_393 1 1 1983 1984 #1983=(1,2304,384)f32 #1984=(1,2304,384)f32 +nn.Linear pnnx_unique_395 1 1 1984 1985 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #1984=(1,2304,384)f32 #1985=(1,2304,192)f32 +pnnx.Expression pnnx_expr_7896 2 1 1981 1985 1986 expr=add(@0,@1) #1981=(1,2304,192)f32 #1985=(1,2304,192)f32 #1986=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_397 1 1 1986 1987 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #1986=(1,2304,192)f32 #1987=(1,2304,192)f32 +Tensor.reshape Tensor.view_1623 1 1 1987 1988 shape=(1,6,8,6,8,192) $input=1987 #1987=(1,2304,192)f32 #1988=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2760 1 1 1988 1989 dims=(0,1,3,2,4,5) $input=1988 #1988=(1,6,8,6,8,192)f32 #1989=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1625 1 1 1989 1990 shape=(36,64,192) $input=1989 #1989=(1,6,6,8,8,192)f32 #1990=(36,64,192)f32 +nn.Linear pnnx_unique_400 1 1 1990 1991 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #1990=(36,64,192)f32 #1991=(36,64,576)f32 +Tensor.reshape Tensor.reshape_550 1 1 1991 1992 shape=(36,64,3,6,32) $input=1991 #1991=(36,64,576)f32 #1992=(36,64,3,6,32)f32 +torch.permute torch.permute_2761 1 1 1992 1993 dims=(2,0,3,1,4) $input=1992 #1992=(36,64,3,6,32)f32 #1993=(3,36,6,64,32)f32 +torch.unbind Tensor.select_824 1 3 1993 1994 1995 1996 dim=0 #1993=(3,36,6,64,32)f32 #1994=(36,6,64,32)f32 #1995=(36,6,64,32)f32 #1996=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_7808 1 1 1994 1997 expr=mul(@0,1.767767e-01) #1994=(36,6,64,32)f32 #1997=(36,6,64,32)f32 +torch.transpose torch.transpose_3097 1 1 1995 1998 dim0=-2 dim1=-1 $input=1995 #1995=(36,6,64,32)f32 #1998=(36,6,32,64)f32 +torch.matmul torch.matmul_2320 2 1 1997 1998 1999 $input=1997 $other=1998 #1997=(36,6,64,32)f32 #1998=(36,6,32,64)f32 #1999=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_11748 0 1 2000 @pnnx_fold_11748=(1,6,64,64)f32 #2000=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_7795 2 1 1999 2000 2001 expr=add(@0,@1) #1999=(36,6,64,64)f32 #2000=(1,6,64,64)f32 #2001=(36,6,64,64)f32 +nn.Softmax pnnx_unique_401 1 1 2001 2002 dim=-1 #2001=(36,6,64,64)f32 #2002=(36,6,64,64)f32 +torch.matmul torch.matmul_2321 2 1 2002 1996 2003 $input=2002 $other=1996 #2002=(36,6,64,64)f32 #1996=(36,6,64,32)f32 #2003=(36,6,64,32)f32 +torch.transpose torch.transpose_3098 1 1 2003 2004 dim0=1 dim1=2 $input=2003 #2003=(36,6,64,32)f32 #2004=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_551 1 1 2004 2005 shape=(36,64,192) $input=2004 #2004=(36,64,6,32)f32 #2005=(36,64,192)f32 +nn.Linear pnnx_unique_403 1 1 2005 2006 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2005=(36,64,192)f32 #2006=(36,64,192)f32 +Tensor.reshape Tensor.view_1629 1 1 2006 2007 shape=(1,6,6,8,8,192) $input=2006 #2006=(36,64,192)f32 #2007=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2763 1 1 2007 2008 dims=(0,1,3,2,4,5) $input=2007 #2007=(1,6,6,8,8,192)f32 #2008=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1631 1 1 2008 2009 shape=(1,2304,192) $input=2008 #2008=(1,6,8,6,8,192)f32 #2009=(1,2304,192)f32 +pnnx.Expression pnnx_expr_7762 2 1 1986 2009 2010 expr=add(@0,@1) #1986=(1,2304,192)f32 #2009=(1,2304,192)f32 #2010=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_405 1 1 2010 2011 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2010=(1,2304,192)f32 #2011=(1,2304,192)f32 +nn.Linear pnnx_unique_406 1 1 2011 2012 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2011=(1,2304,192)f32 #2012=(1,2304,384)f32 +nn.GELU pnnx_unique_407 1 1 2012 2013 #2012=(1,2304,384)f32 #2013=(1,2304,384)f32 +nn.Linear pnnx_unique_409 1 1 2013 2014 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2013=(1,2304,384)f32 #2014=(1,2304,192)f32 +pnnx.Expression pnnx_expr_7760 2 1 2010 2014 2015 expr=add(@0,@1) #2010=(1,2304,192)f32 #2014=(1,2304,192)f32 #2015=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_412 1 1 2015 2016 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2015=(1,2304,192)f32 #2016=(1,2304,192)f32 +Tensor.view Tensor.view_1632 1 1 2016 2017 shape=(1,48,48,192) $input=2016 #2016=(1,2304,192)f32 #2017=(1,48,48,192)f32 +torch.roll torch.roll_2478 1 1 2017 2018 dims=(1,2) shifts=(-4,-4) $input=2017 #2017=(1,48,48,192)f32 #2018=(1,48,48,192)f32 +Tensor.view Tensor.view_1633 1 1 2018 2019 shape=(1,6,8,6,8,192) $input=2018 #2018=(1,48,48,192)f32 #2019=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2764 1 1 2019 2020 dims=(0,1,3,2,4,5) $input=2019 #2019=(1,6,8,6,8,192)f32 #2020=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1635 1 1 2020 2021 shape=(36,64,192) $input=2020 #2020=(1,6,6,8,8,192)f32 #2021=(36,64,192)f32 +nn.Linear pnnx_unique_415 1 1 2021 2022 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2021=(36,64,192)f32 #2022=(36,64,576)f32 +Tensor.reshape Tensor.reshape_552 1 1 2022 2023 shape=(36,64,3,6,32) $input=2022 #2022=(36,64,576)f32 #2023=(36,64,3,6,32)f32 +torch.permute torch.permute_2765 1 1 2023 2024 dims=(2,0,3,1,4) $input=2023 #2023=(36,64,3,6,32)f32 #2024=(3,36,6,64,32)f32 +torch.unbind Tensor.select_827 1 3 2024 2025 2026 2027 dim=0 #2024=(3,36,6,64,32)f32 #2025=(36,6,64,32)f32 #2026=(36,6,64,32)f32 #2027=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_7662 1 1 2025 2028 expr=mul(@0,1.767767e-01) #2025=(36,6,64,32)f32 #2028=(36,6,64,32)f32 +torch.transpose torch.transpose_3099 1 1 2026 2029 dim0=-2 dim1=-1 $input=2026 #2026=(36,6,64,32)f32 #2029=(36,6,32,64)f32 +torch.matmul torch.matmul_2322 2 1 2028 2029 2030 $input=2028 $other=2029 #2028=(36,6,64,32)f32 #2029=(36,6,32,64)f32 #2030=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_11901 0 1 2031 @pnnx_fold_11901=(1,6,64,64)f32 #2031=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_7649 2 1 2030 2031 2032 expr=add(@0,@1) #2030=(36,6,64,64)f32 #2031=(1,6,64,64)f32 #2032=(36,6,64,64)f32 +Tensor.view Tensor.view_1638 1 1 2032 2033 shape=(1,36,6,64,64) $input=2032 #2032=(36,6,64,64)f32 #2033=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_11911 0 1 2034 @pnnx_fold_11911=(1,36,1,64,64)f32 #2034=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_7637 2 1 2033 2034 2035 expr=add(@0,@1) #2033=(1,36,6,64,64)f32 #2034=(1,36,1,64,64)f32 #2035=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1639 1 1 2035 2036 shape=(-1,6,64,64) $input=2035 #2035=(1,36,6,64,64)f32 #2036=(36,6,64,64)f32 +nn.Softmax pnnx_unique_416 1 1 2036 2037 dim=-1 #2036=(36,6,64,64)f32 #2037=(36,6,64,64)f32 +torch.matmul torch.matmul_2323 2 1 2037 2027 2038 $input=2037 $other=2027 #2037=(36,6,64,64)f32 #2027=(36,6,64,32)f32 #2038=(36,6,64,32)f32 +torch.transpose torch.transpose_3100 1 1 2038 2039 dim0=1 dim1=2 $input=2038 #2038=(36,6,64,32)f32 #2039=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_553 1 1 2039 2040 shape=(36,64,192) $input=2039 #2039=(36,64,6,32)f32 #2040=(36,64,192)f32 +nn.Linear pnnx_unique_418 1 1 2040 2041 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2040=(36,64,192)f32 #2041=(36,64,192)f32 +Tensor.reshape Tensor.view_1641 1 1 2041 2042 shape=(1,6,6,8,8,192) $input=2041 #2041=(36,64,192)f32 #2042=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2767 1 1 2042 2043 dims=(0,1,3,2,4,5) $input=2042 #2042=(1,6,6,8,8,192)f32 #2043=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_119 1 1 2043 2044 shape=(1,48,48,-1) $input=2043 #2043=(1,6,8,6,8,192)f32 #2044=(1,48,48,192)f32 +torch.roll torch.roll_2479 1 1 2044 2045 dims=(1,2) shifts=(4,4) $input=2044 #2044=(1,48,48,192)f32 #2045=(1,48,48,192)f32 +Tensor.view Tensor.view_1643 1 1 2045 2046 shape=(1,2304,192) $input=2045 #2045=(1,48,48,192)f32 #2046=(1,2304,192)f32 +pnnx.Expression pnnx_expr_7595 2 1 2015 2046 2047 expr=add(@0,@1) #2015=(1,2304,192)f32 #2046=(1,2304,192)f32 #2047=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_420 1 1 2047 2048 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2047=(1,2304,192)f32 #2048=(1,2304,192)f32 +nn.Linear pnnx_unique_421 1 1 2048 2049 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2048=(1,2304,192)f32 #2049=(1,2304,384)f32 +nn.GELU pnnx_unique_422 1 1 2049 2050 #2049=(1,2304,384)f32 #2050=(1,2304,384)f32 +nn.Linear pnnx_unique_424 1 1 2050 2051 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2050=(1,2304,384)f32 #2051=(1,2304,192)f32 +pnnx.Expression pnnx_expr_7593 2 1 2047 2051 2052 expr=add(@0,@1) #2047=(1,2304,192)f32 #2051=(1,2304,192)f32 #2052=(1,2304,192)f32 +torch.transpose torch.transpose_3101 1 1 2052 2053 dim0=1 dim1=2 $input=2052 #2052=(1,2304,192)f32 #2053=(1,192,2304)f32 +Tensor.view Tensor.view_1644 1 1 2053 2054 shape=(1,192,48,48) $input=2053 #2053=(1,192,2304)f32 #2054=(1,192,48,48)f32 +nn.Conv2d pnnx_unique_426 1 1 2054 2055 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #2054=(1,192,48,48)f32 #2055=(1,192,48,48)f32 +torch.flatten torch.flatten_2194 1 1 2055 2056 end_dim=-1 start_dim=2 $input=2055 #2055=(1,192,48,48)f32 #2056=(1,192,2304)f32 +torch.transpose torch.transpose_3102 1 1 2056 2057 dim0=1 dim1=2 $input=2056 #2056=(1,192,2304)f32 #2057=(1,2304,192)f32 +pnnx.Expression pnnx_expr_7580 2 1 2057 1854 2058 expr=add(@0,@1) #2057=(1,2304,192)f32 #1854=(1,2304,192)f32 #2058=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_427 1 1 2058 2059 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2058=(1,2304,192)f32 #2059=(1,2304,192)f32 +Tensor.reshape Tensor.view_1646 1 1 2059 2060 shape=(1,6,8,6,8,192) $input=2059 #2059=(1,2304,192)f32 #2060=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2768 1 1 2060 2061 dims=(0,1,3,2,4,5) $input=2060 #2060=(1,6,8,6,8,192)f32 #2061=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1648 1 1 2061 2062 shape=(36,64,192) $input=2061 #2061=(1,6,6,8,8,192)f32 #2062=(36,64,192)f32 +nn.Linear pnnx_unique_430 1 1 2062 2063 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2062=(36,64,192)f32 #2063=(36,64,576)f32 +Tensor.reshape Tensor.reshape_554 1 1 2063 2064 shape=(36,64,3,6,32) $input=2063 #2063=(36,64,576)f32 #2064=(36,64,3,6,32)f32 +torch.permute torch.permute_2769 1 1 2064 2065 dims=(2,0,3,1,4) $input=2064 #2064=(36,64,3,6,32)f32 #2065=(3,36,6,64,32)f32 +torch.unbind Tensor.select_830 1 3 2065 2066 2067 2068 dim=0 #2065=(3,36,6,64,32)f32 #2066=(36,6,64,32)f32 #2067=(36,6,64,32)f32 #2068=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_7491 1 1 2066 2069 expr=mul(@0,1.767767e-01) #2066=(36,6,64,32)f32 #2069=(36,6,64,32)f32 +torch.transpose torch.transpose_3103 1 1 2067 2070 dim0=-2 dim1=-1 $input=2067 #2067=(36,6,64,32)f32 #2070=(36,6,32,64)f32 +torch.matmul torch.matmul_2324 2 1 2069 2070 2071 $input=2069 $other=2070 #2069=(36,6,64,32)f32 #2070=(36,6,32,64)f32 #2071=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_12094 0 1 2072 @pnnx_fold_12094=(1,6,64,64)f32 #2072=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_7478 2 1 2071 2072 2073 expr=add(@0,@1) #2071=(36,6,64,64)f32 #2072=(1,6,64,64)f32 #2073=(36,6,64,64)f32 +nn.Softmax pnnx_unique_431 1 1 2073 2074 dim=-1 #2073=(36,6,64,64)f32 #2074=(36,6,64,64)f32 +torch.matmul torch.matmul_2325 2 1 2074 2068 2075 $input=2074 $other=2068 #2074=(36,6,64,64)f32 #2068=(36,6,64,32)f32 #2075=(36,6,64,32)f32 +torch.transpose torch.transpose_3104 1 1 2075 2076 dim0=1 dim1=2 $input=2075 #2075=(36,6,64,32)f32 #2076=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_555 1 1 2076 2077 shape=(36,64,192) $input=2076 #2076=(36,64,6,32)f32 #2077=(36,64,192)f32 +nn.Linear pnnx_unique_433 1 1 2077 2078 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2077=(36,64,192)f32 #2078=(36,64,192)f32 +Tensor.reshape Tensor.view_1652 1 1 2078 2079 shape=(1,6,6,8,8,192) $input=2078 #2078=(36,64,192)f32 #2079=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2771 1 1 2079 2080 dims=(0,1,3,2,4,5) $input=2079 #2079=(1,6,6,8,8,192)f32 #2080=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1654 1 1 2080 2081 shape=(1,2304,192) $input=2080 #2080=(1,6,8,6,8,192)f32 #2081=(1,2304,192)f32 +pnnx.Expression pnnx_expr_7445 2 1 2058 2081 2082 expr=add(@0,@1) #2058=(1,2304,192)f32 #2081=(1,2304,192)f32 #2082=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_435 1 1 2082 2083 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2082=(1,2304,192)f32 #2083=(1,2304,192)f32 +nn.Linear pnnx_unique_436 1 1 2083 2084 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2083=(1,2304,192)f32 #2084=(1,2304,384)f32 +nn.GELU pnnx_unique_437 1 1 2084 2085 #2084=(1,2304,384)f32 #2085=(1,2304,384)f32 +nn.Linear pnnx_unique_439 1 1 2085 2086 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2085=(1,2304,384)f32 #2086=(1,2304,192)f32 +pnnx.Expression pnnx_expr_7443 2 1 2082 2086 2087 expr=add(@0,@1) #2082=(1,2304,192)f32 #2086=(1,2304,192)f32 #2087=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_442 1 1 2087 2088 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2087=(1,2304,192)f32 #2088=(1,2304,192)f32 +Tensor.view Tensor.view_1655 1 1 2088 2089 shape=(1,48,48,192) $input=2088 #2088=(1,2304,192)f32 #2089=(1,48,48,192)f32 +torch.roll torch.roll_2480 1 1 2089 2090 dims=(1,2) shifts=(-4,-4) $input=2089 #2089=(1,48,48,192)f32 #2090=(1,48,48,192)f32 +Tensor.view Tensor.view_1656 1 1 2090 2091 shape=(1,6,8,6,8,192) $input=2090 #2090=(1,48,48,192)f32 #2091=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2772 1 1 2091 2092 dims=(0,1,3,2,4,5) $input=2091 #2091=(1,6,8,6,8,192)f32 #2092=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1658 1 1 2092 2093 shape=(36,64,192) $input=2092 #2092=(1,6,6,8,8,192)f32 #2093=(36,64,192)f32 +nn.Linear pnnx_unique_445 1 1 2093 2094 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2093=(36,64,192)f32 #2094=(36,64,576)f32 +Tensor.reshape Tensor.reshape_556 1 1 2094 2095 shape=(36,64,3,6,32) $input=2094 #2094=(36,64,576)f32 #2095=(36,64,3,6,32)f32 +torch.permute torch.permute_2773 1 1 2095 2096 dims=(2,0,3,1,4) $input=2095 #2095=(36,64,3,6,32)f32 #2096=(3,36,6,64,32)f32 +torch.unbind Tensor.select_833 1 3 2096 2097 2098 2099 dim=0 #2096=(3,36,6,64,32)f32 #2097=(36,6,64,32)f32 #2098=(36,6,64,32)f32 #2099=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_7345 1 1 2097 2100 expr=mul(@0,1.767767e-01) #2097=(36,6,64,32)f32 #2100=(36,6,64,32)f32 +torch.transpose torch.transpose_3105 1 1 2098 2101 dim0=-2 dim1=-1 $input=2098 #2098=(36,6,64,32)f32 #2101=(36,6,32,64)f32 +torch.matmul torch.matmul_2326 2 1 2100 2101 2102 $input=2100 $other=2101 #2100=(36,6,64,32)f32 #2101=(36,6,32,64)f32 #2102=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_12247 0 1 2103 @pnnx_fold_12247=(1,6,64,64)f32 #2103=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_7332 2 1 2102 2103 2104 expr=add(@0,@1) #2102=(36,6,64,64)f32 #2103=(1,6,64,64)f32 #2104=(36,6,64,64)f32 +Tensor.view Tensor.view_1661 1 1 2104 2105 shape=(1,36,6,64,64) $input=2104 #2104=(36,6,64,64)f32 #2105=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_12257 0 1 2106 @pnnx_fold_12257=(1,36,1,64,64)f32 #2106=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_7320 2 1 2105 2106 2107 expr=add(@0,@1) #2105=(1,36,6,64,64)f32 #2106=(1,36,1,64,64)f32 #2107=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1662 1 1 2107 2108 shape=(-1,6,64,64) $input=2107 #2107=(1,36,6,64,64)f32 #2108=(36,6,64,64)f32 +nn.Softmax pnnx_unique_446 1 1 2108 2109 dim=-1 #2108=(36,6,64,64)f32 #2109=(36,6,64,64)f32 +torch.matmul torch.matmul_2327 2 1 2109 2099 2110 $input=2109 $other=2099 #2109=(36,6,64,64)f32 #2099=(36,6,64,32)f32 #2110=(36,6,64,32)f32 +torch.transpose torch.transpose_3106 1 1 2110 2111 dim0=1 dim1=2 $input=2110 #2110=(36,6,64,32)f32 #2111=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_557 1 1 2111 2112 shape=(36,64,192) $input=2111 #2111=(36,64,6,32)f32 #2112=(36,64,192)f32 +nn.Linear pnnx_unique_448 1 1 2112 2113 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2112=(36,64,192)f32 #2113=(36,64,192)f32 +Tensor.reshape Tensor.view_1664 1 1 2113 2114 shape=(1,6,6,8,8,192) $input=2113 #2113=(36,64,192)f32 #2114=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2775 1 1 2114 2115 dims=(0,1,3,2,4,5) $input=2114 #2114=(1,6,6,8,8,192)f32 #2115=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_123 1 1 2115 2116 shape=(1,48,48,-1) $input=2115 #2115=(1,6,8,6,8,192)f32 #2116=(1,48,48,192)f32 +torch.roll torch.roll_2481 1 1 2116 2117 dims=(1,2) shifts=(4,4) $input=2116 #2116=(1,48,48,192)f32 #2117=(1,48,48,192)f32 +Tensor.view Tensor.view_1666 1 1 2117 2118 shape=(1,2304,192) $input=2117 #2117=(1,48,48,192)f32 #2118=(1,2304,192)f32 +pnnx.Expression pnnx_expr_7278 2 1 2087 2118 2119 expr=add(@0,@1) #2087=(1,2304,192)f32 #2118=(1,2304,192)f32 #2119=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_450 1 1 2119 2120 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2119=(1,2304,192)f32 #2120=(1,2304,192)f32 +nn.Linear pnnx_unique_451 1 1 2120 2121 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2120=(1,2304,192)f32 #2121=(1,2304,384)f32 +nn.GELU pnnx_unique_452 1 1 2121 2122 #2121=(1,2304,384)f32 #2122=(1,2304,384)f32 +nn.Linear pnnx_unique_454 1 1 2122 2123 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2122=(1,2304,384)f32 #2123=(1,2304,192)f32 +pnnx.Expression pnnx_expr_7276 2 1 2119 2123 2124 expr=add(@0,@1) #2119=(1,2304,192)f32 #2123=(1,2304,192)f32 #2124=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_456 1 1 2124 2125 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2124=(1,2304,192)f32 #2125=(1,2304,192)f32 +Tensor.reshape Tensor.view_1668 1 1 2125 2126 shape=(1,6,8,6,8,192) $input=2125 #2125=(1,2304,192)f32 #2126=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2776 1 1 2126 2127 dims=(0,1,3,2,4,5) $input=2126 #2126=(1,6,8,6,8,192)f32 #2127=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1670 1 1 2127 2128 shape=(36,64,192) $input=2127 #2127=(1,6,6,8,8,192)f32 #2128=(36,64,192)f32 +nn.Linear pnnx_unique_459 1 1 2128 2129 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2128=(36,64,192)f32 #2129=(36,64,576)f32 +Tensor.reshape Tensor.reshape_558 1 1 2129 2130 shape=(36,64,3,6,32) $input=2129 #2129=(36,64,576)f32 #2130=(36,64,3,6,32)f32 +torch.permute torch.permute_2777 1 1 2130 2131 dims=(2,0,3,1,4) $input=2130 #2130=(36,64,3,6,32)f32 #2131=(3,36,6,64,32)f32 +torch.unbind Tensor.select_836 1 3 2131 2132 2133 2134 dim=0 #2131=(3,36,6,64,32)f32 #2132=(36,6,64,32)f32 #2133=(36,6,64,32)f32 #2134=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_7188 1 1 2132 2135 expr=mul(@0,1.767767e-01) #2132=(36,6,64,32)f32 #2135=(36,6,64,32)f32 +torch.transpose torch.transpose_3107 1 1 2133 2136 dim0=-2 dim1=-1 $input=2133 #2133=(36,6,64,32)f32 #2136=(36,6,32,64)f32 +torch.matmul torch.matmul_2328 2 1 2135 2136 2137 $input=2135 $other=2136 #2135=(36,6,64,32)f32 #2136=(36,6,32,64)f32 #2137=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_12406 0 1 2138 @pnnx_fold_12406=(1,6,64,64)f32 #2138=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_7175 2 1 2137 2138 2139 expr=add(@0,@1) #2137=(36,6,64,64)f32 #2138=(1,6,64,64)f32 #2139=(36,6,64,64)f32 +nn.Softmax pnnx_unique_460 1 1 2139 2140 dim=-1 #2139=(36,6,64,64)f32 #2140=(36,6,64,64)f32 +torch.matmul torch.matmul_2329 2 1 2140 2134 2141 $input=2140 $other=2134 #2140=(36,6,64,64)f32 #2134=(36,6,64,32)f32 #2141=(36,6,64,32)f32 +torch.transpose torch.transpose_3108 1 1 2141 2142 dim0=1 dim1=2 $input=2141 #2141=(36,6,64,32)f32 #2142=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_559 1 1 2142 2143 shape=(36,64,192) $input=2142 #2142=(36,64,6,32)f32 #2143=(36,64,192)f32 +nn.Linear pnnx_unique_462 1 1 2143 2144 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2143=(36,64,192)f32 #2144=(36,64,192)f32 +Tensor.reshape Tensor.view_1674 1 1 2144 2145 shape=(1,6,6,8,8,192) $input=2144 #2144=(36,64,192)f32 #2145=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2779 1 1 2145 2146 dims=(0,1,3,2,4,5) $input=2145 #2145=(1,6,6,8,8,192)f32 #2146=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1676 1 1 2146 2147 shape=(1,2304,192) $input=2146 #2146=(1,6,8,6,8,192)f32 #2147=(1,2304,192)f32 +pnnx.Expression pnnx_expr_7142 2 1 2124 2147 2148 expr=add(@0,@1) #2124=(1,2304,192)f32 #2147=(1,2304,192)f32 #2148=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_464 1 1 2148 2149 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2148=(1,2304,192)f32 #2149=(1,2304,192)f32 +nn.Linear pnnx_unique_465 1 1 2149 2150 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2149=(1,2304,192)f32 #2150=(1,2304,384)f32 +nn.GELU pnnx_unique_466 1 1 2150 2151 #2150=(1,2304,384)f32 #2151=(1,2304,384)f32 +nn.Linear pnnx_unique_468 1 1 2151 2152 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2151=(1,2304,384)f32 #2152=(1,2304,192)f32 +pnnx.Expression pnnx_expr_7140 2 1 2148 2152 2153 expr=add(@0,@1) #2148=(1,2304,192)f32 #2152=(1,2304,192)f32 #2153=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_471 1 1 2153 2154 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2153=(1,2304,192)f32 #2154=(1,2304,192)f32 +Tensor.view Tensor.view_1677 1 1 2154 2155 shape=(1,48,48,192) $input=2154 #2154=(1,2304,192)f32 #2155=(1,48,48,192)f32 +torch.roll torch.roll_2482 1 1 2155 2156 dims=(1,2) shifts=(-4,-4) $input=2155 #2155=(1,48,48,192)f32 #2156=(1,48,48,192)f32 +Tensor.view Tensor.view_1678 1 1 2156 2157 shape=(1,6,8,6,8,192) $input=2156 #2156=(1,48,48,192)f32 #2157=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2780 1 1 2157 2158 dims=(0,1,3,2,4,5) $input=2157 #2157=(1,6,8,6,8,192)f32 #2158=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1680 1 1 2158 2159 shape=(36,64,192) $input=2158 #2158=(1,6,6,8,8,192)f32 #2159=(36,64,192)f32 +nn.Linear pnnx_unique_474 1 1 2159 2160 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2159=(36,64,192)f32 #2160=(36,64,576)f32 +Tensor.reshape Tensor.reshape_560 1 1 2160 2161 shape=(36,64,3,6,32) $input=2160 #2160=(36,64,576)f32 #2161=(36,64,3,6,32)f32 +torch.permute torch.permute_2781 1 1 2161 2162 dims=(2,0,3,1,4) $input=2161 #2161=(36,64,3,6,32)f32 #2162=(3,36,6,64,32)f32 +torch.unbind Tensor.select_839 1 3 2162 2163 2164 2165 dim=0 #2162=(3,36,6,64,32)f32 #2163=(36,6,64,32)f32 #2164=(36,6,64,32)f32 #2165=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_7042 1 1 2163 2166 expr=mul(@0,1.767767e-01) #2163=(36,6,64,32)f32 #2166=(36,6,64,32)f32 +torch.transpose torch.transpose_3109 1 1 2164 2167 dim0=-2 dim1=-1 $input=2164 #2164=(36,6,64,32)f32 #2167=(36,6,32,64)f32 +torch.matmul torch.matmul_2330 2 1 2166 2167 2168 $input=2166 $other=2167 #2166=(36,6,64,32)f32 #2167=(36,6,32,64)f32 #2168=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_12559 0 1 2169 @pnnx_fold_12559=(1,6,64,64)f32 #2169=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_7029 2 1 2168 2169 2170 expr=add(@0,@1) #2168=(36,6,64,64)f32 #2169=(1,6,64,64)f32 #2170=(36,6,64,64)f32 +Tensor.view Tensor.view_1683 1 1 2170 2171 shape=(1,36,6,64,64) $input=2170 #2170=(36,6,64,64)f32 #2171=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_12569 0 1 2172 @pnnx_fold_12569=(1,36,1,64,64)f32 #2172=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_7017 2 1 2171 2172 2173 expr=add(@0,@1) #2171=(1,36,6,64,64)f32 #2172=(1,36,1,64,64)f32 #2173=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1684 1 1 2173 2174 shape=(-1,6,64,64) $input=2173 #2173=(1,36,6,64,64)f32 #2174=(36,6,64,64)f32 +nn.Softmax pnnx_unique_475 1 1 2174 2175 dim=-1 #2174=(36,6,64,64)f32 #2175=(36,6,64,64)f32 +torch.matmul torch.matmul_2331 2 1 2175 2165 2176 $input=2175 $other=2165 #2175=(36,6,64,64)f32 #2165=(36,6,64,32)f32 #2176=(36,6,64,32)f32 +torch.transpose torch.transpose_3110 1 1 2176 2177 dim0=1 dim1=2 $input=2176 #2176=(36,6,64,32)f32 #2177=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_561 1 1 2177 2178 shape=(36,64,192) $input=2177 #2177=(36,64,6,32)f32 #2178=(36,64,192)f32 +nn.Linear pnnx_unique_477 1 1 2178 2179 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2178=(36,64,192)f32 #2179=(36,64,192)f32 +Tensor.reshape Tensor.view_1686 1 1 2179 2180 shape=(1,6,6,8,8,192) $input=2179 #2179=(36,64,192)f32 #2180=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2783 1 1 2180 2181 dims=(0,1,3,2,4,5) $input=2180 #2180=(1,6,6,8,8,192)f32 #2181=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_127 1 1 2181 2182 shape=(1,48,48,-1) $input=2181 #2181=(1,6,8,6,8,192)f32 #2182=(1,48,48,192)f32 +torch.roll torch.roll_2483 1 1 2182 2183 dims=(1,2) shifts=(4,4) $input=2182 #2182=(1,48,48,192)f32 #2183=(1,48,48,192)f32 +Tensor.view Tensor.view_1688 1 1 2183 2184 shape=(1,2304,192) $input=2183 #2183=(1,48,48,192)f32 #2184=(1,2304,192)f32 +pnnx.Expression pnnx_expr_6975 2 1 2153 2184 2185 expr=add(@0,@1) #2153=(1,2304,192)f32 #2184=(1,2304,192)f32 #2185=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_479 1 1 2185 2186 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2185=(1,2304,192)f32 #2186=(1,2304,192)f32 +nn.Linear pnnx_unique_480 1 1 2186 2187 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2186=(1,2304,192)f32 #2187=(1,2304,384)f32 +nn.GELU pnnx_unique_481 1 1 2187 2188 #2187=(1,2304,384)f32 #2188=(1,2304,384)f32 +nn.Linear pnnx_unique_483 1 1 2188 2189 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2188=(1,2304,384)f32 #2189=(1,2304,192)f32 +pnnx.Expression pnnx_expr_6973 2 1 2185 2189 2190 expr=add(@0,@1) #2185=(1,2304,192)f32 #2189=(1,2304,192)f32 #2190=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_485 1 1 2190 2191 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2190=(1,2304,192)f32 #2191=(1,2304,192)f32 +Tensor.reshape Tensor.view_1690 1 1 2191 2192 shape=(1,6,8,6,8,192) $input=2191 #2191=(1,2304,192)f32 #2192=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2784 1 1 2192 2193 dims=(0,1,3,2,4,5) $input=2192 #2192=(1,6,8,6,8,192)f32 #2193=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1692 1 1 2193 2194 shape=(36,64,192) $input=2193 #2193=(1,6,6,8,8,192)f32 #2194=(36,64,192)f32 +nn.Linear pnnx_unique_488 1 1 2194 2195 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2194=(36,64,192)f32 #2195=(36,64,576)f32 +Tensor.reshape Tensor.reshape_562 1 1 2195 2196 shape=(36,64,3,6,32) $input=2195 #2195=(36,64,576)f32 #2196=(36,64,3,6,32)f32 +torch.permute torch.permute_2785 1 1 2196 2197 dims=(2,0,3,1,4) $input=2196 #2196=(36,64,3,6,32)f32 #2197=(3,36,6,64,32)f32 +torch.unbind Tensor.select_842 1 3 2197 2198 2199 2200 dim=0 #2197=(3,36,6,64,32)f32 #2198=(36,6,64,32)f32 #2199=(36,6,64,32)f32 #2200=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_6885 1 1 2198 2201 expr=mul(@0,1.767767e-01) #2198=(36,6,64,32)f32 #2201=(36,6,64,32)f32 +torch.transpose torch.transpose_3111 1 1 2199 2202 dim0=-2 dim1=-1 $input=2199 #2199=(36,6,64,32)f32 #2202=(36,6,32,64)f32 +torch.matmul torch.matmul_2332 2 1 2201 2202 2203 $input=2201 $other=2202 #2201=(36,6,64,32)f32 #2202=(36,6,32,64)f32 #2203=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_12718 0 1 2204 @pnnx_fold_12718=(1,6,64,64)f32 #2204=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_6872 2 1 2203 2204 2205 expr=add(@0,@1) #2203=(36,6,64,64)f32 #2204=(1,6,64,64)f32 #2205=(36,6,64,64)f32 +nn.Softmax pnnx_unique_489 1 1 2205 2206 dim=-1 #2205=(36,6,64,64)f32 #2206=(36,6,64,64)f32 +torch.matmul torch.matmul_2333 2 1 2206 2200 2207 $input=2206 $other=2200 #2206=(36,6,64,64)f32 #2200=(36,6,64,32)f32 #2207=(36,6,64,32)f32 +torch.transpose torch.transpose_3112 1 1 2207 2208 dim0=1 dim1=2 $input=2207 #2207=(36,6,64,32)f32 #2208=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_563 1 1 2208 2209 shape=(36,64,192) $input=2208 #2208=(36,64,6,32)f32 #2209=(36,64,192)f32 +nn.Linear pnnx_unique_491 1 1 2209 2210 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2209=(36,64,192)f32 #2210=(36,64,192)f32 +Tensor.reshape Tensor.view_1696 1 1 2210 2211 shape=(1,6,6,8,8,192) $input=2210 #2210=(36,64,192)f32 #2211=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2787 1 1 2211 2212 dims=(0,1,3,2,4,5) $input=2211 #2211=(1,6,6,8,8,192)f32 #2212=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1698 1 1 2212 2213 shape=(1,2304,192) $input=2212 #2212=(1,6,8,6,8,192)f32 #2213=(1,2304,192)f32 +pnnx.Expression pnnx_expr_6839 2 1 2190 2213 2214 expr=add(@0,@1) #2190=(1,2304,192)f32 #2213=(1,2304,192)f32 #2214=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_493 1 1 2214 2215 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2214=(1,2304,192)f32 #2215=(1,2304,192)f32 +nn.Linear pnnx_unique_494 1 1 2215 2216 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2215=(1,2304,192)f32 #2216=(1,2304,384)f32 +nn.GELU pnnx_unique_495 1 1 2216 2217 #2216=(1,2304,384)f32 #2217=(1,2304,384)f32 +nn.Linear pnnx_unique_497 1 1 2217 2218 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2217=(1,2304,384)f32 #2218=(1,2304,192)f32 +pnnx.Expression pnnx_expr_6837 2 1 2214 2218 2219 expr=add(@0,@1) #2214=(1,2304,192)f32 #2218=(1,2304,192)f32 #2219=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_500 1 1 2219 2220 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2219=(1,2304,192)f32 #2220=(1,2304,192)f32 +Tensor.view Tensor.view_1699 1 1 2220 2221 shape=(1,48,48,192) $input=2220 #2220=(1,2304,192)f32 #2221=(1,48,48,192)f32 +torch.roll torch.roll_2484 1 1 2221 2222 dims=(1,2) shifts=(-4,-4) $input=2221 #2221=(1,48,48,192)f32 #2222=(1,48,48,192)f32 +Tensor.view Tensor.view_1700 1 1 2222 2223 shape=(1,6,8,6,8,192) $input=2222 #2222=(1,48,48,192)f32 #2223=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2788 1 1 2223 2224 dims=(0,1,3,2,4,5) $input=2223 #2223=(1,6,8,6,8,192)f32 #2224=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1702 1 1 2224 2225 shape=(36,64,192) $input=2224 #2224=(1,6,6,8,8,192)f32 #2225=(36,64,192)f32 +nn.Linear pnnx_unique_503 1 1 2225 2226 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2225=(36,64,192)f32 #2226=(36,64,576)f32 +Tensor.reshape Tensor.reshape_564 1 1 2226 2227 shape=(36,64,3,6,32) $input=2226 #2226=(36,64,576)f32 #2227=(36,64,3,6,32)f32 +torch.permute torch.permute_2789 1 1 2227 2228 dims=(2,0,3,1,4) $input=2227 #2227=(36,64,3,6,32)f32 #2228=(3,36,6,64,32)f32 +torch.unbind Tensor.select_845 1 3 2228 2229 2230 2231 dim=0 #2228=(3,36,6,64,32)f32 #2229=(36,6,64,32)f32 #2230=(36,6,64,32)f32 #2231=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_6739 1 1 2229 2232 expr=mul(@0,1.767767e-01) #2229=(36,6,64,32)f32 #2232=(36,6,64,32)f32 +torch.transpose torch.transpose_3113 1 1 2230 2233 dim0=-2 dim1=-1 $input=2230 #2230=(36,6,64,32)f32 #2233=(36,6,32,64)f32 +torch.matmul torch.matmul_2334 2 1 2232 2233 2234 $input=2232 $other=2233 #2232=(36,6,64,32)f32 #2233=(36,6,32,64)f32 #2234=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_12871 0 1 2235 @pnnx_fold_12871=(1,6,64,64)f32 #2235=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_6726 2 1 2234 2235 2236 expr=add(@0,@1) #2234=(36,6,64,64)f32 #2235=(1,6,64,64)f32 #2236=(36,6,64,64)f32 +Tensor.view Tensor.view_1705 1 1 2236 2237 shape=(1,36,6,64,64) $input=2236 #2236=(36,6,64,64)f32 #2237=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_12881 0 1 2238 @pnnx_fold_12881=(1,36,1,64,64)f32 #2238=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_6714 2 1 2237 2238 2239 expr=add(@0,@1) #2237=(1,36,6,64,64)f32 #2238=(1,36,1,64,64)f32 #2239=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1706 1 1 2239 2240 shape=(-1,6,64,64) $input=2239 #2239=(1,36,6,64,64)f32 #2240=(36,6,64,64)f32 +nn.Softmax pnnx_unique_504 1 1 2240 2241 dim=-1 #2240=(36,6,64,64)f32 #2241=(36,6,64,64)f32 +torch.matmul torch.matmul_2335 2 1 2241 2231 2242 $input=2241 $other=2231 #2241=(36,6,64,64)f32 #2231=(36,6,64,32)f32 #2242=(36,6,64,32)f32 +torch.transpose torch.transpose_3114 1 1 2242 2243 dim0=1 dim1=2 $input=2242 #2242=(36,6,64,32)f32 #2243=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_565 1 1 2243 2244 shape=(36,64,192) $input=2243 #2243=(36,64,6,32)f32 #2244=(36,64,192)f32 +nn.Linear pnnx_unique_506 1 1 2244 2245 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2244=(36,64,192)f32 #2245=(36,64,192)f32 +Tensor.reshape Tensor.view_1708 1 1 2245 2246 shape=(1,6,6,8,8,192) $input=2245 #2245=(36,64,192)f32 #2246=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2791 1 1 2246 2247 dims=(0,1,3,2,4,5) $input=2246 #2246=(1,6,6,8,8,192)f32 #2247=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_131 1 1 2247 2248 shape=(1,48,48,-1) $input=2247 #2247=(1,6,8,6,8,192)f32 #2248=(1,48,48,192)f32 +torch.roll torch.roll_2485 1 1 2248 2249 dims=(1,2) shifts=(4,4) $input=2248 #2248=(1,48,48,192)f32 #2249=(1,48,48,192)f32 +Tensor.view Tensor.view_1710 1 1 2249 2250 shape=(1,2304,192) $input=2249 #2249=(1,48,48,192)f32 #2250=(1,2304,192)f32 +pnnx.Expression pnnx_expr_6672 2 1 2219 2250 2251 expr=add(@0,@1) #2219=(1,2304,192)f32 #2250=(1,2304,192)f32 #2251=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_508 1 1 2251 2252 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2251=(1,2304,192)f32 #2252=(1,2304,192)f32 +nn.Linear pnnx_unique_509 1 1 2252 2253 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2252=(1,2304,192)f32 #2253=(1,2304,384)f32 +nn.GELU pnnx_unique_510 1 1 2253 2254 #2253=(1,2304,384)f32 #2254=(1,2304,384)f32 +nn.Linear pnnx_unique_512 1 1 2254 2255 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2254=(1,2304,384)f32 #2255=(1,2304,192)f32 +pnnx.Expression pnnx_expr_6670 2 1 2251 2255 2256 expr=add(@0,@1) #2251=(1,2304,192)f32 #2255=(1,2304,192)f32 #2256=(1,2304,192)f32 +torch.transpose torch.transpose_3115 1 1 2256 2257 dim0=1 dim1=2 $input=2256 #2256=(1,2304,192)f32 #2257=(1,192,2304)f32 +Tensor.view Tensor.view_1711 1 1 2257 2258 shape=(1,192,48,48) $input=2257 #2257=(1,192,2304)f32 #2258=(1,192,48,48)f32 +nn.Conv2d pnnx_unique_514 1 1 2258 2259 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #2258=(1,192,48,48)f32 #2259=(1,192,48,48)f32 +torch.flatten torch.flatten_2195 1 1 2259 2260 end_dim=-1 start_dim=2 $input=2259 #2259=(1,192,48,48)f32 #2260=(1,192,2304)f32 +torch.transpose torch.transpose_3116 1 1 2260 2261 dim0=1 dim1=2 $input=2260 #2260=(1,192,2304)f32 #2261=(1,2304,192)f32 +pnnx.Expression pnnx_expr_6657 2 1 2261 2058 2262 expr=add(@0,@1) #2261=(1,2304,192)f32 #2058=(1,2304,192)f32 #2262=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_515 1 1 2262 2263 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2262=(1,2304,192)f32 #2263=(1,2304,192)f32 +Tensor.reshape Tensor.view_1713 1 1 2263 2264 shape=(1,6,8,6,8,192) $input=2263 #2263=(1,2304,192)f32 #2264=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2792 1 1 2264 2265 dims=(0,1,3,2,4,5) $input=2264 #2264=(1,6,8,6,8,192)f32 #2265=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1715 1 1 2265 2266 shape=(36,64,192) $input=2265 #2265=(1,6,6,8,8,192)f32 #2266=(36,64,192)f32 +nn.Linear pnnx_unique_518 1 1 2266 2267 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2266=(36,64,192)f32 #2267=(36,64,576)f32 +Tensor.reshape Tensor.reshape_566 1 1 2267 2268 shape=(36,64,3,6,32) $input=2267 #2267=(36,64,576)f32 #2268=(36,64,3,6,32)f32 +torch.permute torch.permute_2793 1 1 2268 2269 dims=(2,0,3,1,4) $input=2268 #2268=(36,64,3,6,32)f32 #2269=(3,36,6,64,32)f32 +torch.unbind Tensor.select_848 1 3 2269 2270 2271 2272 dim=0 #2269=(3,36,6,64,32)f32 #2270=(36,6,64,32)f32 #2271=(36,6,64,32)f32 #2272=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_6568 1 1 2270 2273 expr=mul(@0,1.767767e-01) #2270=(36,6,64,32)f32 #2273=(36,6,64,32)f32 +torch.transpose torch.transpose_3117 1 1 2271 2274 dim0=-2 dim1=-1 $input=2271 #2271=(36,6,64,32)f32 #2274=(36,6,32,64)f32 +torch.matmul torch.matmul_2336 2 1 2273 2274 2275 $input=2273 $other=2274 #2273=(36,6,64,32)f32 #2274=(36,6,32,64)f32 #2275=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_13064 0 1 2276 @pnnx_fold_13064=(1,6,64,64)f32 #2276=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_6555 2 1 2275 2276 2277 expr=add(@0,@1) #2275=(36,6,64,64)f32 #2276=(1,6,64,64)f32 #2277=(36,6,64,64)f32 +nn.Softmax pnnx_unique_519 1 1 2277 2278 dim=-1 #2277=(36,6,64,64)f32 #2278=(36,6,64,64)f32 +torch.matmul torch.matmul_2337 2 1 2278 2272 2279 $input=2278 $other=2272 #2278=(36,6,64,64)f32 #2272=(36,6,64,32)f32 #2279=(36,6,64,32)f32 +torch.transpose torch.transpose_3118 1 1 2279 2280 dim0=1 dim1=2 $input=2279 #2279=(36,6,64,32)f32 #2280=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_567 1 1 2280 2281 shape=(36,64,192) $input=2280 #2280=(36,64,6,32)f32 #2281=(36,64,192)f32 +nn.Linear pnnx_unique_521 1 1 2281 2282 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2281=(36,64,192)f32 #2282=(36,64,192)f32 +Tensor.reshape Tensor.view_1719 1 1 2282 2283 shape=(1,6,6,8,8,192) $input=2282 #2282=(36,64,192)f32 #2283=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2795 1 1 2283 2284 dims=(0,1,3,2,4,5) $input=2283 #2283=(1,6,6,8,8,192)f32 #2284=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1721 1 1 2284 2285 shape=(1,2304,192) $input=2284 #2284=(1,6,8,6,8,192)f32 #2285=(1,2304,192)f32 +pnnx.Expression pnnx_expr_6522 2 1 2262 2285 2286 expr=add(@0,@1) #2262=(1,2304,192)f32 #2285=(1,2304,192)f32 #2286=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_523 1 1 2286 2287 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2286=(1,2304,192)f32 #2287=(1,2304,192)f32 +nn.Linear pnnx_unique_524 1 1 2287 2288 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2287=(1,2304,192)f32 #2288=(1,2304,384)f32 +nn.GELU pnnx_unique_525 1 1 2288 2289 #2288=(1,2304,384)f32 #2289=(1,2304,384)f32 +nn.Linear pnnx_unique_527 1 1 2289 2290 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2289=(1,2304,384)f32 #2290=(1,2304,192)f32 +pnnx.Expression pnnx_expr_6520 2 1 2286 2290 2291 expr=add(@0,@1) #2286=(1,2304,192)f32 #2290=(1,2304,192)f32 #2291=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_530 1 1 2291 2292 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2291=(1,2304,192)f32 #2292=(1,2304,192)f32 +Tensor.view Tensor.view_1722 1 1 2292 2293 shape=(1,48,48,192) $input=2292 #2292=(1,2304,192)f32 #2293=(1,48,48,192)f32 +torch.roll torch.roll_2486 1 1 2293 2294 dims=(1,2) shifts=(-4,-4) $input=2293 #2293=(1,48,48,192)f32 #2294=(1,48,48,192)f32 +Tensor.view Tensor.view_1723 1 1 2294 2295 shape=(1,6,8,6,8,192) $input=2294 #2294=(1,48,48,192)f32 #2295=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2796 1 1 2295 2296 dims=(0,1,3,2,4,5) $input=2295 #2295=(1,6,8,6,8,192)f32 #2296=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1725 1 1 2296 2297 shape=(36,64,192) $input=2296 #2296=(1,6,6,8,8,192)f32 #2297=(36,64,192)f32 +nn.Linear pnnx_unique_533 1 1 2297 2298 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2297=(36,64,192)f32 #2298=(36,64,576)f32 +Tensor.reshape Tensor.reshape_568 1 1 2298 2299 shape=(36,64,3,6,32) $input=2298 #2298=(36,64,576)f32 #2299=(36,64,3,6,32)f32 +torch.permute torch.permute_2797 1 1 2299 2300 dims=(2,0,3,1,4) $input=2299 #2299=(36,64,3,6,32)f32 #2300=(3,36,6,64,32)f32 +torch.unbind Tensor.select_851 1 3 2300 2301 2302 2303 dim=0 #2300=(3,36,6,64,32)f32 #2301=(36,6,64,32)f32 #2302=(36,6,64,32)f32 #2303=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_6422 1 1 2301 2304 expr=mul(@0,1.767767e-01) #2301=(36,6,64,32)f32 #2304=(36,6,64,32)f32 +torch.transpose torch.transpose_3119 1 1 2302 2305 dim0=-2 dim1=-1 $input=2302 #2302=(36,6,64,32)f32 #2305=(36,6,32,64)f32 +torch.matmul torch.matmul_2338 2 1 2304 2305 2306 $input=2304 $other=2305 #2304=(36,6,64,32)f32 #2305=(36,6,32,64)f32 #2306=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_13217 0 1 2307 @pnnx_fold_13217=(1,6,64,64)f32 #2307=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_6409 2 1 2306 2307 2308 expr=add(@0,@1) #2306=(36,6,64,64)f32 #2307=(1,6,64,64)f32 #2308=(36,6,64,64)f32 +Tensor.view Tensor.view_1728 1 1 2308 2309 shape=(1,36,6,64,64) $input=2308 #2308=(36,6,64,64)f32 #2309=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_13227 0 1 2310 @pnnx_fold_13227=(1,36,1,64,64)f32 #2310=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_6397 2 1 2309 2310 2311 expr=add(@0,@1) #2309=(1,36,6,64,64)f32 #2310=(1,36,1,64,64)f32 #2311=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1729 1 1 2311 2312 shape=(-1,6,64,64) $input=2311 #2311=(1,36,6,64,64)f32 #2312=(36,6,64,64)f32 +nn.Softmax pnnx_unique_534 1 1 2312 2313 dim=-1 #2312=(36,6,64,64)f32 #2313=(36,6,64,64)f32 +torch.matmul torch.matmul_2339 2 1 2313 2303 2314 $input=2313 $other=2303 #2313=(36,6,64,64)f32 #2303=(36,6,64,32)f32 #2314=(36,6,64,32)f32 +torch.transpose torch.transpose_3120 1 1 2314 2315 dim0=1 dim1=2 $input=2314 #2314=(36,6,64,32)f32 #2315=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_569 1 1 2315 2316 shape=(36,64,192) $input=2315 #2315=(36,64,6,32)f32 #2316=(36,64,192)f32 +nn.Linear pnnx_unique_536 1 1 2316 2317 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2316=(36,64,192)f32 #2317=(36,64,192)f32 +Tensor.reshape Tensor.view_1731 1 1 2317 2318 shape=(1,6,6,8,8,192) $input=2317 #2317=(36,64,192)f32 #2318=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2799 1 1 2318 2319 dims=(0,1,3,2,4,5) $input=2318 #2318=(1,6,6,8,8,192)f32 #2319=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_135 1 1 2319 2320 shape=(1,48,48,-1) $input=2319 #2319=(1,6,8,6,8,192)f32 #2320=(1,48,48,192)f32 +torch.roll torch.roll_2487 1 1 2320 2321 dims=(1,2) shifts=(4,4) $input=2320 #2320=(1,48,48,192)f32 #2321=(1,48,48,192)f32 +Tensor.view Tensor.view_1733 1 1 2321 2322 shape=(1,2304,192) $input=2321 #2321=(1,48,48,192)f32 #2322=(1,2304,192)f32 +pnnx.Expression pnnx_expr_6355 2 1 2291 2322 2323 expr=add(@0,@1) #2291=(1,2304,192)f32 #2322=(1,2304,192)f32 #2323=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_538 1 1 2323 2324 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2323=(1,2304,192)f32 #2324=(1,2304,192)f32 +nn.Linear pnnx_unique_539 1 1 2324 2325 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2324=(1,2304,192)f32 #2325=(1,2304,384)f32 +nn.GELU pnnx_unique_540 1 1 2325 2326 #2325=(1,2304,384)f32 #2326=(1,2304,384)f32 +nn.Linear pnnx_unique_542 1 1 2326 2327 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2326=(1,2304,384)f32 #2327=(1,2304,192)f32 +pnnx.Expression pnnx_expr_6353 2 1 2323 2327 2328 expr=add(@0,@1) #2323=(1,2304,192)f32 #2327=(1,2304,192)f32 #2328=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_544 1 1 2328 2329 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2328=(1,2304,192)f32 #2329=(1,2304,192)f32 +Tensor.reshape Tensor.view_1735 1 1 2329 2330 shape=(1,6,8,6,8,192) $input=2329 #2329=(1,2304,192)f32 #2330=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2800 1 1 2330 2331 dims=(0,1,3,2,4,5) $input=2330 #2330=(1,6,8,6,8,192)f32 #2331=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1737 1 1 2331 2332 shape=(36,64,192) $input=2331 #2331=(1,6,6,8,8,192)f32 #2332=(36,64,192)f32 +nn.Linear pnnx_unique_547 1 1 2332 2333 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2332=(36,64,192)f32 #2333=(36,64,576)f32 +Tensor.reshape Tensor.reshape_570 1 1 2333 2334 shape=(36,64,3,6,32) $input=2333 #2333=(36,64,576)f32 #2334=(36,64,3,6,32)f32 +torch.permute torch.permute_2801 1 1 2334 2335 dims=(2,0,3,1,4) $input=2334 #2334=(36,64,3,6,32)f32 #2335=(3,36,6,64,32)f32 +torch.unbind Tensor.select_854 1 3 2335 2336 2337 2338 dim=0 #2335=(3,36,6,64,32)f32 #2336=(36,6,64,32)f32 #2337=(36,6,64,32)f32 #2338=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_6265 1 1 2336 2339 expr=mul(@0,1.767767e-01) #2336=(36,6,64,32)f32 #2339=(36,6,64,32)f32 +torch.transpose torch.transpose_3121 1 1 2337 2340 dim0=-2 dim1=-1 $input=2337 #2337=(36,6,64,32)f32 #2340=(36,6,32,64)f32 +torch.matmul torch.matmul_2340 2 1 2339 2340 2341 $input=2339 $other=2340 #2339=(36,6,64,32)f32 #2340=(36,6,32,64)f32 #2341=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_13376 0 1 2342 @pnnx_fold_13376=(1,6,64,64)f32 #2342=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_6252 2 1 2341 2342 2343 expr=add(@0,@1) #2341=(36,6,64,64)f32 #2342=(1,6,64,64)f32 #2343=(36,6,64,64)f32 +nn.Softmax pnnx_unique_548 1 1 2343 2344 dim=-1 #2343=(36,6,64,64)f32 #2344=(36,6,64,64)f32 +torch.matmul torch.matmul_2341 2 1 2344 2338 2345 $input=2344 $other=2338 #2344=(36,6,64,64)f32 #2338=(36,6,64,32)f32 #2345=(36,6,64,32)f32 +torch.transpose torch.transpose_3122 1 1 2345 2346 dim0=1 dim1=2 $input=2345 #2345=(36,6,64,32)f32 #2346=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_571 1 1 2346 2347 shape=(36,64,192) $input=2346 #2346=(36,64,6,32)f32 #2347=(36,64,192)f32 +nn.Linear pnnx_unique_550 1 1 2347 2348 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2347=(36,64,192)f32 #2348=(36,64,192)f32 +Tensor.reshape Tensor.view_1741 1 1 2348 2349 shape=(1,6,6,8,8,192) $input=2348 #2348=(36,64,192)f32 #2349=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2803 1 1 2349 2350 dims=(0,1,3,2,4,5) $input=2349 #2349=(1,6,6,8,8,192)f32 #2350=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1743 1 1 2350 2351 shape=(1,2304,192) $input=2350 #2350=(1,6,8,6,8,192)f32 #2351=(1,2304,192)f32 +pnnx.Expression pnnx_expr_6219 2 1 2328 2351 2352 expr=add(@0,@1) #2328=(1,2304,192)f32 #2351=(1,2304,192)f32 #2352=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_552 1 1 2352 2353 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2352=(1,2304,192)f32 #2353=(1,2304,192)f32 +nn.Linear pnnx_unique_553 1 1 2353 2354 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2353=(1,2304,192)f32 #2354=(1,2304,384)f32 +nn.GELU pnnx_unique_554 1 1 2354 2355 #2354=(1,2304,384)f32 #2355=(1,2304,384)f32 +nn.Linear pnnx_unique_556 1 1 2355 2356 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2355=(1,2304,384)f32 #2356=(1,2304,192)f32 +pnnx.Expression pnnx_expr_6217 2 1 2352 2356 2357 expr=add(@0,@1) #2352=(1,2304,192)f32 #2356=(1,2304,192)f32 #2357=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_559 1 1 2357 2358 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2357=(1,2304,192)f32 #2358=(1,2304,192)f32 +Tensor.view Tensor.view_1744 1 1 2358 2359 shape=(1,48,48,192) $input=2358 #2358=(1,2304,192)f32 #2359=(1,48,48,192)f32 +torch.roll torch.roll_2488 1 1 2359 2360 dims=(1,2) shifts=(-4,-4) $input=2359 #2359=(1,48,48,192)f32 #2360=(1,48,48,192)f32 +Tensor.view Tensor.view_1745 1 1 2360 2361 shape=(1,6,8,6,8,192) $input=2360 #2360=(1,48,48,192)f32 #2361=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2804 1 1 2361 2362 dims=(0,1,3,2,4,5) $input=2361 #2361=(1,6,8,6,8,192)f32 #2362=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1747 1 1 2362 2363 shape=(36,64,192) $input=2362 #2362=(1,6,6,8,8,192)f32 #2363=(36,64,192)f32 +nn.Linear pnnx_unique_562 1 1 2363 2364 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2363=(36,64,192)f32 #2364=(36,64,576)f32 +Tensor.reshape Tensor.reshape_572 1 1 2364 2365 shape=(36,64,3,6,32) $input=2364 #2364=(36,64,576)f32 #2365=(36,64,3,6,32)f32 +torch.permute torch.permute_2805 1 1 2365 2366 dims=(2,0,3,1,4) $input=2365 #2365=(36,64,3,6,32)f32 #2366=(3,36,6,64,32)f32 +torch.unbind Tensor.select_857 1 3 2366 2367 2368 2369 dim=0 #2366=(3,36,6,64,32)f32 #2367=(36,6,64,32)f32 #2368=(36,6,64,32)f32 #2369=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_6119 1 1 2367 2370 expr=mul(@0,1.767767e-01) #2367=(36,6,64,32)f32 #2370=(36,6,64,32)f32 +torch.transpose torch.transpose_3123 1 1 2368 2371 dim0=-2 dim1=-1 $input=2368 #2368=(36,6,64,32)f32 #2371=(36,6,32,64)f32 +torch.matmul torch.matmul_2342 2 1 2370 2371 2372 $input=2370 $other=2371 #2370=(36,6,64,32)f32 #2371=(36,6,32,64)f32 #2372=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_13529 0 1 2373 @pnnx_fold_13529=(1,6,64,64)f32 #2373=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_6106 2 1 2372 2373 2374 expr=add(@0,@1) #2372=(36,6,64,64)f32 #2373=(1,6,64,64)f32 #2374=(36,6,64,64)f32 +Tensor.view Tensor.view_1750 1 1 2374 2375 shape=(1,36,6,64,64) $input=2374 #2374=(36,6,64,64)f32 #2375=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_13539 0 1 2376 @pnnx_fold_13539=(1,36,1,64,64)f32 #2376=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_6094 2 1 2375 2376 2377 expr=add(@0,@1) #2375=(1,36,6,64,64)f32 #2376=(1,36,1,64,64)f32 #2377=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1751 1 1 2377 2378 shape=(-1,6,64,64) $input=2377 #2377=(1,36,6,64,64)f32 #2378=(36,6,64,64)f32 +nn.Softmax pnnx_unique_563 1 1 2378 2379 dim=-1 #2378=(36,6,64,64)f32 #2379=(36,6,64,64)f32 +torch.matmul torch.matmul_2343 2 1 2379 2369 2380 $input=2379 $other=2369 #2379=(36,6,64,64)f32 #2369=(36,6,64,32)f32 #2380=(36,6,64,32)f32 +torch.transpose torch.transpose_3124 1 1 2380 2381 dim0=1 dim1=2 $input=2380 #2380=(36,6,64,32)f32 #2381=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_573 1 1 2381 2382 shape=(36,64,192) $input=2381 #2381=(36,64,6,32)f32 #2382=(36,64,192)f32 +nn.Linear pnnx_unique_565 1 1 2382 2383 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2382=(36,64,192)f32 #2383=(36,64,192)f32 +Tensor.reshape Tensor.view_1753 1 1 2383 2384 shape=(1,6,6,8,8,192) $input=2383 #2383=(36,64,192)f32 #2384=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2807 1 1 2384 2385 dims=(0,1,3,2,4,5) $input=2384 #2384=(1,6,6,8,8,192)f32 #2385=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_139 1 1 2385 2386 shape=(1,48,48,-1) $input=2385 #2385=(1,6,8,6,8,192)f32 #2386=(1,48,48,192)f32 +torch.roll torch.roll_2489 1 1 2386 2387 dims=(1,2) shifts=(4,4) $input=2386 #2386=(1,48,48,192)f32 #2387=(1,48,48,192)f32 +Tensor.view Tensor.view_1755 1 1 2387 2388 shape=(1,2304,192) $input=2387 #2387=(1,48,48,192)f32 #2388=(1,2304,192)f32 +pnnx.Expression pnnx_expr_6052 2 1 2357 2388 2389 expr=add(@0,@1) #2357=(1,2304,192)f32 #2388=(1,2304,192)f32 #2389=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_567 1 1 2389 2390 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2389=(1,2304,192)f32 #2390=(1,2304,192)f32 +nn.Linear pnnx_unique_568 1 1 2390 2391 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2390=(1,2304,192)f32 #2391=(1,2304,384)f32 +nn.GELU pnnx_unique_569 1 1 2391 2392 #2391=(1,2304,384)f32 #2392=(1,2304,384)f32 +nn.Linear pnnx_unique_571 1 1 2392 2393 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2392=(1,2304,384)f32 #2393=(1,2304,192)f32 +pnnx.Expression pnnx_expr_6050 2 1 2389 2393 2394 expr=add(@0,@1) #2389=(1,2304,192)f32 #2393=(1,2304,192)f32 #2394=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_573 1 1 2394 2395 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2394=(1,2304,192)f32 #2395=(1,2304,192)f32 +Tensor.reshape Tensor.view_1757 1 1 2395 2396 shape=(1,6,8,6,8,192) $input=2395 #2395=(1,2304,192)f32 #2396=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2808 1 1 2396 2397 dims=(0,1,3,2,4,5) $input=2396 #2396=(1,6,8,6,8,192)f32 #2397=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1759 1 1 2397 2398 shape=(36,64,192) $input=2397 #2397=(1,6,6,8,8,192)f32 #2398=(36,64,192)f32 +nn.Linear pnnx_unique_576 1 1 2398 2399 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2398=(36,64,192)f32 #2399=(36,64,576)f32 +Tensor.reshape Tensor.reshape_574 1 1 2399 2400 shape=(36,64,3,6,32) $input=2399 #2399=(36,64,576)f32 #2400=(36,64,3,6,32)f32 +torch.permute torch.permute_2809 1 1 2400 2401 dims=(2,0,3,1,4) $input=2400 #2400=(36,64,3,6,32)f32 #2401=(3,36,6,64,32)f32 +torch.unbind Tensor.select_860 1 3 2401 2402 2403 2404 dim=0 #2401=(3,36,6,64,32)f32 #2402=(36,6,64,32)f32 #2403=(36,6,64,32)f32 #2404=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_5962 1 1 2402 2405 expr=mul(@0,1.767767e-01) #2402=(36,6,64,32)f32 #2405=(36,6,64,32)f32 +torch.transpose torch.transpose_3125 1 1 2403 2406 dim0=-2 dim1=-1 $input=2403 #2403=(36,6,64,32)f32 #2406=(36,6,32,64)f32 +torch.matmul torch.matmul_2344 2 1 2405 2406 2407 $input=2405 $other=2406 #2405=(36,6,64,32)f32 #2406=(36,6,32,64)f32 #2407=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_13688 0 1 2408 @pnnx_fold_13688=(1,6,64,64)f32 #2408=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_5949 2 1 2407 2408 2409 expr=add(@0,@1) #2407=(36,6,64,64)f32 #2408=(1,6,64,64)f32 #2409=(36,6,64,64)f32 +nn.Softmax pnnx_unique_577 1 1 2409 2410 dim=-1 #2409=(36,6,64,64)f32 #2410=(36,6,64,64)f32 +torch.matmul torch.matmul_2345 2 1 2410 2404 2411 $input=2410 $other=2404 #2410=(36,6,64,64)f32 #2404=(36,6,64,32)f32 #2411=(36,6,64,32)f32 +torch.transpose torch.transpose_3126 1 1 2411 2412 dim0=1 dim1=2 $input=2411 #2411=(36,6,64,32)f32 #2412=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_575 1 1 2412 2413 shape=(36,64,192) $input=2412 #2412=(36,64,6,32)f32 #2413=(36,64,192)f32 +nn.Linear pnnx_unique_579 1 1 2413 2414 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2413=(36,64,192)f32 #2414=(36,64,192)f32 +Tensor.reshape Tensor.view_1763 1 1 2414 2415 shape=(1,6,6,8,8,192) $input=2414 #2414=(36,64,192)f32 #2415=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2811 1 1 2415 2416 dims=(0,1,3,2,4,5) $input=2415 #2415=(1,6,6,8,8,192)f32 #2416=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1765 1 1 2416 2417 shape=(1,2304,192) $input=2416 #2416=(1,6,8,6,8,192)f32 #2417=(1,2304,192)f32 +pnnx.Expression pnnx_expr_5916 2 1 2394 2417 2418 expr=add(@0,@1) #2394=(1,2304,192)f32 #2417=(1,2304,192)f32 #2418=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_581 1 1 2418 2419 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2418=(1,2304,192)f32 #2419=(1,2304,192)f32 +nn.Linear pnnx_unique_582 1 1 2419 2420 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2419=(1,2304,192)f32 #2420=(1,2304,384)f32 +nn.GELU pnnx_unique_583 1 1 2420 2421 #2420=(1,2304,384)f32 #2421=(1,2304,384)f32 +nn.Linear pnnx_unique_585 1 1 2421 2422 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2421=(1,2304,384)f32 #2422=(1,2304,192)f32 +pnnx.Expression pnnx_expr_5914 2 1 2418 2422 2423 expr=add(@0,@1) #2418=(1,2304,192)f32 #2422=(1,2304,192)f32 #2423=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_588 1 1 2423 2424 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2423=(1,2304,192)f32 #2424=(1,2304,192)f32 +Tensor.view Tensor.view_1766 1 1 2424 2425 shape=(1,48,48,192) $input=2424 #2424=(1,2304,192)f32 #2425=(1,48,48,192)f32 +torch.roll torch.roll_2490 1 1 2425 2426 dims=(1,2) shifts=(-4,-4) $input=2425 #2425=(1,48,48,192)f32 #2426=(1,48,48,192)f32 +Tensor.view Tensor.view_1767 1 1 2426 2427 shape=(1,6,8,6,8,192) $input=2426 #2426=(1,48,48,192)f32 #2427=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2812 1 1 2427 2428 dims=(0,1,3,2,4,5) $input=2427 #2427=(1,6,8,6,8,192)f32 #2428=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1769 1 1 2428 2429 shape=(36,64,192) $input=2428 #2428=(1,6,6,8,8,192)f32 #2429=(36,64,192)f32 +nn.Linear pnnx_unique_591 1 1 2429 2430 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2429=(36,64,192)f32 #2430=(36,64,576)f32 +Tensor.reshape Tensor.reshape_576 1 1 2430 2431 shape=(36,64,3,6,32) $input=2430 #2430=(36,64,576)f32 #2431=(36,64,3,6,32)f32 +torch.permute torch.permute_2813 1 1 2431 2432 dims=(2,0,3,1,4) $input=2431 #2431=(36,64,3,6,32)f32 #2432=(3,36,6,64,32)f32 +torch.unbind Tensor.select_863 1 3 2432 2433 2434 2435 dim=0 #2432=(3,36,6,64,32)f32 #2433=(36,6,64,32)f32 #2434=(36,6,64,32)f32 #2435=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_5816 1 1 2433 2436 expr=mul(@0,1.767767e-01) #2433=(36,6,64,32)f32 #2436=(36,6,64,32)f32 +torch.transpose torch.transpose_3127 1 1 2434 2437 dim0=-2 dim1=-1 $input=2434 #2434=(36,6,64,32)f32 #2437=(36,6,32,64)f32 +torch.matmul torch.matmul_2346 2 1 2436 2437 2438 $input=2436 $other=2437 #2436=(36,6,64,32)f32 #2437=(36,6,32,64)f32 #2438=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_13841 0 1 2439 @pnnx_fold_13841=(1,6,64,64)f32 #2439=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_5803 2 1 2438 2439 2440 expr=add(@0,@1) #2438=(36,6,64,64)f32 #2439=(1,6,64,64)f32 #2440=(36,6,64,64)f32 +Tensor.view Tensor.view_1772 1 1 2440 2441 shape=(1,36,6,64,64) $input=2440 #2440=(36,6,64,64)f32 #2441=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_13851 0 1 2442 @pnnx_fold_13851=(1,36,1,64,64)f32 #2442=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_5791 2 1 2441 2442 2443 expr=add(@0,@1) #2441=(1,36,6,64,64)f32 #2442=(1,36,1,64,64)f32 #2443=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1773 1 1 2443 2444 shape=(-1,6,64,64) $input=2443 #2443=(1,36,6,64,64)f32 #2444=(36,6,64,64)f32 +nn.Softmax pnnx_unique_592 1 1 2444 2445 dim=-1 #2444=(36,6,64,64)f32 #2445=(36,6,64,64)f32 +torch.matmul torch.matmul_2347 2 1 2445 2435 2446 $input=2445 $other=2435 #2445=(36,6,64,64)f32 #2435=(36,6,64,32)f32 #2446=(36,6,64,32)f32 +torch.transpose torch.transpose_3128 1 1 2446 2447 dim0=1 dim1=2 $input=2446 #2446=(36,6,64,32)f32 #2447=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_577 1 1 2447 2448 shape=(36,64,192) $input=2447 #2447=(36,64,6,32)f32 #2448=(36,64,192)f32 +nn.Linear pnnx_unique_594 1 1 2448 2449 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2448=(36,64,192)f32 #2449=(36,64,192)f32 +Tensor.reshape Tensor.view_1775 1 1 2449 2450 shape=(1,6,6,8,8,192) $input=2449 #2449=(36,64,192)f32 #2450=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2815 1 1 2450 2451 dims=(0,1,3,2,4,5) $input=2450 #2450=(1,6,6,8,8,192)f32 #2451=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_143 1 1 2451 2452 shape=(1,48,48,-1) $input=2451 #2451=(1,6,8,6,8,192)f32 #2452=(1,48,48,192)f32 +torch.roll torch.roll_2491 1 1 2452 2453 dims=(1,2) shifts=(4,4) $input=2452 #2452=(1,48,48,192)f32 #2453=(1,48,48,192)f32 +Tensor.view Tensor.view_1777 1 1 2453 2454 shape=(1,2304,192) $input=2453 #2453=(1,48,48,192)f32 #2454=(1,2304,192)f32 +pnnx.Expression pnnx_expr_5749 2 1 2423 2454 2455 expr=add(@0,@1) #2423=(1,2304,192)f32 #2454=(1,2304,192)f32 #2455=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_596 1 1 2455 2456 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2455=(1,2304,192)f32 #2456=(1,2304,192)f32 +nn.Linear pnnx_unique_597 1 1 2456 2457 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2456=(1,2304,192)f32 #2457=(1,2304,384)f32 +nn.GELU pnnx_unique_598 1 1 2457 2458 #2457=(1,2304,384)f32 #2458=(1,2304,384)f32 +nn.Linear pnnx_unique_600 1 1 2458 2459 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2458=(1,2304,384)f32 #2459=(1,2304,192)f32 +pnnx.Expression pnnx_expr_5747 2 1 2455 2459 2460 expr=add(@0,@1) #2455=(1,2304,192)f32 #2459=(1,2304,192)f32 #2460=(1,2304,192)f32 +torch.transpose torch.transpose_3129 1 1 2460 2461 dim0=1 dim1=2 $input=2460 #2460=(1,2304,192)f32 #2461=(1,192,2304)f32 +Tensor.view Tensor.view_1778 1 1 2461 2462 shape=(1,192,48,48) $input=2461 #2461=(1,192,2304)f32 #2462=(1,192,48,48)f32 +nn.Conv2d pnnx_unique_602 1 1 2462 2463 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #2462=(1,192,48,48)f32 #2463=(1,192,48,48)f32 +torch.flatten torch.flatten_2196 1 1 2463 2464 end_dim=-1 start_dim=2 $input=2463 #2463=(1,192,48,48)f32 #2464=(1,192,2304)f32 +torch.transpose torch.transpose_3130 1 1 2464 2465 dim0=1 dim1=2 $input=2464 #2464=(1,192,2304)f32 #2465=(1,2304,192)f32 +pnnx.Expression pnnx_expr_5734 2 1 2465 2262 2466 expr=add(@0,@1) #2465=(1,2304,192)f32 #2262=(1,2304,192)f32 #2466=(1,2304,192)f32 +nn.LayerNorm pnnx_unique_603 1 1 2466 2467 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2466=(1,2304,192)f32 #2467=(1,2304,192)f32 +torch.transpose torch.transpose_3131 1 1 2467 2468 dim0=1 dim1=2 $input=2467 #2467=(1,2304,192)f32 #2468=(1,192,2304)f32 +Tensor.view Tensor.view_1779 1 1 2468 2469 shape=(1,192,48,48) $input=2468 #2468=(1,192,2304)f32 #2469=(1,192,48,48)f32 +nn.Conv2d pnnx_unique_604 1 1 2469 2470 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #2469=(1,192,48,48)f32 #2470=(1,192,48,48)f32 +pnnx.Expression pnnx_expr_5724 2 1 2470 7 2471 expr=add(@0,@1) #2470=(1,192,48,48)f32 #7=(1,192,48,48)f32 #2471=(1,192,48,48)f32 +pnnx.Expression pnnx_expr_5721 2 1 2471 1239 2472 expr=sub(@0,@1) #2471=(1,192,48,48)f32 #1239=(1,192,48,48)f32 #2472=(1,192,48,48)f32 +nn.Conv2d conv2d_0 1 1 2472 2473 bias=False dilation=(1,1) groups=1 in_channels=192 kernel_size=(7,7) out_channels=192 padding=(3,3) padding_mode=reflect stride=(1,1) @weight=(192,192,7,7)f32 $input=2472 #2472=(1,192,48,48)f32 #2473=(1,192,48,48)f32 +nn.ReLU manipulator.convblks.0.relu 1 1 2473 2474 #2473=(1,192,48,48)f32 #2474=(1,192,48,48)f32 +pnnx.Expression pnnx_expr_5720 1 1 2474 2475 expr=mul(@0,1.900000e+01) #2474=(1,192,48,48)f32 #2475=(1,192,48,48)f32 +nn.Conv2d conv2d_1 1 1 2475 2476 bias=False dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=reflect stride=(1,1) @weight=(192,192,3,3)f32 $input=2475 #2475=(1,192,48,48)f32 #2476=(1,192,48,48)f32 +nn.Conv2d conv2d_2 1 1 2476 2477 bias=False dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=reflect stride=(1,1) @weight=(192,192,3,3)f32 $input=2476 #2476=(1,192,48,48)f32 #2477=(1,192,48,48)f32 +nn.ReLU manipulator.resblks.0.relu 1 1 2477 2478 #2477=(1,192,48,48)f32 #2478=(1,192,48,48)f32 +nn.Conv2d conv2d_3 1 1 2478 2479 bias=False dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=reflect stride=(1,1) @weight=(192,192,3,3)f32 $input=2478 #2478=(1,192,48,48)f32 #2479=(1,192,48,48)f32 +pnnx.Expression pnnx_expr_5716 3 1 2471 2479 2476 2480 expr=add(@0,add(@1,@2)) #2471=(1,192,48,48)f32 #2479=(1,192,48,48)f32 #2476=(1,192,48,48)f32 #2480=(1,192,48,48)f32 +torch.flatten torch.flatten_2197 1 1 2480 2481 end_dim=-1 start_dim=2 $input=2480 #2480=(1,192,48,48)f32 #2481=(1,192,2304)f32 +torch.transpose torch.transpose_3132 1 1 2481 2482 dim0=1 dim1=2 $input=2481 #2481=(1,192,2304)f32 #2482=(1,2304,192)f32 +nn.LayerNorm patch_embed_mmsa.norm 1 1 2482 2483 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2482=(1,2304,192)f32 #2483=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.0.residual_group.blocks.0.norm1 1 1 2483 2484 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2483=(1,2304,192)f32 #2484=(1,2304,192)f32 +Tensor.reshape Tensor.view_1781 1 1 2484 2485 shape=(1,6,8,6,8,192) $input=2484 #2484=(1,2304,192)f32 #2485=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2816 1 1 2485 2486 dims=(0,1,3,2,4,5) $input=2485 #2485=(1,6,8,6,8,192)f32 #2486=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1783 1 1 2486 2487 shape=(36,64,192) $input=2486 #2486=(1,6,6,8,8,192)f32 #2487=(36,64,192)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.0.attn.qkv 1 1 2487 2488 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2487=(36,64,192)f32 #2488=(36,64,576)f32 +Tensor.reshape Tensor.reshape_578 1 1 2488 2489 shape=(36,64,3,6,32) $input=2488 #2488=(36,64,576)f32 #2489=(36,64,3,6,32)f32 +torch.permute torch.permute_2817 1 1 2489 2490 dims=(2,0,3,1,4) $input=2489 #2489=(36,64,3,6,32)f32 #2490=(3,36,6,64,32)f32 +torch.unbind Tensor.select_866 1 3 2490 2491 2492 2493 dim=0 #2490=(3,36,6,64,32)f32 #2491=(36,6,64,32)f32 #2492=(36,6,64,32)f32 #2493=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_5459 1 1 2491 2494 expr=mul(@0,1.767767e-01) #2491=(36,6,64,32)f32 #2494=(36,6,64,32)f32 +torch.transpose torch.transpose_3133 1 1 2492 2495 dim0=-2 dim1=-1 $input=2492 #2492=(36,6,64,32)f32 #2495=(36,6,32,64)f32 +torch.matmul torch.matmul_2348 2 1 2494 2495 2496 $input=2494 $other=2495 #2494=(36,6,64,32)f32 #2495=(36,6,32,64)f32 #2496=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_14084 0 1 2497 @pnnx_fold_14084=(1,6,64,64)f32 #2497=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_5446 2 1 2496 2497 2498 expr=add(@0,@1) #2496=(36,6,64,64)f32 #2497=(1,6,64,64)f32 #2498=(36,6,64,64)f32 +nn.Softmax layers_mmsa.0.residual_group.blocks.0.attn.softmax 1 1 2498 2499 dim=-1 #2498=(36,6,64,64)f32 #2499=(36,6,64,64)f32 +torch.matmul torch.matmul_2349 2 1 2499 2493 2500 $input=2499 $other=2493 #2499=(36,6,64,64)f32 #2493=(36,6,64,32)f32 #2500=(36,6,64,32)f32 +torch.transpose torch.transpose_3134 1 1 2500 2501 dim0=1 dim1=2 $input=2500 #2500=(36,6,64,32)f32 #2501=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_579 1 1 2501 2502 shape=(36,64,192) $input=2501 #2501=(36,64,6,32)f32 #2502=(36,64,192)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.0.attn.proj 1 1 2502 2503 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2502=(36,64,192)f32 #2503=(36,64,192)f32 +Tensor.reshape Tensor.view_1787 1 1 2503 2504 shape=(1,6,6,8,8,192) $input=2503 #2503=(36,64,192)f32 #2504=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2819 1 1 2504 2505 dims=(0,1,3,2,4,5) $input=2504 #2504=(1,6,6,8,8,192)f32 #2505=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1789 1 1 2505 2506 shape=(1,2304,192) $input=2505 #2505=(1,6,8,6,8,192)f32 #2506=(1,2304,192)f32 +pnnx.Expression pnnx_expr_5413 2 1 2483 2506 2507 expr=add(@0,@1) #2483=(1,2304,192)f32 #2506=(1,2304,192)f32 #2507=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.0.residual_group.blocks.0.norm2 1 1 2507 2508 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2507=(1,2304,192)f32 #2508=(1,2304,192)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.0.mlp.fc1 1 1 2508 2509 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2508=(1,2304,192)f32 #2509=(1,2304,384)f32 +nn.GELU layers_mmsa.0.residual_group.blocks.0.mlp.act 1 1 2509 2510 #2509=(1,2304,384)f32 #2510=(1,2304,384)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.0.mlp.fc2 1 1 2510 2511 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2510=(1,2304,384)f32 #2511=(1,2304,192)f32 +pnnx.Expression pnnx_expr_5411 2 1 2507 2511 2512 expr=add(@0,@1) #2507=(1,2304,192)f32 #2511=(1,2304,192)f32 #2512=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.0.residual_group.blocks.1.norm1 1 1 2512 2513 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2512=(1,2304,192)f32 #2513=(1,2304,192)f32 +Tensor.view Tensor.view_1790 1 1 2513 2514 shape=(1,48,48,192) $input=2513 #2513=(1,2304,192)f32 #2514=(1,48,48,192)f32 +torch.roll torch.roll_2492 1 1 2514 2515 dims=(1,2) shifts=(-4,-4) $input=2514 #2514=(1,48,48,192)f32 #2515=(1,48,48,192)f32 +Tensor.view Tensor.view_1791 1 1 2515 2516 shape=(1,6,8,6,8,192) $input=2515 #2515=(1,48,48,192)f32 #2516=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2820 1 1 2516 2517 dims=(0,1,3,2,4,5) $input=2516 #2516=(1,6,8,6,8,192)f32 #2517=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1793 1 1 2517 2518 shape=(36,64,192) $input=2517 #2517=(1,6,6,8,8,192)f32 #2518=(36,64,192)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.1.attn.qkv 1 1 2518 2519 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2518=(36,64,192)f32 #2519=(36,64,576)f32 +Tensor.reshape Tensor.reshape_580 1 1 2519 2520 shape=(36,64,3,6,32) $input=2519 #2519=(36,64,576)f32 #2520=(36,64,3,6,32)f32 +torch.permute torch.permute_2821 1 1 2520 2521 dims=(2,0,3,1,4) $input=2520 #2520=(36,64,3,6,32)f32 #2521=(3,36,6,64,32)f32 +torch.unbind Tensor.select_869 1 3 2521 2522 2523 2524 dim=0 #2521=(3,36,6,64,32)f32 #2522=(36,6,64,32)f32 #2523=(36,6,64,32)f32 #2524=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_5313 1 1 2522 2525 expr=mul(@0,1.767767e-01) #2522=(36,6,64,32)f32 #2525=(36,6,64,32)f32 +torch.transpose torch.transpose_3135 1 1 2523 2526 dim0=-2 dim1=-1 $input=2523 #2523=(36,6,64,32)f32 #2526=(36,6,32,64)f32 +torch.matmul torch.matmul_2350 2 1 2525 2526 2527 $input=2525 $other=2526 #2525=(36,6,64,32)f32 #2526=(36,6,32,64)f32 #2527=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_14237 0 1 2528 @pnnx_fold_14237=(1,6,64,64)f32 #2528=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_5300 2 1 2527 2528 2529 expr=add(@0,@1) #2527=(36,6,64,64)f32 #2528=(1,6,64,64)f32 #2529=(36,6,64,64)f32 +Tensor.view Tensor.view_1796 1 1 2529 2530 shape=(1,36,6,64,64) $input=2529 #2529=(36,6,64,64)f32 #2530=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_14247 0 1 2531 @pnnx_fold_14247=(1,36,1,64,64)f32 #2531=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_5288 2 1 2530 2531 2532 expr=add(@0,@1) #2530=(1,36,6,64,64)f32 #2531=(1,36,1,64,64)f32 #2532=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1797 1 1 2532 2533 shape=(-1,6,64,64) $input=2532 #2532=(1,36,6,64,64)f32 #2533=(36,6,64,64)f32 +nn.Softmax layers_mmsa.0.residual_group.blocks.1.attn.softmax 1 1 2533 2534 dim=-1 #2533=(36,6,64,64)f32 #2534=(36,6,64,64)f32 +torch.matmul torch.matmul_2351 2 1 2534 2524 2535 $input=2534 $other=2524 #2534=(36,6,64,64)f32 #2524=(36,6,64,32)f32 #2535=(36,6,64,32)f32 +torch.transpose torch.transpose_3136 1 1 2535 2536 dim0=1 dim1=2 $input=2535 #2535=(36,6,64,32)f32 #2536=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_581 1 1 2536 2537 shape=(36,64,192) $input=2536 #2536=(36,64,6,32)f32 #2537=(36,64,192)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.1.attn.proj 1 1 2537 2538 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2537=(36,64,192)f32 #2538=(36,64,192)f32 +Tensor.reshape Tensor.view_1799 1 1 2538 2539 shape=(1,6,6,8,8,192) $input=2538 #2538=(36,64,192)f32 #2539=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2823 1 1 2539 2540 dims=(0,1,3,2,4,5) $input=2539 #2539=(1,6,6,8,8,192)f32 #2540=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_147 1 1 2540 2541 shape=(1,48,48,-1) $input=2540 #2540=(1,6,8,6,8,192)f32 #2541=(1,48,48,192)f32 +torch.roll torch.roll_2493 1 1 2541 2542 dims=(1,2) shifts=(4,4) $input=2541 #2541=(1,48,48,192)f32 #2542=(1,48,48,192)f32 +Tensor.view Tensor.view_1801 1 1 2542 2543 shape=(1,2304,192) $input=2542 #2542=(1,48,48,192)f32 #2543=(1,2304,192)f32 +pnnx.Expression pnnx_expr_5246 2 1 2512 2543 2544 expr=add(@0,@1) #2512=(1,2304,192)f32 #2543=(1,2304,192)f32 #2544=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.0.residual_group.blocks.1.norm2 1 1 2544 2545 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2544=(1,2304,192)f32 #2545=(1,2304,192)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.1.mlp.fc1 1 1 2545 2546 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2545=(1,2304,192)f32 #2546=(1,2304,384)f32 +nn.GELU layers_mmsa.0.residual_group.blocks.1.mlp.act 1 1 2546 2547 #2546=(1,2304,384)f32 #2547=(1,2304,384)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.1.mlp.fc2 1 1 2547 2548 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2547=(1,2304,384)f32 #2548=(1,2304,192)f32 +pnnx.Expression pnnx_expr_5244 2 1 2544 2548 2549 expr=add(@0,@1) #2544=(1,2304,192)f32 #2548=(1,2304,192)f32 #2549=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.0.residual_group.blocks.2.norm1 1 1 2549 2550 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2549=(1,2304,192)f32 #2550=(1,2304,192)f32 +Tensor.reshape Tensor.view_1803 1 1 2550 2551 shape=(1,6,8,6,8,192) $input=2550 #2550=(1,2304,192)f32 #2551=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2824 1 1 2551 2552 dims=(0,1,3,2,4,5) $input=2551 #2551=(1,6,8,6,8,192)f32 #2552=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1805 1 1 2552 2553 shape=(36,64,192) $input=2552 #2552=(1,6,6,8,8,192)f32 #2553=(36,64,192)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.2.attn.qkv 1 1 2553 2554 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2553=(36,64,192)f32 #2554=(36,64,576)f32 +Tensor.reshape Tensor.reshape_582 1 1 2554 2555 shape=(36,64,3,6,32) $input=2554 #2554=(36,64,576)f32 #2555=(36,64,3,6,32)f32 +torch.permute torch.permute_2825 1 1 2555 2556 dims=(2,0,3,1,4) $input=2555 #2555=(36,64,3,6,32)f32 #2556=(3,36,6,64,32)f32 +torch.unbind Tensor.select_872 1 3 2556 2557 2558 2559 dim=0 #2556=(3,36,6,64,32)f32 #2557=(36,6,64,32)f32 #2558=(36,6,64,32)f32 #2559=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_5156 1 1 2557 2560 expr=mul(@0,1.767767e-01) #2557=(36,6,64,32)f32 #2560=(36,6,64,32)f32 +torch.transpose torch.transpose_3137 1 1 2558 2561 dim0=-2 dim1=-1 $input=2558 #2558=(36,6,64,32)f32 #2561=(36,6,32,64)f32 +torch.matmul torch.matmul_2352 2 1 2560 2561 2562 $input=2560 $other=2561 #2560=(36,6,64,32)f32 #2561=(36,6,32,64)f32 #2562=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_14396 0 1 2563 @pnnx_fold_14396=(1,6,64,64)f32 #2563=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_5143 2 1 2562 2563 2564 expr=add(@0,@1) #2562=(36,6,64,64)f32 #2563=(1,6,64,64)f32 #2564=(36,6,64,64)f32 +nn.Softmax layers_mmsa.0.residual_group.blocks.2.attn.softmax 1 1 2564 2565 dim=-1 #2564=(36,6,64,64)f32 #2565=(36,6,64,64)f32 +torch.matmul torch.matmul_2353 2 1 2565 2559 2566 $input=2565 $other=2559 #2565=(36,6,64,64)f32 #2559=(36,6,64,32)f32 #2566=(36,6,64,32)f32 +torch.transpose torch.transpose_3138 1 1 2566 2567 dim0=1 dim1=2 $input=2566 #2566=(36,6,64,32)f32 #2567=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_583 1 1 2567 2568 shape=(36,64,192) $input=2567 #2567=(36,64,6,32)f32 #2568=(36,64,192)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.2.attn.proj 1 1 2568 2569 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2568=(36,64,192)f32 #2569=(36,64,192)f32 +Tensor.reshape Tensor.view_1809 1 1 2569 2570 shape=(1,6,6,8,8,192) $input=2569 #2569=(36,64,192)f32 #2570=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2827 1 1 2570 2571 dims=(0,1,3,2,4,5) $input=2570 #2570=(1,6,6,8,8,192)f32 #2571=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1811 1 1 2571 2572 shape=(1,2304,192) $input=2571 #2571=(1,6,8,6,8,192)f32 #2572=(1,2304,192)f32 +pnnx.Expression pnnx_expr_5110 2 1 2549 2572 2573 expr=add(@0,@1) #2549=(1,2304,192)f32 #2572=(1,2304,192)f32 #2573=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.0.residual_group.blocks.2.norm2 1 1 2573 2574 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2573=(1,2304,192)f32 #2574=(1,2304,192)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.2.mlp.fc1 1 1 2574 2575 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2574=(1,2304,192)f32 #2575=(1,2304,384)f32 +nn.GELU layers_mmsa.0.residual_group.blocks.2.mlp.act 1 1 2575 2576 #2575=(1,2304,384)f32 #2576=(1,2304,384)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.2.mlp.fc2 1 1 2576 2577 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2576=(1,2304,384)f32 #2577=(1,2304,192)f32 +pnnx.Expression pnnx_expr_5108 2 1 2573 2577 2578 expr=add(@0,@1) #2573=(1,2304,192)f32 #2577=(1,2304,192)f32 #2578=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.0.residual_group.blocks.3.norm1 1 1 2578 2579 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2578=(1,2304,192)f32 #2579=(1,2304,192)f32 +Tensor.view Tensor.view_1812 1 1 2579 2580 shape=(1,48,48,192) $input=2579 #2579=(1,2304,192)f32 #2580=(1,48,48,192)f32 +torch.roll torch.roll_2494 1 1 2580 2581 dims=(1,2) shifts=(-4,-4) $input=2580 #2580=(1,48,48,192)f32 #2581=(1,48,48,192)f32 +Tensor.view Tensor.view_1813 1 1 2581 2582 shape=(1,6,8,6,8,192) $input=2581 #2581=(1,48,48,192)f32 #2582=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2828 1 1 2582 2583 dims=(0,1,3,2,4,5) $input=2582 #2582=(1,6,8,6,8,192)f32 #2583=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1815 1 1 2583 2584 shape=(36,64,192) $input=2583 #2583=(1,6,6,8,8,192)f32 #2584=(36,64,192)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.3.attn.qkv 1 1 2584 2585 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2584=(36,64,192)f32 #2585=(36,64,576)f32 +Tensor.reshape Tensor.reshape_584 1 1 2585 2586 shape=(36,64,3,6,32) $input=2585 #2585=(36,64,576)f32 #2586=(36,64,3,6,32)f32 +torch.permute torch.permute_2829 1 1 2586 2587 dims=(2,0,3,1,4) $input=2586 #2586=(36,64,3,6,32)f32 #2587=(3,36,6,64,32)f32 +torch.unbind Tensor.select_875 1 3 2587 2588 2589 2590 dim=0 #2587=(3,36,6,64,32)f32 #2588=(36,6,64,32)f32 #2589=(36,6,64,32)f32 #2590=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_5010 1 1 2588 2591 expr=mul(@0,1.767767e-01) #2588=(36,6,64,32)f32 #2591=(36,6,64,32)f32 +torch.transpose torch.transpose_3139 1 1 2589 2592 dim0=-2 dim1=-1 $input=2589 #2589=(36,6,64,32)f32 #2592=(36,6,32,64)f32 +torch.matmul torch.matmul_2354 2 1 2591 2592 2593 $input=2591 $other=2592 #2591=(36,6,64,32)f32 #2592=(36,6,32,64)f32 #2593=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_14549 0 1 2594 @pnnx_fold_14549=(1,6,64,64)f32 #2594=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_4997 2 1 2593 2594 2595 expr=add(@0,@1) #2593=(36,6,64,64)f32 #2594=(1,6,64,64)f32 #2595=(36,6,64,64)f32 +Tensor.view Tensor.view_1818 1 1 2595 2596 shape=(1,36,6,64,64) $input=2595 #2595=(36,6,64,64)f32 #2596=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_14559 0 1 2597 @pnnx_fold_14559=(1,36,1,64,64)f32 #2597=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_4985 2 1 2596 2597 2598 expr=add(@0,@1) #2596=(1,36,6,64,64)f32 #2597=(1,36,1,64,64)f32 #2598=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1819 1 1 2598 2599 shape=(-1,6,64,64) $input=2598 #2598=(1,36,6,64,64)f32 #2599=(36,6,64,64)f32 +nn.Softmax layers_mmsa.0.residual_group.blocks.3.attn.softmax 1 1 2599 2600 dim=-1 #2599=(36,6,64,64)f32 #2600=(36,6,64,64)f32 +torch.matmul torch.matmul_2355 2 1 2600 2590 2601 $input=2600 $other=2590 #2600=(36,6,64,64)f32 #2590=(36,6,64,32)f32 #2601=(36,6,64,32)f32 +torch.transpose torch.transpose_3140 1 1 2601 2602 dim0=1 dim1=2 $input=2601 #2601=(36,6,64,32)f32 #2602=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_585 1 1 2602 2603 shape=(36,64,192) $input=2602 #2602=(36,64,6,32)f32 #2603=(36,64,192)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.3.attn.proj 1 1 2603 2604 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2603=(36,64,192)f32 #2604=(36,64,192)f32 +Tensor.reshape Tensor.view_1821 1 1 2604 2605 shape=(1,6,6,8,8,192) $input=2604 #2604=(36,64,192)f32 #2605=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2831 1 1 2605 2606 dims=(0,1,3,2,4,5) $input=2605 #2605=(1,6,6,8,8,192)f32 #2606=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_151 1 1 2606 2607 shape=(1,48,48,-1) $input=2606 #2606=(1,6,8,6,8,192)f32 #2607=(1,48,48,192)f32 +torch.roll torch.roll_2495 1 1 2607 2608 dims=(1,2) shifts=(4,4) $input=2607 #2607=(1,48,48,192)f32 #2608=(1,48,48,192)f32 +Tensor.view Tensor.view_1823 1 1 2608 2609 shape=(1,2304,192) $input=2608 #2608=(1,48,48,192)f32 #2609=(1,2304,192)f32 +pnnx.Expression pnnx_expr_4943 2 1 2578 2609 2610 expr=add(@0,@1) #2578=(1,2304,192)f32 #2609=(1,2304,192)f32 #2610=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.0.residual_group.blocks.3.norm2 1 1 2610 2611 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2610=(1,2304,192)f32 #2611=(1,2304,192)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.3.mlp.fc1 1 1 2611 2612 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2611=(1,2304,192)f32 #2612=(1,2304,384)f32 +nn.GELU layers_mmsa.0.residual_group.blocks.3.mlp.act 1 1 2612 2613 #2612=(1,2304,384)f32 #2613=(1,2304,384)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.3.mlp.fc2 1 1 2613 2614 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2613=(1,2304,384)f32 #2614=(1,2304,192)f32 +pnnx.Expression pnnx_expr_4941 2 1 2610 2614 2615 expr=add(@0,@1) #2610=(1,2304,192)f32 #2614=(1,2304,192)f32 #2615=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.0.residual_group.blocks.4.norm1 1 1 2615 2616 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2615=(1,2304,192)f32 #2616=(1,2304,192)f32 +Tensor.reshape Tensor.view_1825 1 1 2616 2617 shape=(1,6,8,6,8,192) $input=2616 #2616=(1,2304,192)f32 #2617=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2832 1 1 2617 2618 dims=(0,1,3,2,4,5) $input=2617 #2617=(1,6,8,6,8,192)f32 #2618=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1827 1 1 2618 2619 shape=(36,64,192) $input=2618 #2618=(1,6,6,8,8,192)f32 #2619=(36,64,192)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.4.attn.qkv 1 1 2619 2620 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2619=(36,64,192)f32 #2620=(36,64,576)f32 +Tensor.reshape Tensor.reshape_586 1 1 2620 2621 shape=(36,64,3,6,32) $input=2620 #2620=(36,64,576)f32 #2621=(36,64,3,6,32)f32 +torch.permute torch.permute_2833 1 1 2621 2622 dims=(2,0,3,1,4) $input=2621 #2621=(36,64,3,6,32)f32 #2622=(3,36,6,64,32)f32 +torch.unbind Tensor.select_878 1 3 2622 2623 2624 2625 dim=0 #2622=(3,36,6,64,32)f32 #2623=(36,6,64,32)f32 #2624=(36,6,64,32)f32 #2625=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_4853 1 1 2623 2626 expr=mul(@0,1.767767e-01) #2623=(36,6,64,32)f32 #2626=(36,6,64,32)f32 +torch.transpose torch.transpose_3141 1 1 2624 2627 dim0=-2 dim1=-1 $input=2624 #2624=(36,6,64,32)f32 #2627=(36,6,32,64)f32 +torch.matmul torch.matmul_2356 2 1 2626 2627 2628 $input=2626 $other=2627 #2626=(36,6,64,32)f32 #2627=(36,6,32,64)f32 #2628=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_14708 0 1 2629 @pnnx_fold_14708=(1,6,64,64)f32 #2629=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_4840 2 1 2628 2629 2630 expr=add(@0,@1) #2628=(36,6,64,64)f32 #2629=(1,6,64,64)f32 #2630=(36,6,64,64)f32 +nn.Softmax layers_mmsa.0.residual_group.blocks.4.attn.softmax 1 1 2630 2631 dim=-1 #2630=(36,6,64,64)f32 #2631=(36,6,64,64)f32 +torch.matmul torch.matmul_2357 2 1 2631 2625 2632 $input=2631 $other=2625 #2631=(36,6,64,64)f32 #2625=(36,6,64,32)f32 #2632=(36,6,64,32)f32 +torch.transpose torch.transpose_3142 1 1 2632 2633 dim0=1 dim1=2 $input=2632 #2632=(36,6,64,32)f32 #2633=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_587 1 1 2633 2634 shape=(36,64,192) $input=2633 #2633=(36,64,6,32)f32 #2634=(36,64,192)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.4.attn.proj 1 1 2634 2635 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2634=(36,64,192)f32 #2635=(36,64,192)f32 +Tensor.reshape Tensor.view_1831 1 1 2635 2636 shape=(1,6,6,8,8,192) $input=2635 #2635=(36,64,192)f32 #2636=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2835 1 1 2636 2637 dims=(0,1,3,2,4,5) $input=2636 #2636=(1,6,6,8,8,192)f32 #2637=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1833 1 1 2637 2638 shape=(1,2304,192) $input=2637 #2637=(1,6,8,6,8,192)f32 #2638=(1,2304,192)f32 +pnnx.Expression pnnx_expr_4807 2 1 2615 2638 2639 expr=add(@0,@1) #2615=(1,2304,192)f32 #2638=(1,2304,192)f32 #2639=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.0.residual_group.blocks.4.norm2 1 1 2639 2640 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2639=(1,2304,192)f32 #2640=(1,2304,192)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.4.mlp.fc1 1 1 2640 2641 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2640=(1,2304,192)f32 #2641=(1,2304,384)f32 +nn.GELU layers_mmsa.0.residual_group.blocks.4.mlp.act 1 1 2641 2642 #2641=(1,2304,384)f32 #2642=(1,2304,384)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.4.mlp.fc2 1 1 2642 2643 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2642=(1,2304,384)f32 #2643=(1,2304,192)f32 +pnnx.Expression pnnx_expr_4805 2 1 2639 2643 2644 expr=add(@0,@1) #2639=(1,2304,192)f32 #2643=(1,2304,192)f32 #2644=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.0.residual_group.blocks.5.norm1 1 1 2644 2645 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2644=(1,2304,192)f32 #2645=(1,2304,192)f32 +Tensor.view Tensor.view_1834 1 1 2645 2646 shape=(1,48,48,192) $input=2645 #2645=(1,2304,192)f32 #2646=(1,48,48,192)f32 +torch.roll torch.roll_2496 1 1 2646 2647 dims=(1,2) shifts=(-4,-4) $input=2646 #2646=(1,48,48,192)f32 #2647=(1,48,48,192)f32 +Tensor.view Tensor.view_1835 1 1 2647 2648 shape=(1,6,8,6,8,192) $input=2647 #2647=(1,48,48,192)f32 #2648=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2836 1 1 2648 2649 dims=(0,1,3,2,4,5) $input=2648 #2648=(1,6,8,6,8,192)f32 #2649=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1837 1 1 2649 2650 shape=(36,64,192) $input=2649 #2649=(1,6,6,8,8,192)f32 #2650=(36,64,192)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.5.attn.qkv 1 1 2650 2651 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2650=(36,64,192)f32 #2651=(36,64,576)f32 +Tensor.reshape Tensor.reshape_588 1 1 2651 2652 shape=(36,64,3,6,32) $input=2651 #2651=(36,64,576)f32 #2652=(36,64,3,6,32)f32 +torch.permute torch.permute_2837 1 1 2652 2653 dims=(2,0,3,1,4) $input=2652 #2652=(36,64,3,6,32)f32 #2653=(3,36,6,64,32)f32 +torch.unbind Tensor.select_881 1 3 2653 2654 2655 2656 dim=0 #2653=(3,36,6,64,32)f32 #2654=(36,6,64,32)f32 #2655=(36,6,64,32)f32 #2656=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_4707 1 1 2654 2657 expr=mul(@0,1.767767e-01) #2654=(36,6,64,32)f32 #2657=(36,6,64,32)f32 +torch.transpose torch.transpose_3143 1 1 2655 2658 dim0=-2 dim1=-1 $input=2655 #2655=(36,6,64,32)f32 #2658=(36,6,32,64)f32 +torch.matmul torch.matmul_2358 2 1 2657 2658 2659 $input=2657 $other=2658 #2657=(36,6,64,32)f32 #2658=(36,6,32,64)f32 #2659=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_14861 0 1 2660 @pnnx_fold_14861=(1,6,64,64)f32 #2660=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_4694 2 1 2659 2660 2661 expr=add(@0,@1) #2659=(36,6,64,64)f32 #2660=(1,6,64,64)f32 #2661=(36,6,64,64)f32 +Tensor.view Tensor.view_1840 1 1 2661 2662 shape=(1,36,6,64,64) $input=2661 #2661=(36,6,64,64)f32 #2662=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_14871 0 1 2663 @pnnx_fold_14871=(1,36,1,64,64)f32 #2663=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_4682 2 1 2662 2663 2664 expr=add(@0,@1) #2662=(1,36,6,64,64)f32 #2663=(1,36,1,64,64)f32 #2664=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1841 1 1 2664 2665 shape=(-1,6,64,64) $input=2664 #2664=(1,36,6,64,64)f32 #2665=(36,6,64,64)f32 +nn.Softmax layers_mmsa.0.residual_group.blocks.5.attn.softmax 1 1 2665 2666 dim=-1 #2665=(36,6,64,64)f32 #2666=(36,6,64,64)f32 +torch.matmul torch.matmul_2359 2 1 2666 2656 2667 $input=2666 $other=2656 #2666=(36,6,64,64)f32 #2656=(36,6,64,32)f32 #2667=(36,6,64,32)f32 +torch.transpose torch.transpose_3144 1 1 2667 2668 dim0=1 dim1=2 $input=2667 #2667=(36,6,64,32)f32 #2668=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_589 1 1 2668 2669 shape=(36,64,192) $input=2668 #2668=(36,64,6,32)f32 #2669=(36,64,192)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.5.attn.proj 1 1 2669 2670 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2669=(36,64,192)f32 #2670=(36,64,192)f32 +Tensor.reshape Tensor.view_1843 1 1 2670 2671 shape=(1,6,6,8,8,192) $input=2670 #2670=(36,64,192)f32 #2671=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2839 1 1 2671 2672 dims=(0,1,3,2,4,5) $input=2671 #2671=(1,6,6,8,8,192)f32 #2672=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_155 1 1 2672 2673 shape=(1,48,48,-1) $input=2672 #2672=(1,6,8,6,8,192)f32 #2673=(1,48,48,192)f32 +torch.roll torch.roll_2497 1 1 2673 2674 dims=(1,2) shifts=(4,4) $input=2673 #2673=(1,48,48,192)f32 #2674=(1,48,48,192)f32 +Tensor.view Tensor.view_1845 1 1 2674 2675 shape=(1,2304,192) $input=2674 #2674=(1,48,48,192)f32 #2675=(1,2304,192)f32 +pnnx.Expression pnnx_expr_4640 2 1 2644 2675 2676 expr=add(@0,@1) #2644=(1,2304,192)f32 #2675=(1,2304,192)f32 #2676=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.0.residual_group.blocks.5.norm2 1 1 2676 2677 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2676=(1,2304,192)f32 #2677=(1,2304,192)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.5.mlp.fc1 1 1 2677 2678 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2677=(1,2304,192)f32 #2678=(1,2304,384)f32 +nn.GELU layers_mmsa.0.residual_group.blocks.5.mlp.act 1 1 2678 2679 #2678=(1,2304,384)f32 #2679=(1,2304,384)f32 +nn.Linear layers_mmsa.0.residual_group.blocks.5.mlp.fc2 1 1 2679 2680 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2679=(1,2304,384)f32 #2680=(1,2304,192)f32 +pnnx.Expression pnnx_expr_4638 2 1 2676 2680 2681 expr=add(@0,@1) #2676=(1,2304,192)f32 #2680=(1,2304,192)f32 #2681=(1,2304,192)f32 +torch.transpose torch.transpose_3145 1 1 2681 2682 dim0=1 dim1=2 $input=2681 #2681=(1,2304,192)f32 #2682=(1,192,2304)f32 +Tensor.view Tensor.view_1846 1 1 2682 2683 shape=(1,192,48,48) $input=2682 #2682=(1,192,2304)f32 #2683=(1,192,48,48)f32 +nn.Conv2d layers_mmsa.0.conv 1 1 2683 2684 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #2683=(1,192,48,48)f32 #2684=(1,192,48,48)f32 +torch.flatten torch.flatten_2198 1 1 2684 2685 end_dim=-1 start_dim=2 $input=2684 #2684=(1,192,48,48)f32 #2685=(1,192,2304)f32 +torch.transpose torch.transpose_3146 1 1 2685 2686 dim0=1 dim1=2 $input=2685 #2685=(1,192,2304)f32 #2686=(1,2304,192)f32 +pnnx.Expression pnnx_expr_4625 2 1 2686 2483 2687 expr=add(@0,@1) #2686=(1,2304,192)f32 #2483=(1,2304,192)f32 #2687=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.1.residual_group.blocks.0.norm1 1 1 2687 2688 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2687=(1,2304,192)f32 #2688=(1,2304,192)f32 +Tensor.reshape Tensor.view_1848 1 1 2688 2689 shape=(1,6,8,6,8,192) $input=2688 #2688=(1,2304,192)f32 #2689=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2840 1 1 2689 2690 dims=(0,1,3,2,4,5) $input=2689 #2689=(1,6,8,6,8,192)f32 #2690=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1850 1 1 2690 2691 shape=(36,64,192) $input=2690 #2690=(1,6,6,8,8,192)f32 #2691=(36,64,192)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.0.attn.qkv 1 1 2691 2692 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2691=(36,64,192)f32 #2692=(36,64,576)f32 +Tensor.reshape Tensor.reshape_590 1 1 2692 2693 shape=(36,64,3,6,32) $input=2692 #2692=(36,64,576)f32 #2693=(36,64,3,6,32)f32 +torch.permute torch.permute_2841 1 1 2693 2694 dims=(2,0,3,1,4) $input=2693 #2693=(36,64,3,6,32)f32 #2694=(3,36,6,64,32)f32 +torch.unbind Tensor.select_884 1 3 2694 2695 2696 2697 dim=0 #2694=(3,36,6,64,32)f32 #2695=(36,6,64,32)f32 #2696=(36,6,64,32)f32 #2697=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_4536 1 1 2695 2698 expr=mul(@0,1.767767e-01) #2695=(36,6,64,32)f32 #2698=(36,6,64,32)f32 +torch.transpose torch.transpose_3147 1 1 2696 2699 dim0=-2 dim1=-1 $input=2696 #2696=(36,6,64,32)f32 #2699=(36,6,32,64)f32 +torch.matmul torch.matmul_2360 2 1 2698 2699 2700 $input=2698 $other=2699 #2698=(36,6,64,32)f32 #2699=(36,6,32,64)f32 #2700=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_15054 0 1 2701 @pnnx_fold_15054=(1,6,64,64)f32 #2701=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_4523 2 1 2700 2701 2702 expr=add(@0,@1) #2700=(36,6,64,64)f32 #2701=(1,6,64,64)f32 #2702=(36,6,64,64)f32 +nn.Softmax layers_mmsa.1.residual_group.blocks.0.attn.softmax 1 1 2702 2703 dim=-1 #2702=(36,6,64,64)f32 #2703=(36,6,64,64)f32 +torch.matmul torch.matmul_2361 2 1 2703 2697 2704 $input=2703 $other=2697 #2703=(36,6,64,64)f32 #2697=(36,6,64,32)f32 #2704=(36,6,64,32)f32 +torch.transpose torch.transpose_3148 1 1 2704 2705 dim0=1 dim1=2 $input=2704 #2704=(36,6,64,32)f32 #2705=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_591 1 1 2705 2706 shape=(36,64,192) $input=2705 #2705=(36,64,6,32)f32 #2706=(36,64,192)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.0.attn.proj 1 1 2706 2707 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2706=(36,64,192)f32 #2707=(36,64,192)f32 +Tensor.reshape Tensor.view_1854 1 1 2707 2708 shape=(1,6,6,8,8,192) $input=2707 #2707=(36,64,192)f32 #2708=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2843 1 1 2708 2709 dims=(0,1,3,2,4,5) $input=2708 #2708=(1,6,6,8,8,192)f32 #2709=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1856 1 1 2709 2710 shape=(1,2304,192) $input=2709 #2709=(1,6,8,6,8,192)f32 #2710=(1,2304,192)f32 +pnnx.Expression pnnx_expr_4490 2 1 2687 2710 2711 expr=add(@0,@1) #2687=(1,2304,192)f32 #2710=(1,2304,192)f32 #2711=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.1.residual_group.blocks.0.norm2 1 1 2711 2712 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2711=(1,2304,192)f32 #2712=(1,2304,192)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.0.mlp.fc1 1 1 2712 2713 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2712=(1,2304,192)f32 #2713=(1,2304,384)f32 +nn.GELU layers_mmsa.1.residual_group.blocks.0.mlp.act 1 1 2713 2714 #2713=(1,2304,384)f32 #2714=(1,2304,384)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.0.mlp.fc2 1 1 2714 2715 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2714=(1,2304,384)f32 #2715=(1,2304,192)f32 +pnnx.Expression pnnx_expr_4488 2 1 2711 2715 2716 expr=add(@0,@1) #2711=(1,2304,192)f32 #2715=(1,2304,192)f32 #2716=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.1.residual_group.blocks.1.norm1 1 1 2716 2717 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2716=(1,2304,192)f32 #2717=(1,2304,192)f32 +Tensor.view Tensor.view_1857 1 1 2717 2718 shape=(1,48,48,192) $input=2717 #2717=(1,2304,192)f32 #2718=(1,48,48,192)f32 +torch.roll torch.roll_2498 1 1 2718 2719 dims=(1,2) shifts=(-4,-4) $input=2718 #2718=(1,48,48,192)f32 #2719=(1,48,48,192)f32 +Tensor.view Tensor.view_1858 1 1 2719 2720 shape=(1,6,8,6,8,192) $input=2719 #2719=(1,48,48,192)f32 #2720=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2844 1 1 2720 2721 dims=(0,1,3,2,4,5) $input=2720 #2720=(1,6,8,6,8,192)f32 #2721=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1860 1 1 2721 2722 shape=(36,64,192) $input=2721 #2721=(1,6,6,8,8,192)f32 #2722=(36,64,192)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.1.attn.qkv 1 1 2722 2723 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2722=(36,64,192)f32 #2723=(36,64,576)f32 +Tensor.reshape Tensor.reshape_592 1 1 2723 2724 shape=(36,64,3,6,32) $input=2723 #2723=(36,64,576)f32 #2724=(36,64,3,6,32)f32 +torch.permute torch.permute_2845 1 1 2724 2725 dims=(2,0,3,1,4) $input=2724 #2724=(36,64,3,6,32)f32 #2725=(3,36,6,64,32)f32 +torch.unbind Tensor.select_887 1 3 2725 2726 2727 2728 dim=0 #2725=(3,36,6,64,32)f32 #2726=(36,6,64,32)f32 #2727=(36,6,64,32)f32 #2728=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_4390 1 1 2726 2729 expr=mul(@0,1.767767e-01) #2726=(36,6,64,32)f32 #2729=(36,6,64,32)f32 +torch.transpose torch.transpose_3149 1 1 2727 2730 dim0=-2 dim1=-1 $input=2727 #2727=(36,6,64,32)f32 #2730=(36,6,32,64)f32 +torch.matmul torch.matmul_2362 2 1 2729 2730 2731 $input=2729 $other=2730 #2729=(36,6,64,32)f32 #2730=(36,6,32,64)f32 #2731=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_15207 0 1 2732 @pnnx_fold_15207=(1,6,64,64)f32 #2732=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_4377 2 1 2731 2732 2733 expr=add(@0,@1) #2731=(36,6,64,64)f32 #2732=(1,6,64,64)f32 #2733=(36,6,64,64)f32 +Tensor.view Tensor.view_1863 1 1 2733 2734 shape=(1,36,6,64,64) $input=2733 #2733=(36,6,64,64)f32 #2734=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_15217 0 1 2735 @pnnx_fold_15217=(1,36,1,64,64)f32 #2735=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_4365 2 1 2734 2735 2736 expr=add(@0,@1) #2734=(1,36,6,64,64)f32 #2735=(1,36,1,64,64)f32 #2736=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1864 1 1 2736 2737 shape=(-1,6,64,64) $input=2736 #2736=(1,36,6,64,64)f32 #2737=(36,6,64,64)f32 +nn.Softmax layers_mmsa.1.residual_group.blocks.1.attn.softmax 1 1 2737 2738 dim=-1 #2737=(36,6,64,64)f32 #2738=(36,6,64,64)f32 +torch.matmul torch.matmul_2363 2 1 2738 2728 2739 $input=2738 $other=2728 #2738=(36,6,64,64)f32 #2728=(36,6,64,32)f32 #2739=(36,6,64,32)f32 +torch.transpose torch.transpose_3150 1 1 2739 2740 dim0=1 dim1=2 $input=2739 #2739=(36,6,64,32)f32 #2740=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_593 1 1 2740 2741 shape=(36,64,192) $input=2740 #2740=(36,64,6,32)f32 #2741=(36,64,192)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.1.attn.proj 1 1 2741 2742 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2741=(36,64,192)f32 #2742=(36,64,192)f32 +Tensor.reshape Tensor.view_1866 1 1 2742 2743 shape=(1,6,6,8,8,192) $input=2742 #2742=(36,64,192)f32 #2743=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2847 1 1 2743 2744 dims=(0,1,3,2,4,5) $input=2743 #2743=(1,6,6,8,8,192)f32 #2744=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_159 1 1 2744 2745 shape=(1,48,48,-1) $input=2744 #2744=(1,6,8,6,8,192)f32 #2745=(1,48,48,192)f32 +torch.roll torch.roll_2499 1 1 2745 2746 dims=(1,2) shifts=(4,4) $input=2745 #2745=(1,48,48,192)f32 #2746=(1,48,48,192)f32 +Tensor.view Tensor.view_1868 1 1 2746 2747 shape=(1,2304,192) $input=2746 #2746=(1,48,48,192)f32 #2747=(1,2304,192)f32 +pnnx.Expression pnnx_expr_4323 2 1 2716 2747 2748 expr=add(@0,@1) #2716=(1,2304,192)f32 #2747=(1,2304,192)f32 #2748=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.1.residual_group.blocks.1.norm2 1 1 2748 2749 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2748=(1,2304,192)f32 #2749=(1,2304,192)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.1.mlp.fc1 1 1 2749 2750 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2749=(1,2304,192)f32 #2750=(1,2304,384)f32 +nn.GELU layers_mmsa.1.residual_group.blocks.1.mlp.act 1 1 2750 2751 #2750=(1,2304,384)f32 #2751=(1,2304,384)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.1.mlp.fc2 1 1 2751 2752 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2751=(1,2304,384)f32 #2752=(1,2304,192)f32 +pnnx.Expression pnnx_expr_4321 2 1 2748 2752 2753 expr=add(@0,@1) #2748=(1,2304,192)f32 #2752=(1,2304,192)f32 #2753=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.1.residual_group.blocks.2.norm1 1 1 2753 2754 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2753=(1,2304,192)f32 #2754=(1,2304,192)f32 +Tensor.reshape Tensor.view_1870 1 1 2754 2755 shape=(1,6,8,6,8,192) $input=2754 #2754=(1,2304,192)f32 #2755=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2848 1 1 2755 2756 dims=(0,1,3,2,4,5) $input=2755 #2755=(1,6,8,6,8,192)f32 #2756=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1872 1 1 2756 2757 shape=(36,64,192) $input=2756 #2756=(1,6,6,8,8,192)f32 #2757=(36,64,192)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.2.attn.qkv 1 1 2757 2758 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2757=(36,64,192)f32 #2758=(36,64,576)f32 +Tensor.reshape Tensor.reshape_594 1 1 2758 2759 shape=(36,64,3,6,32) $input=2758 #2758=(36,64,576)f32 #2759=(36,64,3,6,32)f32 +torch.permute torch.permute_2849 1 1 2759 2760 dims=(2,0,3,1,4) $input=2759 #2759=(36,64,3,6,32)f32 #2760=(3,36,6,64,32)f32 +torch.unbind Tensor.select_890 1 3 2760 2761 2762 2763 dim=0 #2760=(3,36,6,64,32)f32 #2761=(36,6,64,32)f32 #2762=(36,6,64,32)f32 #2763=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_4233 1 1 2761 2764 expr=mul(@0,1.767767e-01) #2761=(36,6,64,32)f32 #2764=(36,6,64,32)f32 +torch.transpose torch.transpose_3151 1 1 2762 2765 dim0=-2 dim1=-1 $input=2762 #2762=(36,6,64,32)f32 #2765=(36,6,32,64)f32 +torch.matmul torch.matmul_2364 2 1 2764 2765 2766 $input=2764 $other=2765 #2764=(36,6,64,32)f32 #2765=(36,6,32,64)f32 #2766=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_15366 0 1 2767 @pnnx_fold_15366=(1,6,64,64)f32 #2767=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_4220 2 1 2766 2767 2768 expr=add(@0,@1) #2766=(36,6,64,64)f32 #2767=(1,6,64,64)f32 #2768=(36,6,64,64)f32 +nn.Softmax layers_mmsa.1.residual_group.blocks.2.attn.softmax 1 1 2768 2769 dim=-1 #2768=(36,6,64,64)f32 #2769=(36,6,64,64)f32 +torch.matmul torch.matmul_2365 2 1 2769 2763 2770 $input=2769 $other=2763 #2769=(36,6,64,64)f32 #2763=(36,6,64,32)f32 #2770=(36,6,64,32)f32 +torch.transpose torch.transpose_3152 1 1 2770 2771 dim0=1 dim1=2 $input=2770 #2770=(36,6,64,32)f32 #2771=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_595 1 1 2771 2772 shape=(36,64,192) $input=2771 #2771=(36,64,6,32)f32 #2772=(36,64,192)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.2.attn.proj 1 1 2772 2773 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2772=(36,64,192)f32 #2773=(36,64,192)f32 +Tensor.reshape Tensor.view_1876 1 1 2773 2774 shape=(1,6,6,8,8,192) $input=2773 #2773=(36,64,192)f32 #2774=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2851 1 1 2774 2775 dims=(0,1,3,2,4,5) $input=2774 #2774=(1,6,6,8,8,192)f32 #2775=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1878 1 1 2775 2776 shape=(1,2304,192) $input=2775 #2775=(1,6,8,6,8,192)f32 #2776=(1,2304,192)f32 +pnnx.Expression pnnx_expr_4187 2 1 2753 2776 2777 expr=add(@0,@1) #2753=(1,2304,192)f32 #2776=(1,2304,192)f32 #2777=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.1.residual_group.blocks.2.norm2 1 1 2777 2778 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2777=(1,2304,192)f32 #2778=(1,2304,192)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.2.mlp.fc1 1 1 2778 2779 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2778=(1,2304,192)f32 #2779=(1,2304,384)f32 +nn.GELU layers_mmsa.1.residual_group.blocks.2.mlp.act 1 1 2779 2780 #2779=(1,2304,384)f32 #2780=(1,2304,384)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.2.mlp.fc2 1 1 2780 2781 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2780=(1,2304,384)f32 #2781=(1,2304,192)f32 +pnnx.Expression pnnx_expr_4185 2 1 2777 2781 2782 expr=add(@0,@1) #2777=(1,2304,192)f32 #2781=(1,2304,192)f32 #2782=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.1.residual_group.blocks.3.norm1 1 1 2782 2783 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2782=(1,2304,192)f32 #2783=(1,2304,192)f32 +Tensor.view Tensor.view_1879 1 1 2783 2784 shape=(1,48,48,192) $input=2783 #2783=(1,2304,192)f32 #2784=(1,48,48,192)f32 +torch.roll torch.roll_2500 1 1 2784 2785 dims=(1,2) shifts=(-4,-4) $input=2784 #2784=(1,48,48,192)f32 #2785=(1,48,48,192)f32 +Tensor.view Tensor.view_1880 1 1 2785 2786 shape=(1,6,8,6,8,192) $input=2785 #2785=(1,48,48,192)f32 #2786=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2852 1 1 2786 2787 dims=(0,1,3,2,4,5) $input=2786 #2786=(1,6,8,6,8,192)f32 #2787=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1882 1 1 2787 2788 shape=(36,64,192) $input=2787 #2787=(1,6,6,8,8,192)f32 #2788=(36,64,192)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.3.attn.qkv 1 1 2788 2789 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2788=(36,64,192)f32 #2789=(36,64,576)f32 +Tensor.reshape Tensor.reshape_596 1 1 2789 2790 shape=(36,64,3,6,32) $input=2789 #2789=(36,64,576)f32 #2790=(36,64,3,6,32)f32 +torch.permute torch.permute_2853 1 1 2790 2791 dims=(2,0,3,1,4) $input=2790 #2790=(36,64,3,6,32)f32 #2791=(3,36,6,64,32)f32 +torch.unbind Tensor.select_893 1 3 2791 2792 2793 2794 dim=0 #2791=(3,36,6,64,32)f32 #2792=(36,6,64,32)f32 #2793=(36,6,64,32)f32 #2794=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_4087 1 1 2792 2795 expr=mul(@0,1.767767e-01) #2792=(36,6,64,32)f32 #2795=(36,6,64,32)f32 +torch.transpose torch.transpose_3153 1 1 2793 2796 dim0=-2 dim1=-1 $input=2793 #2793=(36,6,64,32)f32 #2796=(36,6,32,64)f32 +torch.matmul torch.matmul_2366 2 1 2795 2796 2797 $input=2795 $other=2796 #2795=(36,6,64,32)f32 #2796=(36,6,32,64)f32 #2797=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_15519 0 1 2798 @pnnx_fold_15519=(1,6,64,64)f32 #2798=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_4074 2 1 2797 2798 2799 expr=add(@0,@1) #2797=(36,6,64,64)f32 #2798=(1,6,64,64)f32 #2799=(36,6,64,64)f32 +Tensor.view Tensor.view_1885 1 1 2799 2800 shape=(1,36,6,64,64) $input=2799 #2799=(36,6,64,64)f32 #2800=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_15529 0 1 2801 @pnnx_fold_15529=(1,36,1,64,64)f32 #2801=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_4062 2 1 2800 2801 2802 expr=add(@0,@1) #2800=(1,36,6,64,64)f32 #2801=(1,36,1,64,64)f32 #2802=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1886 1 1 2802 2803 shape=(-1,6,64,64) $input=2802 #2802=(1,36,6,64,64)f32 #2803=(36,6,64,64)f32 +nn.Softmax layers_mmsa.1.residual_group.blocks.3.attn.softmax 1 1 2803 2804 dim=-1 #2803=(36,6,64,64)f32 #2804=(36,6,64,64)f32 +torch.matmul torch.matmul_2367 2 1 2804 2794 2805 $input=2804 $other=2794 #2804=(36,6,64,64)f32 #2794=(36,6,64,32)f32 #2805=(36,6,64,32)f32 +torch.transpose torch.transpose_3154 1 1 2805 2806 dim0=1 dim1=2 $input=2805 #2805=(36,6,64,32)f32 #2806=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_597 1 1 2806 2807 shape=(36,64,192) $input=2806 #2806=(36,64,6,32)f32 #2807=(36,64,192)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.3.attn.proj 1 1 2807 2808 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2807=(36,64,192)f32 #2808=(36,64,192)f32 +Tensor.reshape Tensor.view_1888 1 1 2808 2809 shape=(1,6,6,8,8,192) $input=2808 #2808=(36,64,192)f32 #2809=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2855 1 1 2809 2810 dims=(0,1,3,2,4,5) $input=2809 #2809=(1,6,6,8,8,192)f32 #2810=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_163 1 1 2810 2811 shape=(1,48,48,-1) $input=2810 #2810=(1,6,8,6,8,192)f32 #2811=(1,48,48,192)f32 +torch.roll torch.roll_2501 1 1 2811 2812 dims=(1,2) shifts=(4,4) $input=2811 #2811=(1,48,48,192)f32 #2812=(1,48,48,192)f32 +Tensor.view Tensor.view_1890 1 1 2812 2813 shape=(1,2304,192) $input=2812 #2812=(1,48,48,192)f32 #2813=(1,2304,192)f32 +pnnx.Expression pnnx_expr_4020 2 1 2782 2813 2814 expr=add(@0,@1) #2782=(1,2304,192)f32 #2813=(1,2304,192)f32 #2814=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.1.residual_group.blocks.3.norm2 1 1 2814 2815 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2814=(1,2304,192)f32 #2815=(1,2304,192)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.3.mlp.fc1 1 1 2815 2816 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2815=(1,2304,192)f32 #2816=(1,2304,384)f32 +nn.GELU layers_mmsa.1.residual_group.blocks.3.mlp.act 1 1 2816 2817 #2816=(1,2304,384)f32 #2817=(1,2304,384)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.3.mlp.fc2 1 1 2817 2818 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2817=(1,2304,384)f32 #2818=(1,2304,192)f32 +pnnx.Expression pnnx_expr_4018 2 1 2814 2818 2819 expr=add(@0,@1) #2814=(1,2304,192)f32 #2818=(1,2304,192)f32 #2819=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.1.residual_group.blocks.4.norm1 1 1 2819 2820 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2819=(1,2304,192)f32 #2820=(1,2304,192)f32 +Tensor.reshape Tensor.view_1892 1 1 2820 2821 shape=(1,6,8,6,8,192) $input=2820 #2820=(1,2304,192)f32 #2821=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2856 1 1 2821 2822 dims=(0,1,3,2,4,5) $input=2821 #2821=(1,6,8,6,8,192)f32 #2822=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1894 1 1 2822 2823 shape=(36,64,192) $input=2822 #2822=(1,6,6,8,8,192)f32 #2823=(36,64,192)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.4.attn.qkv 1 1 2823 2824 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2823=(36,64,192)f32 #2824=(36,64,576)f32 +Tensor.reshape Tensor.reshape_598 1 1 2824 2825 shape=(36,64,3,6,32) $input=2824 #2824=(36,64,576)f32 #2825=(36,64,3,6,32)f32 +torch.permute torch.permute_2857 1 1 2825 2826 dims=(2,0,3,1,4) $input=2825 #2825=(36,64,3,6,32)f32 #2826=(3,36,6,64,32)f32 +torch.unbind Tensor.select_896 1 3 2826 2827 2828 2829 dim=0 #2826=(3,36,6,64,32)f32 #2827=(36,6,64,32)f32 #2828=(36,6,64,32)f32 #2829=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_3930 1 1 2827 2830 expr=mul(@0,1.767767e-01) #2827=(36,6,64,32)f32 #2830=(36,6,64,32)f32 +torch.transpose torch.transpose_3155 1 1 2828 2831 dim0=-2 dim1=-1 $input=2828 #2828=(36,6,64,32)f32 #2831=(36,6,32,64)f32 +torch.matmul torch.matmul_2368 2 1 2830 2831 2832 $input=2830 $other=2831 #2830=(36,6,64,32)f32 #2831=(36,6,32,64)f32 #2832=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_15678 0 1 2833 @pnnx_fold_15678=(1,6,64,64)f32 #2833=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_3917 2 1 2832 2833 2834 expr=add(@0,@1) #2832=(36,6,64,64)f32 #2833=(1,6,64,64)f32 #2834=(36,6,64,64)f32 +nn.Softmax layers_mmsa.1.residual_group.blocks.4.attn.softmax 1 1 2834 2835 dim=-1 #2834=(36,6,64,64)f32 #2835=(36,6,64,64)f32 +torch.matmul torch.matmul_2369 2 1 2835 2829 2836 $input=2835 $other=2829 #2835=(36,6,64,64)f32 #2829=(36,6,64,32)f32 #2836=(36,6,64,32)f32 +torch.transpose torch.transpose_3156 1 1 2836 2837 dim0=1 dim1=2 $input=2836 #2836=(36,6,64,32)f32 #2837=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_599 1 1 2837 2838 shape=(36,64,192) $input=2837 #2837=(36,64,6,32)f32 #2838=(36,64,192)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.4.attn.proj 1 1 2838 2839 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2838=(36,64,192)f32 #2839=(36,64,192)f32 +Tensor.reshape Tensor.view_1898 1 1 2839 2840 shape=(1,6,6,8,8,192) $input=2839 #2839=(36,64,192)f32 #2840=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2859 1 1 2840 2841 dims=(0,1,3,2,4,5) $input=2840 #2840=(1,6,6,8,8,192)f32 #2841=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1900 1 1 2841 2842 shape=(1,2304,192) $input=2841 #2841=(1,6,8,6,8,192)f32 #2842=(1,2304,192)f32 +pnnx.Expression pnnx_expr_3884 2 1 2819 2842 2843 expr=add(@0,@1) #2819=(1,2304,192)f32 #2842=(1,2304,192)f32 #2843=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.1.residual_group.blocks.4.norm2 1 1 2843 2844 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2843=(1,2304,192)f32 #2844=(1,2304,192)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.4.mlp.fc1 1 1 2844 2845 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2844=(1,2304,192)f32 #2845=(1,2304,384)f32 +nn.GELU layers_mmsa.1.residual_group.blocks.4.mlp.act 1 1 2845 2846 #2845=(1,2304,384)f32 #2846=(1,2304,384)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.4.mlp.fc2 1 1 2846 2847 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2846=(1,2304,384)f32 #2847=(1,2304,192)f32 +pnnx.Expression pnnx_expr_3882 2 1 2843 2847 2848 expr=add(@0,@1) #2843=(1,2304,192)f32 #2847=(1,2304,192)f32 #2848=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.1.residual_group.blocks.5.norm1 1 1 2848 2849 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2848=(1,2304,192)f32 #2849=(1,2304,192)f32 +Tensor.view Tensor.view_1901 1 1 2849 2850 shape=(1,48,48,192) $input=2849 #2849=(1,2304,192)f32 #2850=(1,48,48,192)f32 +torch.roll torch.roll_2502 1 1 2850 2851 dims=(1,2) shifts=(-4,-4) $input=2850 #2850=(1,48,48,192)f32 #2851=(1,48,48,192)f32 +Tensor.view Tensor.view_1902 1 1 2851 2852 shape=(1,6,8,6,8,192) $input=2851 #2851=(1,48,48,192)f32 #2852=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2860 1 1 2852 2853 dims=(0,1,3,2,4,5) $input=2852 #2852=(1,6,8,6,8,192)f32 #2853=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1904 1 1 2853 2854 shape=(36,64,192) $input=2853 #2853=(1,6,6,8,8,192)f32 #2854=(36,64,192)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.5.attn.qkv 1 1 2854 2855 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2854=(36,64,192)f32 #2855=(36,64,576)f32 +Tensor.reshape Tensor.reshape_600 1 1 2855 2856 shape=(36,64,3,6,32) $input=2855 #2855=(36,64,576)f32 #2856=(36,64,3,6,32)f32 +torch.permute torch.permute_2861 1 1 2856 2857 dims=(2,0,3,1,4) $input=2856 #2856=(36,64,3,6,32)f32 #2857=(3,36,6,64,32)f32 +torch.unbind Tensor.select_899 1 3 2857 2858 2859 2860 dim=0 #2857=(3,36,6,64,32)f32 #2858=(36,6,64,32)f32 #2859=(36,6,64,32)f32 #2860=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_3784 1 1 2858 2861 expr=mul(@0,1.767767e-01) #2858=(36,6,64,32)f32 #2861=(36,6,64,32)f32 +torch.transpose torch.transpose_3157 1 1 2859 2862 dim0=-2 dim1=-1 $input=2859 #2859=(36,6,64,32)f32 #2862=(36,6,32,64)f32 +torch.matmul torch.matmul_2370 2 1 2861 2862 2863 $input=2861 $other=2862 #2861=(36,6,64,32)f32 #2862=(36,6,32,64)f32 #2863=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_15831 0 1 2864 @pnnx_fold_15831=(1,6,64,64)f32 #2864=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_3771 2 1 2863 2864 2865 expr=add(@0,@1) #2863=(36,6,64,64)f32 #2864=(1,6,64,64)f32 #2865=(36,6,64,64)f32 +Tensor.view Tensor.view_1907 1 1 2865 2866 shape=(1,36,6,64,64) $input=2865 #2865=(36,6,64,64)f32 #2866=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_15841 0 1 2867 @pnnx_fold_15841=(1,36,1,64,64)f32 #2867=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_3759 2 1 2866 2867 2868 expr=add(@0,@1) #2866=(1,36,6,64,64)f32 #2867=(1,36,1,64,64)f32 #2868=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1908 1 1 2868 2869 shape=(-1,6,64,64) $input=2868 #2868=(1,36,6,64,64)f32 #2869=(36,6,64,64)f32 +nn.Softmax layers_mmsa.1.residual_group.blocks.5.attn.softmax 1 1 2869 2870 dim=-1 #2869=(36,6,64,64)f32 #2870=(36,6,64,64)f32 +torch.matmul torch.matmul_2371 2 1 2870 2860 2871 $input=2870 $other=2860 #2870=(36,6,64,64)f32 #2860=(36,6,64,32)f32 #2871=(36,6,64,32)f32 +torch.transpose torch.transpose_3158 1 1 2871 2872 dim0=1 dim1=2 $input=2871 #2871=(36,6,64,32)f32 #2872=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_601 1 1 2872 2873 shape=(36,64,192) $input=2872 #2872=(36,64,6,32)f32 #2873=(36,64,192)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.5.attn.proj 1 1 2873 2874 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2873=(36,64,192)f32 #2874=(36,64,192)f32 +Tensor.reshape Tensor.view_1910 1 1 2874 2875 shape=(1,6,6,8,8,192) $input=2874 #2874=(36,64,192)f32 #2875=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2863 1 1 2875 2876 dims=(0,1,3,2,4,5) $input=2875 #2875=(1,6,6,8,8,192)f32 #2876=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_167 1 1 2876 2877 shape=(1,48,48,-1) $input=2876 #2876=(1,6,8,6,8,192)f32 #2877=(1,48,48,192)f32 +torch.roll torch.roll_2503 1 1 2877 2878 dims=(1,2) shifts=(4,4) $input=2877 #2877=(1,48,48,192)f32 #2878=(1,48,48,192)f32 +Tensor.view Tensor.view_1912 1 1 2878 2879 shape=(1,2304,192) $input=2878 #2878=(1,48,48,192)f32 #2879=(1,2304,192)f32 +pnnx.Expression pnnx_expr_3717 2 1 2848 2879 2880 expr=add(@0,@1) #2848=(1,2304,192)f32 #2879=(1,2304,192)f32 #2880=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.1.residual_group.blocks.5.norm2 1 1 2880 2881 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2880=(1,2304,192)f32 #2881=(1,2304,192)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.5.mlp.fc1 1 1 2881 2882 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2881=(1,2304,192)f32 #2882=(1,2304,384)f32 +nn.GELU layers_mmsa.1.residual_group.blocks.5.mlp.act 1 1 2882 2883 #2882=(1,2304,384)f32 #2883=(1,2304,384)f32 +nn.Linear layers_mmsa.1.residual_group.blocks.5.mlp.fc2 1 1 2883 2884 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2883=(1,2304,384)f32 #2884=(1,2304,192)f32 +pnnx.Expression pnnx_expr_3715 2 1 2880 2884 2885 expr=add(@0,@1) #2880=(1,2304,192)f32 #2884=(1,2304,192)f32 #2885=(1,2304,192)f32 +torch.transpose torch.transpose_3159 1 1 2885 2886 dim0=1 dim1=2 $input=2885 #2885=(1,2304,192)f32 #2886=(1,192,2304)f32 +Tensor.view Tensor.view_1913 1 1 2886 2887 shape=(1,192,48,48) $input=2886 #2886=(1,192,2304)f32 #2887=(1,192,48,48)f32 +nn.Conv2d layers_mmsa.1.conv 1 1 2887 2888 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #2887=(1,192,48,48)f32 #2888=(1,192,48,48)f32 +torch.flatten torch.flatten_2199 1 1 2888 2889 end_dim=-1 start_dim=2 $input=2888 #2888=(1,192,48,48)f32 #2889=(1,192,2304)f32 +torch.transpose torch.transpose_3160 1 1 2889 2890 dim0=1 dim1=2 $input=2889 #2889=(1,192,2304)f32 #2890=(1,2304,192)f32 +pnnx.Expression pnnx_expr_3702 2 1 2890 2687 2891 expr=add(@0,@1) #2890=(1,2304,192)f32 #2687=(1,2304,192)f32 #2891=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.2.residual_group.blocks.0.norm1 1 1 2891 2892 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2891=(1,2304,192)f32 #2892=(1,2304,192)f32 +Tensor.reshape Tensor.view_1915 1 1 2892 2893 shape=(1,6,8,6,8,192) $input=2892 #2892=(1,2304,192)f32 #2893=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2864 1 1 2893 2894 dims=(0,1,3,2,4,5) $input=2893 #2893=(1,6,8,6,8,192)f32 #2894=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1917 1 1 2894 2895 shape=(36,64,192) $input=2894 #2894=(1,6,6,8,8,192)f32 #2895=(36,64,192)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.0.attn.qkv 1 1 2895 2896 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2895=(36,64,192)f32 #2896=(36,64,576)f32 +Tensor.reshape Tensor.reshape_602 1 1 2896 2897 shape=(36,64,3,6,32) $input=2896 #2896=(36,64,576)f32 #2897=(36,64,3,6,32)f32 +torch.permute torch.permute_2865 1 1 2897 2898 dims=(2,0,3,1,4) $input=2897 #2897=(36,64,3,6,32)f32 #2898=(3,36,6,64,32)f32 +torch.unbind Tensor.select_902 1 3 2898 2899 2900 2901 dim=0 #2898=(3,36,6,64,32)f32 #2899=(36,6,64,32)f32 #2900=(36,6,64,32)f32 #2901=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_3613 1 1 2899 2902 expr=mul(@0,1.767767e-01) #2899=(36,6,64,32)f32 #2902=(36,6,64,32)f32 +torch.transpose torch.transpose_3161 1 1 2900 2903 dim0=-2 dim1=-1 $input=2900 #2900=(36,6,64,32)f32 #2903=(36,6,32,64)f32 +torch.matmul torch.matmul_2372 2 1 2902 2903 2904 $input=2902 $other=2903 #2902=(36,6,64,32)f32 #2903=(36,6,32,64)f32 #2904=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_16024 0 1 2905 @pnnx_fold_16024=(1,6,64,64)f32 #2905=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_3600 2 1 2904 2905 2906 expr=add(@0,@1) #2904=(36,6,64,64)f32 #2905=(1,6,64,64)f32 #2906=(36,6,64,64)f32 +nn.Softmax layers_mmsa.2.residual_group.blocks.0.attn.softmax 1 1 2906 2907 dim=-1 #2906=(36,6,64,64)f32 #2907=(36,6,64,64)f32 +torch.matmul torch.matmul_2373 2 1 2907 2901 2908 $input=2907 $other=2901 #2907=(36,6,64,64)f32 #2901=(36,6,64,32)f32 #2908=(36,6,64,32)f32 +torch.transpose torch.transpose_3162 1 1 2908 2909 dim0=1 dim1=2 $input=2908 #2908=(36,6,64,32)f32 #2909=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_603 1 1 2909 2910 shape=(36,64,192) $input=2909 #2909=(36,64,6,32)f32 #2910=(36,64,192)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.0.attn.proj 1 1 2910 2911 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2910=(36,64,192)f32 #2911=(36,64,192)f32 +Tensor.reshape Tensor.view_1921 1 1 2911 2912 shape=(1,6,6,8,8,192) $input=2911 #2911=(36,64,192)f32 #2912=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2867 1 1 2912 2913 dims=(0,1,3,2,4,5) $input=2912 #2912=(1,6,6,8,8,192)f32 #2913=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1923 1 1 2913 2914 shape=(1,2304,192) $input=2913 #2913=(1,6,8,6,8,192)f32 #2914=(1,2304,192)f32 +pnnx.Expression pnnx_expr_3567 2 1 2891 2914 2915 expr=add(@0,@1) #2891=(1,2304,192)f32 #2914=(1,2304,192)f32 #2915=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.2.residual_group.blocks.0.norm2 1 1 2915 2916 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2915=(1,2304,192)f32 #2916=(1,2304,192)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.0.mlp.fc1 1 1 2916 2917 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2916=(1,2304,192)f32 #2917=(1,2304,384)f32 +nn.GELU layers_mmsa.2.residual_group.blocks.0.mlp.act 1 1 2917 2918 #2917=(1,2304,384)f32 #2918=(1,2304,384)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.0.mlp.fc2 1 1 2918 2919 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2918=(1,2304,384)f32 #2919=(1,2304,192)f32 +pnnx.Expression pnnx_expr_3565 2 1 2915 2919 2920 expr=add(@0,@1) #2915=(1,2304,192)f32 #2919=(1,2304,192)f32 #2920=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.2.residual_group.blocks.1.norm1 1 1 2920 2921 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2920=(1,2304,192)f32 #2921=(1,2304,192)f32 +Tensor.view Tensor.view_1924 1 1 2921 2922 shape=(1,48,48,192) $input=2921 #2921=(1,2304,192)f32 #2922=(1,48,48,192)f32 +torch.roll torch.roll_2504 1 1 2922 2923 dims=(1,2) shifts=(-4,-4) $input=2922 #2922=(1,48,48,192)f32 #2923=(1,48,48,192)f32 +Tensor.view Tensor.view_1925 1 1 2923 2924 shape=(1,6,8,6,8,192) $input=2923 #2923=(1,48,48,192)f32 #2924=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2868 1 1 2924 2925 dims=(0,1,3,2,4,5) $input=2924 #2924=(1,6,8,6,8,192)f32 #2925=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1927 1 1 2925 2926 shape=(36,64,192) $input=2925 #2925=(1,6,6,8,8,192)f32 #2926=(36,64,192)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.1.attn.qkv 1 1 2926 2927 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2926=(36,64,192)f32 #2927=(36,64,576)f32 +Tensor.reshape Tensor.reshape_604 1 1 2927 2928 shape=(36,64,3,6,32) $input=2927 #2927=(36,64,576)f32 #2928=(36,64,3,6,32)f32 +torch.permute torch.permute_2869 1 1 2928 2929 dims=(2,0,3,1,4) $input=2928 #2928=(36,64,3,6,32)f32 #2929=(3,36,6,64,32)f32 +torch.unbind Tensor.select_905 1 3 2929 2930 2931 2932 dim=0 #2929=(3,36,6,64,32)f32 #2930=(36,6,64,32)f32 #2931=(36,6,64,32)f32 #2932=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_3467 1 1 2930 2933 expr=mul(@0,1.767767e-01) #2930=(36,6,64,32)f32 #2933=(36,6,64,32)f32 +torch.transpose torch.transpose_3163 1 1 2931 2934 dim0=-2 dim1=-1 $input=2931 #2931=(36,6,64,32)f32 #2934=(36,6,32,64)f32 +torch.matmul torch.matmul_2374 2 1 2933 2934 2935 $input=2933 $other=2934 #2933=(36,6,64,32)f32 #2934=(36,6,32,64)f32 #2935=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_16177 0 1 2936 @pnnx_fold_16177=(1,6,64,64)f32 #2936=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_3454 2 1 2935 2936 2937 expr=add(@0,@1) #2935=(36,6,64,64)f32 #2936=(1,6,64,64)f32 #2937=(36,6,64,64)f32 +Tensor.view Tensor.view_1930 1 1 2937 2938 shape=(1,36,6,64,64) $input=2937 #2937=(36,6,64,64)f32 #2938=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_16187 0 1 2939 @pnnx_fold_16187=(1,36,1,64,64)f32 #2939=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_3442 2 1 2938 2939 2940 expr=add(@0,@1) #2938=(1,36,6,64,64)f32 #2939=(1,36,1,64,64)f32 #2940=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1931 1 1 2940 2941 shape=(-1,6,64,64) $input=2940 #2940=(1,36,6,64,64)f32 #2941=(36,6,64,64)f32 +nn.Softmax layers_mmsa.2.residual_group.blocks.1.attn.softmax 1 1 2941 2942 dim=-1 #2941=(36,6,64,64)f32 #2942=(36,6,64,64)f32 +torch.matmul torch.matmul_2375 2 1 2942 2932 2943 $input=2942 $other=2932 #2942=(36,6,64,64)f32 #2932=(36,6,64,32)f32 #2943=(36,6,64,32)f32 +torch.transpose torch.transpose_3164 1 1 2943 2944 dim0=1 dim1=2 $input=2943 #2943=(36,6,64,32)f32 #2944=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_605 1 1 2944 2945 shape=(36,64,192) $input=2944 #2944=(36,64,6,32)f32 #2945=(36,64,192)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.1.attn.proj 1 1 2945 2946 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2945=(36,64,192)f32 #2946=(36,64,192)f32 +Tensor.reshape Tensor.view_1933 1 1 2946 2947 shape=(1,6,6,8,8,192) $input=2946 #2946=(36,64,192)f32 #2947=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2871 1 1 2947 2948 dims=(0,1,3,2,4,5) $input=2947 #2947=(1,6,6,8,8,192)f32 #2948=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_171 1 1 2948 2949 shape=(1,48,48,-1) $input=2948 #2948=(1,6,8,6,8,192)f32 #2949=(1,48,48,192)f32 +torch.roll torch.roll_2505 1 1 2949 2950 dims=(1,2) shifts=(4,4) $input=2949 #2949=(1,48,48,192)f32 #2950=(1,48,48,192)f32 +Tensor.view Tensor.view_1935 1 1 2950 2951 shape=(1,2304,192) $input=2950 #2950=(1,48,48,192)f32 #2951=(1,2304,192)f32 +pnnx.Expression pnnx_expr_3400 2 1 2920 2951 2952 expr=add(@0,@1) #2920=(1,2304,192)f32 #2951=(1,2304,192)f32 #2952=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.2.residual_group.blocks.1.norm2 1 1 2952 2953 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2952=(1,2304,192)f32 #2953=(1,2304,192)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.1.mlp.fc1 1 1 2953 2954 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2953=(1,2304,192)f32 #2954=(1,2304,384)f32 +nn.GELU layers_mmsa.2.residual_group.blocks.1.mlp.act 1 1 2954 2955 #2954=(1,2304,384)f32 #2955=(1,2304,384)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.1.mlp.fc2 1 1 2955 2956 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2955=(1,2304,384)f32 #2956=(1,2304,192)f32 +pnnx.Expression pnnx_expr_3398 2 1 2952 2956 2957 expr=add(@0,@1) #2952=(1,2304,192)f32 #2956=(1,2304,192)f32 #2957=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.2.residual_group.blocks.2.norm1 1 1 2957 2958 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2957=(1,2304,192)f32 #2958=(1,2304,192)f32 +Tensor.reshape Tensor.view_1937 1 1 2958 2959 shape=(1,6,8,6,8,192) $input=2958 #2958=(1,2304,192)f32 #2959=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2872 1 1 2959 2960 dims=(0,1,3,2,4,5) $input=2959 #2959=(1,6,8,6,8,192)f32 #2960=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1939 1 1 2960 2961 shape=(36,64,192) $input=2960 #2960=(1,6,6,8,8,192)f32 #2961=(36,64,192)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.2.attn.qkv 1 1 2961 2962 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2961=(36,64,192)f32 #2962=(36,64,576)f32 +Tensor.reshape Tensor.reshape_606 1 1 2962 2963 shape=(36,64,3,6,32) $input=2962 #2962=(36,64,576)f32 #2963=(36,64,3,6,32)f32 +torch.permute torch.permute_2873 1 1 2963 2964 dims=(2,0,3,1,4) $input=2963 #2963=(36,64,3,6,32)f32 #2964=(3,36,6,64,32)f32 +torch.unbind Tensor.select_908 1 3 2964 2965 2966 2967 dim=0 #2964=(3,36,6,64,32)f32 #2965=(36,6,64,32)f32 #2966=(36,6,64,32)f32 #2967=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_3310 1 1 2965 2968 expr=mul(@0,1.767767e-01) #2965=(36,6,64,32)f32 #2968=(36,6,64,32)f32 +torch.transpose torch.transpose_3165 1 1 2966 2969 dim0=-2 dim1=-1 $input=2966 #2966=(36,6,64,32)f32 #2969=(36,6,32,64)f32 +torch.matmul torch.matmul_2376 2 1 2968 2969 2970 $input=2968 $other=2969 #2968=(36,6,64,32)f32 #2969=(36,6,32,64)f32 #2970=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_16336 0 1 2971 @pnnx_fold_16336=(1,6,64,64)f32 #2971=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_3297 2 1 2970 2971 2972 expr=add(@0,@1) #2970=(36,6,64,64)f32 #2971=(1,6,64,64)f32 #2972=(36,6,64,64)f32 +nn.Softmax layers_mmsa.2.residual_group.blocks.2.attn.softmax 1 1 2972 2973 dim=-1 #2972=(36,6,64,64)f32 #2973=(36,6,64,64)f32 +torch.matmul torch.matmul_2377 2 1 2973 2967 2974 $input=2973 $other=2967 #2973=(36,6,64,64)f32 #2967=(36,6,64,32)f32 #2974=(36,6,64,32)f32 +torch.transpose torch.transpose_3166 1 1 2974 2975 dim0=1 dim1=2 $input=2974 #2974=(36,6,64,32)f32 #2975=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_607 1 1 2975 2976 shape=(36,64,192) $input=2975 #2975=(36,64,6,32)f32 #2976=(36,64,192)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.2.attn.proj 1 1 2976 2977 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #2976=(36,64,192)f32 #2977=(36,64,192)f32 +Tensor.reshape Tensor.view_1943 1 1 2977 2978 shape=(1,6,6,8,8,192) $input=2977 #2977=(36,64,192)f32 #2978=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2875 1 1 2978 2979 dims=(0,1,3,2,4,5) $input=2978 #2978=(1,6,6,8,8,192)f32 #2979=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1945 1 1 2979 2980 shape=(1,2304,192) $input=2979 #2979=(1,6,8,6,8,192)f32 #2980=(1,2304,192)f32 +pnnx.Expression pnnx_expr_3264 2 1 2957 2980 2981 expr=add(@0,@1) #2957=(1,2304,192)f32 #2980=(1,2304,192)f32 #2981=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.2.residual_group.blocks.2.norm2 1 1 2981 2982 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2981=(1,2304,192)f32 #2982=(1,2304,192)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.2.mlp.fc1 1 1 2982 2983 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #2982=(1,2304,192)f32 #2983=(1,2304,384)f32 +nn.GELU layers_mmsa.2.residual_group.blocks.2.mlp.act 1 1 2983 2984 #2983=(1,2304,384)f32 #2984=(1,2304,384)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.2.mlp.fc2 1 1 2984 2985 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #2984=(1,2304,384)f32 #2985=(1,2304,192)f32 +pnnx.Expression pnnx_expr_3262 2 1 2981 2985 2986 expr=add(@0,@1) #2981=(1,2304,192)f32 #2985=(1,2304,192)f32 #2986=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.2.residual_group.blocks.3.norm1 1 1 2986 2987 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #2986=(1,2304,192)f32 #2987=(1,2304,192)f32 +Tensor.view Tensor.view_1946 1 1 2987 2988 shape=(1,48,48,192) $input=2987 #2987=(1,2304,192)f32 #2988=(1,48,48,192)f32 +torch.roll torch.roll_2506 1 1 2988 2989 dims=(1,2) shifts=(-4,-4) $input=2988 #2988=(1,48,48,192)f32 #2989=(1,48,48,192)f32 +Tensor.view Tensor.view_1947 1 1 2989 2990 shape=(1,6,8,6,8,192) $input=2989 #2989=(1,48,48,192)f32 #2990=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2876 1 1 2990 2991 dims=(0,1,3,2,4,5) $input=2990 #2990=(1,6,8,6,8,192)f32 #2991=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1949 1 1 2991 2992 shape=(36,64,192) $input=2991 #2991=(1,6,6,8,8,192)f32 #2992=(36,64,192)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.3.attn.qkv 1 1 2992 2993 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #2992=(36,64,192)f32 #2993=(36,64,576)f32 +Tensor.reshape Tensor.reshape_608 1 1 2993 2994 shape=(36,64,3,6,32) $input=2993 #2993=(36,64,576)f32 #2994=(36,64,3,6,32)f32 +torch.permute torch.permute_2877 1 1 2994 2995 dims=(2,0,3,1,4) $input=2994 #2994=(36,64,3,6,32)f32 #2995=(3,36,6,64,32)f32 +torch.unbind Tensor.select_911 1 3 2995 2996 2997 2998 dim=0 #2995=(3,36,6,64,32)f32 #2996=(36,6,64,32)f32 #2997=(36,6,64,32)f32 #2998=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_3164 1 1 2996 2999 expr=mul(@0,1.767767e-01) #2996=(36,6,64,32)f32 #2999=(36,6,64,32)f32 +torch.transpose torch.transpose_3167 1 1 2997 3000 dim0=-2 dim1=-1 $input=2997 #2997=(36,6,64,32)f32 #3000=(36,6,32,64)f32 +torch.matmul torch.matmul_2378 2 1 2999 3000 3001 $input=2999 $other=3000 #2999=(36,6,64,32)f32 #3000=(36,6,32,64)f32 #3001=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_16489 0 1 3002 @pnnx_fold_16489=(1,6,64,64)f32 #3002=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_3151 2 1 3001 3002 3003 expr=add(@0,@1) #3001=(36,6,64,64)f32 #3002=(1,6,64,64)f32 #3003=(36,6,64,64)f32 +Tensor.view Tensor.view_1952 1 1 3003 3004 shape=(1,36,6,64,64) $input=3003 #3003=(36,6,64,64)f32 #3004=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_16499 0 1 3005 @pnnx_fold_16499=(1,36,1,64,64)f32 #3005=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_3139 2 1 3004 3005 3006 expr=add(@0,@1) #3004=(1,36,6,64,64)f32 #3005=(1,36,1,64,64)f32 #3006=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1953 1 1 3006 3007 shape=(-1,6,64,64) $input=3006 #3006=(1,36,6,64,64)f32 #3007=(36,6,64,64)f32 +nn.Softmax layers_mmsa.2.residual_group.blocks.3.attn.softmax 1 1 3007 3008 dim=-1 #3007=(36,6,64,64)f32 #3008=(36,6,64,64)f32 +torch.matmul torch.matmul_2379 2 1 3008 2998 3009 $input=3008 $other=2998 #3008=(36,6,64,64)f32 #2998=(36,6,64,32)f32 #3009=(36,6,64,32)f32 +torch.transpose torch.transpose_3168 1 1 3009 3010 dim0=1 dim1=2 $input=3009 #3009=(36,6,64,32)f32 #3010=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_609 1 1 3010 3011 shape=(36,64,192) $input=3010 #3010=(36,64,6,32)f32 #3011=(36,64,192)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.3.attn.proj 1 1 3011 3012 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3011=(36,64,192)f32 #3012=(36,64,192)f32 +Tensor.reshape Tensor.view_1955 1 1 3012 3013 shape=(1,6,6,8,8,192) $input=3012 #3012=(36,64,192)f32 #3013=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2879 1 1 3013 3014 dims=(0,1,3,2,4,5) $input=3013 #3013=(1,6,6,8,8,192)f32 #3014=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_175 1 1 3014 3015 shape=(1,48,48,-1) $input=3014 #3014=(1,6,8,6,8,192)f32 #3015=(1,48,48,192)f32 +torch.roll torch.roll_2507 1 1 3015 3016 dims=(1,2) shifts=(4,4) $input=3015 #3015=(1,48,48,192)f32 #3016=(1,48,48,192)f32 +Tensor.view Tensor.view_1957 1 1 3016 3017 shape=(1,2304,192) $input=3016 #3016=(1,48,48,192)f32 #3017=(1,2304,192)f32 +pnnx.Expression pnnx_expr_3097 2 1 2986 3017 3018 expr=add(@0,@1) #2986=(1,2304,192)f32 #3017=(1,2304,192)f32 #3018=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.2.residual_group.blocks.3.norm2 1 1 3018 3019 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3018=(1,2304,192)f32 #3019=(1,2304,192)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.3.mlp.fc1 1 1 3019 3020 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3019=(1,2304,192)f32 #3020=(1,2304,384)f32 +nn.GELU layers_mmsa.2.residual_group.blocks.3.mlp.act 1 1 3020 3021 #3020=(1,2304,384)f32 #3021=(1,2304,384)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.3.mlp.fc2 1 1 3021 3022 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3021=(1,2304,384)f32 #3022=(1,2304,192)f32 +pnnx.Expression pnnx_expr_3095 2 1 3018 3022 3023 expr=add(@0,@1) #3018=(1,2304,192)f32 #3022=(1,2304,192)f32 #3023=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.2.residual_group.blocks.4.norm1 1 1 3023 3024 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3023=(1,2304,192)f32 #3024=(1,2304,192)f32 +Tensor.reshape Tensor.view_1959 1 1 3024 3025 shape=(1,6,8,6,8,192) $input=3024 #3024=(1,2304,192)f32 #3025=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2880 1 1 3025 3026 dims=(0,1,3,2,4,5) $input=3025 #3025=(1,6,8,6,8,192)f32 #3026=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1961 1 1 3026 3027 shape=(36,64,192) $input=3026 #3026=(1,6,6,8,8,192)f32 #3027=(36,64,192)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.4.attn.qkv 1 1 3027 3028 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3027=(36,64,192)f32 #3028=(36,64,576)f32 +Tensor.reshape Tensor.reshape_610 1 1 3028 3029 shape=(36,64,3,6,32) $input=3028 #3028=(36,64,576)f32 #3029=(36,64,3,6,32)f32 +torch.permute torch.permute_2881 1 1 3029 3030 dims=(2,0,3,1,4) $input=3029 #3029=(36,64,3,6,32)f32 #3030=(3,36,6,64,32)f32 +torch.unbind Tensor.select_914 1 3 3030 3031 3032 3033 dim=0 #3030=(3,36,6,64,32)f32 #3031=(36,6,64,32)f32 #3032=(36,6,64,32)f32 #3033=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_3007 1 1 3031 3034 expr=mul(@0,1.767767e-01) #3031=(36,6,64,32)f32 #3034=(36,6,64,32)f32 +torch.transpose torch.transpose_3169 1 1 3032 3035 dim0=-2 dim1=-1 $input=3032 #3032=(36,6,64,32)f32 #3035=(36,6,32,64)f32 +torch.matmul torch.matmul_2380 2 1 3034 3035 3036 $input=3034 $other=3035 #3034=(36,6,64,32)f32 #3035=(36,6,32,64)f32 #3036=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_16648 0 1 3037 @pnnx_fold_16648=(1,6,64,64)f32 #3037=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_2994 2 1 3036 3037 3038 expr=add(@0,@1) #3036=(36,6,64,64)f32 #3037=(1,6,64,64)f32 #3038=(36,6,64,64)f32 +nn.Softmax layers_mmsa.2.residual_group.blocks.4.attn.softmax 1 1 3038 3039 dim=-1 #3038=(36,6,64,64)f32 #3039=(36,6,64,64)f32 +torch.matmul torch.matmul_2381 2 1 3039 3033 3040 $input=3039 $other=3033 #3039=(36,6,64,64)f32 #3033=(36,6,64,32)f32 #3040=(36,6,64,32)f32 +torch.transpose torch.transpose_3170 1 1 3040 3041 dim0=1 dim1=2 $input=3040 #3040=(36,6,64,32)f32 #3041=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_611 1 1 3041 3042 shape=(36,64,192) $input=3041 #3041=(36,64,6,32)f32 #3042=(36,64,192)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.4.attn.proj 1 1 3042 3043 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3042=(36,64,192)f32 #3043=(36,64,192)f32 +Tensor.reshape Tensor.view_1965 1 1 3043 3044 shape=(1,6,6,8,8,192) $input=3043 #3043=(36,64,192)f32 #3044=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2883 1 1 3044 3045 dims=(0,1,3,2,4,5) $input=3044 #3044=(1,6,6,8,8,192)f32 #3045=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1967 1 1 3045 3046 shape=(1,2304,192) $input=3045 #3045=(1,6,8,6,8,192)f32 #3046=(1,2304,192)f32 +pnnx.Expression pnnx_expr_2961 2 1 3023 3046 3047 expr=add(@0,@1) #3023=(1,2304,192)f32 #3046=(1,2304,192)f32 #3047=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.2.residual_group.blocks.4.norm2 1 1 3047 3048 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3047=(1,2304,192)f32 #3048=(1,2304,192)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.4.mlp.fc1 1 1 3048 3049 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3048=(1,2304,192)f32 #3049=(1,2304,384)f32 +nn.GELU layers_mmsa.2.residual_group.blocks.4.mlp.act 1 1 3049 3050 #3049=(1,2304,384)f32 #3050=(1,2304,384)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.4.mlp.fc2 1 1 3050 3051 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3050=(1,2304,384)f32 #3051=(1,2304,192)f32 +pnnx.Expression pnnx_expr_2959 2 1 3047 3051 3052 expr=add(@0,@1) #3047=(1,2304,192)f32 #3051=(1,2304,192)f32 #3052=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.2.residual_group.blocks.5.norm1 1 1 3052 3053 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3052=(1,2304,192)f32 #3053=(1,2304,192)f32 +Tensor.view Tensor.view_1968 1 1 3053 3054 shape=(1,48,48,192) $input=3053 #3053=(1,2304,192)f32 #3054=(1,48,48,192)f32 +torch.roll torch.roll_2508 1 1 3054 3055 dims=(1,2) shifts=(-4,-4) $input=3054 #3054=(1,48,48,192)f32 #3055=(1,48,48,192)f32 +Tensor.view Tensor.view_1969 1 1 3055 3056 shape=(1,6,8,6,8,192) $input=3055 #3055=(1,48,48,192)f32 #3056=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2884 1 1 3056 3057 dims=(0,1,3,2,4,5) $input=3056 #3056=(1,6,8,6,8,192)f32 #3057=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1971 1 1 3057 3058 shape=(36,64,192) $input=3057 #3057=(1,6,6,8,8,192)f32 #3058=(36,64,192)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.5.attn.qkv 1 1 3058 3059 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3058=(36,64,192)f32 #3059=(36,64,576)f32 +Tensor.reshape Tensor.reshape_612 1 1 3059 3060 shape=(36,64,3,6,32) $input=3059 #3059=(36,64,576)f32 #3060=(36,64,3,6,32)f32 +torch.permute torch.permute_2885 1 1 3060 3061 dims=(2,0,3,1,4) $input=3060 #3060=(36,64,3,6,32)f32 #3061=(3,36,6,64,32)f32 +torch.unbind Tensor.select_917 1 3 3061 3062 3063 3064 dim=0 #3061=(3,36,6,64,32)f32 #3062=(36,6,64,32)f32 #3063=(36,6,64,32)f32 #3064=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_2861 1 1 3062 3065 expr=mul(@0,1.767767e-01) #3062=(36,6,64,32)f32 #3065=(36,6,64,32)f32 +torch.transpose torch.transpose_3171 1 1 3063 3066 dim0=-2 dim1=-1 $input=3063 #3063=(36,6,64,32)f32 #3066=(36,6,32,64)f32 +torch.matmul torch.matmul_2382 2 1 3065 3066 3067 $input=3065 $other=3066 #3065=(36,6,64,32)f32 #3066=(36,6,32,64)f32 #3067=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_16801 0 1 3068 @pnnx_fold_16801=(1,6,64,64)f32 #3068=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_2848 2 1 3067 3068 3069 expr=add(@0,@1) #3067=(36,6,64,64)f32 #3068=(1,6,64,64)f32 #3069=(36,6,64,64)f32 +Tensor.view Tensor.view_1974 1 1 3069 3070 shape=(1,36,6,64,64) $input=3069 #3069=(36,6,64,64)f32 #3070=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_16811 0 1 3071 @pnnx_fold_16811=(1,36,1,64,64)f32 #3071=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_2836 2 1 3070 3071 3072 expr=add(@0,@1) #3070=(1,36,6,64,64)f32 #3071=(1,36,1,64,64)f32 #3072=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1975 1 1 3072 3073 shape=(-1,6,64,64) $input=3072 #3072=(1,36,6,64,64)f32 #3073=(36,6,64,64)f32 +nn.Softmax layers_mmsa.2.residual_group.blocks.5.attn.softmax 1 1 3073 3074 dim=-1 #3073=(36,6,64,64)f32 #3074=(36,6,64,64)f32 +torch.matmul torch.matmul_2383 2 1 3074 3064 3075 $input=3074 $other=3064 #3074=(36,6,64,64)f32 #3064=(36,6,64,32)f32 #3075=(36,6,64,32)f32 +torch.transpose torch.transpose_3172 1 1 3075 3076 dim0=1 dim1=2 $input=3075 #3075=(36,6,64,32)f32 #3076=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_613 1 1 3076 3077 shape=(36,64,192) $input=3076 #3076=(36,64,6,32)f32 #3077=(36,64,192)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.5.attn.proj 1 1 3077 3078 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3077=(36,64,192)f32 #3078=(36,64,192)f32 +Tensor.reshape Tensor.view_1977 1 1 3078 3079 shape=(1,6,6,8,8,192) $input=3078 #3078=(36,64,192)f32 #3079=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2887 1 1 3079 3080 dims=(0,1,3,2,4,5) $input=3079 #3079=(1,6,6,8,8,192)f32 #3080=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_179 1 1 3080 3081 shape=(1,48,48,-1) $input=3080 #3080=(1,6,8,6,8,192)f32 #3081=(1,48,48,192)f32 +torch.roll torch.roll_2509 1 1 3081 3082 dims=(1,2) shifts=(4,4) $input=3081 #3081=(1,48,48,192)f32 #3082=(1,48,48,192)f32 +Tensor.view Tensor.view_1979 1 1 3082 3083 shape=(1,2304,192) $input=3082 #3082=(1,48,48,192)f32 #3083=(1,2304,192)f32 +pnnx.Expression pnnx_expr_2794 2 1 3052 3083 3084 expr=add(@0,@1) #3052=(1,2304,192)f32 #3083=(1,2304,192)f32 #3084=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.2.residual_group.blocks.5.norm2 1 1 3084 3085 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3084=(1,2304,192)f32 #3085=(1,2304,192)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.5.mlp.fc1 1 1 3085 3086 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3085=(1,2304,192)f32 #3086=(1,2304,384)f32 +nn.GELU layers_mmsa.2.residual_group.blocks.5.mlp.act 1 1 3086 3087 #3086=(1,2304,384)f32 #3087=(1,2304,384)f32 +nn.Linear layers_mmsa.2.residual_group.blocks.5.mlp.fc2 1 1 3087 3088 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3087=(1,2304,384)f32 #3088=(1,2304,192)f32 +pnnx.Expression pnnx_expr_2792 2 1 3084 3088 3089 expr=add(@0,@1) #3084=(1,2304,192)f32 #3088=(1,2304,192)f32 #3089=(1,2304,192)f32 +torch.transpose torch.transpose_3173 1 1 3089 3090 dim0=1 dim1=2 $input=3089 #3089=(1,2304,192)f32 #3090=(1,192,2304)f32 +Tensor.view Tensor.view_1980 1 1 3090 3091 shape=(1,192,48,48) $input=3090 #3090=(1,192,2304)f32 #3091=(1,192,48,48)f32 +nn.Conv2d layers_mmsa.2.conv 1 1 3091 3092 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #3091=(1,192,48,48)f32 #3092=(1,192,48,48)f32 +torch.flatten torch.flatten_2200 1 1 3092 3093 end_dim=-1 start_dim=2 $input=3092 #3092=(1,192,48,48)f32 #3093=(1,192,2304)f32 +torch.transpose torch.transpose_3174 1 1 3093 3094 dim0=1 dim1=2 $input=3093 #3093=(1,192,2304)f32 #3094=(1,2304,192)f32 +pnnx.Expression pnnx_expr_2779 2 1 3094 2891 3095 expr=add(@0,@1) #3094=(1,2304,192)f32 #2891=(1,2304,192)f32 #3095=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.3.residual_group.blocks.0.norm1 1 1 3095 3096 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3095=(1,2304,192)f32 #3096=(1,2304,192)f32 +Tensor.reshape Tensor.view_1982 1 1 3096 3097 shape=(1,6,8,6,8,192) $input=3096 #3096=(1,2304,192)f32 #3097=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2888 1 1 3097 3098 dims=(0,1,3,2,4,5) $input=3097 #3097=(1,6,8,6,8,192)f32 #3098=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1984 1 1 3098 3099 shape=(36,64,192) $input=3098 #3098=(1,6,6,8,8,192)f32 #3099=(36,64,192)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.0.attn.qkv 1 1 3099 3100 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3099=(36,64,192)f32 #3100=(36,64,576)f32 +Tensor.reshape Tensor.reshape_614 1 1 3100 3101 shape=(36,64,3,6,32) $input=3100 #3100=(36,64,576)f32 #3101=(36,64,3,6,32)f32 +torch.permute torch.permute_2889 1 1 3101 3102 dims=(2,0,3,1,4) $input=3101 #3101=(36,64,3,6,32)f32 #3102=(3,36,6,64,32)f32 +torch.unbind Tensor.select_920 1 3 3102 3103 3104 3105 dim=0 #3102=(3,36,6,64,32)f32 #3103=(36,6,64,32)f32 #3104=(36,6,64,32)f32 #3105=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_2690 1 1 3103 3106 expr=mul(@0,1.767767e-01) #3103=(36,6,64,32)f32 #3106=(36,6,64,32)f32 +torch.transpose torch.transpose_3175 1 1 3104 3107 dim0=-2 dim1=-1 $input=3104 #3104=(36,6,64,32)f32 #3107=(36,6,32,64)f32 +torch.matmul torch.matmul_2384 2 1 3106 3107 3108 $input=3106 $other=3107 #3106=(36,6,64,32)f32 #3107=(36,6,32,64)f32 #3108=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_16994 0 1 3109 @pnnx_fold_16994=(1,6,64,64)f32 #3109=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_2677 2 1 3108 3109 3110 expr=add(@0,@1) #3108=(36,6,64,64)f32 #3109=(1,6,64,64)f32 #3110=(36,6,64,64)f32 +nn.Softmax layers_mmsa.3.residual_group.blocks.0.attn.softmax 1 1 3110 3111 dim=-1 #3110=(36,6,64,64)f32 #3111=(36,6,64,64)f32 +torch.matmul torch.matmul_2385 2 1 3111 3105 3112 $input=3111 $other=3105 #3111=(36,6,64,64)f32 #3105=(36,6,64,32)f32 #3112=(36,6,64,32)f32 +torch.transpose torch.transpose_3176 1 1 3112 3113 dim0=1 dim1=2 $input=3112 #3112=(36,6,64,32)f32 #3113=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_615 1 1 3113 3114 shape=(36,64,192) $input=3113 #3113=(36,64,6,32)f32 #3114=(36,64,192)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.0.attn.proj 1 1 3114 3115 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3114=(36,64,192)f32 #3115=(36,64,192)f32 +Tensor.reshape Tensor.view_1988 1 1 3115 3116 shape=(1,6,6,8,8,192) $input=3115 #3115=(36,64,192)f32 #3116=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2891 1 1 3116 3117 dims=(0,1,3,2,4,5) $input=3116 #3116=(1,6,6,8,8,192)f32 #3117=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_1990 1 1 3117 3118 shape=(1,2304,192) $input=3117 #3117=(1,6,8,6,8,192)f32 #3118=(1,2304,192)f32 +pnnx.Expression pnnx_expr_2644 2 1 3095 3118 3119 expr=add(@0,@1) #3095=(1,2304,192)f32 #3118=(1,2304,192)f32 #3119=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.3.residual_group.blocks.0.norm2 1 1 3119 3120 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3119=(1,2304,192)f32 #3120=(1,2304,192)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.0.mlp.fc1 1 1 3120 3121 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3120=(1,2304,192)f32 #3121=(1,2304,384)f32 +nn.GELU layers_mmsa.3.residual_group.blocks.0.mlp.act 1 1 3121 3122 #3121=(1,2304,384)f32 #3122=(1,2304,384)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.0.mlp.fc2 1 1 3122 3123 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3122=(1,2304,384)f32 #3123=(1,2304,192)f32 +pnnx.Expression pnnx_expr_2642 2 1 3119 3123 3124 expr=add(@0,@1) #3119=(1,2304,192)f32 #3123=(1,2304,192)f32 #3124=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.3.residual_group.blocks.1.norm1 1 1 3124 3125 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3124=(1,2304,192)f32 #3125=(1,2304,192)f32 +Tensor.view Tensor.view_1991 1 1 3125 3126 shape=(1,48,48,192) $input=3125 #3125=(1,2304,192)f32 #3126=(1,48,48,192)f32 +torch.roll torch.roll_2510 1 1 3126 3127 dims=(1,2) shifts=(-4,-4) $input=3126 #3126=(1,48,48,192)f32 #3127=(1,48,48,192)f32 +Tensor.view Tensor.view_1992 1 1 3127 3128 shape=(1,6,8,6,8,192) $input=3127 #3127=(1,48,48,192)f32 #3128=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2892 1 1 3128 3129 dims=(0,1,3,2,4,5) $input=3128 #3128=(1,6,8,6,8,192)f32 #3129=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_1994 1 1 3129 3130 shape=(36,64,192) $input=3129 #3129=(1,6,6,8,8,192)f32 #3130=(36,64,192)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.1.attn.qkv 1 1 3130 3131 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3130=(36,64,192)f32 #3131=(36,64,576)f32 +Tensor.reshape Tensor.reshape_616 1 1 3131 3132 shape=(36,64,3,6,32) $input=3131 #3131=(36,64,576)f32 #3132=(36,64,3,6,32)f32 +torch.permute torch.permute_2893 1 1 3132 3133 dims=(2,0,3,1,4) $input=3132 #3132=(36,64,3,6,32)f32 #3133=(3,36,6,64,32)f32 +torch.unbind Tensor.select_923 1 3 3133 3134 3135 3136 dim=0 #3133=(3,36,6,64,32)f32 #3134=(36,6,64,32)f32 #3135=(36,6,64,32)f32 #3136=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_2544 1 1 3134 3137 expr=mul(@0,1.767767e-01) #3134=(36,6,64,32)f32 #3137=(36,6,64,32)f32 +torch.transpose torch.transpose_3177 1 1 3135 3138 dim0=-2 dim1=-1 $input=3135 #3135=(36,6,64,32)f32 #3138=(36,6,32,64)f32 +torch.matmul torch.matmul_2386 2 1 3137 3138 3139 $input=3137 $other=3138 #3137=(36,6,64,32)f32 #3138=(36,6,32,64)f32 #3139=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_17147 0 1 3140 @pnnx_fold_17147=(1,6,64,64)f32 #3140=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_2531 2 1 3139 3140 3141 expr=add(@0,@1) #3139=(36,6,64,64)f32 #3140=(1,6,64,64)f32 #3141=(36,6,64,64)f32 +Tensor.view Tensor.view_1997 1 1 3141 3142 shape=(1,36,6,64,64) $input=3141 #3141=(36,6,64,64)f32 #3142=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_17157 0 1 3143 @pnnx_fold_17157=(1,36,1,64,64)f32 #3143=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_2519 2 1 3142 3143 3144 expr=add(@0,@1) #3142=(1,36,6,64,64)f32 #3143=(1,36,1,64,64)f32 #3144=(1,36,6,64,64)f32 +Tensor.view Tensor.view_1998 1 1 3144 3145 shape=(-1,6,64,64) $input=3144 #3144=(1,36,6,64,64)f32 #3145=(36,6,64,64)f32 +nn.Softmax layers_mmsa.3.residual_group.blocks.1.attn.softmax 1 1 3145 3146 dim=-1 #3145=(36,6,64,64)f32 #3146=(36,6,64,64)f32 +torch.matmul torch.matmul_2387 2 1 3146 3136 3147 $input=3146 $other=3136 #3146=(36,6,64,64)f32 #3136=(36,6,64,32)f32 #3147=(36,6,64,32)f32 +torch.transpose torch.transpose_3178 1 1 3147 3148 dim0=1 dim1=2 $input=3147 #3147=(36,6,64,32)f32 #3148=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_617 1 1 3148 3149 shape=(36,64,192) $input=3148 #3148=(36,64,6,32)f32 #3149=(36,64,192)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.1.attn.proj 1 1 3149 3150 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3149=(36,64,192)f32 #3150=(36,64,192)f32 +Tensor.reshape Tensor.view_2000 1 1 3150 3151 shape=(1,6,6,8,8,192) $input=3150 #3150=(36,64,192)f32 #3151=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2895 1 1 3151 3152 dims=(0,1,3,2,4,5) $input=3151 #3151=(1,6,6,8,8,192)f32 #3152=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_183 1 1 3152 3153 shape=(1,48,48,-1) $input=3152 #3152=(1,6,8,6,8,192)f32 #3153=(1,48,48,192)f32 +torch.roll torch.roll_2511 1 1 3153 3154 dims=(1,2) shifts=(4,4) $input=3153 #3153=(1,48,48,192)f32 #3154=(1,48,48,192)f32 +Tensor.view Tensor.view_2002 1 1 3154 3155 shape=(1,2304,192) $input=3154 #3154=(1,48,48,192)f32 #3155=(1,2304,192)f32 +pnnx.Expression pnnx_expr_2477 2 1 3124 3155 3156 expr=add(@0,@1) #3124=(1,2304,192)f32 #3155=(1,2304,192)f32 #3156=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.3.residual_group.blocks.1.norm2 1 1 3156 3157 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3156=(1,2304,192)f32 #3157=(1,2304,192)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.1.mlp.fc1 1 1 3157 3158 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3157=(1,2304,192)f32 #3158=(1,2304,384)f32 +nn.GELU layers_mmsa.3.residual_group.blocks.1.mlp.act 1 1 3158 3159 #3158=(1,2304,384)f32 #3159=(1,2304,384)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.1.mlp.fc2 1 1 3159 3160 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3159=(1,2304,384)f32 #3160=(1,2304,192)f32 +pnnx.Expression pnnx_expr_2475 2 1 3156 3160 3161 expr=add(@0,@1) #3156=(1,2304,192)f32 #3160=(1,2304,192)f32 #3161=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.3.residual_group.blocks.2.norm1 1 1 3161 3162 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3161=(1,2304,192)f32 #3162=(1,2304,192)f32 +Tensor.reshape Tensor.view_2004 1 1 3162 3163 shape=(1,6,8,6,8,192) $input=3162 #3162=(1,2304,192)f32 #3163=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2896 1 1 3163 3164 dims=(0,1,3,2,4,5) $input=3163 #3163=(1,6,8,6,8,192)f32 #3164=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_2006 1 1 3164 3165 shape=(36,64,192) $input=3164 #3164=(1,6,6,8,8,192)f32 #3165=(36,64,192)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.2.attn.qkv 1 1 3165 3166 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3165=(36,64,192)f32 #3166=(36,64,576)f32 +Tensor.reshape Tensor.reshape_618 1 1 3166 3167 shape=(36,64,3,6,32) $input=3166 #3166=(36,64,576)f32 #3167=(36,64,3,6,32)f32 +torch.permute torch.permute_2897 1 1 3167 3168 dims=(2,0,3,1,4) $input=3167 #3167=(36,64,3,6,32)f32 #3168=(3,36,6,64,32)f32 +torch.unbind Tensor.select_926 1 3 3168 3169 3170 3171 dim=0 #3168=(3,36,6,64,32)f32 #3169=(36,6,64,32)f32 #3170=(36,6,64,32)f32 #3171=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_2387 1 1 3169 3172 expr=mul(@0,1.767767e-01) #3169=(36,6,64,32)f32 #3172=(36,6,64,32)f32 +torch.transpose torch.transpose_3179 1 1 3170 3173 dim0=-2 dim1=-1 $input=3170 #3170=(36,6,64,32)f32 #3173=(36,6,32,64)f32 +torch.matmul torch.matmul_2388 2 1 3172 3173 3174 $input=3172 $other=3173 #3172=(36,6,64,32)f32 #3173=(36,6,32,64)f32 #3174=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_17306 0 1 3175 @pnnx_fold_17306=(1,6,64,64)f32 #3175=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_2374 2 1 3174 3175 3176 expr=add(@0,@1) #3174=(36,6,64,64)f32 #3175=(1,6,64,64)f32 #3176=(36,6,64,64)f32 +nn.Softmax layers_mmsa.3.residual_group.blocks.2.attn.softmax 1 1 3176 3177 dim=-1 #3176=(36,6,64,64)f32 #3177=(36,6,64,64)f32 +torch.matmul torch.matmul_2389 2 1 3177 3171 3178 $input=3177 $other=3171 #3177=(36,6,64,64)f32 #3171=(36,6,64,32)f32 #3178=(36,6,64,32)f32 +torch.transpose torch.transpose_3180 1 1 3178 3179 dim0=1 dim1=2 $input=3178 #3178=(36,6,64,32)f32 #3179=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_619 1 1 3179 3180 shape=(36,64,192) $input=3179 #3179=(36,64,6,32)f32 #3180=(36,64,192)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.2.attn.proj 1 1 3180 3181 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3180=(36,64,192)f32 #3181=(36,64,192)f32 +Tensor.reshape Tensor.view_2010 1 1 3181 3182 shape=(1,6,6,8,8,192) $input=3181 #3181=(36,64,192)f32 #3182=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2899 1 1 3182 3183 dims=(0,1,3,2,4,5) $input=3182 #3182=(1,6,6,8,8,192)f32 #3183=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_2012 1 1 3183 3184 shape=(1,2304,192) $input=3183 #3183=(1,6,8,6,8,192)f32 #3184=(1,2304,192)f32 +pnnx.Expression pnnx_expr_2341 2 1 3161 3184 3185 expr=add(@0,@1) #3161=(1,2304,192)f32 #3184=(1,2304,192)f32 #3185=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.3.residual_group.blocks.2.norm2 1 1 3185 3186 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3185=(1,2304,192)f32 #3186=(1,2304,192)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.2.mlp.fc1 1 1 3186 3187 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3186=(1,2304,192)f32 #3187=(1,2304,384)f32 +nn.GELU layers_mmsa.3.residual_group.blocks.2.mlp.act 1 1 3187 3188 #3187=(1,2304,384)f32 #3188=(1,2304,384)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.2.mlp.fc2 1 1 3188 3189 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3188=(1,2304,384)f32 #3189=(1,2304,192)f32 +pnnx.Expression pnnx_expr_2339 2 1 3185 3189 3190 expr=add(@0,@1) #3185=(1,2304,192)f32 #3189=(1,2304,192)f32 #3190=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.3.residual_group.blocks.3.norm1 1 1 3190 3191 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3190=(1,2304,192)f32 #3191=(1,2304,192)f32 +Tensor.view Tensor.view_2013 1 1 3191 3192 shape=(1,48,48,192) $input=3191 #3191=(1,2304,192)f32 #3192=(1,48,48,192)f32 +torch.roll torch.roll_2512 1 1 3192 3193 dims=(1,2) shifts=(-4,-4) $input=3192 #3192=(1,48,48,192)f32 #3193=(1,48,48,192)f32 +Tensor.view Tensor.view_2014 1 1 3193 3194 shape=(1,6,8,6,8,192) $input=3193 #3193=(1,48,48,192)f32 #3194=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2900 1 1 3194 3195 dims=(0,1,3,2,4,5) $input=3194 #3194=(1,6,8,6,8,192)f32 #3195=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_2016 1 1 3195 3196 shape=(36,64,192) $input=3195 #3195=(1,6,6,8,8,192)f32 #3196=(36,64,192)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.3.attn.qkv 1 1 3196 3197 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3196=(36,64,192)f32 #3197=(36,64,576)f32 +Tensor.reshape Tensor.reshape_620 1 1 3197 3198 shape=(36,64,3,6,32) $input=3197 #3197=(36,64,576)f32 #3198=(36,64,3,6,32)f32 +torch.permute torch.permute_2901 1 1 3198 3199 dims=(2,0,3,1,4) $input=3198 #3198=(36,64,3,6,32)f32 #3199=(3,36,6,64,32)f32 +torch.unbind Tensor.select_929 1 3 3199 3200 3201 3202 dim=0 #3199=(3,36,6,64,32)f32 #3200=(36,6,64,32)f32 #3201=(36,6,64,32)f32 #3202=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_2241 1 1 3200 3203 expr=mul(@0,1.767767e-01) #3200=(36,6,64,32)f32 #3203=(36,6,64,32)f32 +torch.transpose torch.transpose_3181 1 1 3201 3204 dim0=-2 dim1=-1 $input=3201 #3201=(36,6,64,32)f32 #3204=(36,6,32,64)f32 +torch.matmul torch.matmul_2390 2 1 3203 3204 3205 $input=3203 $other=3204 #3203=(36,6,64,32)f32 #3204=(36,6,32,64)f32 #3205=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_17459 0 1 3206 @pnnx_fold_17459=(1,6,64,64)f32 #3206=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_2228 2 1 3205 3206 3207 expr=add(@0,@1) #3205=(36,6,64,64)f32 #3206=(1,6,64,64)f32 #3207=(36,6,64,64)f32 +Tensor.view Tensor.view_2019 1 1 3207 3208 shape=(1,36,6,64,64) $input=3207 #3207=(36,6,64,64)f32 #3208=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_17469 0 1 3209 @pnnx_fold_17469=(1,36,1,64,64)f32 #3209=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_2216 2 1 3208 3209 3210 expr=add(@0,@1) #3208=(1,36,6,64,64)f32 #3209=(1,36,1,64,64)f32 #3210=(1,36,6,64,64)f32 +Tensor.view Tensor.view_2020 1 1 3210 3211 shape=(-1,6,64,64) $input=3210 #3210=(1,36,6,64,64)f32 #3211=(36,6,64,64)f32 +nn.Softmax layers_mmsa.3.residual_group.blocks.3.attn.softmax 1 1 3211 3212 dim=-1 #3211=(36,6,64,64)f32 #3212=(36,6,64,64)f32 +torch.matmul torch.matmul_2391 2 1 3212 3202 3213 $input=3212 $other=3202 #3212=(36,6,64,64)f32 #3202=(36,6,64,32)f32 #3213=(36,6,64,32)f32 +torch.transpose torch.transpose_3182 1 1 3213 3214 dim0=1 dim1=2 $input=3213 #3213=(36,6,64,32)f32 #3214=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_621 1 1 3214 3215 shape=(36,64,192) $input=3214 #3214=(36,64,6,32)f32 #3215=(36,64,192)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.3.attn.proj 1 1 3215 3216 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3215=(36,64,192)f32 #3216=(36,64,192)f32 +Tensor.reshape Tensor.view_2022 1 1 3216 3217 shape=(1,6,6,8,8,192) $input=3216 #3216=(36,64,192)f32 #3217=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2903 1 1 3217 3218 dims=(0,1,3,2,4,5) $input=3217 #3217=(1,6,6,8,8,192)f32 #3218=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_187 1 1 3218 3219 shape=(1,48,48,-1) $input=3218 #3218=(1,6,8,6,8,192)f32 #3219=(1,48,48,192)f32 +torch.roll torch.roll_2513 1 1 3219 3220 dims=(1,2) shifts=(4,4) $input=3219 #3219=(1,48,48,192)f32 #3220=(1,48,48,192)f32 +Tensor.view Tensor.view_2024 1 1 3220 3221 shape=(1,2304,192) $input=3220 #3220=(1,48,48,192)f32 #3221=(1,2304,192)f32 +pnnx.Expression pnnx_expr_2174 2 1 3190 3221 3222 expr=add(@0,@1) #3190=(1,2304,192)f32 #3221=(1,2304,192)f32 #3222=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.3.residual_group.blocks.3.norm2 1 1 3222 3223 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3222=(1,2304,192)f32 #3223=(1,2304,192)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.3.mlp.fc1 1 1 3223 3224 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3223=(1,2304,192)f32 #3224=(1,2304,384)f32 +nn.GELU layers_mmsa.3.residual_group.blocks.3.mlp.act 1 1 3224 3225 #3224=(1,2304,384)f32 #3225=(1,2304,384)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.3.mlp.fc2 1 1 3225 3226 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3225=(1,2304,384)f32 #3226=(1,2304,192)f32 +pnnx.Expression pnnx_expr_2172 2 1 3222 3226 3227 expr=add(@0,@1) #3222=(1,2304,192)f32 #3226=(1,2304,192)f32 #3227=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.3.residual_group.blocks.4.norm1 1 1 3227 3228 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3227=(1,2304,192)f32 #3228=(1,2304,192)f32 +Tensor.reshape Tensor.view_2026 1 1 3228 3229 shape=(1,6,8,6,8,192) $input=3228 #3228=(1,2304,192)f32 #3229=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2904 1 1 3229 3230 dims=(0,1,3,2,4,5) $input=3229 #3229=(1,6,8,6,8,192)f32 #3230=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_2028 1 1 3230 3231 shape=(36,64,192) $input=3230 #3230=(1,6,6,8,8,192)f32 #3231=(36,64,192)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.4.attn.qkv 1 1 3231 3232 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3231=(36,64,192)f32 #3232=(36,64,576)f32 +Tensor.reshape Tensor.reshape_622 1 1 3232 3233 shape=(36,64,3,6,32) $input=3232 #3232=(36,64,576)f32 #3233=(36,64,3,6,32)f32 +torch.permute torch.permute_2905 1 1 3233 3234 dims=(2,0,3,1,4) $input=3233 #3233=(36,64,3,6,32)f32 #3234=(3,36,6,64,32)f32 +torch.unbind Tensor.select_932 1 3 3234 3235 3236 3237 dim=0 #3234=(3,36,6,64,32)f32 #3235=(36,6,64,32)f32 #3236=(36,6,64,32)f32 #3237=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_2084 1 1 3235 3238 expr=mul(@0,1.767767e-01) #3235=(36,6,64,32)f32 #3238=(36,6,64,32)f32 +torch.transpose torch.transpose_3183 1 1 3236 3239 dim0=-2 dim1=-1 $input=3236 #3236=(36,6,64,32)f32 #3239=(36,6,32,64)f32 +torch.matmul torch.matmul_2392 2 1 3238 3239 3240 $input=3238 $other=3239 #3238=(36,6,64,32)f32 #3239=(36,6,32,64)f32 #3240=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_17618 0 1 3241 @pnnx_fold_17618=(1,6,64,64)f32 #3241=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_2071 2 1 3240 3241 3242 expr=add(@0,@1) #3240=(36,6,64,64)f32 #3241=(1,6,64,64)f32 #3242=(36,6,64,64)f32 +nn.Softmax layers_mmsa.3.residual_group.blocks.4.attn.softmax 1 1 3242 3243 dim=-1 #3242=(36,6,64,64)f32 #3243=(36,6,64,64)f32 +torch.matmul torch.matmul_2393 2 1 3243 3237 3244 $input=3243 $other=3237 #3243=(36,6,64,64)f32 #3237=(36,6,64,32)f32 #3244=(36,6,64,32)f32 +torch.transpose torch.transpose_3184 1 1 3244 3245 dim0=1 dim1=2 $input=3244 #3244=(36,6,64,32)f32 #3245=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_623 1 1 3245 3246 shape=(36,64,192) $input=3245 #3245=(36,64,6,32)f32 #3246=(36,64,192)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.4.attn.proj 1 1 3246 3247 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3246=(36,64,192)f32 #3247=(36,64,192)f32 +Tensor.reshape Tensor.view_2032 1 1 3247 3248 shape=(1,6,6,8,8,192) $input=3247 #3247=(36,64,192)f32 #3248=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2907 1 1 3248 3249 dims=(0,1,3,2,4,5) $input=3248 #3248=(1,6,6,8,8,192)f32 #3249=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_2034 1 1 3249 3250 shape=(1,2304,192) $input=3249 #3249=(1,6,8,6,8,192)f32 #3250=(1,2304,192)f32 +pnnx.Expression pnnx_expr_2038 2 1 3227 3250 3251 expr=add(@0,@1) #3227=(1,2304,192)f32 #3250=(1,2304,192)f32 #3251=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.3.residual_group.blocks.4.norm2 1 1 3251 3252 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3251=(1,2304,192)f32 #3252=(1,2304,192)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.4.mlp.fc1 1 1 3252 3253 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3252=(1,2304,192)f32 #3253=(1,2304,384)f32 +nn.GELU layers_mmsa.3.residual_group.blocks.4.mlp.act 1 1 3253 3254 #3253=(1,2304,384)f32 #3254=(1,2304,384)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.4.mlp.fc2 1 1 3254 3255 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3254=(1,2304,384)f32 #3255=(1,2304,192)f32 +pnnx.Expression pnnx_expr_2036 2 1 3251 3255 3256 expr=add(@0,@1) #3251=(1,2304,192)f32 #3255=(1,2304,192)f32 #3256=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.3.residual_group.blocks.5.norm1 1 1 3256 3257 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3256=(1,2304,192)f32 #3257=(1,2304,192)f32 +Tensor.view Tensor.view_2035 1 1 3257 3258 shape=(1,48,48,192) $input=3257 #3257=(1,2304,192)f32 #3258=(1,48,48,192)f32 +torch.roll torch.roll_2514 1 1 3258 3259 dims=(1,2) shifts=(-4,-4) $input=3258 #3258=(1,48,48,192)f32 #3259=(1,48,48,192)f32 +Tensor.view Tensor.view_2036 1 1 3259 3260 shape=(1,6,8,6,8,192) $input=3259 #3259=(1,48,48,192)f32 #3260=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2908 1 1 3260 3261 dims=(0,1,3,2,4,5) $input=3260 #3260=(1,6,8,6,8,192)f32 #3261=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_2038 1 1 3261 3262 shape=(36,64,192) $input=3261 #3261=(1,6,6,8,8,192)f32 #3262=(36,64,192)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.5.attn.qkv 1 1 3262 3263 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3262=(36,64,192)f32 #3263=(36,64,576)f32 +Tensor.reshape Tensor.reshape_624 1 1 3263 3264 shape=(36,64,3,6,32) $input=3263 #3263=(36,64,576)f32 #3264=(36,64,3,6,32)f32 +torch.permute torch.permute_2909 1 1 3264 3265 dims=(2,0,3,1,4) $input=3264 #3264=(36,64,3,6,32)f32 #3265=(3,36,6,64,32)f32 +torch.unbind Tensor.select_935 1 3 3265 3266 3267 3268 dim=0 #3265=(3,36,6,64,32)f32 #3266=(36,6,64,32)f32 #3267=(36,6,64,32)f32 #3268=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_1938 1 1 3266 3269 expr=mul(@0,1.767767e-01) #3266=(36,6,64,32)f32 #3269=(36,6,64,32)f32 +torch.transpose torch.transpose_3185 1 1 3267 3270 dim0=-2 dim1=-1 $input=3267 #3267=(36,6,64,32)f32 #3270=(36,6,32,64)f32 +torch.matmul torch.matmul_2394 2 1 3269 3270 3271 $input=3269 $other=3270 #3269=(36,6,64,32)f32 #3270=(36,6,32,64)f32 #3271=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_17771 0 1 3272 @pnnx_fold_17771=(1,6,64,64)f32 #3272=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_1925 2 1 3271 3272 3273 expr=add(@0,@1) #3271=(36,6,64,64)f32 #3272=(1,6,64,64)f32 #3273=(36,6,64,64)f32 +Tensor.view Tensor.view_2041 1 1 3273 3274 shape=(1,36,6,64,64) $input=3273 #3273=(36,6,64,64)f32 #3274=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_17781 0 1 3275 @pnnx_fold_17781=(1,36,1,64,64)f32 #3275=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_1913 2 1 3274 3275 3276 expr=add(@0,@1) #3274=(1,36,6,64,64)f32 #3275=(1,36,1,64,64)f32 #3276=(1,36,6,64,64)f32 +Tensor.view Tensor.view_2042 1 1 3276 3277 shape=(-1,6,64,64) $input=3276 #3276=(1,36,6,64,64)f32 #3277=(36,6,64,64)f32 +nn.Softmax layers_mmsa.3.residual_group.blocks.5.attn.softmax 1 1 3277 3278 dim=-1 #3277=(36,6,64,64)f32 #3278=(36,6,64,64)f32 +torch.matmul torch.matmul_2395 2 1 3278 3268 3279 $input=3278 $other=3268 #3278=(36,6,64,64)f32 #3268=(36,6,64,32)f32 #3279=(36,6,64,32)f32 +torch.transpose torch.transpose_3186 1 1 3279 3280 dim0=1 dim1=2 $input=3279 #3279=(36,6,64,32)f32 #3280=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_625 1 1 3280 3281 shape=(36,64,192) $input=3280 #3280=(36,64,6,32)f32 #3281=(36,64,192)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.5.attn.proj 1 1 3281 3282 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3281=(36,64,192)f32 #3282=(36,64,192)f32 +Tensor.reshape Tensor.view_2044 1 1 3282 3283 shape=(1,6,6,8,8,192) $input=3282 #3282=(36,64,192)f32 #3283=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2911 1 1 3283 3284 dims=(0,1,3,2,4,5) $input=3283 #3283=(1,6,6,8,8,192)f32 #3284=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_191 1 1 3284 3285 shape=(1,48,48,-1) $input=3284 #3284=(1,6,8,6,8,192)f32 #3285=(1,48,48,192)f32 +torch.roll torch.roll_2515 1 1 3285 3286 dims=(1,2) shifts=(4,4) $input=3285 #3285=(1,48,48,192)f32 #3286=(1,48,48,192)f32 +Tensor.view Tensor.view_2046 1 1 3286 3287 shape=(1,2304,192) $input=3286 #3286=(1,48,48,192)f32 #3287=(1,2304,192)f32 +pnnx.Expression pnnx_expr_1871 2 1 3256 3287 3288 expr=add(@0,@1) #3256=(1,2304,192)f32 #3287=(1,2304,192)f32 #3288=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.3.residual_group.blocks.5.norm2 1 1 3288 3289 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3288=(1,2304,192)f32 #3289=(1,2304,192)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.5.mlp.fc1 1 1 3289 3290 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3289=(1,2304,192)f32 #3290=(1,2304,384)f32 +nn.GELU layers_mmsa.3.residual_group.blocks.5.mlp.act 1 1 3290 3291 #3290=(1,2304,384)f32 #3291=(1,2304,384)f32 +nn.Linear layers_mmsa.3.residual_group.blocks.5.mlp.fc2 1 1 3291 3292 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3291=(1,2304,384)f32 #3292=(1,2304,192)f32 +pnnx.Expression pnnx_expr_1869 2 1 3288 3292 3293 expr=add(@0,@1) #3288=(1,2304,192)f32 #3292=(1,2304,192)f32 #3293=(1,2304,192)f32 +torch.transpose torch.transpose_3187 1 1 3293 3294 dim0=1 dim1=2 $input=3293 #3293=(1,2304,192)f32 #3294=(1,192,2304)f32 +Tensor.view Tensor.view_2047 1 1 3294 3295 shape=(1,192,48,48) $input=3294 #3294=(1,192,2304)f32 #3295=(1,192,48,48)f32 +nn.Conv2d layers_mmsa.3.conv 1 1 3295 3296 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #3295=(1,192,48,48)f32 #3296=(1,192,48,48)f32 +torch.flatten torch.flatten_2201 1 1 3296 3297 end_dim=-1 start_dim=2 $input=3296 #3296=(1,192,48,48)f32 #3297=(1,192,2304)f32 +torch.transpose torch.transpose_3188 1 1 3297 3298 dim0=1 dim1=2 $input=3297 #3297=(1,192,2304)f32 #3298=(1,2304,192)f32 +pnnx.Expression pnnx_expr_1856 2 1 3298 3095 3299 expr=add(@0,@1) #3298=(1,2304,192)f32 #3095=(1,2304,192)f32 #3299=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.4.residual_group.blocks.0.norm1 1 1 3299 3300 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3299=(1,2304,192)f32 #3300=(1,2304,192)f32 +Tensor.reshape Tensor.view_2049 1 1 3300 3301 shape=(1,6,8,6,8,192) $input=3300 #3300=(1,2304,192)f32 #3301=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2912 1 1 3301 3302 dims=(0,1,3,2,4,5) $input=3301 #3301=(1,6,8,6,8,192)f32 #3302=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_2051 1 1 3302 3303 shape=(36,64,192) $input=3302 #3302=(1,6,6,8,8,192)f32 #3303=(36,64,192)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.0.attn.qkv 1 1 3303 3304 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3303=(36,64,192)f32 #3304=(36,64,576)f32 +Tensor.reshape Tensor.reshape_626 1 1 3304 3305 shape=(36,64,3,6,32) $input=3304 #3304=(36,64,576)f32 #3305=(36,64,3,6,32)f32 +torch.permute torch.permute_2913 1 1 3305 3306 dims=(2,0,3,1,4) $input=3305 #3305=(36,64,3,6,32)f32 #3306=(3,36,6,64,32)f32 +torch.unbind Tensor.select_938 1 3 3306 3307 3308 3309 dim=0 #3306=(3,36,6,64,32)f32 #3307=(36,6,64,32)f32 #3308=(36,6,64,32)f32 #3309=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_1767 1 1 3307 3310 expr=mul(@0,1.767767e-01) #3307=(36,6,64,32)f32 #3310=(36,6,64,32)f32 +torch.transpose torch.transpose_3189 1 1 3308 3311 dim0=-2 dim1=-1 $input=3308 #3308=(36,6,64,32)f32 #3311=(36,6,32,64)f32 +torch.matmul torch.matmul_2396 2 1 3310 3311 3312 $input=3310 $other=3311 #3310=(36,6,64,32)f32 #3311=(36,6,32,64)f32 #3312=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_17964 0 1 3313 @pnnx_fold_17964=(1,6,64,64)f32 #3313=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_1754 2 1 3312 3313 3314 expr=add(@0,@1) #3312=(36,6,64,64)f32 #3313=(1,6,64,64)f32 #3314=(36,6,64,64)f32 +nn.Softmax layers_mmsa.4.residual_group.blocks.0.attn.softmax 1 1 3314 3315 dim=-1 #3314=(36,6,64,64)f32 #3315=(36,6,64,64)f32 +torch.matmul torch.matmul_2397 2 1 3315 3309 3316 $input=3315 $other=3309 #3315=(36,6,64,64)f32 #3309=(36,6,64,32)f32 #3316=(36,6,64,32)f32 +torch.transpose torch.transpose_3190 1 1 3316 3317 dim0=1 dim1=2 $input=3316 #3316=(36,6,64,32)f32 #3317=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_627 1 1 3317 3318 shape=(36,64,192) $input=3317 #3317=(36,64,6,32)f32 #3318=(36,64,192)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.0.attn.proj 1 1 3318 3319 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3318=(36,64,192)f32 #3319=(36,64,192)f32 +Tensor.reshape Tensor.view_2055 1 1 3319 3320 shape=(1,6,6,8,8,192) $input=3319 #3319=(36,64,192)f32 #3320=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2915 1 1 3320 3321 dims=(0,1,3,2,4,5) $input=3320 #3320=(1,6,6,8,8,192)f32 #3321=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_2057 1 1 3321 3322 shape=(1,2304,192) $input=3321 #3321=(1,6,8,6,8,192)f32 #3322=(1,2304,192)f32 +pnnx.Expression pnnx_expr_1721 2 1 3299 3322 3323 expr=add(@0,@1) #3299=(1,2304,192)f32 #3322=(1,2304,192)f32 #3323=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.4.residual_group.blocks.0.norm2 1 1 3323 3324 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3323=(1,2304,192)f32 #3324=(1,2304,192)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.0.mlp.fc1 1 1 3324 3325 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3324=(1,2304,192)f32 #3325=(1,2304,384)f32 +nn.GELU layers_mmsa.4.residual_group.blocks.0.mlp.act 1 1 3325 3326 #3325=(1,2304,384)f32 #3326=(1,2304,384)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.0.mlp.fc2 1 1 3326 3327 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3326=(1,2304,384)f32 #3327=(1,2304,192)f32 +pnnx.Expression pnnx_expr_1719 2 1 3323 3327 3328 expr=add(@0,@1) #3323=(1,2304,192)f32 #3327=(1,2304,192)f32 #3328=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.4.residual_group.blocks.1.norm1 1 1 3328 3329 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3328=(1,2304,192)f32 #3329=(1,2304,192)f32 +Tensor.view Tensor.view_2058 1 1 3329 3330 shape=(1,48,48,192) $input=3329 #3329=(1,2304,192)f32 #3330=(1,48,48,192)f32 +torch.roll torch.roll_2516 1 1 3330 3331 dims=(1,2) shifts=(-4,-4) $input=3330 #3330=(1,48,48,192)f32 #3331=(1,48,48,192)f32 +Tensor.view Tensor.view_2059 1 1 3331 3332 shape=(1,6,8,6,8,192) $input=3331 #3331=(1,48,48,192)f32 #3332=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2916 1 1 3332 3333 dims=(0,1,3,2,4,5) $input=3332 #3332=(1,6,8,6,8,192)f32 #3333=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_2061 1 1 3333 3334 shape=(36,64,192) $input=3333 #3333=(1,6,6,8,8,192)f32 #3334=(36,64,192)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.1.attn.qkv 1 1 3334 3335 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3334=(36,64,192)f32 #3335=(36,64,576)f32 +Tensor.reshape Tensor.reshape_628 1 1 3335 3336 shape=(36,64,3,6,32) $input=3335 #3335=(36,64,576)f32 #3336=(36,64,3,6,32)f32 +torch.permute torch.permute_2917 1 1 3336 3337 dims=(2,0,3,1,4) $input=3336 #3336=(36,64,3,6,32)f32 #3337=(3,36,6,64,32)f32 +torch.unbind Tensor.select_941 1 3 3337 3338 3339 3340 dim=0 #3337=(3,36,6,64,32)f32 #3338=(36,6,64,32)f32 #3339=(36,6,64,32)f32 #3340=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_1621 1 1 3338 3341 expr=mul(@0,1.767767e-01) #3338=(36,6,64,32)f32 #3341=(36,6,64,32)f32 +torch.transpose torch.transpose_3191 1 1 3339 3342 dim0=-2 dim1=-1 $input=3339 #3339=(36,6,64,32)f32 #3342=(36,6,32,64)f32 +torch.matmul torch.matmul_2398 2 1 3341 3342 3343 $input=3341 $other=3342 #3341=(36,6,64,32)f32 #3342=(36,6,32,64)f32 #3343=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_18117 0 1 3344 @pnnx_fold_18117=(1,6,64,64)f32 #3344=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_1608 2 1 3343 3344 3345 expr=add(@0,@1) #3343=(36,6,64,64)f32 #3344=(1,6,64,64)f32 #3345=(36,6,64,64)f32 +Tensor.view Tensor.view_2064 1 1 3345 3346 shape=(1,36,6,64,64) $input=3345 #3345=(36,6,64,64)f32 #3346=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_18127 0 1 3347 @pnnx_fold_18127=(1,36,1,64,64)f32 #3347=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_1596 2 1 3346 3347 3348 expr=add(@0,@1) #3346=(1,36,6,64,64)f32 #3347=(1,36,1,64,64)f32 #3348=(1,36,6,64,64)f32 +Tensor.view Tensor.view_2065 1 1 3348 3349 shape=(-1,6,64,64) $input=3348 #3348=(1,36,6,64,64)f32 #3349=(36,6,64,64)f32 +nn.Softmax layers_mmsa.4.residual_group.blocks.1.attn.softmax 1 1 3349 3350 dim=-1 #3349=(36,6,64,64)f32 #3350=(36,6,64,64)f32 +torch.matmul torch.matmul_2399 2 1 3350 3340 3351 $input=3350 $other=3340 #3350=(36,6,64,64)f32 #3340=(36,6,64,32)f32 #3351=(36,6,64,32)f32 +torch.transpose torch.transpose_3192 1 1 3351 3352 dim0=1 dim1=2 $input=3351 #3351=(36,6,64,32)f32 #3352=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_629 1 1 3352 3353 shape=(36,64,192) $input=3352 #3352=(36,64,6,32)f32 #3353=(36,64,192)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.1.attn.proj 1 1 3353 3354 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3353=(36,64,192)f32 #3354=(36,64,192)f32 +Tensor.reshape Tensor.view_2067 1 1 3354 3355 shape=(1,6,6,8,8,192) $input=3354 #3354=(36,64,192)f32 #3355=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2919 1 1 3355 3356 dims=(0,1,3,2,4,5) $input=3355 #3355=(1,6,6,8,8,192)f32 #3356=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_195 1 1 3356 3357 shape=(1,48,48,-1) $input=3356 #3356=(1,6,8,6,8,192)f32 #3357=(1,48,48,192)f32 +torch.roll torch.roll_2517 1 1 3357 3358 dims=(1,2) shifts=(4,4) $input=3357 #3357=(1,48,48,192)f32 #3358=(1,48,48,192)f32 +Tensor.view Tensor.view_2069 1 1 3358 3359 shape=(1,2304,192) $input=3358 #3358=(1,48,48,192)f32 #3359=(1,2304,192)f32 +pnnx.Expression pnnx_expr_1554 2 1 3328 3359 3360 expr=add(@0,@1) #3328=(1,2304,192)f32 #3359=(1,2304,192)f32 #3360=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.4.residual_group.blocks.1.norm2 1 1 3360 3361 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3360=(1,2304,192)f32 #3361=(1,2304,192)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.1.mlp.fc1 1 1 3361 3362 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3361=(1,2304,192)f32 #3362=(1,2304,384)f32 +nn.GELU layers_mmsa.4.residual_group.blocks.1.mlp.act 1 1 3362 3363 #3362=(1,2304,384)f32 #3363=(1,2304,384)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.1.mlp.fc2 1 1 3363 3364 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3363=(1,2304,384)f32 #3364=(1,2304,192)f32 +pnnx.Expression pnnx_expr_1552 2 1 3360 3364 3365 expr=add(@0,@1) #3360=(1,2304,192)f32 #3364=(1,2304,192)f32 #3365=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.4.residual_group.blocks.2.norm1 1 1 3365 3366 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3365=(1,2304,192)f32 #3366=(1,2304,192)f32 +Tensor.reshape Tensor.view_2071 1 1 3366 3367 shape=(1,6,8,6,8,192) $input=3366 #3366=(1,2304,192)f32 #3367=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2920 1 1 3367 3368 dims=(0,1,3,2,4,5) $input=3367 #3367=(1,6,8,6,8,192)f32 #3368=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_2073 1 1 3368 3369 shape=(36,64,192) $input=3368 #3368=(1,6,6,8,8,192)f32 #3369=(36,64,192)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.2.attn.qkv 1 1 3369 3370 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3369=(36,64,192)f32 #3370=(36,64,576)f32 +Tensor.reshape Tensor.reshape_630 1 1 3370 3371 shape=(36,64,3,6,32) $input=3370 #3370=(36,64,576)f32 #3371=(36,64,3,6,32)f32 +torch.permute torch.permute_2921 1 1 3371 3372 dims=(2,0,3,1,4) $input=3371 #3371=(36,64,3,6,32)f32 #3372=(3,36,6,64,32)f32 +torch.unbind Tensor.select_944 1 3 3372 3373 3374 3375 dim=0 #3372=(3,36,6,64,32)f32 #3373=(36,6,64,32)f32 #3374=(36,6,64,32)f32 #3375=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_1464 1 1 3373 3376 expr=mul(@0,1.767767e-01) #3373=(36,6,64,32)f32 #3376=(36,6,64,32)f32 +torch.transpose torch.transpose_3193 1 1 3374 3377 dim0=-2 dim1=-1 $input=3374 #3374=(36,6,64,32)f32 #3377=(36,6,32,64)f32 +torch.matmul torch.matmul_2400 2 1 3376 3377 3378 $input=3376 $other=3377 #3376=(36,6,64,32)f32 #3377=(36,6,32,64)f32 #3378=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_18276 0 1 3379 @pnnx_fold_18276=(1,6,64,64)f32 #3379=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_1451 2 1 3378 3379 3380 expr=add(@0,@1) #3378=(36,6,64,64)f32 #3379=(1,6,64,64)f32 #3380=(36,6,64,64)f32 +nn.Softmax layers_mmsa.4.residual_group.blocks.2.attn.softmax 1 1 3380 3381 dim=-1 #3380=(36,6,64,64)f32 #3381=(36,6,64,64)f32 +torch.matmul torch.matmul_2401 2 1 3381 3375 3382 $input=3381 $other=3375 #3381=(36,6,64,64)f32 #3375=(36,6,64,32)f32 #3382=(36,6,64,32)f32 +torch.transpose torch.transpose_3194 1 1 3382 3383 dim0=1 dim1=2 $input=3382 #3382=(36,6,64,32)f32 #3383=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_631 1 1 3383 3384 shape=(36,64,192) $input=3383 #3383=(36,64,6,32)f32 #3384=(36,64,192)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.2.attn.proj 1 1 3384 3385 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3384=(36,64,192)f32 #3385=(36,64,192)f32 +Tensor.reshape Tensor.view_2077 1 1 3385 3386 shape=(1,6,6,8,8,192) $input=3385 #3385=(36,64,192)f32 #3386=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2923 1 1 3386 3387 dims=(0,1,3,2,4,5) $input=3386 #3386=(1,6,6,8,8,192)f32 #3387=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_2079 1 1 3387 3388 shape=(1,2304,192) $input=3387 #3387=(1,6,8,6,8,192)f32 #3388=(1,2304,192)f32 +pnnx.Expression pnnx_expr_1418 2 1 3365 3388 3389 expr=add(@0,@1) #3365=(1,2304,192)f32 #3388=(1,2304,192)f32 #3389=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.4.residual_group.blocks.2.norm2 1 1 3389 3390 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3389=(1,2304,192)f32 #3390=(1,2304,192)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.2.mlp.fc1 1 1 3390 3391 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3390=(1,2304,192)f32 #3391=(1,2304,384)f32 +nn.GELU layers_mmsa.4.residual_group.blocks.2.mlp.act 1 1 3391 3392 #3391=(1,2304,384)f32 #3392=(1,2304,384)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.2.mlp.fc2 1 1 3392 3393 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3392=(1,2304,384)f32 #3393=(1,2304,192)f32 +pnnx.Expression pnnx_expr_1416 2 1 3389 3393 3394 expr=add(@0,@1) #3389=(1,2304,192)f32 #3393=(1,2304,192)f32 #3394=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.4.residual_group.blocks.3.norm1 1 1 3394 3395 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3394=(1,2304,192)f32 #3395=(1,2304,192)f32 +Tensor.view Tensor.view_2080 1 1 3395 3396 shape=(1,48,48,192) $input=3395 #3395=(1,2304,192)f32 #3396=(1,48,48,192)f32 +torch.roll torch.roll_2518 1 1 3396 3397 dims=(1,2) shifts=(-4,-4) $input=3396 #3396=(1,48,48,192)f32 #3397=(1,48,48,192)f32 +Tensor.view Tensor.view_2081 1 1 3397 3398 shape=(1,6,8,6,8,192) $input=3397 #3397=(1,48,48,192)f32 #3398=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2924 1 1 3398 3399 dims=(0,1,3,2,4,5) $input=3398 #3398=(1,6,8,6,8,192)f32 #3399=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_2083 1 1 3399 3400 shape=(36,64,192) $input=3399 #3399=(1,6,6,8,8,192)f32 #3400=(36,64,192)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.3.attn.qkv 1 1 3400 3401 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3400=(36,64,192)f32 #3401=(36,64,576)f32 +Tensor.reshape Tensor.reshape_632 1 1 3401 3402 shape=(36,64,3,6,32) $input=3401 #3401=(36,64,576)f32 #3402=(36,64,3,6,32)f32 +torch.permute torch.permute_2925 1 1 3402 3403 dims=(2,0,3,1,4) $input=3402 #3402=(36,64,3,6,32)f32 #3403=(3,36,6,64,32)f32 +torch.unbind Tensor.select_947 1 3 3403 3404 3405 3406 dim=0 #3403=(3,36,6,64,32)f32 #3404=(36,6,64,32)f32 #3405=(36,6,64,32)f32 #3406=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_1318 1 1 3404 3407 expr=mul(@0,1.767767e-01) #3404=(36,6,64,32)f32 #3407=(36,6,64,32)f32 +torch.transpose torch.transpose_3195 1 1 3405 3408 dim0=-2 dim1=-1 $input=3405 #3405=(36,6,64,32)f32 #3408=(36,6,32,64)f32 +torch.matmul torch.matmul_2402 2 1 3407 3408 3409 $input=3407 $other=3408 #3407=(36,6,64,32)f32 #3408=(36,6,32,64)f32 #3409=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_18429 0 1 3410 @pnnx_fold_18429=(1,6,64,64)f32 #3410=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_1305 2 1 3409 3410 3411 expr=add(@0,@1) #3409=(36,6,64,64)f32 #3410=(1,6,64,64)f32 #3411=(36,6,64,64)f32 +Tensor.view Tensor.view_2086 1 1 3411 3412 shape=(1,36,6,64,64) $input=3411 #3411=(36,6,64,64)f32 #3412=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_18439 0 1 3413 @pnnx_fold_18439=(1,36,1,64,64)f32 #3413=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_1293 2 1 3412 3413 3414 expr=add(@0,@1) #3412=(1,36,6,64,64)f32 #3413=(1,36,1,64,64)f32 #3414=(1,36,6,64,64)f32 +Tensor.view Tensor.view_2087 1 1 3414 3415 shape=(-1,6,64,64) $input=3414 #3414=(1,36,6,64,64)f32 #3415=(36,6,64,64)f32 +nn.Softmax layers_mmsa.4.residual_group.blocks.3.attn.softmax 1 1 3415 3416 dim=-1 #3415=(36,6,64,64)f32 #3416=(36,6,64,64)f32 +torch.matmul torch.matmul_2403 2 1 3416 3406 3417 $input=3416 $other=3406 #3416=(36,6,64,64)f32 #3406=(36,6,64,32)f32 #3417=(36,6,64,32)f32 +torch.transpose torch.transpose_3196 1 1 3417 3418 dim0=1 dim1=2 $input=3417 #3417=(36,6,64,32)f32 #3418=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_633 1 1 3418 3419 shape=(36,64,192) $input=3418 #3418=(36,64,6,32)f32 #3419=(36,64,192)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.3.attn.proj 1 1 3419 3420 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3419=(36,64,192)f32 #3420=(36,64,192)f32 +Tensor.reshape Tensor.view_2089 1 1 3420 3421 shape=(1,6,6,8,8,192) $input=3420 #3420=(36,64,192)f32 #3421=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2927 1 1 3421 3422 dims=(0,1,3,2,4,5) $input=3421 #3421=(1,6,6,8,8,192)f32 #3422=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_199 1 1 3422 3423 shape=(1,48,48,-1) $input=3422 #3422=(1,6,8,6,8,192)f32 #3423=(1,48,48,192)f32 +torch.roll torch.roll_2519 1 1 3423 3424 dims=(1,2) shifts=(4,4) $input=3423 #3423=(1,48,48,192)f32 #3424=(1,48,48,192)f32 +Tensor.view Tensor.view_2091 1 1 3424 3425 shape=(1,2304,192) $input=3424 #3424=(1,48,48,192)f32 #3425=(1,2304,192)f32 +pnnx.Expression pnnx_expr_1251 2 1 3394 3425 3426 expr=add(@0,@1) #3394=(1,2304,192)f32 #3425=(1,2304,192)f32 #3426=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.4.residual_group.blocks.3.norm2 1 1 3426 3427 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3426=(1,2304,192)f32 #3427=(1,2304,192)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.3.mlp.fc1 1 1 3427 3428 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3427=(1,2304,192)f32 #3428=(1,2304,384)f32 +nn.GELU layers_mmsa.4.residual_group.blocks.3.mlp.act 1 1 3428 3429 #3428=(1,2304,384)f32 #3429=(1,2304,384)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.3.mlp.fc2 1 1 3429 3430 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3429=(1,2304,384)f32 #3430=(1,2304,192)f32 +pnnx.Expression pnnx_expr_1249 2 1 3426 3430 3431 expr=add(@0,@1) #3426=(1,2304,192)f32 #3430=(1,2304,192)f32 #3431=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.4.residual_group.blocks.4.norm1 1 1 3431 3432 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3431=(1,2304,192)f32 #3432=(1,2304,192)f32 +Tensor.reshape Tensor.view_2093 1 1 3432 3433 shape=(1,6,8,6,8,192) $input=3432 #3432=(1,2304,192)f32 #3433=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2928 1 1 3433 3434 dims=(0,1,3,2,4,5) $input=3433 #3433=(1,6,8,6,8,192)f32 #3434=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_2095 1 1 3434 3435 shape=(36,64,192) $input=3434 #3434=(1,6,6,8,8,192)f32 #3435=(36,64,192)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.4.attn.qkv 1 1 3435 3436 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3435=(36,64,192)f32 #3436=(36,64,576)f32 +Tensor.reshape Tensor.reshape_634 1 1 3436 3437 shape=(36,64,3,6,32) $input=3436 #3436=(36,64,576)f32 #3437=(36,64,3,6,32)f32 +torch.permute torch.permute_2929 1 1 3437 3438 dims=(2,0,3,1,4) $input=3437 #3437=(36,64,3,6,32)f32 #3438=(3,36,6,64,32)f32 +torch.unbind Tensor.select_950 1 3 3438 3439 3440 3441 dim=0 #3438=(3,36,6,64,32)f32 #3439=(36,6,64,32)f32 #3440=(36,6,64,32)f32 #3441=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_1161 1 1 3439 3442 expr=mul(@0,1.767767e-01) #3439=(36,6,64,32)f32 #3442=(36,6,64,32)f32 +torch.transpose torch.transpose_3197 1 1 3440 3443 dim0=-2 dim1=-1 $input=3440 #3440=(36,6,64,32)f32 #3443=(36,6,32,64)f32 +torch.matmul torch.matmul_2404 2 1 3442 3443 3444 $input=3442 $other=3443 #3442=(36,6,64,32)f32 #3443=(36,6,32,64)f32 #3444=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_18588 0 1 3445 @pnnx_fold_18588=(1,6,64,64)f32 #3445=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_1148 2 1 3444 3445 3446 expr=add(@0,@1) #3444=(36,6,64,64)f32 #3445=(1,6,64,64)f32 #3446=(36,6,64,64)f32 +nn.Softmax layers_mmsa.4.residual_group.blocks.4.attn.softmax 1 1 3446 3447 dim=-1 #3446=(36,6,64,64)f32 #3447=(36,6,64,64)f32 +torch.matmul torch.matmul_2405 2 1 3447 3441 3448 $input=3447 $other=3441 #3447=(36,6,64,64)f32 #3441=(36,6,64,32)f32 #3448=(36,6,64,32)f32 +torch.transpose torch.transpose_3198 1 1 3448 3449 dim0=1 dim1=2 $input=3448 #3448=(36,6,64,32)f32 #3449=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_635 1 1 3449 3450 shape=(36,64,192) $input=3449 #3449=(36,64,6,32)f32 #3450=(36,64,192)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.4.attn.proj 1 1 3450 3451 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3450=(36,64,192)f32 #3451=(36,64,192)f32 +Tensor.reshape Tensor.view_2099 1 1 3451 3452 shape=(1,6,6,8,8,192) $input=3451 #3451=(36,64,192)f32 #3452=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2931 1 1 3452 3453 dims=(0,1,3,2,4,5) $input=3452 #3452=(1,6,6,8,8,192)f32 #3453=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_2101 1 1 3453 3454 shape=(1,2304,192) $input=3453 #3453=(1,6,8,6,8,192)f32 #3454=(1,2304,192)f32 +pnnx.Expression pnnx_expr_1115 2 1 3431 3454 3455 expr=add(@0,@1) #3431=(1,2304,192)f32 #3454=(1,2304,192)f32 #3455=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.4.residual_group.blocks.4.norm2 1 1 3455 3456 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3455=(1,2304,192)f32 #3456=(1,2304,192)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.4.mlp.fc1 1 1 3456 3457 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3456=(1,2304,192)f32 #3457=(1,2304,384)f32 +nn.GELU layers_mmsa.4.residual_group.blocks.4.mlp.act 1 1 3457 3458 #3457=(1,2304,384)f32 #3458=(1,2304,384)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.4.mlp.fc2 1 1 3458 3459 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3458=(1,2304,384)f32 #3459=(1,2304,192)f32 +pnnx.Expression pnnx_expr_1113 2 1 3455 3459 3460 expr=add(@0,@1) #3455=(1,2304,192)f32 #3459=(1,2304,192)f32 #3460=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.4.residual_group.blocks.5.norm1 1 1 3460 3461 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3460=(1,2304,192)f32 #3461=(1,2304,192)f32 +Tensor.view Tensor.view_2102 1 1 3461 3462 shape=(1,48,48,192) $input=3461 #3461=(1,2304,192)f32 #3462=(1,48,48,192)f32 +torch.roll torch.roll_2520 1 1 3462 3463 dims=(1,2) shifts=(-4,-4) $input=3462 #3462=(1,48,48,192)f32 #3463=(1,48,48,192)f32 +Tensor.view Tensor.view_2103 1 1 3463 3464 shape=(1,6,8,6,8,192) $input=3463 #3463=(1,48,48,192)f32 #3464=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2932 1 1 3464 3465 dims=(0,1,3,2,4,5) $input=3464 #3464=(1,6,8,6,8,192)f32 #3465=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_2105 1 1 3465 3466 shape=(36,64,192) $input=3465 #3465=(1,6,6,8,8,192)f32 #3466=(36,64,192)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.5.attn.qkv 1 1 3466 3467 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3466=(36,64,192)f32 #3467=(36,64,576)f32 +Tensor.reshape Tensor.reshape_636 1 1 3467 3468 shape=(36,64,3,6,32) $input=3467 #3467=(36,64,576)f32 #3468=(36,64,3,6,32)f32 +torch.permute torch.permute_2933 1 1 3468 3469 dims=(2,0,3,1,4) $input=3468 #3468=(36,64,3,6,32)f32 #3469=(3,36,6,64,32)f32 +torch.unbind Tensor.select_953 1 3 3469 3470 3471 3472 dim=0 #3469=(3,36,6,64,32)f32 #3470=(36,6,64,32)f32 #3471=(36,6,64,32)f32 #3472=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_1015 1 1 3470 3473 expr=mul(@0,1.767767e-01) #3470=(36,6,64,32)f32 #3473=(36,6,64,32)f32 +torch.transpose torch.transpose_3199 1 1 3471 3474 dim0=-2 dim1=-1 $input=3471 #3471=(36,6,64,32)f32 #3474=(36,6,32,64)f32 +torch.matmul torch.matmul_2406 2 1 3473 3474 3475 $input=3473 $other=3474 #3473=(36,6,64,32)f32 #3474=(36,6,32,64)f32 #3475=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_18741 0 1 3476 @pnnx_fold_18741=(1,6,64,64)f32 #3476=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_1002 2 1 3475 3476 3477 expr=add(@0,@1) #3475=(36,6,64,64)f32 #3476=(1,6,64,64)f32 #3477=(36,6,64,64)f32 +Tensor.view Tensor.view_2108 1 1 3477 3478 shape=(1,36,6,64,64) $input=3477 #3477=(36,6,64,64)f32 #3478=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_18751 0 1 3479 @pnnx_fold_18751=(1,36,1,64,64)f32 #3479=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_990 2 1 3478 3479 3480 expr=add(@0,@1) #3478=(1,36,6,64,64)f32 #3479=(1,36,1,64,64)f32 #3480=(1,36,6,64,64)f32 +Tensor.view Tensor.view_2109 1 1 3480 3481 shape=(-1,6,64,64) $input=3480 #3480=(1,36,6,64,64)f32 #3481=(36,6,64,64)f32 +nn.Softmax layers_mmsa.4.residual_group.blocks.5.attn.softmax 1 1 3481 3482 dim=-1 #3481=(36,6,64,64)f32 #3482=(36,6,64,64)f32 +torch.matmul torch.matmul_2407 2 1 3482 3472 3483 $input=3482 $other=3472 #3482=(36,6,64,64)f32 #3472=(36,6,64,32)f32 #3483=(36,6,64,32)f32 +torch.transpose torch.transpose_3200 1 1 3483 3484 dim0=1 dim1=2 $input=3483 #3483=(36,6,64,32)f32 #3484=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_637 1 1 3484 3485 shape=(36,64,192) $input=3484 #3484=(36,64,6,32)f32 #3485=(36,64,192)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.5.attn.proj 1 1 3485 3486 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3485=(36,64,192)f32 #3486=(36,64,192)f32 +Tensor.reshape Tensor.view_2111 1 1 3486 3487 shape=(1,6,6,8,8,192) $input=3486 #3486=(36,64,192)f32 #3487=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2935 1 1 3487 3488 dims=(0,1,3,2,4,5) $input=3487 #3487=(1,6,6,8,8,192)f32 #3488=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_203 1 1 3488 3489 shape=(1,48,48,-1) $input=3488 #3488=(1,6,8,6,8,192)f32 #3489=(1,48,48,192)f32 +torch.roll torch.roll_2521 1 1 3489 3490 dims=(1,2) shifts=(4,4) $input=3489 #3489=(1,48,48,192)f32 #3490=(1,48,48,192)f32 +Tensor.view Tensor.view_2113 1 1 3490 3491 shape=(1,2304,192) $input=3490 #3490=(1,48,48,192)f32 #3491=(1,2304,192)f32 +pnnx.Expression pnnx_expr_948 2 1 3460 3491 3492 expr=add(@0,@1) #3460=(1,2304,192)f32 #3491=(1,2304,192)f32 #3492=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.4.residual_group.blocks.5.norm2 1 1 3492 3493 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3492=(1,2304,192)f32 #3493=(1,2304,192)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.5.mlp.fc1 1 1 3493 3494 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3493=(1,2304,192)f32 #3494=(1,2304,384)f32 +nn.GELU layers_mmsa.4.residual_group.blocks.5.mlp.act 1 1 3494 3495 #3494=(1,2304,384)f32 #3495=(1,2304,384)f32 +nn.Linear layers_mmsa.4.residual_group.blocks.5.mlp.fc2 1 1 3495 3496 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3495=(1,2304,384)f32 #3496=(1,2304,192)f32 +pnnx.Expression pnnx_expr_946 2 1 3492 3496 3497 expr=add(@0,@1) #3492=(1,2304,192)f32 #3496=(1,2304,192)f32 #3497=(1,2304,192)f32 +torch.transpose torch.transpose_3201 1 1 3497 3498 dim0=1 dim1=2 $input=3497 #3497=(1,2304,192)f32 #3498=(1,192,2304)f32 +Tensor.view Tensor.view_2114 1 1 3498 3499 shape=(1,192,48,48) $input=3498 #3498=(1,192,2304)f32 #3499=(1,192,48,48)f32 +nn.Conv2d layers_mmsa.4.conv 1 1 3499 3500 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #3499=(1,192,48,48)f32 #3500=(1,192,48,48)f32 +torch.flatten torch.flatten_2202 1 1 3500 3501 end_dim=-1 start_dim=2 $input=3500 #3500=(1,192,48,48)f32 #3501=(1,192,2304)f32 +torch.transpose torch.transpose_3202 1 1 3501 3502 dim0=1 dim1=2 $input=3501 #3501=(1,192,2304)f32 #3502=(1,2304,192)f32 +pnnx.Expression pnnx_expr_933 2 1 3502 3299 3503 expr=add(@0,@1) #3502=(1,2304,192)f32 #3299=(1,2304,192)f32 #3503=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.5.residual_group.blocks.0.norm1 1 1 3503 3504 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3503=(1,2304,192)f32 #3504=(1,2304,192)f32 +Tensor.reshape Tensor.view_2116 1 1 3504 3505 shape=(1,6,8,6,8,192) $input=3504 #3504=(1,2304,192)f32 #3505=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2936 1 1 3505 3506 dims=(0,1,3,2,4,5) $input=3505 #3505=(1,6,8,6,8,192)f32 #3506=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_2118 1 1 3506 3507 shape=(36,64,192) $input=3506 #3506=(1,6,6,8,8,192)f32 #3507=(36,64,192)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.0.attn.qkv 1 1 3507 3508 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3507=(36,64,192)f32 #3508=(36,64,576)f32 +Tensor.reshape Tensor.reshape_638 1 1 3508 3509 shape=(36,64,3,6,32) $input=3508 #3508=(36,64,576)f32 #3509=(36,64,3,6,32)f32 +torch.permute torch.permute_2937 1 1 3509 3510 dims=(2,0,3,1,4) $input=3509 #3509=(36,64,3,6,32)f32 #3510=(3,36,6,64,32)f32 +torch.unbind Tensor.select_956 1 3 3510 3511 3512 3513 dim=0 #3510=(3,36,6,64,32)f32 #3511=(36,6,64,32)f32 #3512=(36,6,64,32)f32 #3513=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_844 1 1 3511 3514 expr=mul(@0,1.767767e-01) #3511=(36,6,64,32)f32 #3514=(36,6,64,32)f32 +torch.transpose torch.transpose_3203 1 1 3512 3515 dim0=-2 dim1=-1 $input=3512 #3512=(36,6,64,32)f32 #3515=(36,6,32,64)f32 +torch.matmul torch.matmul_2408 2 1 3514 3515 3516 $input=3514 $other=3515 #3514=(36,6,64,32)f32 #3515=(36,6,32,64)f32 #3516=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_18934 0 1 3517 @pnnx_fold_18934=(1,6,64,64)f32 #3517=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_831 2 1 3516 3517 3518 expr=add(@0,@1) #3516=(36,6,64,64)f32 #3517=(1,6,64,64)f32 #3518=(36,6,64,64)f32 +nn.Softmax layers_mmsa.5.residual_group.blocks.0.attn.softmax 1 1 3518 3519 dim=-1 #3518=(36,6,64,64)f32 #3519=(36,6,64,64)f32 +torch.matmul torch.matmul_2409 2 1 3519 3513 3520 $input=3519 $other=3513 #3519=(36,6,64,64)f32 #3513=(36,6,64,32)f32 #3520=(36,6,64,32)f32 +torch.transpose torch.transpose_3204 1 1 3520 3521 dim0=1 dim1=2 $input=3520 #3520=(36,6,64,32)f32 #3521=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_639 1 1 3521 3522 shape=(36,64,192) $input=3521 #3521=(36,64,6,32)f32 #3522=(36,64,192)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.0.attn.proj 1 1 3522 3523 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3522=(36,64,192)f32 #3523=(36,64,192)f32 +Tensor.reshape Tensor.view_2122 1 1 3523 3524 shape=(1,6,6,8,8,192) $input=3523 #3523=(36,64,192)f32 #3524=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2939 1 1 3524 3525 dims=(0,1,3,2,4,5) $input=3524 #3524=(1,6,6,8,8,192)f32 #3525=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_2124 1 1 3525 3526 shape=(1,2304,192) $input=3525 #3525=(1,6,8,6,8,192)f32 #3526=(1,2304,192)f32 +pnnx.Expression pnnx_expr_798 2 1 3503 3526 3527 expr=add(@0,@1) #3503=(1,2304,192)f32 #3526=(1,2304,192)f32 #3527=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.5.residual_group.blocks.0.norm2 1 1 3527 3528 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3527=(1,2304,192)f32 #3528=(1,2304,192)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.0.mlp.fc1 1 1 3528 3529 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3528=(1,2304,192)f32 #3529=(1,2304,384)f32 +nn.GELU layers_mmsa.5.residual_group.blocks.0.mlp.act 1 1 3529 3530 #3529=(1,2304,384)f32 #3530=(1,2304,384)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.0.mlp.fc2 1 1 3530 3531 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3530=(1,2304,384)f32 #3531=(1,2304,192)f32 +pnnx.Expression pnnx_expr_796 2 1 3527 3531 3532 expr=add(@0,@1) #3527=(1,2304,192)f32 #3531=(1,2304,192)f32 #3532=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.5.residual_group.blocks.1.norm1 1 1 3532 3533 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3532=(1,2304,192)f32 #3533=(1,2304,192)f32 +Tensor.view Tensor.view_2125 1 1 3533 3534 shape=(1,48,48,192) $input=3533 #3533=(1,2304,192)f32 #3534=(1,48,48,192)f32 +torch.roll torch.roll_2522 1 1 3534 3535 dims=(1,2) shifts=(-4,-4) $input=3534 #3534=(1,48,48,192)f32 #3535=(1,48,48,192)f32 +Tensor.view Tensor.view_2126 1 1 3535 3536 shape=(1,6,8,6,8,192) $input=3535 #3535=(1,48,48,192)f32 #3536=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2940 1 1 3536 3537 dims=(0,1,3,2,4,5) $input=3536 #3536=(1,6,8,6,8,192)f32 #3537=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_2128 1 1 3537 3538 shape=(36,64,192) $input=3537 #3537=(1,6,6,8,8,192)f32 #3538=(36,64,192)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.1.attn.qkv 1 1 3538 3539 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3538=(36,64,192)f32 #3539=(36,64,576)f32 +Tensor.reshape Tensor.reshape_640 1 1 3539 3540 shape=(36,64,3,6,32) $input=3539 #3539=(36,64,576)f32 #3540=(36,64,3,6,32)f32 +torch.permute torch.permute_2941 1 1 3540 3541 dims=(2,0,3,1,4) $input=3540 #3540=(36,64,3,6,32)f32 #3541=(3,36,6,64,32)f32 +torch.unbind Tensor.select_959 1 3 3541 3542 3543 3544 dim=0 #3541=(3,36,6,64,32)f32 #3542=(36,6,64,32)f32 #3543=(36,6,64,32)f32 #3544=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_698 1 1 3542 3545 expr=mul(@0,1.767767e-01) #3542=(36,6,64,32)f32 #3545=(36,6,64,32)f32 +torch.transpose torch.transpose_3205 1 1 3543 3546 dim0=-2 dim1=-1 $input=3543 #3543=(36,6,64,32)f32 #3546=(36,6,32,64)f32 +torch.matmul torch.matmul_2410 2 1 3545 3546 3547 $input=3545 $other=3546 #3545=(36,6,64,32)f32 #3546=(36,6,32,64)f32 #3547=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_19087 0 1 3548 @pnnx_fold_19087=(1,6,64,64)f32 #3548=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_685 2 1 3547 3548 3549 expr=add(@0,@1) #3547=(36,6,64,64)f32 #3548=(1,6,64,64)f32 #3549=(36,6,64,64)f32 +Tensor.view Tensor.view_2131 1 1 3549 3550 shape=(1,36,6,64,64) $input=3549 #3549=(36,6,64,64)f32 #3550=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_19097 0 1 3551 @pnnx_fold_19097=(1,36,1,64,64)f32 #3551=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_673 2 1 3550 3551 3552 expr=add(@0,@1) #3550=(1,36,6,64,64)f32 #3551=(1,36,1,64,64)f32 #3552=(1,36,6,64,64)f32 +Tensor.view Tensor.view_2132 1 1 3552 3553 shape=(-1,6,64,64) $input=3552 #3552=(1,36,6,64,64)f32 #3553=(36,6,64,64)f32 +nn.Softmax layers_mmsa.5.residual_group.blocks.1.attn.softmax 1 1 3553 3554 dim=-1 #3553=(36,6,64,64)f32 #3554=(36,6,64,64)f32 +torch.matmul torch.matmul_2411 2 1 3554 3544 3555 $input=3554 $other=3544 #3554=(36,6,64,64)f32 #3544=(36,6,64,32)f32 #3555=(36,6,64,32)f32 +torch.transpose torch.transpose_3206 1 1 3555 3556 dim0=1 dim1=2 $input=3555 #3555=(36,6,64,32)f32 #3556=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_641 1 1 3556 3557 shape=(36,64,192) $input=3556 #3556=(36,64,6,32)f32 #3557=(36,64,192)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.1.attn.proj 1 1 3557 3558 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3557=(36,64,192)f32 #3558=(36,64,192)f32 +Tensor.reshape Tensor.view_2134 1 1 3558 3559 shape=(1,6,6,8,8,192) $input=3558 #3558=(36,64,192)f32 #3559=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2943 1 1 3559 3560 dims=(0,1,3,2,4,5) $input=3559 #3559=(1,6,6,8,8,192)f32 #3560=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_207 1 1 3560 3561 shape=(1,48,48,-1) $input=3560 #3560=(1,6,8,6,8,192)f32 #3561=(1,48,48,192)f32 +torch.roll torch.roll_2523 1 1 3561 3562 dims=(1,2) shifts=(4,4) $input=3561 #3561=(1,48,48,192)f32 #3562=(1,48,48,192)f32 +Tensor.view Tensor.view_2136 1 1 3562 3563 shape=(1,2304,192) $input=3562 #3562=(1,48,48,192)f32 #3563=(1,2304,192)f32 +pnnx.Expression pnnx_expr_631 2 1 3532 3563 3564 expr=add(@0,@1) #3532=(1,2304,192)f32 #3563=(1,2304,192)f32 #3564=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.5.residual_group.blocks.1.norm2 1 1 3564 3565 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3564=(1,2304,192)f32 #3565=(1,2304,192)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.1.mlp.fc1 1 1 3565 3566 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3565=(1,2304,192)f32 #3566=(1,2304,384)f32 +nn.GELU layers_mmsa.5.residual_group.blocks.1.mlp.act 1 1 3566 3567 #3566=(1,2304,384)f32 #3567=(1,2304,384)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.1.mlp.fc2 1 1 3567 3568 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3567=(1,2304,384)f32 #3568=(1,2304,192)f32 +pnnx.Expression pnnx_expr_629 2 1 3564 3568 3569 expr=add(@0,@1) #3564=(1,2304,192)f32 #3568=(1,2304,192)f32 #3569=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.5.residual_group.blocks.2.norm1 1 1 3569 3570 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3569=(1,2304,192)f32 #3570=(1,2304,192)f32 +Tensor.reshape Tensor.view_2138 1 1 3570 3571 shape=(1,6,8,6,8,192) $input=3570 #3570=(1,2304,192)f32 #3571=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2944 1 1 3571 3572 dims=(0,1,3,2,4,5) $input=3571 #3571=(1,6,8,6,8,192)f32 #3572=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_2140 1 1 3572 3573 shape=(36,64,192) $input=3572 #3572=(1,6,6,8,8,192)f32 #3573=(36,64,192)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.2.attn.qkv 1 1 3573 3574 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3573=(36,64,192)f32 #3574=(36,64,576)f32 +Tensor.reshape Tensor.reshape_642 1 1 3574 3575 shape=(36,64,3,6,32) $input=3574 #3574=(36,64,576)f32 #3575=(36,64,3,6,32)f32 +torch.permute torch.permute_2945 1 1 3575 3576 dims=(2,0,3,1,4) $input=3575 #3575=(36,64,3,6,32)f32 #3576=(3,36,6,64,32)f32 +torch.unbind Tensor.select_962 1 3 3576 3577 3578 3579 dim=0 #3576=(3,36,6,64,32)f32 #3577=(36,6,64,32)f32 #3578=(36,6,64,32)f32 #3579=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_541 1 1 3577 3580 expr=mul(@0,1.767767e-01) #3577=(36,6,64,32)f32 #3580=(36,6,64,32)f32 +torch.transpose torch.transpose_3207 1 1 3578 3581 dim0=-2 dim1=-1 $input=3578 #3578=(36,6,64,32)f32 #3581=(36,6,32,64)f32 +torch.matmul torch.matmul_2412 2 1 3580 3581 3582 $input=3580 $other=3581 #3580=(36,6,64,32)f32 #3581=(36,6,32,64)f32 #3582=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_19246 0 1 3583 @pnnx_fold_19246=(1,6,64,64)f32 #3583=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_528 2 1 3582 3583 3584 expr=add(@0,@1) #3582=(36,6,64,64)f32 #3583=(1,6,64,64)f32 #3584=(36,6,64,64)f32 +nn.Softmax layers_mmsa.5.residual_group.blocks.2.attn.softmax 1 1 3584 3585 dim=-1 #3584=(36,6,64,64)f32 #3585=(36,6,64,64)f32 +torch.matmul torch.matmul_2413 2 1 3585 3579 3586 $input=3585 $other=3579 #3585=(36,6,64,64)f32 #3579=(36,6,64,32)f32 #3586=(36,6,64,32)f32 +torch.transpose torch.transpose_3208 1 1 3586 3587 dim0=1 dim1=2 $input=3586 #3586=(36,6,64,32)f32 #3587=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_643 1 1 3587 3588 shape=(36,64,192) $input=3587 #3587=(36,64,6,32)f32 #3588=(36,64,192)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.2.attn.proj 1 1 3588 3589 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3588=(36,64,192)f32 #3589=(36,64,192)f32 +Tensor.reshape Tensor.view_2144 1 1 3589 3590 shape=(1,6,6,8,8,192) $input=3589 #3589=(36,64,192)f32 #3590=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2947 1 1 3590 3591 dims=(0,1,3,2,4,5) $input=3590 #3590=(1,6,6,8,8,192)f32 #3591=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_2146 1 1 3591 3592 shape=(1,2304,192) $input=3591 #3591=(1,6,8,6,8,192)f32 #3592=(1,2304,192)f32 +pnnx.Expression pnnx_expr_495 2 1 3569 3592 3593 expr=add(@0,@1) #3569=(1,2304,192)f32 #3592=(1,2304,192)f32 #3593=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.5.residual_group.blocks.2.norm2 1 1 3593 3594 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3593=(1,2304,192)f32 #3594=(1,2304,192)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.2.mlp.fc1 1 1 3594 3595 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3594=(1,2304,192)f32 #3595=(1,2304,384)f32 +nn.GELU layers_mmsa.5.residual_group.blocks.2.mlp.act 1 1 3595 3596 #3595=(1,2304,384)f32 #3596=(1,2304,384)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.2.mlp.fc2 1 1 3596 3597 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3596=(1,2304,384)f32 #3597=(1,2304,192)f32 +pnnx.Expression pnnx_expr_493 2 1 3593 3597 3598 expr=add(@0,@1) #3593=(1,2304,192)f32 #3597=(1,2304,192)f32 #3598=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.5.residual_group.blocks.3.norm1 1 1 3598 3599 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3598=(1,2304,192)f32 #3599=(1,2304,192)f32 +Tensor.view Tensor.view_2147 1 1 3599 3600 shape=(1,48,48,192) $input=3599 #3599=(1,2304,192)f32 #3600=(1,48,48,192)f32 +torch.roll torch.roll_2524 1 1 3600 3601 dims=(1,2) shifts=(-4,-4) $input=3600 #3600=(1,48,48,192)f32 #3601=(1,48,48,192)f32 +Tensor.view Tensor.view_2148 1 1 3601 3602 shape=(1,6,8,6,8,192) $input=3601 #3601=(1,48,48,192)f32 #3602=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2948 1 1 3602 3603 dims=(0,1,3,2,4,5) $input=3602 #3602=(1,6,8,6,8,192)f32 #3603=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_2150 1 1 3603 3604 shape=(36,64,192) $input=3603 #3603=(1,6,6,8,8,192)f32 #3604=(36,64,192)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.3.attn.qkv 1 1 3604 3605 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3604=(36,64,192)f32 #3605=(36,64,576)f32 +Tensor.reshape Tensor.reshape_644 1 1 3605 3606 shape=(36,64,3,6,32) $input=3605 #3605=(36,64,576)f32 #3606=(36,64,3,6,32)f32 +torch.permute torch.permute_2949 1 1 3606 3607 dims=(2,0,3,1,4) $input=3606 #3606=(36,64,3,6,32)f32 #3607=(3,36,6,64,32)f32 +torch.unbind Tensor.select_965 1 3 3607 3608 3609 3610 dim=0 #3607=(3,36,6,64,32)f32 #3608=(36,6,64,32)f32 #3609=(36,6,64,32)f32 #3610=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_395 1 1 3608 3611 expr=mul(@0,1.767767e-01) #3608=(36,6,64,32)f32 #3611=(36,6,64,32)f32 +torch.transpose torch.transpose_3209 1 1 3609 3612 dim0=-2 dim1=-1 $input=3609 #3609=(36,6,64,32)f32 #3612=(36,6,32,64)f32 +torch.matmul torch.matmul_2414 2 1 3611 3612 3613 $input=3611 $other=3612 #3611=(36,6,64,32)f32 #3612=(36,6,32,64)f32 #3613=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_19399 0 1 3614 @pnnx_fold_19399=(1,6,64,64)f32 #3614=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_382 2 1 3613 3614 3615 expr=add(@0,@1) #3613=(36,6,64,64)f32 #3614=(1,6,64,64)f32 #3615=(36,6,64,64)f32 +Tensor.view Tensor.view_2153 1 1 3615 3616 shape=(1,36,6,64,64) $input=3615 #3615=(36,6,64,64)f32 #3616=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_19409 0 1 3617 @pnnx_fold_19409=(1,36,1,64,64)f32 #3617=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_370 2 1 3616 3617 3618 expr=add(@0,@1) #3616=(1,36,6,64,64)f32 #3617=(1,36,1,64,64)f32 #3618=(1,36,6,64,64)f32 +Tensor.view Tensor.view_2154 1 1 3618 3619 shape=(-1,6,64,64) $input=3618 #3618=(1,36,6,64,64)f32 #3619=(36,6,64,64)f32 +nn.Softmax layers_mmsa.5.residual_group.blocks.3.attn.softmax 1 1 3619 3620 dim=-1 #3619=(36,6,64,64)f32 #3620=(36,6,64,64)f32 +torch.matmul torch.matmul_2415 2 1 3620 3610 3621 $input=3620 $other=3610 #3620=(36,6,64,64)f32 #3610=(36,6,64,32)f32 #3621=(36,6,64,32)f32 +torch.transpose torch.transpose_3210 1 1 3621 3622 dim0=1 dim1=2 $input=3621 #3621=(36,6,64,32)f32 #3622=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_645 1 1 3622 3623 shape=(36,64,192) $input=3622 #3622=(36,64,6,32)f32 #3623=(36,64,192)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.3.attn.proj 1 1 3623 3624 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3623=(36,64,192)f32 #3624=(36,64,192)f32 +Tensor.reshape Tensor.view_2156 1 1 3624 3625 shape=(1,6,6,8,8,192) $input=3624 #3624=(36,64,192)f32 #3625=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2951 1 1 3625 3626 dims=(0,1,3,2,4,5) $input=3625 #3625=(1,6,6,8,8,192)f32 #3626=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_211 1 1 3626 3627 shape=(1,48,48,-1) $input=3626 #3626=(1,6,8,6,8,192)f32 #3627=(1,48,48,192)f32 +torch.roll torch.roll_2525 1 1 3627 3628 dims=(1,2) shifts=(4,4) $input=3627 #3627=(1,48,48,192)f32 #3628=(1,48,48,192)f32 +Tensor.view Tensor.view_2158 1 1 3628 3629 shape=(1,2304,192) $input=3628 #3628=(1,48,48,192)f32 #3629=(1,2304,192)f32 +pnnx.Expression pnnx_expr_328 2 1 3598 3629 3630 expr=add(@0,@1) #3598=(1,2304,192)f32 #3629=(1,2304,192)f32 #3630=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.5.residual_group.blocks.3.norm2 1 1 3630 3631 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3630=(1,2304,192)f32 #3631=(1,2304,192)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.3.mlp.fc1 1 1 3631 3632 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3631=(1,2304,192)f32 #3632=(1,2304,384)f32 +nn.GELU layers_mmsa.5.residual_group.blocks.3.mlp.act 1 1 3632 3633 #3632=(1,2304,384)f32 #3633=(1,2304,384)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.3.mlp.fc2 1 1 3633 3634 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3633=(1,2304,384)f32 #3634=(1,2304,192)f32 +pnnx.Expression pnnx_expr_326 2 1 3630 3634 3635 expr=add(@0,@1) #3630=(1,2304,192)f32 #3634=(1,2304,192)f32 #3635=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.5.residual_group.blocks.4.norm1 1 1 3635 3636 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3635=(1,2304,192)f32 #3636=(1,2304,192)f32 +Tensor.reshape Tensor.view_2160 1 1 3636 3637 shape=(1,6,8,6,8,192) $input=3636 #3636=(1,2304,192)f32 #3637=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2952 1 1 3637 3638 dims=(0,1,3,2,4,5) $input=3637 #3637=(1,6,8,6,8,192)f32 #3638=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_2162 1 1 3638 3639 shape=(36,64,192) $input=3638 #3638=(1,6,6,8,8,192)f32 #3639=(36,64,192)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.4.attn.qkv 1 1 3639 3640 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3639=(36,64,192)f32 #3640=(36,64,576)f32 +Tensor.reshape Tensor.reshape_646 1 1 3640 3641 shape=(36,64,3,6,32) $input=3640 #3640=(36,64,576)f32 #3641=(36,64,3,6,32)f32 +torch.permute torch.permute_2953 1 1 3641 3642 dims=(2,0,3,1,4) $input=3641 #3641=(36,64,3,6,32)f32 #3642=(3,36,6,64,32)f32 +torch.unbind Tensor.select_968 1 3 3642 3643 3644 3645 dim=0 #3642=(3,36,6,64,32)f32 #3643=(36,6,64,32)f32 #3644=(36,6,64,32)f32 #3645=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_238 1 1 3643 3646 expr=mul(@0,1.767767e-01) #3643=(36,6,64,32)f32 #3646=(36,6,64,32)f32 +torch.transpose torch.transpose_3211 1 1 3644 3647 dim0=-2 dim1=-1 $input=3644 #3644=(36,6,64,32)f32 #3647=(36,6,32,64)f32 +torch.matmul torch.matmul_2416 2 1 3646 3647 3648 $input=3646 $other=3647 #3646=(36,6,64,32)f32 #3647=(36,6,32,64)f32 #3648=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_19558 0 1 3649 @pnnx_fold_19558=(1,6,64,64)f32 #3649=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_225 2 1 3648 3649 3650 expr=add(@0,@1) #3648=(36,6,64,64)f32 #3649=(1,6,64,64)f32 #3650=(36,6,64,64)f32 +nn.Softmax layers_mmsa.5.residual_group.blocks.4.attn.softmax 1 1 3650 3651 dim=-1 #3650=(36,6,64,64)f32 #3651=(36,6,64,64)f32 +torch.matmul torch.matmul_2417 2 1 3651 3645 3652 $input=3651 $other=3645 #3651=(36,6,64,64)f32 #3645=(36,6,64,32)f32 #3652=(36,6,64,32)f32 +torch.transpose torch.transpose_3212 1 1 3652 3653 dim0=1 dim1=2 $input=3652 #3652=(36,6,64,32)f32 #3653=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_647 1 1 3653 3654 shape=(36,64,192) $input=3653 #3653=(36,64,6,32)f32 #3654=(36,64,192)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.4.attn.proj 1 1 3654 3655 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3654=(36,64,192)f32 #3655=(36,64,192)f32 +Tensor.reshape Tensor.view_2166 1 1 3655 3656 shape=(1,6,6,8,8,192) $input=3655 #3655=(36,64,192)f32 #3656=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2955 1 1 3656 3657 dims=(0,1,3,2,4,5) $input=3656 #3656=(1,6,6,8,8,192)f32 #3657=(1,6,8,6,8,192)f32 +Tensor.reshape Tensor.view_2168 1 1 3657 3658 shape=(1,2304,192) $input=3657 #3657=(1,6,8,6,8,192)f32 #3658=(1,2304,192)f32 +pnnx.Expression pnnx_expr_192 2 1 3635 3658 3659 expr=add(@0,@1) #3635=(1,2304,192)f32 #3658=(1,2304,192)f32 #3659=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.5.residual_group.blocks.4.norm2 1 1 3659 3660 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3659=(1,2304,192)f32 #3660=(1,2304,192)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.4.mlp.fc1 1 1 3660 3661 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3660=(1,2304,192)f32 #3661=(1,2304,384)f32 +nn.GELU layers_mmsa.5.residual_group.blocks.4.mlp.act 1 1 3661 3662 #3661=(1,2304,384)f32 #3662=(1,2304,384)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.4.mlp.fc2 1 1 3662 3663 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3662=(1,2304,384)f32 #3663=(1,2304,192)f32 +pnnx.Expression pnnx_expr_190 2 1 3659 3663 3664 expr=add(@0,@1) #3659=(1,2304,192)f32 #3663=(1,2304,192)f32 #3664=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.5.residual_group.blocks.5.norm1 1 1 3664 3665 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3664=(1,2304,192)f32 #3665=(1,2304,192)f32 +Tensor.view Tensor.view_2169 1 1 3665 3666 shape=(1,48,48,192) $input=3665 #3665=(1,2304,192)f32 #3666=(1,48,48,192)f32 +torch.roll torch.roll_2526 1 1 3666 3667 dims=(1,2) shifts=(-4,-4) $input=3666 #3666=(1,48,48,192)f32 #3667=(1,48,48,192)f32 +Tensor.view Tensor.view_2170 1 1 3667 3668 shape=(1,6,8,6,8,192) $input=3667 #3667=(1,48,48,192)f32 #3668=(1,6,8,6,8,192)f32 +torch.permute torch.permute_2956 1 1 3668 3669 dims=(0,1,3,2,4,5) $input=3668 #3668=(1,6,8,6,8,192)f32 #3669=(1,6,6,8,8,192)f32 +Tensor.reshape Tensor.view_2172 1 1 3669 3670 shape=(36,64,192) $input=3669 #3669=(1,6,6,8,8,192)f32 #3670=(36,64,192)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.5.attn.qkv 1 1 3670 3671 bias=True in_features=192 out_features=576 @bias=(576)f32 @weight=(576,192)f32 #3670=(36,64,192)f32 #3671=(36,64,576)f32 +Tensor.reshape Tensor.reshape_648 1 1 3671 3672 shape=(36,64,3,6,32) $input=3671 #3671=(36,64,576)f32 #3672=(36,64,3,6,32)f32 +torch.permute torch.permute_2957 1 1 3672 3673 dims=(2,0,3,1,4) $input=3672 #3672=(36,64,3,6,32)f32 #3673=(3,36,6,64,32)f32 +torch.unbind Tensor.select_971 1 3 3673 3674 3675 3676 dim=0 #3673=(3,36,6,64,32)f32 #3674=(36,6,64,32)f32 #3675=(36,6,64,32)f32 #3676=(36,6,64,32)f32 +pnnx.Expression pnnx_expr_92 1 1 3674 3677 expr=mul(@0,1.767767e-01) #3674=(36,6,64,32)f32 #3677=(36,6,64,32)f32 +torch.transpose torch.transpose_3213 1 1 3675 3678 dim0=-2 dim1=-1 $input=3675 #3675=(36,6,64,32)f32 #3678=(36,6,32,64)f32 +torch.matmul torch.matmul_2418 2 1 3677 3678 3679 $input=3677 $other=3678 #3677=(36,6,64,32)f32 #3678=(36,6,32,64)f32 #3679=(36,6,64,64)f32 +pnnx.Attribute pnnx_fold_19711 0 1 3680 @pnnx_fold_19711=(1,6,64,64)f32 #3680=(1,6,64,64)f32 +pnnx.Expression pnnx_expr_79 2 1 3679 3680 3681 expr=add(@0,@1) #3679=(36,6,64,64)f32 #3680=(1,6,64,64)f32 #3681=(36,6,64,64)f32 +Tensor.view Tensor.view_2175 1 1 3681 3682 shape=(1,36,6,64,64) $input=3681 #3681=(36,6,64,64)f32 #3682=(1,36,6,64,64)f32 +pnnx.Attribute pnnx_fold_19721 0 1 3683 @pnnx_fold_19721=(1,36,1,64,64)f32 #3683=(1,36,1,64,64)f32 +pnnx.Expression pnnx_expr_67 2 1 3682 3683 3684 expr=add(@0,@1) #3682=(1,36,6,64,64)f32 #3683=(1,36,1,64,64)f32 #3684=(1,36,6,64,64)f32 +Tensor.view Tensor.view_2176 1 1 3684 3685 shape=(-1,6,64,64) $input=3684 #3684=(1,36,6,64,64)f32 #3685=(36,6,64,64)f32 +nn.Softmax layers_mmsa.5.residual_group.blocks.5.attn.softmax 1 1 3685 3686 dim=-1 #3685=(36,6,64,64)f32 #3686=(36,6,64,64)f32 +torch.matmul torch.matmul_2419 2 1 3686 3676 3687 $input=3686 $other=3676 #3686=(36,6,64,64)f32 #3676=(36,6,64,32)f32 #3687=(36,6,64,32)f32 +torch.transpose torch.transpose_3214 1 1 3687 3688 dim0=1 dim1=2 $input=3687 #3687=(36,6,64,32)f32 #3688=(36,64,6,32)f32 +Tensor.reshape Tensor.reshape_649 1 1 3688 3689 shape=(36,64,192) $input=3688 #3688=(36,64,6,32)f32 #3689=(36,64,192)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.5.attn.proj 1 1 3689 3690 bias=True in_features=192 out_features=192 @bias=(192)f32 @weight=(192,192)f32 #3689=(36,64,192)f32 #3690=(36,64,192)f32 +Tensor.reshape Tensor.view_2178 1 1 3690 3691 shape=(1,6,6,8,8,192) $input=3690 #3690=(36,64,192)f32 #3691=(1,6,6,8,8,192)f32 +torch.permute torch.permute_2959 1 1 3691 3692 dims=(0,1,3,2,4,5) $input=3691 #3691=(1,6,6,8,8,192)f32 #3692=(1,6,8,6,8,192)f32 +Tensor.reshape view_shape_215 1 1 3692 3693 shape=(1,48,48,-1) $input=3692 #3692=(1,6,8,6,8,192)f32 #3693=(1,48,48,192)f32 +torch.roll torch.roll_2527 1 1 3693 3694 dims=(1,2) shifts=(4,4) $input=3693 #3693=(1,48,48,192)f32 #3694=(1,48,48,192)f32 +Tensor.view Tensor.view_2180 1 1 3694 3695 shape=(1,2304,192) $input=3694 #3694=(1,48,48,192)f32 #3695=(1,2304,192)f32 +pnnx.Expression pnnx_expr_25 2 1 3664 3695 3696 expr=add(@0,@1) #3664=(1,2304,192)f32 #3695=(1,2304,192)f32 #3696=(1,2304,192)f32 +nn.LayerNorm layers_mmsa.5.residual_group.blocks.5.norm2 1 1 3696 3697 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3696=(1,2304,192)f32 #3697=(1,2304,192)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.5.mlp.fc1 1 1 3697 3698 bias=True in_features=192 out_features=384 @bias=(384)f32 @weight=(384,192)f32 #3697=(1,2304,192)f32 #3698=(1,2304,384)f32 +nn.GELU layers_mmsa.5.residual_group.blocks.5.mlp.act 1 1 3698 3699 #3698=(1,2304,384)f32 #3699=(1,2304,384)f32 +nn.Linear layers_mmsa.5.residual_group.blocks.5.mlp.fc2 1 1 3699 3700 bias=True in_features=384 out_features=192 @bias=(192)f32 @weight=(192,384)f32 #3699=(1,2304,384)f32 #3700=(1,2304,192)f32 +pnnx.Expression pnnx_expr_23 2 1 3696 3700 3701 expr=add(@0,@1) #3696=(1,2304,192)f32 #3700=(1,2304,192)f32 #3701=(1,2304,192)f32 +torch.transpose torch.transpose_3215 1 1 3701 3702 dim0=1 dim1=2 $input=3701 #3701=(1,2304,192)f32 #3702=(1,192,2304)f32 +Tensor.view Tensor.view_2181 1 1 3702 3703 shape=(1,192,48,48) $input=3702 #3702=(1,192,2304)f32 #3703=(1,192,48,48)f32 +nn.Conv2d layers_mmsa.5.conv 1 1 3703 3704 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #3703=(1,192,48,48)f32 #3704=(1,192,48,48)f32 +torch.flatten torch.flatten_2203 1 1 3704 3705 end_dim=-1 start_dim=2 $input=3704 #3704=(1,192,48,48)f32 #3705=(1,192,2304)f32 +torch.transpose torch.transpose_3216 1 1 3705 3706 dim0=1 dim1=2 $input=3705 #3705=(1,192,2304)f32 #3706=(1,2304,192)f32 +pnnx.Expression pnnx_expr_10 2 1 3706 3503 3707 expr=add(@0,@1) #3706=(1,2304,192)f32 #3503=(1,2304,192)f32 #3707=(1,2304,192)f32 +nn.LayerNorm norm_mmsa 1 1 3707 3708 elementwise_affine=True eps=1.000000e-05 normalized_shape=(192) @bias=(192)f32 @weight=(192)f32 #3707=(1,2304,192)f32 #3708=(1,2304,192)f32 +torch.transpose torch.transpose_3217 1 1 3708 3709 dim0=1 dim1=2 $input=3708 #3708=(1,2304,192)f32 #3709=(1,192,2304)f32 +Tensor.view Tensor.view_2182 1 1 3709 3710 shape=(1,192,48,48) $input=3709 #3709=(1,192,2304)f32 #3710=(1,192,48,48)f32 +nn.Conv2d conv_after_body_mmsa 1 1 3710 3711 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=192 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(192)f32 @weight=(192,192,3,3)f32 #3710=(1,192,48,48)f32 #3711=(1,192,48,48)f32 +pnnx.Expression pnnx_expr_0 2 1 3711 2480 3712 expr=add(@0,@1) #3711=(1,192,48,48)f32 #2480=(1,192,48,48)f32 #3712=(1,192,48,48)f32 +nn.ConvTranspose2d upsample_conv 1 1 3712 3713 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(8,8) out_channels=192 output_padding=(0,0) padding=(0,0) stride=(8,8) @bias=(192)f32 @weight=(192,192,8,8)f32 #3712=(1,192,48,48)f32 #3713=(1,192,384,384)f32 +nn.Conv2d conv_last 1 1 3713 3714 bias=True dilation=(1,1) groups=1 in_channels=192 kernel_size=(3,3) out_channels=3 padding=(1,1) padding_mode=zeros stride=(1,1) @bias=(3)f32 @weight=(3,192,3,3)f32 #3713=(1,192,384,384)f32 #3714=(1,3,384,384)f32 +pnnx.Attribute pnnx_fold_2234 0 1 3715 @pnnx_fold_2234=(1,192,48,48)f32 #3715=(1,192,48,48)f32 +prim::TupleConstruct pnnx_21200 4 1 3714 1239 2471 3715 3716 #3714=(1,3,384,384)f32 #1239=(1,192,48,48)f32 #2471=(1,192,48,48)f32 #3715=(1,192,48,48)f32 +pnnx.Output pnnx_output_0 1 0 3716