download
history
blame
contribute
delete
1.4 GB
Detected Pickle imports (150)
- "__torch__.torch.nn.modules.normalization.___torch_mangle_71.GroupNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_130.Linear",
- "__torch__.torch.nn.modules.normalization.GroupNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_47.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_68.GroupNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_79.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_92.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_94.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_6.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_32.GroupNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_102.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_13.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_70.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_0.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_43.LayerNorm",
- "__torch__.torch.nn.modules.activation.___torch_mangle_23.MultiheadAttention",
- "__torch__.torch.nn.modules.activation.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_119.GroupNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_74.GroupNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_15.Conv2d",
- "__torch__.diffusion_embJITnative.diffusion_emb",
- "__torch__.torch.nn.modules.linear.___torch_mangle_2.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_75.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_64.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_63.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_5.GroupNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_27.LayerNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_132.Conv2d",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_19.GroupNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_31.Conv2d",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_76.LayerNorm",
- "__torch__.torch.nn.modules.conv.Conv2d",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14.GroupNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_97.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_25.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_85.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_67.Conv2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_41.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_42.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_131.GroupNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_22.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_16.Linear",
- "torch.FloatStorage",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_112.GroupNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_116.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_66.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_61.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_95.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_21.LayerNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_72.Conv2d",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_10.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_17.GroupNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_36.Conv2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_56.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_133.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_55.GroupNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_69.Conv2d",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_93.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_12.Linear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_123.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_53.GroupNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_121.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_60.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_106.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_110.Conv2d",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_46.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_28.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_82.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_44.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_81.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_3.GroupNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_117.GroupNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_4.Conv2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_88.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_20.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_91.GroupNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_125.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_52.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_128.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_127.LayerNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_115.Conv2d",
- "__torch__.torch.nn.modules.conv.___torch_mangle_135.Conv2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_30.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_90.Conv2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_129.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_134.GroupNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_83.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_39.Linear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_26.MultiheadAttention",
- "__torch__.torch.nn.modules.conv.___torch_mangle_87.Conv2d",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_50.GroupNorm",
- "__torch__.torch.nn.modules.activation.___torch_mangle_9.MultiheadAttention",
- "__torch__.torch.nn.modules.conv.___torch_mangle_105.Conv2d",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_38.GroupNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_35.GroupNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_65.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_108.Conv2d",
- "__torch__.torch.nn.modules.activation.___torch_mangle_78.MultiheadAttention",
- "__torch__.torch.nn.modules.conv.___torch_mangle_18.Conv2d",
- "__torch__.torch.nn.modules.activation.___torch_mangle_62.MultiheadAttention",
- "collections.OrderedDict",
- "__torch__.torch.nn.modules.conv.___torch_mangle_118.Conv2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_80.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_101.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_114.GroupNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_40.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_104.GroupNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_51.Conv2d",
- "__torch__.torch.nn.modules.linear.Linear",
- "__torch__.torch.nn.modules.linear.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_8.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_107.GroupNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_33.Conv2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_34.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_73.Conv2d",
- "__torch__.torch.nn.modules.activation.___torch_mangle_59.MultiheadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_77.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.normalization.LayerNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_54.Conv2d",
- "__torch__.torch.nn.modules.conv.___torch_mangle_103.Conv2d",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_24.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_96.LayerNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_37.Conv2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_84.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_111.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_109.GroupNorm",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_124.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_89.GroupNorm",
- "__torch__.torch.nn.modules.activation.___torch_mangle_98.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_7.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_99.LayerNorm",
- "__torch__.torch.nn.modules.activation.___torch_mangle_45.MultiheadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_120.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_48.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_58.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_49.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_113.Conv2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_11.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_57.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_122.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_126.MultiheadAttention",
- "__torch__.torch.nn.modules.conv.___torch_mangle_1.Conv2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_100.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_86.GroupNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_29.Linear",
- "torch.DoubleStorage",
- "torch.LongStorage",
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict"
Git LFS Details
- SHA256: e5ceed1dac4aee955843ee4feda36b0bdc8fc77cbe1988a5d6f18ab96eb66d21
- Pointer size: 135 Bytes
- Size of remote file: 1.4 GB
Git Large File Storage (LFS) replaces large files with text pointers inside Git, while storing the file contents on a remote server. More info.