commit files to HF hub
543a9ae
-
1.38 kB
initial commit
-
752 Bytes
commit files to HF hub
-
585 Bytes
commit files to HF hub
deberta_v1_checkpoint_1.pt
Detected Pickle imports (27)
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2SelfOutput",
- "collections.OrderedDict",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2PredictionHeadTransform",
- "transformers.activations.GELUActivation",
- "torch.nn.modules.normalization.LayerNorm",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2ForMaskedLM",
- "transformers.models.deberta_v2.modeling_deberta_v2.DisentangledSelfAttention",
- "torch.LongStorage",
- "transformers.models.deberta_v2.configuration_deberta_v2.DebertaV2Config",
- "torch._utils._rebuild_tensor_v2",
- "torch.FloatStorage",
- "torch.nn.modules.sparse.Embedding",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Intermediate",
- "torch._C._nn.gelu",
- "torch.nn.modules.container.ModuleList",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2LMPredictionHead",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Encoder",
- "torch._utils._rebuild_parameter",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2OnlyMLMHead",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Embeddings",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Attention",
- "torch.nn.modules.linear.Linear",
- "__builtin__.set",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Model",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Layer",
- "transformers.models.deberta_v2.modeling_deberta_v2.StableDropout",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Output"
How to fix it?
297 MB
commit files to HF hub
deberta_v1_checkpoint_2.pt
Detected Pickle imports (27)
- "torch.nn.modules.container.ModuleList",
- "torch.LongStorage",
- "transformers.models.deberta_v2.configuration_deberta_v2.DebertaV2Config",
- "torch.FloatStorage",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2LMPredictionHead",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2PredictionHeadTransform",
- "transformers.activations.GELUActivation",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2OnlyMLMHead",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2ForMaskedLM",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Embeddings",
- "torch.nn.modules.sparse.Embedding",
- "torch._utils._rebuild_parameter",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Attention",
- "torch.nn.modules.linear.Linear",
- "torch._C._nn.gelu",
- "transformers.models.deberta_v2.modeling_deberta_v2.DisentangledSelfAttention",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Output",
- "collections.OrderedDict",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Model",
- "torch._utils._rebuild_tensor_v2",
- "torch.nn.modules.normalization.LayerNorm",
- "__builtin__.set",
- "transformers.models.deberta_v2.modeling_deberta_v2.StableDropout",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Intermediate",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Layer",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Encoder",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2SelfOutput"
How to fix it?
297 MB
commit files to HF hub
deberta_v1_checkpoint_3.pt
Detected Pickle imports (27)
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Model",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2SelfOutput",
- "torch._C._nn.gelu",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2OnlyMLMHead",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2PredictionHeadTransform",
- "transformers.models.deberta_v2.modeling_deberta_v2.DisentangledSelfAttention",
- "torch._utils._rebuild_tensor_v2",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Layer",
- "transformers.activations.GELUActivation",
- "torch.FloatStorage",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2LMPredictionHead",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Embeddings",
- "torch.nn.modules.linear.Linear",
- "torch.LongStorage",
- "torch._utils._rebuild_parameter",
- "__builtin__.set",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Intermediate",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Attention",
- "collections.OrderedDict",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Encoder",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2ForMaskedLM",
- "torch.nn.modules.sparse.Embedding",
- "torch.nn.modules.container.ModuleList",
- "transformers.models.deberta_v2.configuration_deberta_v2.DebertaV2Config",
- "transformers.models.deberta_v2.modeling_deberta_v2.StableDropout",
- "torch.nn.modules.normalization.LayerNorm",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Output"
How to fix it?
297 MB
commit files to HF hub
deberta_v1_checkpoint_4.pt
Detected Pickle imports (27)
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Model",
- "transformers.models.deberta_v2.configuration_deberta_v2.DebertaV2Config",
- "torch._utils._rebuild_tensor_v2",
- "torch.nn.modules.linear.Linear",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2PredictionHeadTransform",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Encoder",
- "__builtin__.set",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2ForMaskedLM",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Embeddings",
- "torch._C._nn.gelu",
- "transformers.activations.GELUActivation",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2LMPredictionHead",
- "collections.OrderedDict",
- "torch.nn.modules.normalization.LayerNorm",
- "torch.FloatStorage",
- "torch.nn.modules.container.ModuleList",
- "torch.LongStorage",
- "transformers.models.deberta_v2.modeling_deberta_v2.DisentangledSelfAttention",
- "torch.nn.modules.sparse.Embedding",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Layer",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2SelfOutput",
- "transformers.models.deberta_v2.modeling_deberta_v2.StableDropout",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2OnlyMLMHead",
- "torch._utils._rebuild_parameter",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Intermediate",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Output",
- "transformers.models.deberta_v2.modeling_deberta_v2.DebertaV2Attention"
How to fix it?
297 MB
commit files to HF hub
-
360 kB
commit files to HF hub
-
297 MB
commit files to HF hub
-
5.07 MB
commit files to HF hub
-
279 Bytes
commit files to HF hub
-
480 Bytes
commit files to HF hub