dima806 commited on
Commit
66a06b8
1 Parent(s): 0c13410

Upload folder using huggingface_hub

Browse files
checkpoint-54160/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-cased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "id2label": {
12
+ "0": "NOT ENGAGING TITLE",
13
+ "1": "ENGAGING TITLE"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "max_position_embeddings": 512,
17
+ "model_type": "distilbert",
18
+ "n_heads": 12,
19
+ "n_layers": 6,
20
+ "output_past": true,
21
+ "pad_token_id": 0,
22
+ "problem_type": "single_label_classification",
23
+ "qa_dropout": 0.1,
24
+ "seq_classif_dropout": 0.2,
25
+ "sinusoidal_pos_embds": false,
26
+ "tie_weights_": true,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.33.0",
29
+ "vocab_size": 28996
30
+ }
checkpoint-54160/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8535918ecf8cf8bec47f7ae2c3eadb70aa540a461d523e43243b40b14b9435f6
3
+ size 526351109
checkpoint-54160/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4e4088d5de17cc5b3424ad8cf8603025daf6ea728f7ac35c11c12d9e744ba37
3
+ size 263167661
checkpoint-54160/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9341dc9bdf4e97658fa823b1e7c1a009256a346e9ed99f63e56cefbd6a7170a4
3
+ size 14575
checkpoint-54160/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55967bc10c16de91ae29050de757b294700e3ea82fd41c4aa57614b0368b2b90
3
+ size 627
checkpoint-54160/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-54160/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5dd1a1f2e98a0e0ca785cd748fe58dc5f1d46745a8b7ea48dfe273245cf4bfa
3
+ size 4027
config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-cased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "id2label": {
12
+ "0": "NOT ENGAGING TITLE",
13
+ "1": "ENGAGING TITLE"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "max_position_embeddings": 512,
17
+ "model_type": "distilbert",
18
+ "n_heads": 12,
19
+ "n_layers": 6,
20
+ "output_past": true,
21
+ "pad_token_id": 0,
22
+ "problem_type": "single_label_classification",
23
+ "qa_dropout": 0.1,
24
+ "seq_classif_dropout": 0.2,
25
+ "sinusoidal_pos_embds": false,
26
+ "tie_weights_": true,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.33.0",
29
+ "vocab_size": 28996
30
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4e4088d5de17cc5b3424ad8cf8603025daf6ea728f7ac35c11c12d9e744ba37
3
+ size 263167661
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5dd1a1f2e98a0e0ca785cd748fe58dc5f1d46745a8b7ea48dfe273245cf4bfa
3
+ size 4027
vocab.txt ADDED
The diff for this file is too large to render. See raw diff