ericntay commited on
Commit
0d0b58f
1 Parent(s): 7c61f1a

Training in progress, step 500

Browse files
config.json CHANGED
@@ -1,13 +1,14 @@
1
  {
2
- "_name_or_path": "distilbert-base-uncased",
3
- "activation": "gelu",
4
  "architectures": [
5
- "DistilBertForSequenceClassification"
6
  ],
7
- "attention_dropout": 0.1,
8
- "dim": 768,
9
- "dropout": 0.1,
10
- "hidden_dim": 3072,
 
 
11
  "id2label": {
12
  "0": "LABEL_0",
13
  "1": "LABEL_1",
@@ -17,6 +18,7 @@
17
  "5": "LABEL_5"
18
  },
19
  "initializer_range": 0.02,
 
20
  "label2id": {
21
  "LABEL_0": 0,
22
  "LABEL_1": 1,
@@ -25,17 +27,17 @@
25
  "LABEL_4": 4,
26
  "LABEL_5": 5
27
  },
 
28
  "max_position_embeddings": 512,
29
- "model_type": "distilbert",
30
- "n_heads": 12,
31
- "n_layers": 6,
32
  "pad_token_id": 0,
 
33
  "problem_type": "single_label_classification",
34
- "qa_dropout": 0.1,
35
- "seq_classif_dropout": 0.2,
36
- "sinusoidal_pos_embds": false,
37
- "tie_weights_": true,
38
  "torch_dtype": "float32",
39
  "transformers_version": "4.19.4",
40
- "vocab_size": 30522
 
 
41
  }
 
1
  {
2
+ "_name_or_path": "bert-base-cased",
 
3
  "architectures": [
4
+ "BertForSequenceClassification"
5
  ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
  "id2label": {
13
  "0": "LABEL_0",
14
  "1": "LABEL_1",
 
18
  "5": "LABEL_5"
19
  },
20
  "initializer_range": 0.02,
21
+ "intermediate_size": 3072,
22
  "label2id": {
23
  "LABEL_0": 0,
24
  "LABEL_1": 1,
 
27
  "LABEL_4": 4,
28
  "LABEL_5": 5
29
  },
30
+ "layer_norm_eps": 1e-12,
31
  "max_position_embeddings": 512,
32
+ "model_type": "bert",
33
+ "num_attention_heads": 12,
34
+ "num_hidden_layers": 12,
35
  "pad_token_id": 0,
36
+ "position_embedding_type": "absolute",
37
  "problem_type": "single_label_classification",
 
 
 
 
38
  "torch_dtype": "float32",
39
  "transformers_version": "4.19.4",
40
+ "type_vocab_size": 2,
41
+ "use_cache": true,
42
+ "vocab_size": 28996
43
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:65ff3a6b2334feb1882372193c074500c47059513055889b16ea0cb9d5a0bc95
3
- size 267866609
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cb01a1a6d73f908c765511e135a5a4df936bbf1024a2cc05face50e509e002b
3
+ size 433330541
runs/Jun13_15-19-05_a30db051be23/1655134648.392652/events.out.tfevents.1655134648.a30db051be23.1311.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5069f4a2b2333c64643a58113e750e55a076df0b08e10c5037269caad812ec3
3
+ size 5175
runs/Jun13_15-19-05_a30db051be23/1655134786.8272123/events.out.tfevents.1655134786.a30db051be23.1311.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:135f9da47ba8af46c35fc0b373ed8528ea182851fbd9b1c9282554c645a33b92
3
+ size 5175
runs/Jun13_15-19-05_a30db051be23/1655135035.3613205/events.out.tfevents.1655135035.a30db051be23.1311.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a81da043227f634f07d704b9087392cc52655c760672d287fda07c87cdc0f46
3
+ size 5175
runs/Jun13_15-19-05_a30db051be23/events.out.tfevents.1655133583.a30db051be23.1311.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8eea5140f048e41901b36542c6c17e3265a86bf3241e7c9f633c284499a7f2ab
3
- size 4360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9463c378aed1a96a14e0023fa4e6f8f42c600145da99becb2d33d455bec29fe0
3
+ size 5083
runs/Jun13_15-19-05_a30db051be23/events.out.tfevents.1655134648.a30db051be23.1311.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92239372bd84a9516e0386a787d69442e177fa638648d6ae4775dec4002ee362
3
+ size 3768
runs/Jun13_15-19-05_a30db051be23/events.out.tfevents.1655134786.a30db051be23.1311.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3006f79ebda8909ba636981ff7bb09a9d3d7d80d9461a23bf0afefb6dd668c54
3
+ size 3925
runs/Jun13_15-19-05_a30db051be23/events.out.tfevents.1655135035.a30db051be23.1311.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e057c132063523af81b618cccfb88e6b3bfd94dad69da5b7549eb290e1595f5b
3
+ size 4248
runs/Jun13_15-52-03_a30db051be23/1655135531.5349307/events.out.tfevents.1655135531.a30db051be23.1311.9 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5eaa48ba2cfbb9b05faabafd1a769b7f4a2718c6f3c9f57a89cddd58afae273
3
+ size 5175
runs/Jun13_15-52-03_a30db051be23/events.out.tfevents.1655135531.a30db051be23.1311.8 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d0d683bae7e678f62a043a1857f7e5f6d79053fe8ca12b73b41f62b8a89ffac
3
+ size 3788
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "distilbert-base-uncased", "tokenizer_class": "DistilBertTokenizer"}
 
1
+ {"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "bert-base-cased", "tokenizer_class": "BertTokenizer"}
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:63dd9b185ba543b9275cb9a3fa23f5ac94f613b5d0f74434d43630a266715fa9
3
  size 3183
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8822d4d68a316fc85d662b9cc5c52c3413ac699654d6e7817ed8bf9461f45a1
3
  size 3183
vocab.txt CHANGED
The diff for this file is too large to render. See raw diff