KushalRamaiya commited on
Commit
8c80dc1
1 Parent(s): 21d7439

Training in progress, step 500

Browse files
config.json CHANGED
@@ -32,6 +32,6 @@
32
  "sinusoidal_pos_embds": false,
33
  "tie_weights_": true,
34
  "torch_dtype": "float32",
35
- "transformers_version": "4.11.3",
36
  "vocab_size": 30522
37
  }
 
32
  "sinusoidal_pos_embds": false,
33
  "tie_weights_": true,
34
  "torch_dtype": "float32",
35
+ "transformers_version": "4.22.2",
36
  "vocab_size": 30522
37
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:056c3bc5f519f6e3c9968f45a441c9a691e5676df7817857f0c575da5241659b
3
- size 267866225
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8a65626cce9434fb9b3fea63d705aac3c65a3f946428dc79c75c1f4a509eea3
3
+ size 267860465
runs/Sep29_12-43-37_0d9ee5f4a97c/events.out.tfevents.1664455423.0d9ee5f4a97c.32379.8 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a58d59ff2c53d2777cf91f9250c1a5b8eaf07d7dd4f63f566e9bdba5fd33627c
3
- size 4644
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fecbc423ca1ae49dd6089b1b4db9f59bb6b15e6f49f684dee7f73043b210cbe8
3
+ size 5893
runs/Sep29_14-22-34_0d9ee5f4a97c/1664461404.5819356/events.out.tfevents.1664461404.0d9ee5f4a97c.188083.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3df4e08f6f7078b636c69ca9c0459623b09f902b7b30c08ab9fd9a942c6d649
3
+ size 5564
runs/Sep29_14-22-34_0d9ee5f4a97c/events.out.tfevents.1664461404.0d9ee5f4a97c.188083.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d094b23bf317b353112061523daf1c7db8554fa786cfb5f2f0316aa049c6417
3
+ size 6355
runs/Sep29_14-35-49_0d9ee5f4a97c/1664462170.1361766/events.out.tfevents.1664462170.0d9ee5f4a97c.188083.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd855cfd58256f799b0b12e97e418d7f6262ddb3dd2949e53bf4c65e7d897f30
3
+ size 5564
runs/Sep29_14-35-49_0d9ee5f4a97c/events.out.tfevents.1664462170.0d9ee5f4a97c.188083.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea60881f89ae30272779fe9ea0fff374b649615b0caa820186d13aafc11a82fa
3
+ size 4211
special_tokens_map.json CHANGED
@@ -1 +1,7 @@
1
- {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1 +1,14 @@
1
- {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "distilbert-base-uncased", "tokenizer_class": "DistilBertTokenizer"}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_lower_case": true,
4
+ "mask_token": "[MASK]",
5
+ "model_max_length": 512,
6
+ "name_or_path": "distilbert-base-uncased",
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
9
+ "special_tokens_map_file": null,
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "DistilBertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f40b11d4b3f3420e8605f0287f1854959de121c7656cac458ce07ec25ae3d10d
3
- size 2863
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f84a2fcbbab592642377be4381cf4e1ceb7baa6d89bd3856b0c69be7c95554a
3
+ size 3439