Commit
·
cb50ccc
1
Parent(s):
2ce4dc2
Training in progress, step 500
Browse files- config.json +1 -1
- pytorch_model.bin +2 -2
- runs/Feb21_12-46-00_c77a7abfd0e8/1676983635.2731745/events.out.tfevents.1676983635.c77a7abfd0e8.487.1 +3 -0
- runs/Feb21_12-46-00_c77a7abfd0e8/events.out.tfevents.1676983635.c77a7abfd0e8.487.0 +3 -0
- special_tokens_map.json +7 -0
- tokenizer.json +0 -0
- tokenizer_config.json +14 -0
- training_args.bin +2 -2
- vocab.txt +0 -0
config.json
CHANGED
@@ -80,6 +80,6 @@
|
|
80 |
"sinusoidal_pos_embds": false,
|
81 |
"tie_weights_": true,
|
82 |
"torch_dtype": "float32",
|
83 |
-
"transformers_version": "4.
|
84 |
"vocab_size": 30522
|
85 |
}
|
|
|
80 |
"sinusoidal_pos_embds": false,
|
81 |
"tie_weights_": true,
|
82 |
"torch_dtype": "float32",
|
83 |
+
"transformers_version": "4.26.1",
|
84 |
"vocab_size": 30522
|
85 |
}
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:19424ed24da12c023d56f789d211e4072e8898e193049073f2215eb1df0ab86d
|
3 |
+
size 267935469
|
runs/Feb21_12-46-00_c77a7abfd0e8/1676983635.2731745/events.out.tfevents.1676983635.c77a7abfd0e8.487.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dab016576712840ae28562eab3600b8f50837ebfa6d598d852297b0bc47644db
|
3 |
+
size 5707
|
runs/Feb21_12-46-00_c77a7abfd0e8/events.out.tfevents.1676983635.c77a7abfd0e8.487.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7a8ab61aa825f63981ebb96d82b1948bba40b025b17786f2b1733895cd7f599a
|
3 |
+
size 4956
|
special_tokens_map.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"cls_token": "[CLS]",
|
3 |
+
"mask_token": "[MASK]",
|
4 |
+
"pad_token": "[PAD]",
|
5 |
+
"sep_token": "[SEP]",
|
6 |
+
"unk_token": "[UNK]"
|
7 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"cls_token": "[CLS]",
|
3 |
+
"do_lower_case": true,
|
4 |
+
"mask_token": "[MASK]",
|
5 |
+
"model_max_length": 512,
|
6 |
+
"name_or_path": "distilbert-base-uncased",
|
7 |
+
"pad_token": "[PAD]",
|
8 |
+
"sep_token": "[SEP]",
|
9 |
+
"special_tokens_map_file": null,
|
10 |
+
"strip_accents": null,
|
11 |
+
"tokenize_chinese_chars": true,
|
12 |
+
"tokenizer_class": "DistilBertTokenizer",
|
13 |
+
"unk_token": "[UNK]"
|
14 |
+
}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:22a0da41b689868699c951e90991f305cadd15e009bef51b47f3abd8845815b0
|
3 |
+
size 3515
|
vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|