Training in progress, step 500
Browse files- .gitignore +1 -0
- config.json +0 -0
- pytorch_model.bin +3 -0
- runs/Jun08_08-47-48_ab3cccd7188e/1654678106.315368/events.out.tfevents.1654678106.ab3cccd7188e.81.1 +3 -0
- runs/Jun08_08-47-48_ab3cccd7188e/events.out.tfevents.1654678106.ab3cccd7188e.81.0 +3 -0
- sentencepiece.bpe.model +3 -0
- special_tokens_map.json +1 -0
- tokenizer_config.json +1 -0
- training_args.bin +3 -0
.gitignore
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
checkpoint-*/
|
config.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a1d9e56aaa947ebd8378920541362cee24993a519146f756af77b3d5bd082f52
|
3 |
+
size 455875437
|
runs/Jun08_08-47-48_ab3cccd7188e/1654678106.315368/events.out.tfevents.1654678106.ab3cccd7188e.81.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e33d3e81d3560e75a3224723c041ff034757a939f92c267134919cc7952df592
|
3 |
+
size 5211
|
runs/Jun08_08-47-48_ab3cccd7188e/events.out.tfevents.1654678106.ab3cccd7188e.81.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:669b1ff002d4653d63f9b2fc366a20a8860e26deca1b92e78f2e632369dc1cd1
|
3 |
+
size 215648
|
sentencepiece.bpe.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:988bc5a00281c6d210a5d34bd143d0363741a432fefe741bf71e61b1869d4314
|
3 |
+
size 810912
|
special_tokens_map.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true}, "additional_special_tokens": ["<s>NOTUSED", "</s>NOTUSED"]}
|
tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "additional_special_tokens": ["<s>NOTUSED", "</s>NOTUSED"], "sp_model_kwargs": {}, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "camembert-base", "tokenizer_class": "CamembertTokenizer"}
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eb41ba92a6083b1db10bce9b923c860a097567cb3200fb16c16fbc480196b571
|
3 |
+
size 3247
|