Training in progress epoch 0
Browse files- README.md +5 -5
- config.json +1 -1
- generation_config.json +1 -1
- logs/train/events.out.tfevents.1681508244.ip-172-26-13-131.62230.0.v2 +3 -0
- logs/validation/events.out.tfevents.1681510986.ip-172-26-13-131.62230.1.v2 +3 -0
- merges.txt +1 -1
- tf_model.h5 +1 -1
- tokenizer.json +1 -0
- tokenizer_config.json +1 -1
README.md
CHANGED
@@ -14,8 +14,8 @@ probably proofread and complete it, then remove this comment. -->
|
|
14 |
|
15 |
This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on an unknown dataset.
|
16 |
It achieves the following results on the evaluation set:
|
17 |
-
- Train Loss: 3.
|
18 |
-
- Validation Loss: 3.
|
19 |
- Epoch: 0
|
20 |
|
21 |
## Model description
|
@@ -42,12 +42,12 @@ The following hyperparameters were used during training:
|
|
42 |
|
43 |
| Train Loss | Validation Loss | Epoch |
|
44 |
|:----------:|:---------------:|:-----:|
|
45 |
-
| 3.
|
46 |
|
47 |
|
48 |
### Framework versions
|
49 |
|
50 |
-
- Transformers 4.
|
51 |
- TensorFlow 2.12.0
|
52 |
- Datasets 2.11.0
|
53 |
-
- Tokenizers 0.13.
|
|
|
14 |
|
15 |
This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on an unknown dataset.
|
16 |
It achieves the following results on the evaluation set:
|
17 |
+
- Train Loss: 3.8585
|
18 |
+
- Validation Loss: 3.6762
|
19 |
- Epoch: 0
|
20 |
|
21 |
## Model description
|
|
|
42 |
|
43 |
| Train Loss | Validation Loss | Epoch |
|
44 |
|:----------:|:---------------:|:-----:|
|
45 |
+
| 3.8585 | 3.6762 | 0 |
|
46 |
|
47 |
|
48 |
### Framework versions
|
49 |
|
50 |
+
- Transformers 4.28.0
|
51 |
- TensorFlow 2.12.0
|
52 |
- Datasets 2.11.0
|
53 |
+
- Tokenizers 0.13.3
|
config.json
CHANGED
@@ -39,7 +39,7 @@
|
|
39 |
"max_length": 50
|
40 |
}
|
41 |
},
|
42 |
-
"transformers_version": "4.
|
43 |
"use_cache": true,
|
44 |
"vocab_size": 50257
|
45 |
}
|
|
|
39 |
"max_length": 50
|
40 |
}
|
41 |
},
|
42 |
+
"transformers_version": "4.28.0",
|
43 |
"use_cache": true,
|
44 |
"vocab_size": 50257
|
45 |
}
|
generation_config.json
CHANGED
@@ -2,5 +2,5 @@
|
|
2 |
"_from_model_config": true,
|
3 |
"bos_token_id": 50256,
|
4 |
"eos_token_id": 50256,
|
5 |
-
"transformers_version": "4.
|
6 |
}
|
|
|
2 |
"_from_model_config": true,
|
3 |
"bos_token_id": 50256,
|
4 |
"eos_token_id": 50256,
|
5 |
+
"transformers_version": "4.28.0"
|
6 |
}
|
logs/train/events.out.tfevents.1681508244.ip-172-26-13-131.62230.0.v2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:21a97c3769fa5882708521778c5c93daa42902816b5ef69a6dc2d70f8ec916b7
|
3 |
+
size 958050
|
logs/validation/events.out.tfevents.1681510986.ip-172-26-13-131.62230.1.v2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7c1c471fa98218c39a0b2fa6cc492ca82c16340836e05ff5c1eb5b97619c1a2a
|
3 |
+
size 232
|
merges.txt
CHANGED
@@ -1,4 +1,4 @@
|
|
1 |
-
#version: 0.2
|
2 |
Ġ t
|
3 |
Ġ a
|
4 |
h e
|
|
|
1 |
+
#version: 0.2
|
2 |
Ġ t
|
3 |
Ġ a
|
4 |
h e
|
tf_model.h5
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 327745472
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bef552580849791532a9c36b2031c3a3945a06c9a55131e50e88526ad0f5939a
|
3 |
size 327745472
|
tokenizer.json
CHANGED
@@ -39,6 +39,7 @@
|
|
39 |
"continuing_subword_prefix": "",
|
40 |
"end_of_word_suffix": "",
|
41 |
"fuse_unk": false,
|
|
|
42 |
"vocab": {
|
43 |
"!": 0,
|
44 |
"\"": 1,
|
|
|
39 |
"continuing_subword_prefix": "",
|
40 |
"end_of_word_suffix": "",
|
41 |
"fuse_unk": false,
|
42 |
+
"byte_fallback": false,
|
43 |
"vocab": {
|
44 |
"!": 0,
|
45 |
"\"": 1,
|
tokenizer_config.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
"add_prefix_space": false,
|
3 |
"bos_token": "<|endoftext|>",
|
|
|
4 |
"eos_token": "<|endoftext|>",
|
5 |
"model_max_length": 1024,
|
6 |
-
"special_tokens_map_file": null,
|
7 |
"tokenizer_class": "GPT2Tokenizer",
|
8 |
"unk_token": "<|endoftext|>"
|
9 |
}
|
|
|
1 |
{
|
2 |
"add_prefix_space": false,
|
3 |
"bos_token": "<|endoftext|>",
|
4 |
+
"clean_up_tokenization_spaces": true,
|
5 |
"eos_token": "<|endoftext|>",
|
6 |
"model_max_length": 1024,
|
|
|
7 |
"tokenizer_class": "GPT2Tokenizer",
|
8 |
"unk_token": "<|endoftext|>"
|
9 |
}
|