rwang5688 commited on
Commit
6fc2c66
1 Parent(s): 87edd0b

Training in progress epoch 0

Browse files
README.md CHANGED
@@ -14,8 +14,8 @@ probably proofread and complete it, then remove this comment. -->
14
 
15
  This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
- - Train Loss: 3.8593
18
- - Validation Loss: 3.6735
19
  - Epoch: 0
20
 
21
  ## Model description
@@ -42,12 +42,12 @@ The following hyperparameters were used during training:
42
 
43
  | Train Loss | Validation Loss | Epoch |
44
  |:----------:|:---------------:|:-----:|
45
- | 3.8593 | 3.6735 | 0 |
46
 
47
 
48
  ### Framework versions
49
 
50
- - Transformers 4.27.4
51
  - TensorFlow 2.12.0
52
  - Datasets 2.11.0
53
- - Tokenizers 0.13.2
 
14
 
15
  This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Train Loss: 3.8585
18
+ - Validation Loss: 3.6762
19
  - Epoch: 0
20
 
21
  ## Model description
 
42
 
43
  | Train Loss | Validation Loss | Epoch |
44
  |:----------:|:---------------:|:-----:|
45
+ | 3.8585 | 3.6762 | 0 |
46
 
47
 
48
  ### Framework versions
49
 
50
+ - Transformers 4.28.0
51
  - TensorFlow 2.12.0
52
  - Datasets 2.11.0
53
+ - Tokenizers 0.13.3
config.json CHANGED
@@ -39,7 +39,7 @@
39
  "max_length": 50
40
  }
41
  },
42
- "transformers_version": "4.27.4",
43
  "use_cache": true,
44
  "vocab_size": 50257
45
  }
 
39
  "max_length": 50
40
  }
41
  },
42
+ "transformers_version": "4.28.0",
43
  "use_cache": true,
44
  "vocab_size": 50257
45
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 50256,
4
  "eos_token_id": 50256,
5
- "transformers_version": "4.27.4"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 50256,
4
  "eos_token_id": 50256,
5
+ "transformers_version": "4.28.0"
6
  }
logs/train/events.out.tfevents.1681508244.ip-172-26-13-131.62230.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21a97c3769fa5882708521778c5c93daa42902816b5ef69a6dc2d70f8ec916b7
3
+ size 958050
logs/validation/events.out.tfevents.1681510986.ip-172-26-13-131.62230.1.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c1c471fa98218c39a0b2fa6cc492ca82c16340836e05ff5c1eb5b97619c1a2a
3
+ size 232
merges.txt CHANGED
@@ -1,4 +1,4 @@
1
- #version: 0.2 - Trained by `huggingface/tokenizers`
2
  Ġ t
3
  Ġ a
4
  h e
 
1
+ #version: 0.2
2
  Ġ t
3
  Ġ a
4
  h e
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:540a0f1919b1c239d645d82a6d6fc8ad78e69be78c3155e23abee64fcdda42e6
3
  size 327745472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bef552580849791532a9c36b2031c3a3945a06c9a55131e50e88526ad0f5939a
3
  size 327745472
tokenizer.json CHANGED
@@ -39,6 +39,7 @@
39
  "continuing_subword_prefix": "",
40
  "end_of_word_suffix": "",
41
  "fuse_unk": false,
 
42
  "vocab": {
43
  "!": 0,
44
  "\"": 1,
 
39
  "continuing_subword_prefix": "",
40
  "end_of_word_suffix": "",
41
  "fuse_unk": false,
42
+ "byte_fallback": false,
43
  "vocab": {
44
  "!": 0,
45
  "\"": 1,
tokenizer_config.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "add_prefix_space": false,
3
  "bos_token": "<|endoftext|>",
 
4
  "eos_token": "<|endoftext|>",
5
  "model_max_length": 1024,
6
- "special_tokens_map_file": null,
7
  "tokenizer_class": "GPT2Tokenizer",
8
  "unk_token": "<|endoftext|>"
9
  }
 
1
  {
2
  "add_prefix_space": false,
3
  "bos_token": "<|endoftext|>",
4
+ "clean_up_tokenization_spaces": true,
5
  "eos_token": "<|endoftext|>",
6
  "model_max_length": 1024,
 
7
  "tokenizer_class": "GPT2Tokenizer",
8
  "unk_token": "<|endoftext|>"
9
  }