ifyou819 commited on
Commit
a3120ef
1 Parent(s): 6f83df5

End of training

Browse files
Files changed (5) hide show
  1. README.md +12 -4
  2. config.json +3 -3
  3. generation_config.json +1 -1
  4. model.safetensors +1 -1
  5. training_args.bin +2 -2
README.md CHANGED
@@ -1,8 +1,10 @@
1
  ---
2
  license: mit
3
- base_model: openai-community/gpt2
4
  tags:
5
  - generated_from_trainer
 
 
6
  model-index:
7
  - name: results
8
  results: []
@@ -13,7 +15,9 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # results
15
 
16
- This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2) on an unknown dataset.
 
 
17
 
18
  ## Model description
19
 
@@ -42,11 +46,15 @@ The following hyperparameters were used during training:
42
 
43
  ### Training results
44
 
 
 
 
 
45
 
46
 
47
  ### Framework versions
48
 
49
- - Transformers 4.41.2
50
  - Pytorch 2.1.2
51
  - Datasets 2.18.0
52
- - Tokenizers 0.19.1
 
1
  ---
2
  license: mit
3
+ base_model: gpt2
4
  tags:
5
  - generated_from_trainer
6
+ datasets:
7
+ - multi_news
8
  model-index:
9
  - name: results
10
  results: []
 
15
 
16
  # results
17
 
18
+ This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on the multi_news dataset.
19
+ It achieves the following results on the evaluation set:
20
+ - Loss: 2.0572
21
 
22
  ## Model description
23
 
 
46
 
47
  ### Training results
48
 
49
+ | Training Loss | Epoch | Step | Validation Loss |
50
+ |:-------------:|:-----:|:----:|:---------------:|
51
+ | No log | 1.0 | 250 | 2.0936 |
52
+ | 2.1486 | 2.0 | 500 | 2.0572 |
53
 
54
 
55
  ### Framework versions
56
 
57
+ - Transformers 4.39.3
58
  - Pytorch 2.1.2
59
  - Datasets 2.18.0
60
+ - Tokenizers 0.15.2
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "openai-community/gpt2",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPT2LMHeadModel"
@@ -29,11 +29,11 @@
29
  "task_specific_params": {
30
  "text-generation": {
31
  "do_sample": true,
32
- "max_length": 128
33
  }
34
  },
35
  "torch_dtype": "float32",
36
- "transformers_version": "4.41.2",
37
  "use_cache": true,
38
  "vocab_size": 50257
39
  }
 
1
  {
2
+ "_name_or_path": "gpt2",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPT2LMHeadModel"
 
29
  "task_specific_params": {
30
  "text-generation": {
31
  "do_sample": true,
32
+ "max_length": 50
33
  }
34
  },
35
  "torch_dtype": "float32",
36
+ "transformers_version": "4.39.3",
37
  "use_cache": true,
38
  "vocab_size": 50257
39
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 50256,
4
  "eos_token_id": 50256,
5
- "transformers_version": "4.41.2"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 50256,
4
  "eos_token_id": 50256,
5
+ "transformers_version": "4.39.3"
6
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bdc4c3e7ed9d833122aff41c8fcffbef0becf272b4a37443b54b2e79b780fb11
3
  size 497774208
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a62887d7120b320f9c0a62410df0fbb84579b3dd3adb6638b443006d81294c9e
3
  size 497774208
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d4ca1f9b97a529bfe5c6e592b026bc73b65e553f25cc19abbe24d8b5323668e0
3
- size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b852357a1488e812a42d7ececd189fd38a36bdfddc526869b4559ff6385b3b0
3
+ size 4856