AlekseyKorshuk commited on
Commit
6df531f
1 Parent(s): 2baf059

huggingartists

Browse files
README.md CHANGED
@@ -45,15 +45,15 @@ from datasets import load_dataset
45
  dataset = load_dataset("huggingartists/aaron-watson")
46
  ```
47
 
48
- [Explore the data](https://wandb.ai/huggingartists/huggingartists/runs/2hy5w75w/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline.
49
 
50
  ## Training procedure
51
 
52
  The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on Aaron Watson's lyrics.
53
 
54
- Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/huggingartists/huggingartists/runs/2h0fs8qq) for full transparency and reproducibility.
55
 
56
- At the end of training, [the final model](https://wandb.ai/huggingartists/huggingartists/runs/2h0fs8qq/artifacts) is logged and versioned.
57
 
58
  ## How to use
59
 
45
  dataset = load_dataset("huggingartists/aaron-watson")
46
  ```
47
 
48
+ [Explore the data](https://wandb.ai/huggingartists/huggingartists/runs/204aifr7/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline.
49
 
50
  ## Training procedure
51
 
52
  The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on Aaron Watson's lyrics.
53
 
54
+ Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/huggingartists/huggingartists/runs/20vl7rt5) for full transparency and reproducibility.
55
 
56
+ At the end of training, [the final model](https://wandb.ai/huggingartists/huggingartists/runs/20vl7rt5/artifacts) is logged and versioned.
57
 
58
  ## How to use
59
 
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "gpt2",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPT2LMHeadModel"
1
  {
2
+ "_name_or_path": "huggingartists/aaron-watson",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPT2LMHeadModel"
evaluation.txt CHANGED
@@ -1 +1 @@
1
- {"eval_loss": 2.996838331222534, "eval_runtime": 1.5779, "eval_samples_per_second": 21.547, "eval_steps_per_second": 3.169, "epoch": 2.0}
1
+ {"eval_loss": 2.9697048664093018, "eval_runtime": 1.8075, "eval_samples_per_second": 20.47, "eval_steps_per_second": 2.766, "epoch": 5.0}
flax_model.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bed746cf4a023760b7e67a6ff2a80d140a2f44937ce586a9dad8d46e04a4b32c
3
  size 497764120
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75c4ed6e042b687530615246e336ead94b468f8bffd1f840eeb3535c66e14912
3
  size 497764120
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4dd6b9dedd4d3504b8efcee5e7b4135aa72ece84eb0a07df5d59820f745486db
3
  size 995603825
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af00edd8315e8240fc11cc4bbe4fd8812f8e9944d63e1e804ed4135ef7995423
3
  size 995603825
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ad5b45a12c5b8b9ea6e74cd1f0e45fefae3e6e3e41e981df0e65db608dbd91a4
3
  size 510403817
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ae3248a48d48f36b9ecb728e6afed45c93ad66bd3f5adf5a346cb1e380479b5
3
  size 510403817
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ae45c233b14eda3cf7879d31899e37130a0d6bf44f948bbd5db7707ce01455fa
3
  size 14503
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6e12081284488e0302f0f15cc7b9e867b93ab823a7f5b31fa14b05b508db300
3
  size 14503
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c40e0b165562c00350699054d9d108e03d2359770326bafc9d90a379a5ec4c14
3
  size 623
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d625cddcadf735c6fc48a4f3f1d506c16673a1b6749611364d76e3a20b91fce
3
  size 623
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"unk_token": "<|endoftext|>", "bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "gpt2", "tokenizer_class": "GPT2Tokenizer"}
1
+ {"unk_token": "<|endoftext|>", "bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "huggingartists/aaron-watson", "tokenizer_class": "GPT2Tokenizer"}
trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "best_metric": 2.996838331222534,
3
- "best_model_checkpoint": "output/aaron-watson/checkpoint-50",
4
- "epoch": 2.0,
5
- "global_step": 50,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -82,11 +82,49 @@
82
  "eval_samples_per_second": 21.512,
83
  "eval_steps_per_second": 3.163,
84
  "step": 50
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
85
  }
86
  ],
87
- "max_steps": 50,
88
- "num_train_epochs": 2,
89
- "total_flos": 51474530304000.0,
90
  "trial_name": null,
91
  "trial_params": null
92
  }
1
  {
2
+ "best_metric": 2.9697048664093018,
3
+ "best_model_checkpoint": "output/aaron-watson/checkpoint-75",
4
+ "epoch": 3.0,
5
+ "global_step": 75,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
82
  "eval_samples_per_second": 21.512,
83
  "eval_steps_per_second": 3.163,
84
  "step": 50
85
+ },
86
+ {
87
+ "epoch": 2.2,
88
+ "learning_rate": 0.00012409856581412136,
89
+ "loss": 3.0342,
90
+ "step": 55
91
+ },
92
+ {
93
+ "epoch": 2.4,
94
+ "learning_rate": 8.979856581412141e-05,
95
+ "loss": 2.8495,
96
+ "step": 60
97
+ },
98
+ {
99
+ "epoch": 2.6,
100
+ "learning_rate": 4.7401434185878625e-05,
101
+ "loss": 2.895,
102
+ "step": 65
103
+ },
104
+ {
105
+ "epoch": 2.8,
106
+ "learning_rate": 1.310143418587862e-05,
107
+ "loss": 2.8209,
108
+ "step": 70
109
+ },
110
+ {
111
+ "epoch": 3.0,
112
+ "learning_rate": 0.0,
113
+ "loss": 2.7285,
114
+ "step": 75
115
+ },
116
+ {
117
+ "epoch": 3.0,
118
+ "eval_loss": 2.9697048664093018,
119
+ "eval_runtime": 1.7446,
120
+ "eval_samples_per_second": 21.208,
121
+ "eval_steps_per_second": 2.866,
122
+ "step": 75
123
  }
124
  ],
125
+ "max_steps": 125,
126
+ "num_train_epochs": 5,
127
+ "total_flos": 76689211392000.0,
128
  "trial_name": null,
129
  "trial_params": null
130
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fc58aa350caebd0c275c3294305ba485a147a6157454835c10b9e04bb0858638
3
  size 2671
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66237f886a59d41c0e749b69ee85bc3f1c09d6f131ed3be21ee7e5ab55356657
3
  size 2671