pszemraj commited on
Commit
fb7d1f7
1 Parent(s): 0c788ea

End of training

Browse files
all_results.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 4.0,
3
+ "eval_loss": 1.6668897867202759,
4
+ "eval_runtime": 102.7014,
5
+ "eval_samples": 13319,
6
+ "eval_samples_per_second": 129.687,
7
+ "eval_steps_per_second": 16.212,
8
+ "perplexity": 5.295671489170355,
9
+ "train_loss": 1.9931427570304485,
10
+ "train_runtime": 30194.781,
11
+ "train_samples": 253694,
12
+ "train_samples_per_second": 33.608,
13
+ "train_steps_per_second": 0.131
14
+ }
config.json CHANGED
@@ -21,7 +21,7 @@
21
  "num_hidden_layers": 24,
22
  "pad_token_id": 1,
23
  "prefix": "</s>",
24
- "torch_dtype": "float32",
25
  "transformers_version": "4.25.0.dev0",
26
  "use_cache": true,
27
  "vocab_size": 50272,
 
21
  "num_hidden_layers": 24,
22
  "pad_token_id": 1,
23
  "prefix": "</s>",
24
+ "torch_dtype": "bfloat16",
25
  "transformers_version": "4.25.0.dev0",
26
  "use_cache": true,
27
  "vocab_size": 50272,
eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 4.0,
3
+ "eval_loss": 1.6668897867202759,
4
+ "eval_runtime": 102.7014,
5
+ "eval_samples": 13319,
6
+ "eval_samples_per_second": 129.687,
7
+ "eval_steps_per_second": 16.212,
8
+ "perplexity": 5.295671489170355
9
+ }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2648d976595b9b0cfbad08b91da10fda8b0c838456789fe0b413346c60ba274b
3
- size 1324917277
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:743eb512852251be08a75f53fcbdc54da6810c302f91b9f948513d6064ebc538
3
+ size 662524445
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 4.0,
3
+ "train_loss": 1.9931427570304485,
4
+ "train_runtime": 30194.781,
5
+ "train_samples": 253694,
6
+ "train_samples_per_second": 33.608,
7
+ "train_steps_per_second": 0.131
8
+ }
trainer_state.json ADDED
The diff for this file is too large to render. See raw diff