File size: 2,674 Bytes
f1e9374
 
 
 
 
 
 
 
efac3a8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
---
tags:
- pytorch_model_hub_mixin
- model_hub_mixin
---

This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
- Library: [More Information Needed]
- Docs: [More Information Needed]


done
Some non-default generation parameters are set in the model config. These should go into a GenerationConfig file (https://huggingface.co/docs/transformers/generation_strategies#save-a-custom-decoding-strategy-with-your-model) instead. This warning will be raised to an exception in v4.41.
Non-default generation parameters: {'max_length': 62, 'min_length': 11, 'early_stopping': True, 'num_beams': 6, 'no_repeat_ngram_size': 3, 'forced_eos_token_id': 2}
done
Some non-default generation parameters are set in the model config. These should go into a GenerationConfig file (https://huggingface.co/docs/transformers/generation_strategies#save-a-custom-decoding-strategy-with-your-model) instead. This warning will be raised to an exception in v4.41.
Non-default generation parameters: {'max_length': 62, 'min_length': 11, 'early_stopping': True, 'num_beams': 6, 'no_repeat_ngram_size': 3, 'forced_eos_token_id': 2}
done
Some non-default generation parameters are set in the model config. These should go into a GenerationConfig file (https://huggingface.co/docs/transformers/generation_strategies#save-a-custom-decoding-strategy-with-your-model) instead. This warning will be raised to an exception in v4.41.
Non-default generation parameters: {'max_length': 62, 'min_length': 11, 'early_stopping': True, 'num_beams': 6, 'no_repeat_ngram_size': 3, 'forced_eos_token_id': 2}
There were missing keys in the checkpoint model loaded: ['model.encoder.embed_tokens.weight', 'model.decoder.embed_tokens.weight', 'lm_head.weight'].
TrainOutput(global_step=25746, training_loss=2.2170493731102003, metrics={'train_runtime': 18355.3967, 'train_samples_per_second': 11.221, 'train_steps_per_second': 1.403, 'total_flos': 1.4870878331849933e+17, 'train_loss': 2.2170493731102003, 'epoch': 2.999825225750073})

 [25746/25746 5:05:54, Epoch 2/3]
Epoch	Training Loss	Validation Loss	Rouge1	Rouge2	Rougel	Rougelsum	Gen Len
0	2.478600	4.186782	33.523800	10.600200	24.960500	26.367300	35.008900
2	1.953600	4.762401	34.154400	11.034800	25.465400	26.935200	35.101500


 [967/967 22:11]
done
{'eval_loss': 4.169002056121826, 'eval_rouge1': 33.2377, 'eval_rouge2': 10.4108, 'eval_rougeL': 24.8623, 'eval_rougeLsum': 26.2076, 'eval_gen_len': 35.0672, 'eval_runtime': 1349.1495, 'eval_samples_per_second': 2.867, 'eval_steps_per_second': 0.717, 'epoch': 2.999825225750073}