Edit model card

This model has been pushed to the Hub using the PytorchModelHubMixin integration:

  • Library: [More Information Needed]
  • Docs: [More Information Needed]

done Some non-default generation parameters are set in the model config. These should go into a GenerationConfig file (https://huggingface.co/docs/transformers/generation_strategies#save-a-custom-decoding-strategy-with-your-model) instead. This warning will be raised to an exception in v4.41. Non-default generation parameters: {'max_length': 62, 'min_length': 11, 'early_stopping': True, 'num_beams': 6, 'no_repeat_ngram_size': 3, 'forced_eos_token_id': 2} done Some non-default generation parameters are set in the model config. These should go into a GenerationConfig file (https://huggingface.co/docs/transformers/generation_strategies#save-a-custom-decoding-strategy-with-your-model) instead. This warning will be raised to an exception in v4.41. Non-default generation parameters: {'max_length': 62, 'min_length': 11, 'early_stopping': True, 'num_beams': 6, 'no_repeat_ngram_size': 3, 'forced_eos_token_id': 2} done Some non-default generation parameters are set in the model config. These should go into a GenerationConfig file (https://huggingface.co/docs/transformers/generation_strategies#save-a-custom-decoding-strategy-with-your-model) instead. This warning will be raised to an exception in v4.41. Non-default generation parameters: {'max_length': 62, 'min_length': 11, 'early_stopping': True, 'num_beams': 6, 'no_repeat_ngram_size': 3, 'forced_eos_token_id': 2} There were missing keys in the checkpoint model loaded: ['model.encoder.embed_tokens.weight', 'model.decoder.embed_tokens.weight', 'lm_head.weight']. TrainOutput(global_step=25746, training_loss=2.2170493731102003, metrics={'train_runtime': 18355.3967, 'train_samples_per_second': 11.221, 'train_steps_per_second': 1.403, 'total_flos': 1.4870878331849933e+17, 'train_loss': 2.2170493731102003, 'epoch': 2.999825225750073})

[25746/25746 5:05:54, Epoch 2/3] Epoch Training Loss Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len 0 2.478600 4.186782 33.523800 10.600200 24.960500 26.367300 35.008900 2 1.953600 4.762401 34.154400 11.034800 25.465400 26.935200 35.101500

[967/967 22:11] done {'eval_loss': 4.169002056121826, 'eval_rouge1': 33.2377, 'eval_rouge2': 10.4108, 'eval_rougeL': 24.8623, 'eval_rougeLsum': 26.2076, 'eval_gen_len': 35.0672, 'eval_runtime': 1349.1495, 'eval_samples_per_second': 2.867, 'eval_steps_per_second': 0.717, 'epoch': 2.999825225750073}

Downloads last month
9
Safetensors
Model size
426 params
Tensor type
F32
·
Inference API
This model can be loaded on Inference API (serverless).