gemma2b-summarize-gpt4o
Collection
9 items
•
Updated
This model is a fine-tuned version of google/gemma-2b on the llama-duo/synth_summarize_dataset_dedup dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
2.9975 | 0.8 | 2 | 3.1990 |
2.8744 | 2.0 | 5 | 3.0048 |
2.8744 | 2.8 | 7 | 2.8992 |
2.3833 | 4.0 | 10 | 2.8237 |
2.3833 | 4.8 | 12 | 2.7889 |
2.1436 | 6.0 | 15 | 2.7588 |
2.1436 | 6.8 | 17 | 2.7501 |
2.0522 | 8.0 | 20 | 2.7415 |
Base model
google/gemma-2b