Edit model card

Gemma2b_V03_BRONCO_CARDIO_SUMMARY_CATALOG

This model is a fine-tuned version of google/gemma-2b on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 0.2659
  • Num Input Tokens Seen: 32680580

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 4
  • eval_batch_size: 4
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_steps: 50
  • num_epochs: 5

Training results

Training Loss Epoch Step Validation Loss Input Tokens Seen
0.5621 0.2500 1370 0.5018 1573488
0.4518 0.4999 2740 0.4115 3273680
0.386 0.7499 4110 0.3684 4936100
0.364 0.9998 5480 0.3308 6511060
0.2709 1.2498 6850 0.3131 8278928
0.2422 1.4997 8220 0.2944 9895248
0.2264 1.7497 9590 0.2707 11490512
0.235 1.9996 10960 0.2554 13071912
0.1685 2.2496 12330 0.2606 14810372
0.1538 2.4995 13700 0.2519 16231800
0.1566 2.7495 15070 0.2459 17815564
0.1548 2.9995 16440 0.2387 19581608
0.1182 3.2494 17810 0.2599 21161040
0.1106 3.4994 19180 0.2576 22698788
0.1176 3.7493 20550 0.2550 24438800
0.1175 3.9993 21920 0.2558 26121308
0.1032 4.2492 23290 0.2645 27832740
0.0998 4.4992 24660 0.2656 29363700
0.102 4.7491 26030 0.2658 31053092
0.1019 4.9991 27400 0.2659 32672364

Framework versions

  • Transformers 4.40.2
  • Pytorch 2.2.2+cu121
  • Datasets 2.19.0
  • Tokenizers 0.19.1
Downloads last month
30
Safetensors
Model size
2.51B params
Tensor type
BF16
·

Finetuned from

Datasets used to train BachelorThesis/Gemma-2b_V03_BRONCO_CARDIO_SUMMARY_CATALOG

Collection including BachelorThesis/Gemma-2b_V03_BRONCO_CARDIO_SUMMARY_CATALOG