CodeQwen1.5-7B-Chat_components_dataset_size_52_epochs_10_2024-06-13_03-26-06_22016823
This model is a fine-tuned version of Qwen/CodeQwen1.5-7B-Chat on the None dataset. It achieves the following results on the evaluation set:
- Loss: 0.2601
- Accuracy: 0.494
- Chrf: 0.846
- Bleu: 0.777
- Sacrebleu: 0.8
- Rouge1: 0.851
- Rouge2: 0.742
- Rougel: 0.835
- Rougelsum: 0.846
- Meteor: 0.833
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.001
- train_batch_size: 1
- eval_batch_size: 1
- seed: 3407
- distributed_type: multi-GPU
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-06
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 52
- training_steps: 520
- mixed_precision_training: Native AMP
Training results
Training Loss | Epoch | Step | Validation Loss | Accuracy | Chrf | Bleu | Sacrebleu | Rouge1 | Rouge2 | Rougel | Rougelsum | Meteor |
---|---|---|---|---|---|---|---|---|---|---|---|---|
0.086 | 0.2063 | 52 | 0.4647 | 0.496 | 0.771 | 0.678 | 0.7 | 0.725 | 0.544 | 0.682 | 0.716 | 0.72 |
0.0084 | 0.4127 | 104 | 0.3979 | 0.495 | 0.796 | 0.71 | 0.7 | 0.809 | 0.643 | 0.768 | 0.801 | 0.784 |
0.004 | 0.6190 | 156 | 0.3448 | 0.495 | 0.812 | 0.723 | 0.7 | 0.82 | 0.68 | 0.795 | 0.814 | 0.796 |
0.0011 | 0.8254 | 208 | 0.3082 | 0.494 | 0.828 | 0.751 | 0.8 | 0.83 | 0.702 | 0.806 | 0.823 | 0.823 |
0.0024 | 1.0317 | 260 | 0.3036 | 0.494 | 0.833 | 0.754 | 0.8 | 0.833 | 0.702 | 0.806 | 0.826 | 0.82 |
0.019 | 1.2381 | 312 | 0.2854 | 0.495 | 0.838 | 0.765 | 0.8 | 0.839 | 0.713 | 0.818 | 0.834 | 0.822 |
0.0011 | 1.4444 | 364 | 0.2765 | 0.495 | 0.842 | 0.77 | 0.8 | 0.844 | 0.722 | 0.825 | 0.838 | 0.834 |
0.0021 | 1.6508 | 416 | 0.2650 | 0.495 | 0.846 | 0.774 | 0.8 | 0.853 | 0.741 | 0.835 | 0.847 | 0.823 |
0.0006 | 1.8571 | 468 | 0.2643 | 0.495 | 0.844 | 0.774 | 0.8 | 0.848 | 0.732 | 0.831 | 0.843 | 0.832 |
0.0797 | 2.0635 | 520 | 0.2601 | 0.494 | 0.846 | 0.777 | 0.8 | 0.851 | 0.742 | 0.835 | 0.846 | 0.833 |
Framework versions
- PEFT 0.7.1
- Transformers 4.41.2
- Pytorch 2.3.0+cu121
- Datasets 2.19.1
- Tokenizers 0.19.1
- Downloads last month
- 1
Model tree for vdavidr/CodeQwen1.5-7B-Chat_components_dataset_size_52_epochs_10_2024-06-13_03-26-06_22016823
Base model
Qwen/CodeQwen1.5-7B-Chat