mcanoglu commited on
Commit
5e31a24
1 Parent(s): bc4b188

End of training

Browse files
README.md CHANGED
@@ -4,7 +4,6 @@ tags:
4
  - generated_from_trainer
5
  metrics:
6
  - accuracy
7
- - f1
8
  - precision
9
  - recall
10
  model-index:
@@ -19,11 +18,11 @@ should probably proofread and complete it, then remove this comment. -->
19
 
20
  This model is a fine-tuned version of [bigcode/starcoderbase-1b](https://huggingface.co/bigcode/starcoderbase-1b) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
- - Loss: 0.5149
23
- - Accuracy: 0.7523
24
- - F1: 0.7482
25
- - Precision: 0.7430
26
- - Recall: 0.7533
27
 
28
  ## Model description
29
 
@@ -43,28 +42,30 @@ More information needed
43
 
44
  The following hyperparameters were used during training:
45
  - learning_rate: 2e-05
46
- - train_batch_size: 2
47
- - eval_batch_size: 2
48
  - seed: 4711
49
- - gradient_accumulation_steps: 16
50
  - total_train_batch_size: 32
51
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
  - lr_scheduler_type: linear
53
- - num_epochs: 3
54
  - mixed_precision_training: Native AMP
55
 
56
  ### Training results
57
 
58
- | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall |
59
- |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:---------:|:------:|
60
- | 0.7604 | 1.0 | 996 | 0.5379 | 0.7144 | 0.6627 | 0.7829 | 0.5745 |
61
- | 0.4649 | 2.0 | 1992 | 0.4524 | 0.7480 | 0.7585 | 0.7129 | 0.8104 |
62
- | 0.318 | 3.0 | 2988 | 0.5149 | 0.7523 | 0.7482 | 0.7430 | 0.7533 |
 
 
63
 
64
 
65
  ### Framework versions
66
 
67
- - Transformers 4.36.2
68
- - Pytorch 2.1.2+cu121
69
- - Datasets 2.16.1
70
- - Tokenizers 0.15.0
 
4
  - generated_from_trainer
5
  metrics:
6
  - accuracy
 
7
  - precision
8
  - recall
9
  model-index:
 
18
 
19
  This model is a fine-tuned version of [bigcode/starcoderbase-1b](https://huggingface.co/bigcode/starcoderbase-1b) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.9591
22
+ - Accuracy: 0.7666
23
+ - Roc Auc: 0.7662
24
+ - Precision: 0.7657
25
+ - Recall: 0.7523
26
 
27
  ## Model description
28
 
 
42
 
43
  The following hyperparameters were used during training:
44
  - learning_rate: 2e-05
45
+ - train_batch_size: 8
46
+ - eval_batch_size: 8
47
  - seed: 4711
48
+ - gradient_accumulation_steps: 4
49
  - total_train_batch_size: 32
50
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
  - lr_scheduler_type: linear
52
+ - num_epochs: 5
53
  - mixed_precision_training: Native AMP
54
 
55
  ### Training results
56
 
57
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | Roc Auc | Precision | Recall |
58
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:-------:|:---------:|:------:|
59
+ | 0.7596 | 1.0 | 996 | 0.5406 | 0.6852 | 0.6897 | 0.6264 | 0.8813 |
60
+ | 0.4855 | 2.0 | 1993 | 0.4691 | 0.7377 | 0.7396 | 0.6954 | 0.8237 |
61
+ | 0.3547 | 3.0 | 2989 | 0.4832 | 0.7480 | 0.7479 | 0.7410 | 0.7441 |
62
+ | 0.2463 | 4.0 | 3986 | 0.5966 | 0.7628 | 0.7646 | 0.7196 | 0.8428 |
63
+ | 0.1633 | 5.0 | 4980 | 0.9591 | 0.7666 | 0.7662 | 0.7657 | 0.7523 |
64
 
65
 
66
  ### Framework versions
67
 
68
+ - Transformers 4.37.2
69
+ - Pytorch 2.2.0+cu121
70
+ - Datasets 2.17.1
71
+ - Tokenizers 0.15.2
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e11dcea9a5c14ef2b08a21b7b13adc4cac323b81829b3216840bcac2a158fe38
3
  size 4548876216
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2127f62c21723d8676fa2b820ff6f99a031af22d0564351b7e6ac1b760951084
3
  size 4548876216
runs/Feb21_06-14-19_nglczrkt3t/events.out.tfevents.1708496059.nglczrkt3t.174.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ade619510e9d9bd0b3e3e7870a6a474b667ebea39b87b66de94b470f69c64474
3
- size 8690
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c28a99e931174c5a3c3d3ab70bc0dc0014b9a441d93d5cf4d87abdf02f0c6b35
3
+ size 9044