hieunguyenminh commited on
Commit
7e93387
1 Parent(s): 60344a9

End of training

Browse files
Files changed (1) hide show
  1. README.md +5 -3
README.md CHANGED
@@ -2,6 +2,8 @@
2
  license: mit
3
  library_name: peft
4
  tags:
 
 
5
  - generated_from_trainer
6
  base_model: TheBloke/zephyr-7B-beta-GPTQ
7
  model-index:
@@ -34,12 +36,12 @@ More information needed
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 0.0002
37
- - train_batch_size: 8
38
  - eval_batch_size: 8
39
  - seed: 42
40
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
41
  - lr_scheduler_type: cosine
42
- - training_steps: 30
43
  - mixed_precision_training: Native AMP
44
 
45
  ### Training results
@@ -51,5 +53,5 @@ The following hyperparameters were used during training:
51
  - PEFT 0.7.1
52
  - Transformers 4.36.2
53
  - Pytorch 2.1.0+cu118
54
- - Datasets 2.15.0
55
  - Tokenizers 0.15.0
 
2
  license: mit
3
  library_name: peft
4
  tags:
5
+ - trl
6
+ - sft
7
  - generated_from_trainer
8
  base_model: TheBloke/zephyr-7B-beta-GPTQ
9
  model-index:
 
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 0.0002
39
+ - train_batch_size: 32
40
  - eval_batch_size: 8
41
  - seed: 42
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: cosine
44
+ - training_steps: 300
45
  - mixed_precision_training: Native AMP
46
 
47
  ### Training results
 
53
  - PEFT 0.7.1
54
  - Transformers 4.36.2
55
  - Pytorch 2.1.0+cu118
56
+ - Datasets 2.16.0
57
  - Tokenizers 0.15.0