kkchaulagain commited on
Commit
323aeaf
1 Parent(s): 366c382

End of training

Browse files
Files changed (1) hide show
  1. README.md +15 -13
README.md CHANGED
@@ -1,15 +1,15 @@
1
- ---
2
- base_model: llava-hf/llava-1.5-7b-hf
3
- library_name: peft
4
- license: llama2
5
- tags:
6
- - trl
7
- - sft
8
- - generated_from_trainer
9
- model-index:
10
- - name: llava-1.5-7b-hf-ft-mix-vsft
11
- results: []
12
- ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
@@ -36,9 +36,11 @@ More information needed
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 1.4e-05
39
- - train_batch_size: 8
40
  - eval_batch_size: 8
41
  - seed: 42
 
 
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: linear
44
  - num_epochs: 1
 
1
+ ---
2
+ base_model: llava-hf/llava-1.5-7b-hf
3
+ library_name: peft
4
+ license: llama2
5
+ tags:
6
+ - trl
7
+ - sft
8
+ - generated_from_trainer
9
+ model-index:
10
+ - name: llava-1.5-7b-hf-ft-mix-vsft
11
+ results: []
12
+ ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
 
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 1.4e-05
39
+ - train_batch_size: 15
40
  - eval_batch_size: 8
41
  - seed: 42
42
+ - gradient_accumulation_steps: 2
43
+ - total_train_batch_size: 30
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
  - num_epochs: 1