hari02 commited on
Commit
07aafe2
1 Parent(s): 2e5f7f5

End of training

Browse files
README.md CHANGED
@@ -4,7 +4,7 @@ tags:
4
  - trl
5
  - sft
6
  - generated_from_trainer
7
- base_model: llava-hf/llava-1.5-7b-hf
8
  model-index:
9
  - name: llava-1.5-7b-hf-med
10
  results: []
@@ -15,7 +15,7 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # llava-1.5-7b-hf-med
17
 
18
- This model is a fine-tuned version of [llava-hf/llava-1.5-7b-hf](https://huggingface.co/llava-hf/llava-1.5-7b-hf) on the None dataset.
19
 
20
  ## Model description
21
 
@@ -35,22 +35,18 @@ More information needed
35
 
36
  The following hyperparameters were used during training:
37
  - learning_rate: 1.4e-05
38
- - train_batch_size: 8
39
  - eval_batch_size: 8
40
  - seed: 42
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
- - num_epochs: 7
44
  - mixed_precision_training: Native AMP
45
 
46
- ### Training results
47
-
48
-
49
-
50
  ### Framework versions
51
 
52
  - PEFT 0.10.0
53
- - Transformers 4.40.1
54
- - Pytorch 2.1.2
55
- - Datasets 2.18.0
56
  - Tokenizers 0.19.1
 
4
  - trl
5
  - sft
6
  - generated_from_trainer
7
+ base_model: llava_21
8
  model-index:
9
  - name: llava-1.5-7b-hf-med
10
  results: []
 
15
 
16
  # llava-1.5-7b-hf-med
17
 
18
+ This model was trained from scratch on the None dataset.
19
 
20
  ## Model description
21
 
 
35
 
36
  The following hyperparameters were used during training:
37
  - learning_rate: 1.4e-05
38
+ - train_batch_size: 2
39
  - eval_batch_size: 8
40
  - seed: 42
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
+ - num_epochs: 1
44
  - mixed_precision_training: Native AMP
45
 
 
 
 
 
46
  ### Framework versions
47
 
48
  - PEFT 0.10.0
49
+ - Transformers 4.41.0.dev0
50
+ - Pytorch 2.3.0+cu121
51
+ - Datasets 2.19.0
52
  - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -4,7 +4,7 @@
4
  "base_model_class": "LlavaForConditionalGeneration",
5
  "parent_library": "transformers.models.llava.modeling_llava"
6
  },
7
- "base_model_name_or_path": "llava-hf/llava-1.5-7b-hf",
8
  "bias": "none",
9
  "fan_in_fan_out": false,
10
  "inference_mode": true,
@@ -13,8 +13,8 @@
13
  "layers_pattern": null,
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
- "lora_alpha": 16,
17
- "lora_dropout": 0.0,
18
  "megatron_config": null,
19
  "megatron_core": "megatron.core",
20
  "modules_to_save": null,
@@ -23,19 +23,9 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "down_proj",
27
- "fc1",
28
  "q_proj",
29
- "fc2",
30
- "lm_head",
31
- "linear_1",
32
- "up_proj",
33
- "k_proj",
34
- "gate_proj",
35
- "out_proj",
36
- "linear_2",
37
- "o_proj",
38
- "v_proj"
39
  ],
40
  "task_type": null,
41
  "use_dora": false,
 
4
  "base_model_class": "LlavaForConditionalGeneration",
5
  "parent_library": "transformers.models.llava.modeling_llava"
6
  },
7
+ "base_model_name_or_path": "llava_21",
8
  "bias": "none",
9
  "fan_in_fan_out": false,
10
  "inference_mode": true,
 
13
  "layers_pattern": null,
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_dropout": 0.05,
18
  "megatron_config": null,
19
  "megatron_core": "megatron.core",
20
  "modules_to_save": null,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
 
26
  "q_proj",
27
+ "v_proj",
28
+ "k_proj"
 
 
 
 
 
 
 
 
29
  ],
30
  "task_type": null,
31
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1d078f9ad4f57d6398492cbf74fcfec22ef6899f3cbc6a9a265f8b7f27e2cffe
3
- size 621143576
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4eba1d1f6b79d77e9a1c87d7bd24fc279216dd80e221c7123468f1c28870607
3
+ size 29936104
tokenizer_config.json CHANGED
@@ -44,6 +44,7 @@
44
  }
45
  },
46
  "bos_token": "<s>",
 
47
  "clean_up_tokenization_spaces": false,
48
  "eos_token": "</s>",
49
  "legacy": false,
 
44
  }
45
  },
46
  "bos_token": "<s>",
47
+ "chat_template": "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.USER: <image> {{ messages['question'] }} ASSISTANT: {{ messages['answer'] }}{{eos_token}}",
48
  "clean_up_tokenization_spaces": false,
49
  "eos_token": "</s>",
50
  "legacy": false,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3ba08ee91fa367aef9023e12eba1e71c794d30577bfd3ed1d898775fd7ec640e
3
- size 5048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efb93992f6f82189a888c13c6ec2648a4251642a1b6fe37169bdb644fee48c28
3
+ size 4984