santiadavani commited on
Commit
c94e4d6
1 Parent(s): d1ecc07

Model save

Browse files
README.md CHANGED
@@ -1,11 +1,12 @@
1
  ---
2
- base_model: meta-llama/Llama-2-7b-chat-hf
3
  tags:
4
  - trl
5
  - sft
6
  - generated_from_trainer
7
  datasets:
8
  - generator
 
9
  model-index:
10
  - name: alpaca-gpt4-conversation-llama2-7b-chat
11
  results: []
@@ -43,7 +44,7 @@ The following hyperparameters were used during training:
43
  - total_train_batch_size: 16
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
- - num_epochs: 100
47
 
48
  ### Training results
49
 
@@ -51,7 +52,8 @@ The following hyperparameters were used during training:
51
 
52
  ### Framework versions
53
 
54
- - Transformers 4.37.2
 
55
  - Pytorch 2.1.2+cu121
56
  - Datasets 2.15.0
57
- - Tokenizers 0.15.0
 
1
  ---
2
+ library_name: peft
3
  tags:
4
  - trl
5
  - sft
6
  - generated_from_trainer
7
  datasets:
8
  - generator
9
+ base_model: meta-llama/Llama-2-7b-chat-hf
10
  model-index:
11
  - name: alpaca-gpt4-conversation-llama2-7b-chat
12
  results: []
 
44
  - total_train_batch_size: 16
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
+ - num_epochs: 1
48
 
49
  ### Training results
50
 
 
52
 
53
  ### Framework versions
54
 
55
+ - PEFT 0.7.1
56
+ - Transformers 4.38.1
57
  - Pytorch 2.1.2+cu121
58
  - Datasets 2.15.0
59
+ - Tokenizers 0.15.0
adapter_config.json CHANGED
@@ -4,7 +4,7 @@
4
  "base_model_name_or_path": "meta-llama/Llama-2-7b-chat-hf",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
- "inference_mode": false,
8
  "init_lora_weights": true,
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
@@ -19,8 +19,8 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "q_proj",
23
- "v_proj"
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
 
4
  "base_model_name_or_path": "meta-llama/Llama-2-7b-chat-hf",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
  "init_lora_weights": true,
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "v_proj",
23
+ "q_proj"
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e1585b5aaa370316380af250d6628d3ca5f0d49bc8a371ec98a52802203038ba
3
- size 4211128
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4622c6b4897441a5a0f4841ab611d68363bd4b7a51f5001b25b92b8aa3f5a36c
3
+ size 2114040
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:79b2fbaf4efc5c4ab1418a44b1dca5c36286645f476e02f87d60f3a85c668c5d
3
- size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:922cb154a1f60c1c4631f3c6b2706039d831eec84253a5a6e603924ef7444a0e
3
+ size 4920