henilp105 commited on
Commit
7a539a4
1 Parent(s): f63693c

Model save

Browse files
README.md CHANGED
@@ -1,11 +1,11 @@
1
  ---
2
- license: apache-2.0
3
  library_name: peft
4
  tags:
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
- base_model: mistralai/Mistral-7B-v0.1
9
  datasets:
10
  - generator
11
  model-index:
@@ -18,7 +18,7 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  # zephyr-7b-sft-qlora
20
 
21
- This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the generator dataset.
22
 
23
  ## Model description
24
 
 
1
  ---
2
+ license: llama2
3
  library_name: peft
4
  tags:
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
+ base_model: meta-llama/Llama-2-7b-chat-hf
9
  datasets:
10
  - generator
11
  model-index:
 
18
 
19
  # zephyr-7b-sft-qlora
20
 
21
+ This model is a fine-tuned version of [meta-llama/Llama-2-7b-chat-hf](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) on the generator dataset.
22
 
23
  ## Model description
24
 
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "mistralai/Mistral-7B-v0.1",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -22,11 +22,11 @@
22
  "target_modules": [
23
  "o_proj",
24
  "up_proj",
25
- "q_proj",
26
- "down_proj",
27
  "gate_proj",
 
28
  "k_proj",
29
- "v_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "meta-llama/Llama-2-7b-chat-hf",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
22
  "target_modules": [
23
  "o_proj",
24
  "up_proj",
 
 
25
  "gate_proj",
26
+ "v_proj",
27
  "k_proj",
28
+ "q_proj",
29
+ "down_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1969cac1b6c71b1b86df0cd92ab327717d7c189a3e86766d1eee4dd47c2c9caa
3
- size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:590e4d1c89929b0908eefbc3478046b41d18f13fa5ad140ea2ca1b66a967fdac
3
+ size 159967880
runs/Jul10_14-35-28_b822afc3e210/events.out.tfevents.1720622378.b822afc3e210.6052.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ef2f7fe851aab8fa117ed35255346bee5216ab216ada5e0a1705db9d7e84b86
3
+ size 4184
step_0/README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  library_name: peft
3
- base_model: mistralai/Mistral-7B-v0.1
4
  ---
5
 
6
  # Model Card for Model ID
 
1
  ---
2
  library_name: peft
3
+ base_model: meta-llama/Llama-2-7b-chat-hf
4
  ---
5
 
6
  # Model Card for Model ID
step_0/adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "mistralai/Mistral-7B-v0.1",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -22,11 +22,11 @@
22
  "target_modules": [
23
  "o_proj",
24
  "up_proj",
25
- "q_proj",
26
- "down_proj",
27
  "gate_proj",
 
28
  "k_proj",
29
- "v_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "meta-llama/Llama-2-7b-chat-hf",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
22
  "target_modules": [
23
  "o_proj",
24
  "up_proj",
 
 
25
  "gate_proj",
26
+ "v_proj",
27
  "k_proj",
28
+ "q_proj",
29
+ "down_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
step_0/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1969cac1b6c71b1b86df0cd92ab327717d7c189a3e86766d1eee4dd47c2c9caa
3
- size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:590e4d1c89929b0908eefbc3478046b41d18f13fa5ad140ea2ca1b66a967fdac
3
+ size 159967880
step_0/tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
step_0/tokenizer.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
- size 493443
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
step_0/tokenizer_config.json CHANGED
@@ -27,7 +27,6 @@
27
  "special": true
28
  }
29
  },
30
- "additional_special_tokens": [],
31
  "bos_token": "<s>",
32
  "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
33
  "clean_up_tokenization_spaces": false,
@@ -35,8 +34,8 @@
35
  "legacy": false,
36
  "model_max_length": 2048,
37
  "pad_token": "</s>",
 
38
  "sp_model_kwargs": {},
39
- "spaces_between_special_tokens": false,
40
  "tokenizer_class": "LlamaTokenizer",
41
  "unk_token": "<unk>",
42
  "use_default_system_prompt": false
 
27
  "special": true
28
  }
29
  },
 
30
  "bos_token": "<s>",
31
  "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
32
  "clean_up_tokenization_spaces": false,
 
34
  "legacy": false,
35
  "model_max_length": 2048,
36
  "pad_token": "</s>",
37
+ "padding_side": "right",
38
  "sp_model_kwargs": {},
 
39
  "tokenizer_class": "LlamaTokenizer",
40
  "unk_token": "<unk>",
41
  "use_default_system_prompt": false
step_0/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:75b79bceb55adaaacd7b137a075a580a343454adecd6ec2c2622e16fe54c3ea8
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c37de8a80617c24d55b30a4ca9c08d99731043ee2d8d7b1657591b9541a00f17
3
  size 5176
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
- size 493443
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json CHANGED
@@ -27,7 +27,6 @@
27
  "special": true
28
  }
29
  },
30
- "additional_special_tokens": [],
31
  "bos_token": "<s>",
32
  "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
33
  "clean_up_tokenization_spaces": false,
@@ -35,8 +34,8 @@
35
  "legacy": false,
36
  "model_max_length": 2048,
37
  "pad_token": "</s>",
 
38
  "sp_model_kwargs": {},
39
- "spaces_between_special_tokens": false,
40
  "tokenizer_class": "LlamaTokenizer",
41
  "unk_token": "<unk>",
42
  "use_default_system_prompt": false
 
27
  "special": true
28
  }
29
  },
 
30
  "bos_token": "<s>",
31
  "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
32
  "clean_up_tokenization_spaces": false,
 
34
  "legacy": false,
35
  "model_max_length": 2048,
36
  "pad_token": "</s>",
37
+ "padding_side": "right",
38
  "sp_model_kwargs": {},
 
39
  "tokenizer_class": "LlamaTokenizer",
40
  "unk_token": "<unk>",
41
  "use_default_system_prompt": false
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:75b79bceb55adaaacd7b137a075a580a343454adecd6ec2c2622e16fe54c3ea8
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c37de8a80617c24d55b30a4ca9c08d99731043ee2d8d7b1657591b9541a00f17
3
  size 5176