RayBernard commited on
Commit
2f74234
·
verified ·
1 Parent(s): ec7aa8d

Model save

Browse files
README.md CHANGED
@@ -6,8 +6,6 @@ tags:
6
  - sft
7
  - generated_from_trainer
8
  base_model: meta-llama/Meta-Llama-3-8B-Instruct
9
- datasets:
10
- - generator
11
  model-index:
12
  - name: llama-3-8B-Instruct-ft
13
  results: []
@@ -18,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  # llama-3-8B-Instruct-ft
20
 
21
- This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) on the generator dataset.
22
 
23
  ## Model description
24
 
@@ -49,10 +47,6 @@ The following hyperparameters were used during training:
49
  - training_steps: 50
50
  - mixed_precision_training: Native AMP
51
 
52
- ### Training results
53
-
54
-
55
-
56
  ### Framework versions
57
 
58
  - PEFT 0.11.1
 
6
  - sft
7
  - generated_from_trainer
8
  base_model: meta-llama/Meta-Llama-3-8B-Instruct
 
 
9
  model-index:
10
  - name: llama-3-8B-Instruct-ft
11
  results: []
 
16
 
17
  # llama-3-8B-Instruct-ft
18
 
19
+ This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) on an unknown dataset.
20
 
21
  ## Model description
22
 
 
47
  - training_steps: 50
48
  - mixed_precision_training: Native AMP
49
 
 
 
 
 
50
  ### Framework versions
51
 
52
  - PEFT 0.11.1
adapter_config.json CHANGED
@@ -20,9 +20,9 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "output_proj",
24
  "v_proj",
25
- "q_proj"
 
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "v_proj",
24
+ "q_proj",
25
+ "output_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b2f4f656bc27573d4fe5273ef6a205ae4a16ada9e167bbb2a208d120044582d1
3
  size 4311795296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6612fac503fb25391a30af892c3abaf1845ea2bcdaa82ee2b0df7906ad579eb6
3
  size 4311795296
special_tokens_map.json CHANGED
@@ -13,5 +13,11 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": "[PAD]"
 
 
 
 
 
 
17
  }
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a405e611c914882d409c5f452127183581fcc197a05ba420b211903fa2fb4a97
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a15f78c8ffb8f3abf4e3dffe77f14cc8f0dbbd2cea086134065e125d0bb19da8
3
  size 5368