pkbiswas commited on
Commit
e3db789
1 Parent(s): 0a63efe

End of training

Browse files
README.md CHANGED
@@ -5,6 +5,8 @@ datasets:
5
  library_name: peft
6
  license: llama3.1
7
  tags:
 
 
8
  - generated_from_trainer
9
  model-index:
10
  - name: Llama-3.1-8B-Summarization-QLoRa
@@ -18,7 +20,7 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [meta-llama/Meta-Llama-3.1-8B](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B) on the scitldr dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 2.3876
22
 
23
  ## Model description
24
 
@@ -51,9 +53,9 @@ The following hyperparameters were used during training:
51
 
52
  | Training Loss | Epoch | Step | Validation Loss |
53
  |:-------------:|:------:|:----:|:---------------:|
54
- | 2.2018 | 0.5020 | 500 | 2.3023 |
55
- | 2.2004 | 1.0040 | 1000 | 2.2920 |
56
- | 1.683 | 1.5060 | 1500 | 2.3876 |
57
 
58
 
59
  ### Framework versions
 
5
  library_name: peft
6
  license: llama3.1
7
  tags:
8
+ - trl
9
+ - sft
10
  - generated_from_trainer
11
  model-index:
12
  - name: Llama-3.1-8B-Summarization-QLoRa
 
20
 
21
  This model is a fine-tuned version of [meta-llama/Meta-Llama-3.1-8B](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B) on the scitldr dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 2.3829
24
 
25
  ## Model description
26
 
 
53
 
54
  | Training Loss | Epoch | Step | Validation Loss |
55
  |:-------------:|:------:|:----:|:---------------:|
56
+ | 2.2034 | 0.5020 | 500 | 2.3051 |
57
+ | 2.2011 | 1.0040 | 1000 | 2.2903 |
58
+ | 1.6885 | 1.5060 | 1500 | 2.3829 |
59
 
60
 
61
  ### Framework versions
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
 
23
  "up_proj",
24
- "down_proj",
25
  "v_proj",
26
- "k_proj",
27
- "gate_proj",
28
- "q_proj",
29
- "o_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "gate_proj",
24
+ "k_proj",
25
+ "o_proj",
26
  "up_proj",
 
27
  "v_proj",
28
+ "down_proj",
29
+ "q_proj"
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a382b2c8016e588a6c5f72e865c1cdab775e6a9226289bdb80d7a68d2d8c59a6
3
  size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55a9b3ccd5b6785884cdc83f55630d248479a61c05d9de2cac6323e87ca838bb
3
  size 167832240
runs/Sep05_16-02-15_967fd41d9bcb/events.out.tfevents.1725552161.967fd41d9bcb.41484.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea5182ed4f3311b6316bcc39bbb8b77a5ffda092947a4afbb3a615d6c423e395
3
+ size 9036
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b41c747be2ed63e04cabc2dc84ce736b2327c644803ad2e2834e25362de72bde
3
- size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f01c48a6772e4cc38c20e14de85c548e0b804406a635c2273c5c2364cd8ec0ef
3
+ size 5432