hermeschen1116 commited on
Commit
b1c593a
1 Parent(s): 5112586

End of training

Browse files
README.md ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: unsloth/llama-2-7b-bnb-4bit
3
+ library_name: peft
4
+ license: apache-2.0
5
+ tags:
6
+ - trl
7
+ - sft
8
+ - unsloth
9
+ - generated_from_trainer
10
+ model-index:
11
+ - name: response_generator_for_emotion_chat_bot
12
+ results: []
13
+ ---
14
+
15
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
+ should probably proofread and complete it, then remove this comment. -->
17
+
18
+ # response_generator_for_emotion_chat_bot
19
+
20
+ This model is a fine-tuned version of [unsloth/llama-2-7b-bnb-4bit](https://huggingface.co/unsloth/llama-2-7b-bnb-4bit) on an unknown dataset.
21
+
22
+ ## Model description
23
+
24
+ More information needed
25
+
26
+ ## Intended uses & limitations
27
+
28
+ More information needed
29
+
30
+ ## Training and evaluation data
31
+
32
+ More information needed
33
+
34
+ ## Training procedure
35
+
36
+ ### Training hyperparameters
37
+
38
+ The following hyperparameters were used during training:
39
+ - learning_rate: 0.0002
40
+ - train_batch_size: 4
41
+ - eval_batch_size: 8
42
+ - seed: 42
43
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
+ - lr_scheduler_type: constant
45
+ - lr_scheduler_warmup_ratio: 0.03
46
+ - num_epochs: 1
47
+
48
+ ### Framework versions
49
+
50
+ - PEFT 0.11.1
51
+ - Transformers 4.41.2
52
+ - Pytorch 2.3.0+cu121
53
+ - Datasets 2.20.0
54
+ - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -15,21 +15,21 @@
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": [
18
- "embed_tokens",
19
- "lm_head"
20
  ],
21
  "peft_type": "LORA",
22
- "r": 16,
23
  "rank_pattern": {},
24
  "revision": "unsloth",
25
  "target_modules": [
26
- "o_proj",
27
- "gate_proj",
28
- "v_proj",
29
- "q_proj",
30
  "up_proj",
 
 
31
  "down_proj",
32
- "k_proj"
 
 
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
 
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": [
18
+ "lm_head",
19
+ "embed_tokens"
20
  ],
21
  "peft_type": "LORA",
22
+ "r": 8,
23
  "rank_pattern": {},
24
  "revision": "unsloth",
25
  "target_modules": [
 
 
 
 
26
  "up_proj",
27
+ "v_proj",
28
+ "o_proj",
29
  "down_proj",
30
+ "k_proj",
31
+ "q_proj",
32
+ "gate_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5ef477971bf2e97171eff988659f1f3a6be8c1bbabe614d0064f782e49331901
3
- size 1733078400
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e33740cf840bcfdb116fbeca8b9356c6534619a884b57c85950643e7865e95d9
3
+ size 1653123632
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dba76fa35584b8f6400f76fc2657c6449140eb0fc1470551d8e6da4f5fb7e7b7
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c7df05d620f636d89b16a79736c9367ba18b1a8b0a255003afdef713dccd26d
3
  size 5176