yangyx30678 commited on
Commit
c021e9b
1 Parent(s): bfeb2ea

update: readme

Browse files
Files changed (1) hide show
  1. README.md +70 -0
README.md ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: unsloth/llama-2-7b-bnb-4bit
3
+ library_name: peft
4
+ license: apache-2.0
5
+ datasets: hermeschen1116/daily_dialog_for_RG
6
+ tags:
7
+ - trl
8
+ - unsloth
9
+ model-index:
10
+ - name: response_generator_for_emotion_chat_bot
11
+ results: []
12
+ language:
13
+ - en
14
+ pipeline_tag: text-generation
15
+ metrics:
16
+ - accuracy
17
+ - f1-score
18
+ ---
19
+
20
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
21
+ should probably proofread and complete it, then remove this comment. -->
22
+
23
+ # Response Generator for [Emotion Chat Bot](https://github.com/hermeschen1116/chat-bot)
24
+
25
+ This model is a fine-tuned version of [unsloth/llama-2-7b-bnb-4bit](https://huggingface.co/unsloth/llama-2-7b-bnb-4bit) on [hermeschen1116/daily_dialog_for_RG](https://huggingface.co/datasets/hermeschen1116/daily_dialog_for_RG), self modified version of [daily_dialog](li2017dailydialog/daily_dialog).
26
+
27
+ ## Model description
28
+
29
+ More information needed
30
+
31
+ ## Intended uses & limitations
32
+
33
+ More information needed
34
+
35
+ ## Training and evaluation data
36
+
37
+ More information needed
38
+
39
+ ## Training procedure
40
+
41
+ ### Training hyperparameters
42
+
43
+ The following hyperparameters were used during training:
44
+ - system_prompt: ""
45
+ - learning_rate: 0.0002
46
+ - weight_decay: 0.001
47
+ - max_grad_norm: 0.3
48
+ - warmup_ratio: 0.03
49
+ - max_steps: -1
50
+ - train_batch_size: 4
51
+ - seed: 42
52
+ - optimizer: paged_adamw_32bit with betas=(0.9,0.999) and epsilon=1e-08
53
+ - lr_scheduler_type: constant
54
+ - lr_scheduler_warmup_ratio: 0.03
55
+ - num_epochs: 1
56
+ - init_lora_weights: true
57
+ - lora_rank: 16
58
+ - lora_alpha: 16
59
+ - lora_dropout: 0.1
60
+ - use_rslora: true
61
+
62
+ ### Framework versions
63
+
64
+ - PEFT 0.11.1
65
+ - Transformers 4.41.2
66
+ - Pytorch 2.3.0+cu121
67
+ - Datasets 2.20.0
68
+ - Tokenizers 0.19.1
69
+ - Trl 0.8.6
70
+ - Bitsandbytes