hipnologo commited on
Commit
ef0b67d
1 Parent(s): 6e19851

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +55 -3
README.md CHANGED
@@ -1,10 +1,50 @@
1
  ---
2
  library_name: peft
3
  ---
4
- ## Training procedure
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
 
 
 
 
 
 
6
 
7
- The following `bitsandbytes` quantization config was used during training:
 
 
8
  - load_in_8bit: False
9
  - load_in_4bit: True
10
  - llm_int8_threshold: 6.0
@@ -14,7 +54,19 @@ The following `bitsandbytes` quantization config was used during training:
14
  - bnb_4bit_quant_type: nf4
15
  - bnb_4bit_use_double_quant: True
16
  - bnb_4bit_compute_dtype: bfloat16
17
- ### Framework versions
18
 
 
19
 
20
  - PEFT 0.4.0.dev0
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  library_name: peft
3
  ---
4
+ language: en
5
+ thumbnail:
6
+ tags:
7
+ - peft
8
+ - text-generation
9
+ - chatbot
10
+ - ecommerce
11
+ - fine-tuned
12
+ pipeline_tag: text-generation
13
+ license: apache-2.0
14
+ datasets:
15
+ - kaggle
16
+ metrics:
17
+ -
18
+ ---
19
+
20
+ # Falcon 7B LLM Fine Tune Model
21
+
22
+ ## Model description
23
+
24
+ This model is a fine-tuned version of the `tiiuae/falcon-7b` model using the QLoRa library and the PEFT library. It was fine-tuned on the [Ecommerce-FAQ-Chatbot-Dataset](https://kaggle.com/datasets/saadmakhdoom/ecommerce-faq-chatbot-dataset) from Kaggle.
25
+
26
+ ## Intended uses & limitations
27
+
28
+ #### How to use
29
+
30
+ ```python
31
+ from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
32
+ import torch
33
+
34
+ model_id = "hipnologo/Falcon-7B-FineTune-Chatbot"
35
+
36
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
37
+ model = AutoModelForCausalLM.from_pretrained(model_id)
38
 
39
+ # generate text
40
+ input_prompt = "Hello, Bot!"
41
+ input_ids = tokenizer.encode(input_prompt, return_tensors='pt')
42
+ output = model.generate(input_ids)
43
+ output_text = tokenizer.decode(output[:, input_ids.shape[-1]:][0], skip_special_tokens=True)
44
 
45
+ ## Training procedure
46
+
47
+ The model was fine-tuned on the [Ecommerce-FAQ-Chatbot-Dataset](https://kaggle.com/datasets/saadmakhdoom/ecommerce-faq-chatbot-dataset) using the `bitsandbytes` quantization config:
48
  - load_in_8bit: False
49
  - load_in_4bit: True
50
  - llm_int8_threshold: 6.0
 
54
  - bnb_4bit_quant_type: nf4
55
  - bnb_4bit_use_double_quant: True
56
  - bnb_4bit_compute_dtype: bfloat16
 
57
 
58
+ ### Framework versions
59
 
60
  - PEFT 0.4.0.dev0
61
+
62
+ ## Evaluation results
63
+
64
+ The model was trained for 80 steps, with the training loss decreasing from 0.184 to nearly 0. The final training loss was 0.03094411873175886.
65
+
66
+ - Trainable params: 2359296
67
+ - All params: 3611104128
68
+ - Trainable%: 0.06533447711203746
69
+
70
+ ## License
71
+
72
+ This model is licensed under Apache 2.0. Please see the [LICENSE](https://www.apache.org/licenses/LICENSE-2.0) for more information.