hipnologo commited on
Commit
fb0d381
1 Parent(s): ef0b67d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -24
README.md CHANGED
@@ -1,20 +1,13 @@
1
  ---
2
  library_name: peft
3
- ---
4
- language: en
5
- thumbnail:
6
- tags:
7
- - peft
8
- - text-generation
9
- - chatbot
10
- - ecommerce
11
- - fine-tuned
12
- pipeline_tag: text-generation
13
  license: apache-2.0
14
  datasets:
15
- - kaggle
16
- metrics:
17
- -
 
 
 
18
  ---
19
 
20
  # Falcon 7B LLM Fine Tune Model
@@ -41,19 +34,20 @@ input_prompt = "Hello, Bot!"
41
  input_ids = tokenizer.encode(input_prompt, return_tensors='pt')
42
  output = model.generate(input_ids)
43
  output_text = tokenizer.decode(output[:, input_ids.shape[-1]:][0], skip_special_tokens=True)
 
44
 
45
  ## Training procedure
46
 
47
  The model was fine-tuned on the [Ecommerce-FAQ-Chatbot-Dataset](https://kaggle.com/datasets/saadmakhdoom/ecommerce-faq-chatbot-dataset) using the `bitsandbytes` quantization config:
48
- - load_in_8bit: False
49
- - load_in_4bit: True
50
- - llm_int8_threshold: 6.0
51
- - llm_int8_skip_modules: None
52
- - llm_int8_enable_fp32_cpu_offload: False
53
- - llm_int8_has_fp16_weight: False
54
- - bnb_4bit_quant_type: nf4
55
- - bnb_4bit_use_double_quant: True
56
- - bnb_4bit_compute_dtype: bfloat16
57
 
58
  ### Framework versions
59
 
@@ -61,7 +55,7 @@ The model was fine-tuned on the [Ecommerce-FAQ-Chatbot-Dataset](https://kaggle.c
61
 
62
  ## Evaluation results
63
 
64
- The model was trained for 80 steps, with the training loss decreasing from 0.184 to nearly 0. The final training loss was 0.03094411873175886.
65
 
66
  - Trainable params: 2359296
67
  - All params: 3611104128
@@ -69,4 +63,4 @@ The model was trained for 80 steps, with the training loss decreasing from 0.184
69
 
70
  ## License
71
 
72
- This model is licensed under Apache 2.0. Please see the [LICENSE](https://www.apache.org/licenses/LICENSE-2.0) for more information.
 
1
  ---
2
  library_name: peft
 
 
 
 
 
 
 
 
 
 
3
  license: apache-2.0
4
  datasets:
5
+ - dltdojo/ecommerce-faq-chatbot-dataset
6
+ language:
7
+ - en
8
+ pipeline_tag: text-generation
9
+ tags:
10
+ - text-generation-inference
11
  ---
12
 
13
  # Falcon 7B LLM Fine Tune Model
 
34
  input_ids = tokenizer.encode(input_prompt, return_tensors='pt')
35
  output = model.generate(input_ids)
36
  output_text = tokenizer.decode(output[:, input_ids.shape[-1]:][0], skip_special_tokens=True)
37
+ ```
38
 
39
  ## Training procedure
40
 
41
  The model was fine-tuned on the [Ecommerce-FAQ-Chatbot-Dataset](https://kaggle.com/datasets/saadmakhdoom/ecommerce-faq-chatbot-dataset) using the `bitsandbytes` quantization config:
42
+ - load_in_8bit: `False`
43
+ - load_in_4bit: `True`
44
+ - llm_int8_threshold: `6.0`
45
+ - llm_int8_skip_modules: `None`
46
+ - llm_int8_enable_fp32_cpu_offload: `False`
47
+ - llm_int8_has_fp16_weight: `False`
48
+ - bnb_4bit_quant_type: `nf4`
49
+ - bnb_4bit_use_double_quant: `True`
50
+ - bnb_4bit_compute_dtype: `bfloat16`
51
 
52
  ### Framework versions
53
 
 
55
 
56
  ## Evaluation results
57
 
58
+ The model was trained for 80 steps, with the training loss decreasing from 0.184 to nearly 0. The final training loss was `0.03094411873175886`.
59
 
60
  - Trainable params: 2359296
61
  - All params: 3611104128
 
63
 
64
  ## License
65
 
66
+ This model is licensed under Apache 2.0. Please see the [LICENSE](https://www.apache.org/licenses/LICENSE-2.0) for more information.