MaziyarPanahi commited on
Commit
56750f8
1 Parent(s): 07923f6

Update README.md (#1)

Browse files

- Update README.md (24a17449797bac71210a9d123a3299cdf18fb7f3)

Files changed (1) hide show
  1. README.md +36 -9
README.md CHANGED
@@ -4,10 +4,16 @@ library_name: peft
4
  tags:
5
  - axolotl
6
  - generated_from_trainer
 
 
 
7
  base_model: NousResearch/Nous-Hermes-2-Mixtral-8x7B-SFT
8
  model-index:
9
  - name: Nous-Hermes-2-Mixtral-8x7B-SFT-Wikihow
10
  results: []
 
 
 
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -23,21 +29,12 @@ model_type: MixtralForCausalLM
23
  tokenizer_type: LlamaTokenizer
24
  trust_remote_code: true
25
 
26
- # hub_model_id: MaziyarPanahi/Nous-Hermes-2-Mixtral-8x7B-SFT-Function-Calling
27
  hub_model_id: MaziyarPanahi/Nous-Hermes-2-Mixtral-8x7B-SFT-Wikihow
28
  hf_use_auth_token: true
29
 
30
  load_in_4bit: true
31
  strict: false
32
 
33
- # datasets:
34
- # - path: Arist12/EABF-ShareGPT-Long-3.5k
35
- # type: sharegpt
36
- # conversation: chatml
37
- # datasets:
38
- # - path: hypervariance/function-calling-sharegpt
39
- # type: sharegpt
40
- # conversation: chatml
41
  datasets:
42
  - path: HuggingFaceTB/cosmopedia
43
  name:
@@ -131,6 +128,36 @@ This model is a fine-tuned version of [NousResearch/Nous-Hermes-2-Mixtral-8x7B-S
131
  It achieves the following results on the evaluation set:
132
  - Loss: 0.4559
133
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
134
  ## Model description
135
 
136
  More information needed
 
4
  tags:
5
  - axolotl
6
  - generated_from_trainer
7
+ - mixtral
8
+ - nous_hermes
9
+ - wikihow
10
  base_model: NousResearch/Nous-Hermes-2-Mixtral-8x7B-SFT
11
  model-index:
12
  - name: Nous-Hermes-2-Mixtral-8x7B-SFT-Wikihow
13
  results: []
14
+ datasets:
15
+ - HuggingFaceTB/cosmopedia
16
+ pipeline_tag: text-generation
17
  ---
18
 
19
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
29
  tokenizer_type: LlamaTokenizer
30
  trust_remote_code: true
31
 
 
32
  hub_model_id: MaziyarPanahi/Nous-Hermes-2-Mixtral-8x7B-SFT-Wikihow
33
  hf_use_auth_token: true
34
 
35
  load_in_4bit: true
36
  strict: false
37
 
 
 
 
 
 
 
 
 
38
  datasets:
39
  - path: HuggingFaceTB/cosmopedia
40
  name:
 
128
  It achieves the following results on the evaluation set:
129
  - Loss: 0.4559
130
 
131
+ ## How to use
132
+
133
+ **PEFT**
134
+ ```python
135
+ from peft import PeftModel, PeftConfig
136
+ from transformers import AutoModelForCausalLM
137
+
138
+ model_id = "MaziyarPanahi/Nous-Hermes-2-Mixtral-8x7B-SFT-Wikihow"
139
+
140
+ config = PeftConfig.from_pretrained(model_id)
141
+ model = AutoModelForCausalLM.from_pretrained("NousResearch/Nous-Hermes-2-Mixtral-8x7B-SFT")
142
+ model = PeftModel.from_pretrained(model, model_id)
143
+ ```
144
+
145
+ **Transformers**
146
+ ```python
147
+ # Use a pipeline as a high-level helper
148
+ from transformers import pipeline
149
+
150
+ model_id = "MaziyarPanahi/Nous-Hermes-2-Mixtral-8x7B-SFT-Wikihow"
151
+
152
+ pipe = pipeline("text-generation", model=model_id)
153
+
154
+ # Load model directly
155
+ from transformers import AutoTokenizer, AutoModelForCausalLM
156
+
157
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
158
+ model = AutoModelForCausalLM.from_pretrained(model_id)
159
+ ```
160
+
161
  ## Model description
162
 
163
  More information needed