mgv99 commited on
Commit
84bf7c0
1 Parent(s): ebcfb68

Upload folder using huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +37 -41
README.md CHANGED
@@ -1,31 +1,27 @@
1
  ---
2
- {}
 
 
 
 
 
 
 
 
 
 
3
  ---
4
- ---
5
- license:
6
- base_model:
7
- - psmathur/orca_mini_v3_13b
8
- - WizardLM/WizardLM-13B-V1.2
9
- - garage-bAInd/Platypus2-13B
10
- tags:
11
- - merge
12
- - mergekit
13
- - lazymergekit
14
- - psmathur/orca_mini_v3_13b
15
- - WizardLM/WizardLM-13B-V1.2
16
- - garage-bAInd/Platypus2-13B
17
- ---
18
 
19
- # holaa
20
 
21
- holaa is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
22
- * [psmathur/orca_mini_v3_13b](https://huggingface.co/psmathur/orca_mini_v3_13b)
23
- * [WizardLM/WizardLM-13B-V1.2](https://huggingface.co/WizardLM/WizardLM-13B-V1.2)
24
- * [garage-bAInd/Platypus2-13B](https://huggingface.co/garage-bAInd/Platypus2-13B)
25
 
26
- ## 🧩 Configuration
27
 
28
- ```yamlmodels:
29
  - model: psmathur/orca_mini_v3_13b
30
  parameters:
31
  weight: 1.0
@@ -38,27 +34,27 @@
38
  merge_method: linear
39
  dtype: float16```
40
 
41
- ## 💻 Usage
42
 
43
- ```python
44
- !pip install -qU transformers accelerate
45
 
46
- from transformers import AutoTokenizer
47
- import transformers
48
- import torch
49
 
50
- model = "mgv99/holaa"
51
- messages = [{"role": "user", "content": "What is a large language model?"}]
52
 
53
- tokenizer = AutoTokenizer.from_pretrained(model)
54
- prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
55
- pipeline = transformers.pipeline(
56
- "text-generation",
57
- model=model,
58
- torch_dtype=torch.float16,
59
- device_map="auto",
60
- )
61
 
62
- outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
63
- print(outputs[0]["generated_text"])
64
- ```
 
1
  ---
2
+ tags:
3
+ - merge
4
+ - mergekit
5
+ - lazymergekit
6
+ - psmathur/orca_mini_v3_13b
7
+ - WizardLM/WizardLM-13B-V1.2
8
+ - garage-bAInd/Platypus2-13B
9
+ base_model:
10
+ - psmathur/orca_mini_v3_13b
11
+ - WizardLM/WizardLM-13B-V1.2
12
+ - garage-bAInd/Platypus2-13B
13
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
 
15
+ # holaa
16
 
17
+ holaa is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
18
+ * [psmathur/orca_mini_v3_13b](https://huggingface.co/psmathur/orca_mini_v3_13b)
19
+ * [WizardLM/WizardLM-13B-V1.2](https://huggingface.co/WizardLM/WizardLM-13B-V1.2)
20
+ * [garage-bAInd/Platypus2-13B](https://huggingface.co/garage-bAInd/Platypus2-13B)
21
 
22
+ ## 🧩 Configuration
23
 
24
+ ```yamlmodels:
25
  - model: psmathur/orca_mini_v3_13b
26
  parameters:
27
  weight: 1.0
 
34
  merge_method: linear
35
  dtype: float16```
36
 
37
+ ## 💻 Usage
38
 
39
+ ```python
40
+ !pip install -qU transformers accelerate
41
 
42
+ from transformers import AutoTokenizer
43
+ import transformers
44
+ import torch
45
 
46
+ model = "mgv99/holaa"
47
+ messages = [{"role": "user", "content": "What is a large language model?"}]
48
 
49
+ tokenizer = AutoTokenizer.from_pretrained(model)
50
+ prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
51
+ pipeline = transformers.pipeline(
52
+ "text-generation",
53
+ model=model,
54
+ torch_dtype=torch.float16,
55
+ device_map="auto",
56
+ )
57
 
58
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
59
+ print(outputs[0]["generated_text"])
60
+ ```