ibivibiv commited on
Commit
dcfb65b
1 Parent(s): 5e2f54d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -1
README.md CHANGED
@@ -9,14 +9,31 @@ language:
9
 
10
  Fine tune of [Smaug 72b v0.1](https://huggingface.co/abacusai/Smaug-72B-v0.1) using an alpaca data set I have handy. The data is of planning and reasoning, which I use to help allow a model to break down a set of asks into a logical plan. For some odd reason it bumps the mmlu and winogrande? I would have expected the ARC to go up over those two, but this is often more of an artform than a science at times. All thanks to [Albacus.AI](https://huggingface.co/abacusai) for sharing their work.
11
 
 
 
12
  ![img](./alpaca_dragon.png)
13
 
14
 
15
 
 
16
  ## How to Get Started with the Model
17
 
18
  Use the code below to get started with the model.
19
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
 
21
  ## Evaluation
22
 
@@ -82,7 +99,7 @@ Use the code below to get started with the model.
82
  | hendrycksTest-us_foreign_policy | 94.00 |
83
  | hendrycksTest-virology | 57.23 |
84
  | hendrycksTest-world_religions | 89.47 |
85
- | truthfulqa:mc | - |
86
  | winogrande | 86.03 |
87
  | gsm8k | 77.63 |
88
 
 
9
 
10
  Fine tune of [Smaug 72b v0.1](https://huggingface.co/abacusai/Smaug-72B-v0.1) using an alpaca data set I have handy. The data is of planning and reasoning, which I use to help allow a model to break down a set of asks into a logical plan. For some odd reason it bumps the mmlu and winogrande? I would have expected the ARC to go up over those two, but this is often more of an artform than a science at times. All thanks to [Albacus.AI](https://huggingface.co/abacusai) for sharing their work.
11
 
12
+ I used the same dataset in training one of my owl series [Strix Rufipes 70B](https://huggingface.co/ibivibiv/strix-rufipes-70b), which has worked well for planning out development tasks and other technical work.
13
+
14
  ![img](./alpaca_dragon.png)
15
 
16
 
17
 
18
+
19
  ## How to Get Started with the Model
20
 
21
  Use the code below to get started with the model.
22
 
23
+ ```
24
+ # Load model directly
25
+ from transformers import AutoTokenizer, AutoModelForCausalLM
26
+
27
+ tokenizer = AutoTokenizer.from_pretrained("ibivibiv/alpaca-dragon-72b-v1")
28
+ model = AutoModelForCausalLM.from_pretrained("ibivibiv/alpaca-dragon-72b-v1")
29
+
30
+ inputs = tokenizer("### Instruction: Create a plan for developing the game of snake in python using pygame.\n### Response:\n", return_tensors="pt", return_attention_mask=False)
31
+
32
+ outputs = model.generate(**inputs, max_length=200)
33
+ text = tokenizer.batch_decode(outputs)[0]
34
+ print(text)
35
+ ```
36
+
37
 
38
  ## Evaluation
39
 
 
99
  | hendrycksTest-us_foreign_policy | 94.00 |
100
  | hendrycksTest-virology | 57.23 |
101
  | hendrycksTest-world_religions | 89.47 |
102
+ | truthfulqa:mc | 72.6 |
103
  | winogrande | 86.03 |
104
  | gsm8k | 77.63 |
105