alimosavian
commited on
Commit
•
01f0399
1
Parent(s):
b207e9f
Update README.md
Browse files
README.md
CHANGED
@@ -15,7 +15,7 @@ language:
|
|
15 |
|
16 |
This is the first release of a series of Swedish large language models we call "Lynx". Micro is a small model (2 billion params), but punches way above its weight!
|
17 |
|
18 |
-
Lynx micro is a fine-tune of Google DeepMind Gemma 2B, scores just below GPT-3.5 Turbo on Scandeval
|
19 |
|
20 |
We believe that this is a really good model (for its size), but keep in mind that it is still a small model and hasn't memorized as much as larger models tend to do.
|
21 |
|
@@ -27,40 +27,6 @@ We believe that this is a really good model (for its size), but keep in mind tha
|
|
27 |
- **Finetuned from model:** [Gemma 2B, 1.1 instruct](https://huggingface.co/google/gemma-1.1-2b-it)
|
28 |
|
29 |
|
30 |
-
## Uses
|
31 |
-
|
32 |
-
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
33 |
-
|
34 |
-
### Direct Use
|
35 |
-
|
36 |
-
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
37 |
-
|
38 |
-
[More Information Needed]
|
39 |
-
|
40 |
-
### Downstream Use [optional]
|
41 |
-
|
42 |
-
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
|
43 |
-
|
44 |
-
[More Information Needed]
|
45 |
-
|
46 |
-
### Out-of-Scope Use
|
47 |
-
|
48 |
-
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
|
49 |
-
|
50 |
-
[More Information Needed]
|
51 |
-
|
52 |
-
## Bias, Risks, and Limitations
|
53 |
-
|
54 |
-
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
|
55 |
-
|
56 |
-
[More Information Needed]
|
57 |
-
|
58 |
-
### Recommendations
|
59 |
-
|
60 |
-
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
|
61 |
-
|
62 |
-
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
|
63 |
-
|
64 |
## How to Get Started with the Model
|
65 |
|
66 |
```python
|
@@ -128,49 +94,14 @@ r = pipe(
|
|
128 |
|
129 |
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
130 |
|
131 |
-
#### Speeds, Sizes, Times [optional]
|
132 |
-
|
133 |
-
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
134 |
|
135 |
[More Information Needed]
|
136 |
|
137 |
## Evaluation
|
138 |
|
139 |
-
|
140 |
|
141 |
-
### Testing Data, Factors & Metrics
|
142 |
|
143 |
-
#### Testing Data
|
144 |
-
|
145 |
-
<!-- This should link to a Dataset Card if possible. -->
|
146 |
-
|
147 |
-
[More Information Needed]
|
148 |
-
|
149 |
-
#### Factors
|
150 |
-
|
151 |
-
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
|
152 |
-
|
153 |
-
[More Information Needed]
|
154 |
-
|
155 |
-
#### Metrics
|
156 |
-
|
157 |
-
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
|
158 |
-
|
159 |
-
[More Information Needed]
|
160 |
-
|
161 |
-
### Results
|
162 |
-
|
163 |
-
[More Information Needed]
|
164 |
-
|
165 |
-
#### Summary
|
166 |
-
|
167 |
-
|
168 |
-
|
169 |
-
## Model Examination [optional]
|
170 |
-
|
171 |
-
<!-- Relevant interpretability work for the model goes here -->
|
172 |
-
|
173 |
-
[More Information Needed]
|
174 |
|
175 |
## Environmental Impact
|
176 |
|
|
|
15 |
|
16 |
This is the first release of a series of Swedish large language models we call "Lynx". Micro is a small model (2 billion params), but punches way above its weight!
|
17 |
|
18 |
+
Lynx micro is a fine-tune of Google DeepMind Gemma 2B, scores just below GPT-3.5 Turbo on [Scandeval](https://www.scandeval.com). In fact, the only non OpenAI model (currently) topping the Swedish NLG board on scandeval is a fine-tune of Llama-3 by AI Sweden based on our data recipe.
|
19 |
|
20 |
We believe that this is a really good model (for its size), but keep in mind that it is still a small model and hasn't memorized as much as larger models tend to do.
|
21 |
|
|
|
27 |
- **Finetuned from model:** [Gemma 2B, 1.1 instruct](https://huggingface.co/google/gemma-1.1-2b-it)
|
28 |
|
29 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
30 |
## How to Get Started with the Model
|
31 |
|
32 |
```python
|
|
|
94 |
|
95 |
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
96 |
|
|
|
|
|
|
|
97 |
|
98 |
[More Information Needed]
|
99 |
|
100 |
## Evaluation
|
101 |
|
102 |
+
The model has been evaluated on [Scandeval](https://www.scandeval.com) swedish subset.
|
103 |
|
|
|
104 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
105 |
|
106 |
## Environmental Impact
|
107 |
|