alimosavian commited on
Commit
01f0399
1 Parent(s): b207e9f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -71
README.md CHANGED
@@ -15,7 +15,7 @@ language:
15
 
16
  This is the first release of a series of Swedish large language models we call "Lynx". Micro is a small model (2 billion params), but punches way above its weight!
17
 
18
- Lynx micro is a fine-tune of Google DeepMind Gemma 2B, scores just below GPT-3.5 Turbo on Scandeval (https://www.scandeval.com). In fact, the only non OpenAI model (currently) topping the Swedish NLG board on scandeval is a fine-tune of Llama-3 by AI Sweden based on our data recipe.
19
 
20
  We believe that this is a really good model (for its size), but keep in mind that it is still a small model and hasn't memorized as much as larger models tend to do.
21
 
@@ -27,40 +27,6 @@ We believe that this is a really good model (for its size), but keep in mind tha
27
  - **Finetuned from model:** [Gemma 2B, 1.1 instruct](https://huggingface.co/google/gemma-1.1-2b-it)
28
 
29
 
30
- ## Uses
31
-
32
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
33
-
34
- ### Direct Use
35
-
36
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
37
-
38
- [More Information Needed]
39
-
40
- ### Downstream Use [optional]
41
-
42
- <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
43
-
44
- [More Information Needed]
45
-
46
- ### Out-of-Scope Use
47
-
48
- <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
49
-
50
- [More Information Needed]
51
-
52
- ## Bias, Risks, and Limitations
53
-
54
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
55
-
56
- [More Information Needed]
57
-
58
- ### Recommendations
59
-
60
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
61
-
62
- Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
63
-
64
  ## How to Get Started with the Model
65
 
66
  ```python
@@ -128,49 +94,14 @@ r = pipe(
128
 
129
  - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
130
 
131
- #### Speeds, Sizes, Times [optional]
132
-
133
- <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
134
 
135
  [More Information Needed]
136
 
137
  ## Evaluation
138
 
139
- <!-- This section describes the evaluation protocols and provides the results. -->
140
 
141
- ### Testing Data, Factors & Metrics
142
 
143
- #### Testing Data
144
-
145
- <!-- This should link to a Dataset Card if possible. -->
146
-
147
- [More Information Needed]
148
-
149
- #### Factors
150
-
151
- <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
152
-
153
- [More Information Needed]
154
-
155
- #### Metrics
156
-
157
- <!-- These are the evaluation metrics being used, ideally with a description of why. -->
158
-
159
- [More Information Needed]
160
-
161
- ### Results
162
-
163
- [More Information Needed]
164
-
165
- #### Summary
166
-
167
-
168
-
169
- ## Model Examination [optional]
170
-
171
- <!-- Relevant interpretability work for the model goes here -->
172
-
173
- [More Information Needed]
174
 
175
  ## Environmental Impact
176
 
 
15
 
16
  This is the first release of a series of Swedish large language models we call "Lynx". Micro is a small model (2 billion params), but punches way above its weight!
17
 
18
+ Lynx micro is a fine-tune of Google DeepMind Gemma 2B, scores just below GPT-3.5 Turbo on [Scandeval](https://www.scandeval.com). In fact, the only non OpenAI model (currently) topping the Swedish NLG board on scandeval is a fine-tune of Llama-3 by AI Sweden based on our data recipe.
19
 
20
  We believe that this is a really good model (for its size), but keep in mind that it is still a small model and hasn't memorized as much as larger models tend to do.
21
 
 
27
  - **Finetuned from model:** [Gemma 2B, 1.1 instruct](https://huggingface.co/google/gemma-1.1-2b-it)
28
 
29
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
30
  ## How to Get Started with the Model
31
 
32
  ```python
 
94
 
95
  - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
 
 
 
 
97
 
98
  [More Information Needed]
99
 
100
  ## Evaluation
101
 
102
+ The model has been evaluated on [Scandeval](https://www.scandeval.com) swedish subset.
103
 
 
104
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
105
 
106
  ## Environmental Impact
107