File size: 1,077 Bytes
b7d9c4f
 
440ff8f
 
 
 
b7d9c4f
 
 
 
77dd25f
8e5081a
751775c
 
 
8e5081a
 
45f9485
8e5081a
 
 
 
 
 
 
 
 
 
751775c
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
---
library_name: transformers
base_model:
- AI-Sweden-Models/gpt-sw3-6.7b-v2
datasets:
- skvarre/sv-instruct-v1
---

## Model Details

Finetune of [gpt-sw3-6.7b-v2](https://huggingface.co/AI-Sweden-Models/gpt-sw3-6.7b-v2) using **LoRA** with 4-bit quantization. Adapters merged with base model, however with ```bfloat16``` precision tensors. 

## Usage:

This is a finetune experiment. How to use will be provided later. 

## Evaluation 
[ScandEval](https://scandeval.com/swedish-nlg/) benchmarks:
| Dataset      | Performance (Metric 1 / Metric 2)    |
|:------------:|:------------------------------------:|
|    swerec    | 74.95 ± 1.17 / 61.38 ± 1.37          |
|     suc3     | 30.75 ± 4.11 / 25.69 ± 4.83          |
|   scala-sv   | 8.96 ± 2.09 / 51.50 ± 2.94           |
|  scandiqa-sv | 50.71 ± 0.99 / 56.76 ± 0.89          |
|     swedn    | 64.37 ± 0.72 / 18.25 ± 0.29          |
|    mmlu-sv   | 5.45 ± 0.91 / 28.14 ± 0.82           |
| hellaswag-sv | 27.95 ± 0.73 / 4.19 ± 0.94           |
|     speed    | 5322.20 ± 1132.75 / 1280.06 ± 408.08 |