File size: 1,496 Bytes
2863d67
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ba8ef5f
2863d67
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ba8ef5f
2863d67
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
---
language:
- sv
license: llama3.1
library_name: transformers
tags:
- unsloth
datasets:
- neph1/bellman-7b-finetune
- neph1/codefeedback-swedish
---

# Model Card for Bellman

This version of bellman is finetuned from llama-3.1-instruct-8b.
It's finetuned for prompt question answering, based on a dataset created from 
Swedish wikipedia, with a lot of Sweden-centric questions.
New from previous versions is questions from a translated code-feedback dataset, as well as a number of stories. It's not great at generating stories, 
but better than previosly.

Please note, the HuggingFace inference api is probably trying to load the adapter (lora) which isn't going to work.

![image/png](https://cdn-uploads.huggingface.co/production/uploads/653cd3049107029eb004f968/IDGX3d9lGe6yx-yHjsrav.png)

## Model Details

Training run on 240724:

Step 	Training Loss 	Validation Loss<br>
25 	    1.352200 	1.034565<br>
50 	    1.033600 	1.009348<br>
75 	    1.022400 	0.996665<br>
100 	1.002900 	0.988050<br>
125 	1.014600 	0.981633<br>
150 	1.006300 	0.975584<br>
175 	0.988800 	0.970966<br>
200 	0.985300 	0.967037<br>
225 	0.992400 	0.964120<br>
250 	0.950000 	0.962472<br>
275 	0.931000 	0.960848<br>
300 	0.932000 	0.958946 <-- picked checkpoint <br>

### Model Description


- **Developed by:** Me
- **Funded by:** Me
- **Model type:** Instruct
- **Language(s) (NLP):** Swedish
- **License:** llama-3.1
- **Finetuned from model:** Llama3.1 Instruct 8b

## Model Card Contact

rickard@mindemia.com