RichardErkhov commited on
Commit
680ad49
1 Parent(s): e3206e4

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +118 -0
README.md ADDED
@@ -0,0 +1,118 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ Llama-3-8B-instruct - GGUF
11
+ - Model creator: https://huggingface.co/AI-Sweden-Models/
12
+ - Original model: https://huggingface.co/AI-Sweden-Models/Llama-3-8B-instruct/
13
+
14
+
15
+ | Name | Quant method | Size |
16
+ | ---- | ---- | ---- |
17
+ | [Llama-3-8B-instruct.Q2_K.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.Q2_K.gguf) | Q2_K | 2.96GB |
18
+ | [Llama-3-8B-instruct.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.IQ3_XS.gguf) | IQ3_XS | 3.28GB |
19
+ | [Llama-3-8B-instruct.IQ3_S.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.IQ3_S.gguf) | IQ3_S | 3.43GB |
20
+ | [Llama-3-8B-instruct.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.Q3_K_S.gguf) | Q3_K_S | 3.41GB |
21
+ | [Llama-3-8B-instruct.IQ3_M.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.IQ3_M.gguf) | IQ3_M | 3.52GB |
22
+ | [Llama-3-8B-instruct.Q3_K.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.Q3_K.gguf) | Q3_K | 3.74GB |
23
+ | [Llama-3-8B-instruct.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.Q3_K_M.gguf) | Q3_K_M | 3.74GB |
24
+ | [Llama-3-8B-instruct.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.Q3_K_L.gguf) | Q3_K_L | 4.03GB |
25
+ | [Llama-3-8B-instruct.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.IQ4_XS.gguf) | IQ4_XS | 4.18GB |
26
+ | [Llama-3-8B-instruct.Q4_0.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.Q4_0.gguf) | Q4_0 | 4.34GB |
27
+ | [Llama-3-8B-instruct.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.IQ4_NL.gguf) | IQ4_NL | 4.38GB |
28
+ | [Llama-3-8B-instruct.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.Q4_K_S.gguf) | Q4_K_S | 4.37GB |
29
+ | [Llama-3-8B-instruct.Q4_K.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.Q4_K.gguf) | Q4_K | 4.58GB |
30
+ | [Llama-3-8B-instruct.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.Q4_K_M.gguf) | Q4_K_M | 4.58GB |
31
+ | [Llama-3-8B-instruct.Q4_1.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.Q4_1.gguf) | Q4_1 | 4.78GB |
32
+ | [Llama-3-8B-instruct.Q5_0.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.Q5_0.gguf) | Q5_0 | 5.21GB |
33
+ | [Llama-3-8B-instruct.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.Q5_K_S.gguf) | Q5_K_S | 5.21GB |
34
+ | [Llama-3-8B-instruct.Q5_K.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.Q5_K.gguf) | Q5_K | 5.34GB |
35
+ | [Llama-3-8B-instruct.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.Q5_K_M.gguf) | Q5_K_M | 5.34GB |
36
+ | [Llama-3-8B-instruct.Q5_1.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.Q5_1.gguf) | Q5_1 | 5.65GB |
37
+ | [Llama-3-8B-instruct.Q6_K.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.Q6_K.gguf) | Q6_K | 6.14GB |
38
+ | [Llama-3-8B-instruct.Q8_0.gguf](https://huggingface.co/RichardErkhov/AI-Sweden-Models_-_Llama-3-8B-instruct-gguf/blob/main/Llama-3-8B-instruct.Q8_0.gguf) | Q8_0 | 7.95GB |
39
+
40
+
41
+
42
+
43
+ Original model description:
44
+ ---
45
+ language:
46
+ - sv
47
+ - da
48
+ - 'no'
49
+ license: llama3
50
+ base_model: AI-Sweden-Models/Llama-3-8B
51
+ ---
52
+
53
+ # Checkpoint 1
54
+
55
+ ## Training setup
56
+ The training was perfomed on the [LUMI supercomputer](https://lumi-supercomputer.eu/) within the [DeployAI EU project](https://www.ai.se/en/project/deployai).
57
+ Based of the base model [AI-Sweden-Models/Llama-3-8B](https://huggingface.co/AI-Sweden-Models/Llama-3-8B).
58
+
59
+ ## Dataset
60
+ A data recipe by: [42 Labs](https://huggingface.co/four-two-labs)
61
+ ![](https://huggingface.co/AI-Sweden-Models/Llama-3-8B-instruct/resolve/main/42logo.png?download=true)
62
+
63
+ ## How to use
64
+
65
+ ```python
66
+ import transformers
67
+ import torch
68
+
69
+ model_id = "AI-Sweden-Models/Llama-3-8B-instruct"
70
+
71
+ pipeline = transformers.pipeline(
72
+ "text-generation",
73
+ model=model_id,
74
+ model_kwargs={"torch_dtype": torch.bfloat16},
75
+ device_map="auto",
76
+ )
77
+
78
+ messages = [
79
+ {"role": "system", "content": "Du är en hjälpsam assistant som svarar klokt och vänligt."},
80
+ {"role": "user", "content": "Hur gör man pannkakor? Och vad behöver man handla? Undrar också vad 5+6 är.."},
81
+ ]
82
+
83
+ terminators = [
84
+ pipeline.tokenizer.eos_token_id,
85
+ pipeline.tokenizer.convert_tokens_to_ids("<|eot_id|>")
86
+ ]
87
+
88
+ outputs = pipeline(
89
+ messages,
90
+ max_new_tokens=256,
91
+ eos_token_id=terminators,
92
+ do_sample=True,
93
+ temperature=0.6,
94
+ top_p=0.9,
95
+ )
96
+ print(outputs[0]["generated_text"][-1])
97
+ ```
98
+ ```python
99
+ >>> "För att göra pannkakor behöver du följande ingredienser:
100
+ - 1 kopp vetemjöl
101
+ - 1 tesked bakpulver
102
+ - 1/4 tesked salt
103
+ - 1 kopp mjölk
104
+ - 1 stort ägg
105
+ - 2 matskedar smält smör eller olja
106
+
107
+ För att börja, blanda vetemjölet, bakpulvret och saltet i en bunke. I en annan skål, vispa ihop mjölken, ägget och smöret eller oljan.
108
+ Tillsätt de våta ingredienserna till de torra ingredienserna och blanda tills det är väl blandat.
109
+ Låt smeten vila i cirka 10 minuter.
110
+
111
+ För att göra pannkakorna, värm en non-stick-panna eller stekpanna över medelvärme.
112
+ När den är varm, häll smeten på pannan och grädda tills kanterna börjar torka ut och toppen är fast.
113
+ Vänd pannkakan med en stekspade och grädda den andra sidan tills den är gyllenbrun.
114
+ Upprepa med resten av smeten.
115
+
116
+ När det gäller 5+6 är svaret 11."
117
+ ```
118
+