NorskGPT-Llama-3-8b-v0.1

This model is a Norwegian variant of Meta-Llama-3-8B, fine-tuned on a carefully selected mix of Norwegian instruction pairs. The model is tuned to understand and generate text in Norwegain.

Intended Use

This model is free to use for personal and research use. However a commercial license is required for commerical applications. This model can be used as an assistant-like chat. Try it out :)

Prompt Template

<|im_start|>system
Du er NorskGPT ....<|im_end|>
<|im_start|>user
Hei<|im_end|>
<|im_start|>assistant
Hei, hva kan jeg hjelpe deg med?<|im_end|>

Sample script

from transformers import AutoModelForCausalLM, AutoTokenizer

device = "cuda" # the device to load the model onto

model_name = "bineric/NorskGPT-Llama3-8b"
model = AutoModelForCausalLM.from_pretrained(model_name)
tokenizer = AutoTokenizer.from_pretrained(model_name)

messages = [
    {"role": "user", "content": "Du er NorskGPT - en AI bot som hjelper brukeren med Γ₯ svare pΓ₯ spΓΈrsmΓ₯l?"},
    {"role": "assistant", "content": "Hei, jeg er NorskGPT, hva kan jeg hjelpe deg med?"},
    {"role": "user", "content": "Fortell meg om Oslo"}
]

encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt")

model_inputs = encodeds.to(device)
model.to(device)

generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True)
decoded = tokenizer.batch_decode(generated_ids)
print(decoded[0])

Limitations

  • This is an LLM, not a knowledge model. It can not be expected to have more information about Norway than the base model.
  • It will generally preform better on tasks that involves summarization, question answering and chat, than on tasks that requires more knowledge about Norway, specific domains, or tasks where the model can answer freely.
  • The model is released as is, and would in most cases need prompt tuning to achieve optimal results.

License

Attribution-NonCommercial-ShareAlike 4.0 International

This model is free to use for personal and research use. However a commercial license is required for commerical applications.

You are free to:

Share β€” copy and redistribute the material in any medium or format
Adapt β€” remix, transform, and build upon the material
The licensor cannot revoke these freedoms as long as you follow the license terms.

Under the following terms:

Attribution β€” You must give appropriate credit , provide a link to the license, and indicate if changes were made . You may do so in any reasonable manner, but not in any way that suggests the licensor endorses you or your use.
NonCommercial β€” You may not use the material for commercial purposes .
ShareAlike β€” If you remix, transform, or build upon the material, you must distribute your contributions under the same license as the original.
No additional restrictions β€” You may not apply legal terms or technological measures that legally restrict others from doing anything the license permits.
Downloads last month
168
Safetensors
Model size
8.03B params
Tensor type
BF16
Β·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for bineric/NorskGPT-Llama3-8b

Finetuned
(372)
this model
Finetunes
2 models
Merges
2 models
Quantizations
4 models

Spaces using bineric/NorskGPT-Llama3-8b 6