You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Model Details

The Dorna models are a family of decoder-only models, specifically trained/fine-tuned on Persian data, developed by Part AI. As a new release, an 8B instruct model from this family is being made available. Dorna2-Llama3.1-8B-Instruct is built using the Meta Llama 3.1 Instruct model.

How to use

To test and use model freely on Hugging Face Spaces click here!

You can also run conversational inference using the Transformers Auto classes with the generate() function. Let's look at an example.

import torch
import transformers
from transformers import AutoTokenizer, AutoModelForCausalLM


tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModelForCausalLM.from_pretrained(
    model_path,
    torch_dtype=torch.bfloat16,
    device_map="auto",
)

messages = [
    {"role": "system",
     "content": "You are a helpful Persian assistant. Please answer questions in the asked language."},
    {"role": "user", "content": "کاغذ A4 بزرگ تر است یا A5؟"},
]

input_ids = tokenizer.apply_chat_template(
    messages,
    add_generation_prompt=True,
    return_tensors="pt"
).to(model.device)

terminators = [
    tokenizer.eos_token_id,
    tokenizer.convert_tokens_to_ids("<|eot_id|>")
]

outputs = model.generate(
    input_ids,
    max_new_tokens=256,
    eos_token_id=terminators,
    do_sample=True,
    temperature=0.3,
    top_p=0.85,
)
response = outputs[0][input_ids.shape[-1]:]
print(tokenizer.decode(response, skip_special_tokens=True))

You can also use the notebook below to test the model in Google Colab.

Colab Code

Evaluation

Comparative Evaluation

This evaluation compares Dorna2-Llama3.1-8B-Instruct, Llama3.1-8B-Instruct, and other fine-tuned Llama3.1-8B models. For broader comparisons among various large language models (LLMs), please refer to the Open Persian LLM Leaderboard, which provides a comprehensive evaluation across multiple LLMs.

Tasks and Evaluation Framework

Five specialized tasks have been carefully curated to evaluate and benchmark the models. Each task has been designed to challenge different aspects of the models' capabilities. These tasks include:

  • Part Multiple Choice: Focuses on common knowledge and reasoning in a multiple-choice format.
  • ARC Easy: Tests on easy-level general knowledge.
  • ARC Challenge: Assesses models on harder questions requiring advanced reasoning.
  • MMLU Pro: Covers professional-level exams.
  • AUT Multiple Choice Persian: Specialized Persian-language examination.

Each dataset is entirely in Persian, offering a unique and robust testing ground for LLMs in non-English settings. Collectively, the datasets contain over 40k samples, spanning diverse linguistic and technical challenges such as Common Knowledge, Reasoning, Summarization, and Specialized Examinations.

Evaluation Results

Model Average Accuracy Part Multiple Choice ARC Easy ARC Challenge MMLU Pro AUT Multiple Choice Persian
PartAI/Dorna2-Llama3.1-8B-Instruct 50.72 34.48 79.59 64.42 21.47 53.64
O1-OPEN/OpenO1-LLama-8B-v0.1 50.22 34.66 77.87 63.08 21.24 54.24
meta-llama/Llama-3.1-8B-Instruct 50.14 36.68 78.40 60.40 21.00 54.24
NousResearch/Hermes-3-Llama-3.1-8B 48.77 35.01 77.01 58.39 21.00 52.46
Skywork/Skywork-o1-Open-Llama-3.1-8B 34.15 27.02 47.12 41.61 14.55 40.43

Contact us

If you have any questions regarding this model, you can reach us via the community on Hugging Face.

Downloads last month
484
Safetensors
Model size
8.03B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for PartAI/Dorna2-Llama3.1-8B-Instruct

Finetuned
(581)
this model
Quantizations
4 models

Space using PartAI/Dorna2-Llama3.1-8B-Instruct 1

Collection including PartAI/Dorna2-Llama3.1-8B-Instruct