Edit model card

Math_Arabic_Llama-3.2-3B-Instruct

Model Description

Math_Arabic_Llama-3.2-3B-Instruct is a fine-tuned version of the Llama-3.2-3B-Instruct model, tailored for solving mathematical problems in Arabic. The model was trained using the Arabic LLaMA Math Dataset, which includes a wide range of mathematical problems in natural language (Arabic). This model is ideal for educational applications, tutoring, and systems that require automatic math problem-solving in Arabic.

Model Details

Training Data

The model was fine-tuned on the Arabic LLaMA Math Dataset, which consists of 12,496 examples covering various mathematical topics, such as:

  • Basic Arithmetic
  • Algebra
  • Geometry
  • Probability
  • Combinatorics

Each example in the dataset includes:

  • Instruction: The problem statement in Arabic
  • Solution: The solution to the problem in Arabic

Intended Use

Primary Use Cases:

  • Solving mathematical problems in Arabic
  • Educational applications
  • Tutoring systems for Arabic-speaking students
  • Mathematical reasoning tasks in Arabic

How to Use

You can use the model in Python with the Hugging Face transformers library:

from transformers import AutoModelForCausalLM, AutoTokenizer

tokenizer = AutoTokenizer.from_pretrained("Jr23xd23/Math_Arabic_Llama-3.2-3B-Instruct")
model = AutoModelForCausalLM.from_pretrained("Jr23xd23/Math_Arabic_Llama-3.2-3B-Instruct")

# Example: Solving a math problem in Arabic
input_text = "ما هو مجموع الزوايا في مثلث؟" # What is the sum of angles in a triangle?
inputs = tokenizer(input_text, return_tensors="pt")
output = model.generate(**inputs, max_length=100)
print(tokenizer.decode(output[0], skip_special_tokens=True))

Limitations

  • The model is not designed for non-mathematical language tasks.
  • Performance may degrade when applied to highly complex mathematical problems beyond the scope of the training dataset.
  • The model's outputs should be verified for critical applications.

License

This model is licensed under the Apache 2.0 License.

Citation

If you use this model in your research or projects, please cite it as follows:

@model{Math_Arabic_Llama_3.2_3B_Instruct,
  title={Math_Arabic_Llama-3.2-3B-Instruct},
  author={Jr23xd23},
  year={2024},
  publisher={Hugging Face},
  url={https://huggingface.co/Jr23xd23/Math_Arabic_Llama-3.2-3B-Instruct},
}

Acknowledgements

Special thanks to the creators of the Arabic LLaMA Math Dataset for providing a rich resource for fine-tuning the model.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Examples
Unable to determine this model's library. Check the docs .

Model tree for Jr23xd23/Math_Arabic_Llama-3.2-3B-Instruct

Finetuned
(55)
this model

Dataset used to train Jr23xd23/Math_Arabic_Llama-3.2-3B-Instruct