Model Card
Model Description
This is a Large Language Model (LLM) trained on a subset of the dataset "mlabonne/orpo-dpo-mix-40k".
Evaluation Results
Hellaswag
Metric | Value |
---|---|
Accuracy | 0.4517 |
How to Use
To use this model, simply download the checkpoint and load it into your preferred deep learning framework.
- Downloads last month
- 5
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for d4niel92/llama-3.2-1B-orpo
Base model
meta-llama/Llama-3.2-1B