Someshfengde's picture
Upload ./README.md with huggingface_hub
c8559c1 verified
## Instruction Tuning LLAMA3
This repo uses the `torchtune` for instruction tuning the llama3 pretrained model on mathematical tasks using LORA.
### Wandb report link
https://wandb.ai/som/torchtune_llama3?nw=nwusersom
## Instruction_tuned Model
https://huggingface.co/Someshfengde/llama-3-instruction-tuned-AIMO
### Original metallama model
https://huggingface.co/meta-llama/Meta-Llama-3-8B
## For running this project
```
> pip install poetry
> poetry install
```
Further commands over shell terminal
### To download the model
```
tune download meta-llama/Meta-Llama-3-8B \
--output-dir llama3-8b-hf \
--hf-token <HF_TOKEN>
```
**To start instruction tuning with lora and torchtune**
```
tune run lora_finetune_single_device --config ./lora_finetune_single_device.yaml
```
### To quantize the model
```
tune run quantize --config ./quantization_config.yaml
```
### To generate inference from model.
```
tune run generate --config ./generation_config.yaml \
prompt="what is 2 + 2."
```
## Dataset used
https://huggingface.co/datasets/Someshfengde/AIMO_dataset
### Evaluations
**To run evaluations**
```
tune run eleuther_eval --config ./eval_config.yaml
```
### TruthfulQA: 0.42
![alt text](images/image.png)
### MMLU Abstract Algebra: 0.35
![alt text](images/image-1.png)
### MATHQA: 0.33
![alt text](images/image-2.png)
### Agieval_sat_math: 0.31
![alt text](images/image-3.png)