--- base_model: meta-llama/Llama-3.2-3B-Instruct datasets: - tatsu-lab/alpaca language: en tags: - torchtune --- # my_cool_model This model is a finetuned version of [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct) on the [tatsu-lab/alpaca](https://huggingface.co/tatsu-lab/alpaca) dataset. # Model description More information needed # Training and evaluation results More information needed # Training procedure This model was trained using the [torchtune](https://github.com/pytorch/torchtune) library using the following command: ```bash ppo_full_finetune_single_device.py \ --config ./target/7B_full_ppo_low_memory_single_device.yaml \ device=cuda \ metric_logger._component_=torchtune.utils.metric_logging.WandBLogger \ metric_logger.project=torchtune_ppo \ forward_batch_size=2 \ batch_size=64 \ ppo_batch_size=32 \ gradient_accumulation_steps=16 \ compile=True \ optimizer._component_=bitsandbytes.optim.PagedAdamW \ optimizer.lr=3e-4 ``` # Framework versions - torchtune 0.0.0 - torchao 0.5.0 - datasets 2.20.0 - sentencepiece 0.2.0