Alpaca-Lora-Swe 7B

Alpaca-Lora-Swe-7b is a LLaMA-7B model fine-tuned on the translated Stanford Alpaca dataset to follow the 🇸🇪 Swedish instructions

This model was trained for 5 epoch with a combined dataset of english + swedish and the original Alpca-Lora prompt using the following command:

python3 finetune.py --base_model='./llama-7b'  --output_dir='./lora-alpaca' --resume_from_checkpoint true --micro_batch_size=14 --num_epochs=3 --cutoff_len=512 --group_by_length --output_dir='./lora-alpaca-swe'  --data_path='alpaca_gpt4_combined.json'  --num_epochs 5  --lora_target_modules='[q_proj,k_proj,v_proj,o_proj]'

The training run logs are available at https://wandb.ai/jeremy-cochoy/huggingface/runs/896ntg42

For more information, please visit the Github repo: https://github.com/jeremycochoy/alpaca-lora-swe

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Dataset used to train jeremyc/alpaca-lora-swe-7b