|
--- |
|
library_name: transformers |
|
tags: |
|
- unsloth |
|
license: llama3 |
|
datasets: |
|
- mii-community/ultrafeedback-preferences-translated-ita |
|
- efederici/alpaca-vs-alpaca-orpo-dpo |
|
--- |
|
|
|
# Model Card for Model ID |
|
|
|
This is llama-3-8b ORPO finetuning for the italian language over a concatenation of two datasets: |
|
- [mii-community/ultrafeedback-preferences-translated-ita](https://huggingface.co/datasets/mii-community/ultrafeedback-preferences-translated-ita) |
|
- [efederici/alpaca-vs-alpaca-orpo-dpo](https://huggingface.co/datasets/efederici/alpaca-vs-alpaca-orpo-dpo) |
|
|
|
The other two differences with `diegobit/llama-3-8b-Instruct-bnb-4bit-ita-orpo` are: |
|
- the starting model, not instruct, `astronomer/Llama-3-8B-Special-Tokens-Adjusted` instead of `unsloth/llama-3-8b-Instruct-bnb-4bit` |
|
- no loading in 4bits |
|
- given the increased need of GPU memory, the sequence max length used for finetuning is 4096 |
|
|
|
## Model Details |
|
|
|
### Model Description |
|
|
|
- **Developed by:** Diego Giorgini |
|
- **Funded by:** AI Technologies SRL - www.aitechnologies.it |
|
- **Language(s) (NLP):** Italian |
|
- **License:** llama3 |
|
- **Finetuned from model:** astronomer/Llama-3-8B-Special-Tokens-Adjusted |
|
|
|
## Training Details |
|
|
|
### Environment |
|
|
|
unsloth: 2024.5 |
|
torch: 2.2 |
|
|
|
### Training Data |
|
|
|
- `mii-community/ultrafeedback-preferences-translated-ita` is a selection of 55k rows of the ultrafeedback dataset, translated into italian with argotranslate. |
|
- `efederici/alpaca-vs-alpaca-orpo-dpo`: The Alpaca vs. Alpaca dataset is a curated blend of the Alpaca dataset and the Alpaca GPT-4 dataset, both available on HuggingFace Datasets. It uses the standard GPT dataset as the 'rejected' answer, steering the model towards the GPT-4 answer, which is considered as the 'chosen' one. |
|
|
|
### Training Procedure |
|
|
|
#### Preprocessing [optional] |
|
|
|
- No preprocessing has been performed, except for formatting with the llama3 chat_template from unsloth: |
|
|
|
```tokenizer = get_chat_template(tokenizer, chat_template = "llama-3")``` |
|
|
|
#### Training Hyperparameters |
|
|
|
- **Training regime:** bf16 |
|
|
|
- **Model loading parameters:** |
|
|
|
``` |
|
max_seq_length = 4096 |
|
dtype = None |
|
load_in_4bit = False |
|
``` |
|
|
|
- **PEFT parameters:** |
|
|
|
``` |
|
r = 64 |
|
lora_alpha = 64 |
|
lora_dropout = 0 |
|
bias = "none" |
|
random_state = 3407 |
|
use_rslora = False |
|
loftq_config = None |
|
``` |
|
|
|
- **ORPOConfig parameters:** |
|
|
|
``` |
|
max_length = 4096 |
|
max_prompt_length = max_seq_length//2 |
|
max_completion_length = max_seq_length//2 |
|
warmup_ratio = 0.1 |
|
weight_decay = 0.01 |
|
per_device_train_batch_size = 1 |
|
gradient_accumulation_steps = 16 |
|
learning_rate=8e-6 |
|
beta = 0.1 |
|
optim = "paged_adamw_8bit" |
|
lr_scheduler_type = "linear" |
|
num_train_epochs = 1 |
|
``` |
|
|
|
#### Speeds, Sizes, Times |
|
|
|
19h on an A100-40GB |
|
|
|
## Model Card Contact |
|
|
|
diego.giorgini@icloud.com |