|
{ |
|
"model_name": "llama2_lora", |
|
"finetuning_config": { |
|
"learning_rate": 5e-05, |
|
"gradient_accumulation_steps": 1, |
|
"batch_size": 1, |
|
"weight_decay": 0.01, |
|
"warmup_steps": 50, |
|
"eval_steps": 5000, |
|
"save_steps": 5000, |
|
"max_length": 512, |
|
"num_train_epochs": 3, |
|
"logging_steps": 10, |
|
"max_grad_norm": 2.0, |
|
"save_total_limit": 4, |
|
"optimizer_name": "adamw", |
|
"output_dir": "./checkpoints/llama2-lora_wow-mctx_f1-top1" |
|
}, |
|
"generation_config": { |
|
"penalty_alpha": 0.6, |
|
"top_k": 4, |
|
"max_new_tokens": 256, |
|
"do_sample": false, |
|
"top_p": null |
|
} |
|
} |