| { |
| "adapter_path": "/Users/giacomo/coding/snaply/writing_assistant_fine_tuning/output/adapters", |
| "alpha": 512, |
| "batch_size": 4, |
| "config": "/Users/giacomo/coding/snaply/writing_assistant_fine_tuning/output/lora_config.yaml", |
| "data": "/Users/giacomo/coding/snaply/writing_assistant_fine_tuning/data", |
| "dropout": 0.05, |
| "fine_tune_type": "lora", |
| "grad_accumulation_steps": 4, |
| "grad_checkpoint": false, |
| "iters": 3500, |
| "learning_rate": 2.5e-05, |
| "lora_parameters": { |
| "rank": 8, |
| "dropout": 0.0, |
| "scale": 20.0 |
| }, |
| "lr_schedule": null, |
| "mask_prompt": true, |
| "max_seq_length": 2048, |
| "model": "LiquidAI/LFM2.5-1.2B-Instruct-MLX-8bit", |
| "num_layers": 16, |
| "optimizer": "adam", |
| "optimizer_config": { |
| "adam": {}, |
| "adamw": {}, |
| "muon": {}, |
| "sgd": {}, |
| "adafactor": {} |
| }, |
| "project_name": null, |
| "rank": 256, |
| "report_to": null, |
| "resume_adapter_file": null, |
| "save_every": 50, |
| "seed": 0, |
| "steps_per_eval": 50, |
| "steps_per_report": 10, |
| "test": false, |
| "test_batches": 500, |
| "train": true, |
| "val_batches": 18 |
| } |