{ "model_name": "TinyLlama/TinyLlama-1.1B-Chat-v1.0", "model_type": "llama", "training_params": { "learning_rate": 2e-4, "num_train_epochs": 3, "per_device_train_batch_size": 4, "gradient_accumulation_steps": 4, "warmup_ratio": 0.03, "lr_scheduler_type": "cosine", "max_grad_norm": 0.3, "weight_decay": 0.01 }, "lora_config": { "r": 8, "lora_alpha": 16, "target_modules": ["q_proj", "k_proj", "v_proj", "o_proj", "gate_proj", "up_proj", "down_proj"], "bias": "none", "task_type": "CAUSAL_LM" }, "generation_config": { "max_length": 200, "temperature": 0.7, "top_p": 0.9, "top_k": 40, "repetition_penalty": 1.1 }, "training_framework": { "name": "unsloth", "quantization": "4bit", "use_flash_attention": true, "use_peft": true } }