| { | |
| "model": "logos23-gemma2_2b", | |
| "family": "gemma2_2b", | |
| "family_name": "Gemma 2 2B", | |
| "base_model": "google/gemma-2-2b", | |
| "base_model_quantized": "google/gemma-2-2b", | |
| "method": "LoRA (bf16)", | |
| "framework": "unsloth", | |
| "lora_rank": 64, | |
| "lora_alpha": 64, | |
| "lora_target_modules": [ | |
| "q_proj", | |
| "k_proj", | |
| "v_proj", | |
| "o_proj", | |
| "gate_proj", | |
| "up_proj", | |
| "down_proj" | |
| ], | |
| "epochs": 3, | |
| "effective_batch_size": 16, | |
| "load_in_4bit": false, | |
| "learning_rate": 0.0002, | |
| "lr_scheduler": "cosine", | |
| "max_seq_length": 2048, | |
| "dataset": "logos22_nothink.jsonl", | |
| "dataset_size": 895, | |
| "train_on_responses_only": true, | |
| "think_blocks": "stripped (no-think variant)", | |
| "final_loss": 1.2898975720717794, | |
| "runtime_seconds": 221.1209 | |
| } |