File size: 889 Bytes
b8b0aaa
d116aef
7ebc9c5
 
 
c5ce72e
 
3108483
c5ce72e
01fcaaf
7ebc9c5
dc813e2
f1c37f9
18ea001
 
 
 
01fcaaf
dc813e2
 
7ebc9c5
 
 
 
b8b0aaa
 
 
 
 
01fcaaf
b8b0aaa
7a117c1
 
 
 
b8b0aaa
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
{
  "model": {
    "general": {
      "max_seq_length": 128,
      "seed": 42,
      "num_train_epochs": null,
      "max_steps": 60,
      "dataset_text_field": "text",
      "model_name": null,
      "repository": null
    },
    "peft": {
      "r": 16,
      "alpha": 16,
      "dropout": 0,
      "bias": "none",
      "rslora": false,
      "loftq_config": null
    },
    "sft": {
      "per_device_train_batch_size": 2,
      "gradient_accumulation_steps": 4,
      "learning_rate": 2e-4,
      "warmup_steps": 5,
      "logging_steps": 1,
      "optim": "adamw_8bit",
      "weight_decay": 0.01,
      "lr_scheduler_type": "linear",
      "dataset_num_proc": 2,
      "packing": false
    },
    "choices": ["unsloth/llama-3-8b-bnb-4bit", 
                "unsloth/Qwen2-7B", 
                "unsloth/gemma-7b-bnb-4bit",
                "unsloth/mistral-7b-v0.3-bnb-4bit"]
  }
}