mikecovlee commited on
Commit
5faabe5
1 Parent(s): 7e80c53

Upload 3 files

Browse files
Files changed (3) hide show
  1. adapter_config.json +17 -0
  2. adapter_model.bin +3 -0
  3. mlora-train.json +39 -0
adapter_config.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bias": "none",
3
+ "peft_type": "MIXLORA",
4
+ "task_type": "CAUSAL_LM",
5
+ "r": 8,
6
+ "lora_alpha": 16,
7
+ "lora_dropout": 0.05,
8
+ "target_modules": [
9
+ "w1_proj",
10
+ "w2_proj",
11
+ "w3_proj"
12
+ ],
13
+ "routing_strategy": "mixtral",
14
+ "num_experts": 8,
15
+ "act_fn": "silu",
16
+ "top_k": 3
17
+ }
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ad12b855eddc19cb40331ea5d031130f690f7078a6bdec829bd1e5e98d73596
3
+ size 375936965
mlora-train.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cutoff_len": 1024,
3
+ "save_step": 2000,
4
+ "early_stop_test_step": 2000,
5
+ "train_lora_candidate_num": 1,
6
+ "train_lora_simultaneously_num": 1,
7
+ "train_strategy": "optim",
8
+ "lora": [
9
+ {
10
+ "name": "alpaca-mixlora-7b",
11
+ "optim": "adamw",
12
+ "lr": 3e-4,
13
+ "batch_size": 16,
14
+ "micro_batch_size": 4,
15
+ "test_batch_size": 64,
16
+ "num_epochs": 2,
17
+ "r": 8,
18
+ "lora_alpha": 16,
19
+ "lora_dropout": 0.05,
20
+ "target_modules": {
21
+ "q_proj": false,
22
+ "k_proj": false,
23
+ "v_proj": false,
24
+ "o_proj": false,
25
+ "w1_proj": true,
26
+ "w2_proj": true,
27
+ "w3_proj": true
28
+ },
29
+ "routing_strategy": "mixtral",
30
+ "num_experts": 8,
31
+ "top_k": 3,
32
+ "act_fn": "silu",
33
+ "data": "yahma/alpaca-cleaned",
34
+ "prompt": "template/alpaca.json",
35
+ "group_by_length": false,
36
+ "expand_side": "right"
37
+ }
38
+ ]
39
+ }