| { | |
| "cutoff_len": 1900, | |
| "save_step": 500, | |
| "train_lora_candidate_num": 1, | |
| "train_lora_simultaneously_num": 1, | |
| "train_strategy": "optim", | |
| "evaluate_result": "./result", | |
| "lora": [ | |
| { | |
| "name": "c2cite++", | |
| "task_name": "synsci++", | |
| "optim": "adamw", | |
| "scheduler_type": "constant", | |
| "warmup_steps": 0, | |
| "lr": 1e-4, | |
| "batch_size": 1, | |
| "micro_batch_size": 1, | |
| "evaluate_batch_size": 1, | |
| "num_epochs": 4, | |
| "r": 8, | |
| "lora_alpha": 16, | |
| "lora_dropout": 0.05, | |
| "target_modules": { | |
| "q_proj": true, | |
| "k_proj": true, | |
| "v_proj": true, | |
| "o_proj": true, | |
| "gate_proj": true, | |
| "down_proj": true, | |
| "up_proj": true | |
| }, | |
| "routing_strategy": "", | |
| "group_by_length": false, | |
| "atten_mat_coin": 0.1, | |
| "router_coin": 0.1, | |
| "cite_coin": 0.2 | |
| } | |
| ] | |
| } |