AF0815 commited on
Commit
03ff6d6
·
verified ·
1 Parent(s): 9751884

Upload merged Qwen3-4B-Instruct-2507 model (auto-generated README)

Browse files
README.md CHANGED
@@ -17,7 +17,6 @@ tags:
17
  - agentbench
18
  - alfworld
19
  - dbbench
20
- - db-oversampling
21
  ---
22
 
23
  # qwen3-4b-agentbench-dbalf-lora
@@ -49,13 +48,13 @@ Loss is applied to **all assistant turns** in the trajectory, enabling the model
49
  - Mixing ratio (pre-merge target): **DB:ALF = 1:1**
50
 
51
  ### DB Oversampling (category-aware)
52
- Enabled: **True**
53
 
54
  DB category weights used during training-data preparation:
55
 
56
- - counting: 6
57
- - comparison: 4
58
- - ranking: 2
59
  - select: 1
60
  - insert: 1
61
  - update: 1
@@ -65,10 +64,10 @@ DB category weights used during training-data preparation:
65
 
66
  - Base model: Qwen/Qwen3-4B-Instruct-2507
67
  - Method: LoRA (full precision base)
68
- - Max sequence length: 2048
69
- - Epochs: 2
70
- - Learning rate: 2e-06
71
- - LoRA: r=64, alpha=128, dropout=0.0
72
  - Per-device train batch size: 2
73
  - Gradient accumulation: 4
74
 
@@ -80,7 +79,7 @@ from peft import PeftModel
80
  import torch
81
 
82
  base = "Qwen/Qwen3-4B-Instruct-2507"
83
- adapter = "your_id/your-repo"
84
 
85
  tokenizer = AutoTokenizer.from_pretrained(base)
86
  model = AutoModelForCausalLM.from_pretrained(
 
17
  - agentbench
18
  - alfworld
19
  - dbbench
 
20
  ---
21
 
22
  # qwen3-4b-agentbench-dbalf-lora
 
48
  - Mixing ratio (pre-merge target): **DB:ALF = 1:1**
49
 
50
  ### DB Oversampling (category-aware)
51
+ Enabled: **False**
52
 
53
  DB category weights used during training-data preparation:
54
 
55
+ - counting: 1
56
+ - comparison: 1
57
+ - ranking: 1
58
  - select: 1
59
  - insert: 1
60
  - update: 1
 
64
 
65
  - Base model: Qwen/Qwen3-4B-Instruct-2507
66
  - Method: LoRA (full precision base)
67
+ - Max sequence length: 4096
68
+ - Epochs: 1
69
+ - Learning rate: 2e-04
70
+ - LoRA: r=32, alpha=64, dropout=0.05
71
  - Per-device train batch size: 2
72
  - Gradient accumulation: 4
73
 
 
79
  import torch
80
 
81
  base = "Qwen/Qwen3-4B-Instruct-2507"
82
+ adapter = "your_id/your-model-name"
83
 
84
  tokenizer = AutoTokenizer.from_pretrained(base)
85
  model = AutoModelForCausalLM.from_pretrained(
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ce2ce8c52da46bf7471c0d2eef812a191d5be71b1f82e2539a20dafd00148e4f
3
  size 4967215360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b4f9428ff718d9cb13a3c2edc84fc40c13618001bd765b1ddb70cd26dc7b465
3
  size 4967215360
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:353c3f776c17902ac161c9d1ab001e4929ffcffd4c9626ce12a4a4362de0ac7f
3
  size 3077766632
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f556fb2a665d7393d46117557f24b9ce0ff12b3f0577e160b9de892b3f11973
3
  size 3077766632