Alvar10 commited on
Commit
58145e6
·
verified ·
1 Parent(s): 3830d20

Upload GRPO LoRA adapter for Qwen3-4B (non-thinking, math)

Browse files
Files changed (2) hide show
  1. adapter_config.json +3 -3
  2. adapter_model.safetensors +1 -1
adapter_config.json CHANGED
@@ -34,12 +34,12 @@
34
  "rank_pattern": {},
35
  "revision": null,
36
  "target_modules": [
 
 
37
  "v_proj",
 
38
  "gate_proj",
39
- "q_proj",
40
  "k_proj",
41
- "o_proj",
42
- "up_proj",
43
  "down_proj"
44
  ],
45
  "target_parameters": null,
 
34
  "rank_pattern": {},
35
  "revision": null,
36
  "target_modules": [
37
+ "o_proj",
38
+ "q_proj",
39
  "v_proj",
40
+ "up_proj",
41
  "gate_proj",
 
42
  "k_proj",
 
 
43
  "down_proj"
44
  ],
45
  "target_parameters": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a974d0ca375868e8163363e6ffe0f7864646df5eaac1c89a9f7a93620f4db23
3
  size 66127776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0f97299338b5c54b4adce7e0b4fc986874da3c2c171054e4445461069b5c298
3
  size 66127776