DatPySci commited on
Commit
d8844aa
·
verified ·
1 Parent(s): 1843f32

Upload folder using huggingface_hub

Browse files
models/Qwen2.5-Math-1.5B-omega-GRPO-step300/config.json CHANGED
@@ -3,7 +3,7 @@
3
  "Qwen2ForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
6
- "dtype": "float32",
7
  "eos_token_id": 151643,
8
  "hidden_act": "silu",
9
  "hidden_size": 1536,
 
3
  "Qwen2ForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
6
+ "dtype": "bfloat16",
7
  "eos_token_id": 151643,
8
  "hidden_act": "silu",
9
  "hidden_size": 1536,
models/Qwen2.5-Math-1.5B-omega-GRPO-step300/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a75d4f2b64c47c6964f910afc722fbb7e1e9fe44ac3cfeb48fc40f15f8a92c58
3
+ size 3554214752