ArunkumarVR commited on
Commit
df506da
ยท
verified ยท
1 Parent(s): c89abb5

Update Model Card: Official Commercial Release

Browse files
Files changed (1) hide show
  1. README.md +34 -43
README.md CHANGED
@@ -4,49 +4,50 @@ language:
4
  - en
5
  pipeline_tag: text-generation
6
  tags:
7
- - qwen3
8
  - reasoning
9
- - long-context
10
- - distillation
11
- - math
12
  - enterprise
13
- - research
 
14
  base_model: Qwen/Qwen3-4B
 
15
  ---
16
 
17
  # DeepBrainz-R1-4B-16K
18
 
19
- **DeepBrainz-R1-4B-16K** is a high-performance reasoning model in the **DeepBrainz-R series**, designed for structured problem-solving, analysis, and enterprise research workflows.
20
 
21
- It is distilled from the **Qwen3-32B** teacher model into a compact **4B** architecture using **Online Policy Distillation (OPD)**, emphasizing reasoning quality and instruction robustness over a **16K context window**.
22
 
23
  ---
24
 
25
- ## Model Highlights
26
 
27
- - **4B Parameters**: Optimized balance of performance and inference cost.
28
- - **16K Context Length**: Capable of processing medium-to-long documents and reasoning chains.
29
- - **Distilled Precision**: Trained via NeMo-RL OPD from a **Qwen3-32B** teacher.
30
- - **Architecture**: Standard Qwen3 (Dense), optimized for modern GPU inference.
 
31
 
32
  ---
33
 
34
- ## Intended Use
35
 
36
- - **Complex Reasoning**: Multi-step math, logic puzzles, and code analysis.
37
- - **Agentic Workflows**: Reliable planning and tool use within 16K context.
38
- - **Research**: Investigating distillation scaling laws (32B $\to$ 4B).
39
- - **Efficient Deployment**: Fits easily on consumer GPUs and edge servers.
40
 
41
- *Note: This model is optimized for reasoning tasks. For general conversational chit-chat, we recommend applying a specific instruction template.*
42
 
43
  ---
44
 
45
- ## Usage
46
 
47
  ```python
48
  from transformers import AutoModelForCausalLM, AutoTokenizer
49
- import torch
50
 
51
  model_id = "DeepBrainz/DeepBrainz-R1-4B-16K"
52
 
@@ -57,46 +58,36 @@ model = AutoModelForCausalLM.from_pretrained(
57
  device_map="auto"
58
  )
59
 
60
- # Example: Math Reasoning
61
- prompt = "Solve step by step: If 3x + 7 = 22, what is x?"
62
  inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
63
 
64
- outputs = model.generate(
65
- **inputs,
66
- max_new_tokens=512,
67
- temperature=0.6,
68
- top_p=0.95,
69
- do_sample=True
70
- )
71
-
72
  print(tokenizer.decode(outputs[0], skip_special_tokens=True))
73
  ```
74
 
75
  ---
76
 
77
- ## Training Summary
78
 
79
- The model was produced using a **multi-stage optimization process** involving large-scale supervision and iterative refinement to improve reasoning quality and robustness.
80
 
81
- - **Teacher**: Qwen3-32B (Dense)
82
- - **Student**: Qwen3-4B
83
- - **Method**: Online Policy Distillation (OPD)
84
- - **Context**: 16,384 tokens
85
 
86
  ---
87
 
88
- ## Limitations
89
 
90
- Performance depends on task complexity and inference configuration. While significantly stronger than smaller models, it may still hallucinate on obscure facts compared to 30B+ models.
91
 
92
  ---
93
 
94
- ## License
95
 
96
- Apache 2.0
97
 
98
  ---
99
 
100
- ## About DeepBrainz
101
-
102
- DeepBrainz builds reasoning-first AI systems focused on efficiency, structure, and real-world problem-solving.
 
 
4
  - en
5
  pipeline_tag: text-generation
6
  tags:
7
+ - deepbrainz
8
  - reasoning
9
+ - mathematics
10
+ - code
 
11
  - enterprise
12
+ - 4b
13
+ - long-context
14
  base_model: Qwen/Qwen3-4B
15
+ library_name: transformers
16
  ---
17
 
18
  # DeepBrainz-R1-4B-16K
19
 
20
+ **DeepBrainz-R1-4B-16K** is a compact, high-performance reasoning model engineered by **DeepBrainz AI & Labs**. Designed for scalability and efficiency, it specializes in structured chain-of-thought reasoning, mathematical problem solving, and logical analysis.
21
 
22
+ This model is part of the **DeepBrainz-R1 Series**, built to deliver frontier-class reasoning capabilities in cost-effective parameter sizes.
23
 
24
  ---
25
 
26
+ ## ๐Ÿš€ Model Highlights
27
 
28
+ - **Parameter Count:** ~4B
29
+ - **Context Window:** 16,384 tokens
30
+ - **Specialization:** STEM Reasoning, Logic, Code Analysis
31
+ - **Architecture:** Optimized Dense Transformer (Qwen2.5/3 Compatible)
32
+ - **Deployment:** Ready for vLLM, TGI, and local inference
33
 
34
  ---
35
 
36
+ ## ๐ŸŽฏ Intended Use Cases
37
 
38
+ - **Agentic Workflows:** Reliability in multi-step planning tasks.
39
+ - **Math & Science:** Solving complex word problems and equations.
40
+ - **Code Generation:** Writing and debugging algorithms.
41
+ - **Structured Data Extraction:** Parsing and reasoning over unstructured text.
42
 
43
+ > **Note:** This is a base reasoning model. For conversational chat, we recommend using a specific instruct template or fine-tuning on your domain data.
44
 
45
  ---
46
 
47
+ ## ๐Ÿ’ป Usage
48
 
49
  ```python
50
  from transformers import AutoModelForCausalLM, AutoTokenizer
 
51
 
52
  model_id = "DeepBrainz/DeepBrainz-R1-4B-16K"
53
 
 
58
  device_map="auto"
59
  )
60
 
61
+ prompt = "Analyze the time complexity of the following algorithm:"
 
62
  inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
63
 
64
+ outputs = model.generate(**inputs, max_new_tokens=256)
 
 
 
 
 
 
 
65
  print(tokenizer.decode(outputs[0], skip_special_tokens=True))
66
  ```
67
 
68
  ---
69
 
70
+ ## ๐Ÿ—๏ธ Technical Summary
71
 
72
+ The model was produced using a **multi-stage optimization process** involving large-scale supervision and iterative refinement. It is designed to maximize reasoning quality while maintaining instruction robustness.
73
 
74
+ *Specific training methodologies and dataset compositions are proprietary.*
 
 
 
75
 
76
  ---
77
 
78
+ ## ๐Ÿ›ก๏ธ Limitations & Safety
79
 
80
+ While this model demonstrates strong reasoning capabilities, it may still produce inaccurate information ("hallucinations"). Users should implement appropriate guardrails for production deployments.
81
 
82
  ---
83
 
84
+ ## ๐Ÿ“œ License
85
 
86
+ This model is released under the **Apache 2.0** license, allowing for academic and commercial use.
87
 
88
  ---
89
 
90
+ <div align="center">
91
+ <b>DeepBrainz AI & Labs</b><br>
92
+ <i>Advancing General Intelligence through Scalable Reasoning</i>
93
+ </div>