Welly-code commited on
Commit
0589a8e
·
verified ·
1 Parent(s): fb176f8

Delete README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +0 -173
README.md DELETED
@@ -1,173 +0,0 @@
1
- ---
2
- base_model: Qwen/Qwen2.5-Coder-3B-Instruct
3
- datasets:
4
- - nvidia/Nemotron-Agentic-v1
5
- - my-ai-stack/Stack-4.0-Dataset
6
- pipeline_tag: text-generation
7
- license: apache-2.0
8
- tags:
9
- - code-generation
10
- - agentic-ai
11
- - tool-use
12
- - lora
13
- - qwen
14
- - python
15
- - coding-assistant
16
- - transformers
17
- - peft
18
- - 3b-parameter-model
19
- model_index:
20
- - name: Stack X Ultimate
21
- results:
22
- - task:
23
- type: text-generation
24
- description: Agentic coding assistant with tool-use capabilities
25
- dataset:
26
- type: openai/openai_humaneval
27
- name: HumanEval
28
- metrics:
29
- - type: pass@1
30
- value: TBD
31
- ---
32
-
33
- # Stack X Ultimate
34
-
35
- **A state-of-the-art agentic coding model built on Qwen2.5-Coder-3B-Instruct**
36
-
37
- Stack X is a LoRA adapter trained on a curated mix of real agentic conversations, designed to make open-weight models better at multi-step tool use, code generation, and complex reasoning tasks.
38
-
39
- ---
40
-
41
- ## Model Details
42
-
43
- - **Base Model:** Qwen/Qwen2.5-Coder-3B-Instruct
44
- - **Architecture:** Transformer (3B parameters)
45
- - **Training Type:** QLoRA (LoRA rank 32, 7 modules targeted)
46
- - **Trained by:** Walid Sobhie via OpenClaw agentic pipeline
47
- - **Framework:** Hugging Face Transformers + PEFT + PyTorch bf16
48
- - **Training Hardware:** NVIDIA V100-SXM2-16GB (GCP spot instance)
49
- - **Training Steps:** 3,000 steps (curriculum sorted, cosine LR decay)
50
- - **Effective Batch Size:** 16 (gradient accumulation)
51
- - **Max Context:** 1,536 tokens
52
-
53
- ---
54
-
55
- ## Training Data
56
-
57
- | Source | Description | Count |
58
- |--------|-------------|-------|
59
- | NVIDIA Nemotron Agentic | Real multi-step tool calling conversations | ~7,000 |
60
- | Stack-4.0 Smart | High-complexity agentic tasks | ~10,000 |
61
- | Stack-4.0 Tools | Diverse tool-use patterns | ~10,000 |
62
- | **Total (deduped)** | **After deduplication** | **~6,100** |
63
-
64
- Training data was filtered, deduplicated, and sorted by complexity (curriculum learning) before training.
65
-
66
- ---
67
-
68
- ## Capabilities
69
-
70
- Stack X is designed to excel at:
71
-
72
- - **Multi-step tool use** — chains multiple tool calls with proper reasoning
73
- - **Code generation** — Python, JavaScript, shell, and more
74
- - **Debugging** — finds and explains bugs with fixes
75
- - **Math & reasoning** — step-by-step calculation and problem solving
76
- - **Research tasks** — information retrieval and synthesis
77
-
78
- ---
79
-
80
- ## Usage
81
-
82
- ### With PEFT (recommended — preserves base model)
83
-
84
- ```python
85
- from transformers import AutoTokenizer, AutoModelForCausalLM
86
- from peft import PeftModel
87
-
88
- BASE = "Qwen/Qwen2.5-Coder-3B-Instruct"
89
- ADAPTER = "my-ai-stack/Stack-X-Ultimate"
90
-
91
- tokenizer = AutoTokenizer.from_pretrained(BASE, trust_remote_code=True)
92
- tokenizer.pad_token = tokenizer.eos_token
93
-
94
- base = AutoModelForCausalLM.from_pretrained(BASE, torch_dtype="bfloat16", device_map="auto")
95
- model = PeftModel.from_pretrained(base, ADAPTER)
96
-
97
- # Chat
98
- messages = [{"role": "user", "content": "Use the calculate tool to find sqrt(144)"}]
99
- text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
100
- inputs = tokenizer(text, return_tensors="pt").to(model.device)
101
- outputs = model.generate(**inputs, max_new_tokens=256)
102
- print(tokenizer.decode(outputs[0], skip_special_tokens=True))
103
- ```
104
-
105
- ### Merged (full model)
106
-
107
- ```python
108
- # See: my-ai-stack/Stack-X-Ultimate-Merged
109
- from transformers import AutoTokenizer, AutoModelForCausalLM
110
-
111
- model = AutoModelForCausalLM.from_pretrained("my-ai-stack/Stack-X-Ultimate-Merged", torch_dtype="bfloat16", device_map="auto")
112
- tokenizer = AutoTokenizer.from_pretrained("my-ai-stack/Stack-X-Ultimate-Merged")
113
- ```
114
-
115
- ---
116
-
117
- ## Performance
118
-
119
- | Benchmark | Score |
120
- |-----------|-------|
121
- | HumanEval (0-shot) | TBD |
122
- | Agentic tool call | TBD |
123
- | Reasoning (commonsense) | TBD |
124
-
125
- *Evaluation results will be posted after training completes.*
126
-
127
- ---
128
-
129
- ## Limitations
130
-
131
- - LoRA adapter requires compatible base model (Qwen2.5-Coder-3B-Instruct)
132
- - Max context 1,536 tokens — not suitable for very long documents
133
- - Trained primarily in English — other language performance may vary
134
- - Tool use limited to the patterns seen in training data
135
-
136
- ---
137
-
138
- ## Training Recipe
139
-
140
- ```
141
- Base model: Qwen/Qwen2.5-Coder-3B-Instruct
142
- LoRA rank: 32 (59M trainable params)
143
- LoRA alpha: 64
144
- Target modules: q_proj, k_proj, v_proj, o_proj,
145
- gate_proj, up_proj, down_proj
146
- Learning rate: 2e-4 (cosine decay)
147
- Warmup: 150 steps
148
- Batch size: 1 × gradient_accumulation=16
149
- Optimizer: AdamW (bf16)
150
- Max grad norm: 0.5
151
- Weight decay: 0.1
152
- Mixed precision: bf16
153
- Gradient checkpointing: enabled
154
- ```
155
-
156
- ---
157
-
158
- ## Citation
159
-
160
- ```bibtex
161
- @misc{stackx2026,
162
- title={Stack X Ultimate},
163
- author={Walid Sobhie},
164
- year={2026},
165
- url={https://huggingface.co/my-ai-stack/Stack-X-Ultimate}
166
- }
167
- ```
168
-
169
- ---
170
-
171
- ## Disclaimer
172
-
173
- This model is provided as-is. Training was performed automatically via an OpenClaw agentic pipeline. Results may vary. Not reviewed for safety in production deployments.