AxionLab-official commited on
Commit
f81f042
ยท
verified ยท
1 Parent(s): 9694ce5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +145 -86
README.md CHANGED
@@ -7,87 +7,124 @@ base_model:
7
  - AxionLab-official/MiniBot-0.9M-Base
8
  ---
9
 
10
- ## ๐Ÿง  MiniBot-0.9M-Instruct
11
 
12
- Instruction-tuned GPT-2 style language model (~900K parameters) optimized for Portuguese conversational tasks.
13
 
14
- ## ๐Ÿ“Œ Model Overview
 
 
 
15
 
16
- MiniBot-0.9M-Instruct is an instruction-tuned version of MiniBot-0.9M-Base, designed to better follow prompts, respond to user inputs, and generate more coherent conversational outputs in Portuguese.
 
 
 
 
17
 
18
  Built on a GPT-2 architecture (~0.9M parameters), this model was fine-tuned on conversational and instruction-style data to improve usability in real-world interactions.
19
 
20
- ๐ŸŽฏ Key Characteristics
21
- ๐Ÿ‡ง๐Ÿ‡ท Language: Portuguese (primary)
22
- ๐Ÿง  Architecture: GPT-2 style (decoder-only Transformer)
23
- ๐Ÿ”ค Embeddings: GPT-2 compatible
24
- ๐Ÿ“‰ Parameters: ~900K
25
- โš™๏ธ Base Model: MiniBot-0.9M-Base
26
- ๐ŸŽฏ Fine-tuning: Instruction tuning (supervised)
27
- โœ… Alignment: Basic prompt-following behavior
28
- ๐Ÿง  What Changed from Base?
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
 
30
- Compared to the base model:
31
 
32
- Feature Base Instruct
33
- Prompt understanding โŒ โœ…
34
- Conversational flow โš ๏ธ โœ…
35
- Instruction following โŒ โœ…
36
- Coherence Baixa Melhorada
37
- Usability Experimental Practical
38
 
39
- ๐Ÿ‘‰ The model is now significantly more usable in chat scenarios.
 
 
 
40
 
41
- ๐Ÿ—๏ธ Architecture
42
 
43
- Same core as base:
44
 
45
- Decoder-only Transformer (GPT-2 style)
46
- Token + positional embeddings
47
- Self-attention + MLP blocks
48
- Autoregressive generation
49
 
50
- No architectural changes โ€” only behavioral improvement via fine-tuning.
51
 
52
- ๐Ÿ“š Fine-Tuning
53
- Dataset
 
 
 
54
 
55
- The model was fine-tuned on a Portuguese instruction-style conversational dataset, including:
56
 
57
- Perguntas e respostas
58
- Instruรงรตes simples
59
- Chat estilo assistente
60
- Roleplay bรกsico
61
- Conversas naturais
62
- Format
63
  User: Me explique o que รฉ gravidade
64
  Bot: A gravidade รฉ a forรงa que atrai objetos com massa...
65
- Strategy
66
- Supervised fine-tuning (SFT)
67
- Pattern learning for instruction-following
68
- No RLHF or preference optimization
69
- ๐Ÿ’ก Capabilities
70
 
71
- โœ… Strengths:
 
 
 
72
 
73
- Seguir instruรงรตes simples
74
- Responder perguntas bรกsicas
75
- Conversar de forma mais natural
76
- Melhor coerรชncia em respostas curtas
77
- Estrutura de diรกlogo mais consistente
78
 
79
- โŒ Limitations:
 
 
 
 
80
 
81
- Raciocรญnio ainda limitado
82
- Pode errar fatos
83
- Nรฃo mantรฉm contexto longo
84
- Sensรญvel a prompts mal estruturados
85
 
86
- ๐Ÿ‘‰ Mesmo com instruct tuning, ainda รฉ um modelo extremamente pequeno.
 
 
 
87
 
88
- ๐Ÿš€ Usage
89
- Hugging Face Transformers
90
- ```Python
 
 
 
 
 
 
 
 
 
 
 
 
91
  from transformers import AutoTokenizer, AutoModelForCausalLM
92
 
93
  model_name = "AxionLab-official/MiniBot-0.9M-Instruct"
@@ -103,47 +140,69 @@ outputs = model.generate(
103
  max_new_tokens=80,
104
  temperature=0.7,
105
  top_p=0.9,
106
- do_sample=True
107
  )
108
 
109
  print(tokenizer.decode(outputs[0], skip_special_tokens=True))
110
  ```
111
 
112
- โš™๏ธ Recommended Settings
113
 
114
- Para melhor qualidade:
 
 
 
 
 
115
 
116
- temperature: 0.6 โ€“ 0.8
117
- top_p: 0.85 โ€“ 0.95
118
- do_sample: True
119
- max_new_tokens: 40 โ€“ 100
120
 
121
- ๐Ÿ‘‰ Instruct models tendem a performar melhor com menos aleatoriedade.
 
 
122
 
123
- ๐Ÿงช Intended Use
124
- ๐Ÿ’ฌ Chatbots leves em portuguรชs
125
- ๐ŸŽฎ NPCs e jogos
126
- ๐Ÿง  Testes de fine-tuning
127
- ๐Ÿ“š Educaรงรฃo em NLP
128
- โšก Aplicaรงรตes locais (CPU-only)
129
- โš ๏ธ Limitations
 
130
 
131
- Modelo extremamente pequeno
132
- Sem alinhamento robusto
133
- Pode gerar respostas incorretas
134
- Nรฃo adequado para produรงรฃo crรญtica
135
 
136
- ๐Ÿ”ฎ Future Work
137
- ๐Ÿง  Reasoning-tuned version (MiniBot-Reason)
138
- ๐Ÿ“ˆ Scaling para 1Mโ€“10M parรขmetros
139
- ๐Ÿ“š Dataset mais diverso
140
- ๐Ÿค– Melhor alinhamento de respostas
141
- ๐Ÿงฉ Tool-use experiments
142
 
143
- ๐Ÿ“œ License
 
 
 
 
 
144
 
145
- MIT
146
 
147
- ๐Ÿ‘ค Author
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
148
 
149
- Developed by AxionLab
 
 
 
7
  - AxionLab-official/MiniBot-0.9M-Base
8
  ---
9
 
10
+ # ๐Ÿง  MiniBot-0.9M-Instruct
11
 
12
+ > **Instruction-tuned GPT-2 style language model (~900K parameters) optimized for Portuguese conversational tasks.**
13
 
14
+ [![Model](https://img.shields.io/badge/๐Ÿค—%20Hugging%20Face-MiniBot--0.9M--Instruct-yellow)](https://huggingface.co/AxionLab-official/MiniBot-0.9M-Instruct)
15
+ [![License](https://img.shields.io/badge/License-MIT-green.svg)](https://opensource.org/licenses/MIT)
16
+ [![Language](https://img.shields.io/badge/Language-Portuguese-blue)](https://huggingface.co/AxionLab-official/MiniBot-0.9M-Instruct)
17
+ [![Parameters](https://img.shields.io/badge/Parameters-~900K-orange)](https://huggingface.co/AxionLab-official/MiniBot-0.9M-Instruct)
18
 
19
+ ---
20
+
21
+ ## ๐Ÿ“Œ Overview
22
+
23
+ **MiniBot-0.9M-Instruct** is the instruction-tuned version of [MiniBot-0.9M-Base](https://huggingface.co/AxionLab-official/MiniBot-0.9M-Base), designed to follow prompts more accurately, respond to user inputs, and generate more coherent conversational outputs in **Portuguese**.
24
 
25
  Built on a GPT-2 architecture (~0.9M parameters), this model was fine-tuned on conversational and instruction-style data to improve usability in real-world interactions.
26
 
27
+ ---
28
+
29
+ ## ๐ŸŽฏ Key Characteristics
30
+
31
+ | Attribute | Detail |
32
+ |---|---|
33
+ | ๐Ÿ‡ง๐Ÿ‡ท **Language** | Portuguese (primary) |
34
+ | ๐Ÿง  **Architecture** | GPT-2 style (Transformer decoder-only) |
35
+ | ๐Ÿ”ค **Embeddings** | GPT-2 compatible |
36
+ | ๐Ÿ“‰ **Parameters** | ~900K |
37
+ | โš™๏ธ **Base Model** | MiniBot-0.9M-Base |
38
+ | ๐ŸŽฏ **Fine-tuning** | Instruction tuning (supervised) |
39
+ | โœ… **Alignment** | Basic prompt-following behavior |
40
+
41
+ ---
42
+
43
+ ## ๐Ÿง  What Changed from Base?
44
+
45
+ Instruction tuning introduced significant behavioral improvements with no architectural changes:
46
+
47
+ | Feature | Base | Instruct |
48
+ |---|---|---|
49
+ | Prompt understanding | โŒ | โœ… |
50
+ | Conversational flow | โš ๏ธ Partial | โœ… |
51
+ | Instruction following | โŒ | โœ… |
52
+ | Overall coherence | Low | Improved |
53
+ | Practical usability | Experimental | Functional |
54
 
55
+ > ๐Ÿ’ก The model is now significantly more usable in chat scenarios.
56
 
57
+ ---
58
+
59
+ ## ๐Ÿ—๏ธ Architecture
60
+
61
+ The core architecture remains identical to the base model:
 
62
 
63
+ - **Decoder-only Transformer** (GPT-2 style)
64
+ - Token embeddings + positional embeddings
65
+ - Self-attention + MLP blocks
66
+ - Autoregressive generation
67
 
68
+ No structural changes were made โ€” only behavioral improvement through fine-tuning.
69
 
70
+ ---
71
 
72
+ ## ๐Ÿ“š Fine-Tuning Dataset
 
 
 
73
 
74
+ The model was fine-tuned on a Portuguese instruction-style conversational dataset composed of:
75
 
76
+ - ๐Ÿ’ฌ Questions and answers
77
+ - ๐Ÿ“‹ Simple instructions
78
+ - ๐Ÿค– Assistant-style chat
79
+ - ๐ŸŽญ Basic roleplay
80
+ - ๐Ÿ—ฃ๏ธ Natural conversations
81
 
82
+ **Expected format:**
83
 
84
+ ```
 
 
 
 
 
85
  User: Me explique o que รฉ gravidade
86
  Bot: A gravidade รฉ a forรงa que atrai objetos com massa...
87
+ ```
 
 
 
 
88
 
89
+ **Training strategy:**
90
+ - Supervised Fine-Tuning (SFT)
91
+ - Pattern learning for instruction-following
92
+ - No RLHF or preference optimization
93
 
94
+ ---
95
+
96
+ ## ๐Ÿ’ก Capabilities
97
+
98
+ ### โœ… Strengths
99
 
100
+ - Following simple instructions
101
+ - Answering basic questions
102
+ - Conversing more naturally
103
+ - Higher coherence in short responses
104
+ - More consistent dialogue structure
105
 
106
+ ### โŒ Limitations
 
 
 
107
 
108
+ - Reasoning is still limited
109
+ - May generate incorrect facts
110
+ - Does not retain long context
111
+ - Sensitive to poorly structured prompts
112
 
113
+ > โš ๏ธ Even with instruction tuning, this remains an extremely small model. Adjust expectations accordingly.
114
+
115
+ ---
116
+
117
+ ## ๐Ÿš€ Getting Started
118
+
119
+ ### Installation
120
+
121
+ ```bash
122
+ pip install transformers torch
123
+ ```
124
+
125
+ ### Usage with Hugging Face Transformers
126
+
127
+ ```python
128
  from transformers import AutoTokenizer, AutoModelForCausalLM
129
 
130
  model_name = "AxionLab-official/MiniBot-0.9M-Instruct"
 
140
  max_new_tokens=80,
141
  temperature=0.7,
142
  top_p=0.9,
143
+ do_sample=True,
144
  )
145
 
146
  print(tokenizer.decode(outputs[0], skip_special_tokens=True))
147
  ```
148
 
149
+ ### โš™๏ธ Recommended Settings
150
 
151
+ | Parameter | Recommended Value | Description |
152
+ |---|---|---|
153
+ | `temperature` | `0.6 โ€“ 0.8` | Controls randomness |
154
+ | `top_p` | `0.85 โ€“ 0.95` | Nucleus sampling |
155
+ | `do_sample` | `True` | Enable sampling |
156
+ | `max_new_tokens` | `40 โ€“ 100` | Response length |
157
 
158
+ > ๐Ÿ’ก Instruct models tend to perform better at lower temperatures. Try values around `0.65` for more accurate and focused responses.
 
 
 
159
 
160
+ ---
161
+
162
+ ## ๐Ÿงช Intended Use Cases
163
 
164
+ | Use Case | Suitability |
165
+ |---|---|
166
+ | ๐Ÿ’ฌ Lightweight Portuguese chatbots | โœ… Ideal |
167
+ | ๐ŸŽฎ NPCs and games | โœ… Ideal |
168
+ | ๐Ÿง  Fine-tuning experiments | โœ… Ideal |
169
+ | ๐Ÿ“š NLP education | โœ… Ideal |
170
+ | โšก Local / CPU-only applications | โœ… Ideal |
171
+ | ๐Ÿญ Critical production environments | โŒ Not recommended |
172
 
173
+ ---
 
 
 
174
 
175
+ ## โš ๏ธ Disclaimer
 
 
 
 
 
176
 
177
+ - Extremely small model (~900K parameters)
178
+ - No robust alignment (no RLHF)
179
+ - May generate incorrect or nonsensical responses
180
+ - **Not suitable for critical production environments**
181
+
182
+ ---
183
 
184
+ ## ๐Ÿ”ฎ Future Work
185
 
186
+ - [ ] ๐Ÿง  Reasoning-tuned version (`MiniBot-Reason`)
187
+ - [ ] ๐Ÿ“ˆ Scaling to 1Mโ€“10M parameters
188
+ - [ ] ๐Ÿ“š Larger and more diverse dataset
189
+ - [ ] ๐Ÿค– Improved response alignment
190
+ - [ ] ๐Ÿงฉ Tool-use experiments
191
+
192
+ ---
193
+
194
+ ## ๐Ÿ“œ License
195
+
196
+ Distributed under the **MIT License**. See [`LICENSE`](LICENSE) for more details.
197
+
198
+ ---
199
+
200
+ ## ๐Ÿ‘ค Author
201
+
202
+ Developed by **[AxionLab](https://huggingface.co/AxionLab-official)** ๐Ÿ”ฌ
203
+
204
+ ---
205
 
206
+ <div align="center">
207
+ <sub>MiniBot-0.9M-Instruct ยท AxionLab ยท MIT License</sub>
208
+ </div>