khtsly commited on
Commit
25ae94c
·
verified ·
1 Parent(s): b5f2c93

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +28 -12
README.md CHANGED
@@ -1,22 +1,38 @@
1
  ---
 
 
 
 
 
2
  tags:
 
 
 
 
 
 
 
 
3
  - gguf
4
  - llama.cpp
5
- - unsloth
6
  - vision-language-model
 
7
  ---
8
 
9
- # Mini-Coder-GGUF : GGUF
 
 
 
 
 
 
 
 
10
 
11
- This model was finetuned and converted to GGUF format using [Unsloth](https://github.com/unslothai/unsloth).
 
 
12
 
13
- **Example usage**:
14
- - For text only LLMs: `llama-cli -hf khtsly/Mini-Coder-GGUF --jinja`
15
- - For multimodal models: `llama-mtmd-cli -hf khtsly/Mini-Coder-GGUF --jinja`
16
 
17
- ## Available Model files:
18
- - `Qwen3.5-9B.Q6_K.gguf`
19
- - `Qwen3.5-9B.Q8_0.gguf`
20
- - `Qwen3.5-9B.BF16-mmproj.gguf`
21
- This was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth)
22
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
1
  ---
2
+ language:
3
+ - en
4
+ - zh
5
+ license: apache-2.0
6
+ base_model: Qwen/Qwen3.5-9B
7
  tags:
8
+ - unsloth
9
+ - qwen
10
+ - qwen3.5
11
+ - reasoning
12
+ - chain-of-thought
13
+ - lora
14
+ - luau
15
+ - roblox
16
  - gguf
17
  - llama.cpp
 
18
  - vision-language-model
19
+ pipeline_tag: image-text-to-text
20
  ---
21
 
22
+ # Mini-Coder
23
+
24
+ Mini-Coder is build on top of Qwen3.5-9B model with Continual Pretraining (CPT), we feed ~500k high-quality curated luau samples to improves the luau coding tasks capability.
25
+
26
+ We also inject over 14k samples from open-source of claude 4.6 distillations with a fews additional samples for Supervised-Finetuning (SFT) to improves the model reasoning, We also see the average consumed tokens has drastically reduced.
27
+
28
+ It's fine-tuned efficiently using LoRA (16-bit) and rsLoRA with Rank (r) set to 64 and Alpha (α) set to 128, ensuring strong adaptation and retention of new complex logic, it were trained specifically to handle up to 32,768 (32k) tokens of maximum output (recommended).
29
+
30
+ # Uploaded finetuned model
31
 
32
+ - **Developed by:** khtsly
33
+ - **License:** apache-2.0
34
+ - **Finetuned from model :** khtsly/Coder-9B
35
 
36
+ This qwen3_5_text model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
 
 
37
 
38
+ [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)