pedrodev2026 commited on
Commit
e6328b6
·
verified ·
1 Parent(s): bafa4aa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +86 -1
README.md CHANGED
@@ -9,4 +9,89 @@ tags:
9
  - coder
10
  - code
11
  - microcoder
12
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  - coder
10
  - code
11
  - microcoder
12
+ ---
13
+ # Microcoder 1.5B
14
+
15
+ **Microcoder 1.5B** is a code-focused language model fine-tuned from [Qwen 2.5 Coder 1.5B Instruct](https://huggingface.co/Qwen/Qwen2.5-Coder-1.5B-Instruct) using LoRA (Low-Rank Adaptation) on curated code datasets. It is designed for code generation, completion, and instruction-following tasks in a lightweight, efficient package.
16
+
17
+ ---
18
+
19
+ ## Model Details
20
+
21
+ | Property | Value |
22
+ |------------------|--------------------------------------------|
23
+ | **Base Model** | Qwen 2.5 Coder 1.5B Instruct |
24
+ | **Fine-tuning** | LoRA |
25
+ | **Parameters** | ~1.5B |
26
+ | **License** | BSD 3-Clause |
27
+ | **Language** | English (primary), multilingual code |
28
+ | **Task** | Code generation, completion, instruction following |
29
+
30
+ ---
31
+
32
+ ## Benchmarks
33
+
34
+ | Benchmark | Metric | Score |
35
+ |--------------------|----------|--------------|
36
+ | HumanEval | pass@1 | **59.15%** |
37
+
38
+ > HumanEval results were obtained using the model in **GGUF format** with **Q5_K_M quantization**. Results may vary slightly with other formats or quantization levels.
39
+
40
+ ---
41
+
42
+ ## Usage
43
+
44
+ > **Important:** You must use `apply_chat_template` when formatting inputs. Passing raw text directly to the tokenizer will produce incorrect results.
45
+
46
+ ```python
47
+ from transformers import AutoTokenizer, AutoModelForCausalLM
48
+
49
+ model_id = "your-org/microcoder-1.5b"
50
+
51
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
52
+ model = AutoModelForCausalLM.from_pretrained(model_id)
53
+
54
+ messages = [
55
+ {
56
+ "role": "user",
57
+ "content": "Write a Python function that returns the nth Fibonacci number."
58
+ }
59
+ ]
60
+
61
+ input_text = tokenizer.apply_chat_template(
62
+ messages,
63
+ tokenize=False,
64
+ add_generation_prompt=True
65
+ )
66
+
67
+ inputs = tokenizer(input_text, return_tensors="pt")
68
+ outputs = model.generate(**inputs, max_new_tokens=256)
69
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
70
+ ```
71
+
72
+ ---
73
+
74
+ ## Training Details
75
+
76
+ Microcoder 1.5B was fine-tuned using LoRA on top of Qwen 2.5 Coder 1.5B Instruct. The training focused on code-heavy datasets covering multiple programming languages and problem-solving scenarios, aiming to improve instruction-following and code correctness at a small model scale.
77
+
78
+ ---
79
+
80
+ ## Credits
81
+
82
+ - **Model credits** — see [`MODEL_CREDITS.md`](./MODEL_CREDITS.md)
83
+ - **Dataset credits** — see [`DATASET_CREDITS.md`](./DATASET_CREDITS.md)
84
+
85
+ ---
86
+
87
+ ## License
88
+
89
+ The Microcoder 1.5B model weights and associated code in this repository are released under the **BSD 3-Clause License**. See [`LICENSE`](./LICENSE) for details.
90
+
91
+ Note that the base model (Qwen 2.5 Coder 1.5B Instruct) and the datasets used for fine-tuning are subject to their own respective licenses, as detailed in the credit files above.
92
+
93
+ ---
94
+
95
+ ## Notice
96
+
97
+ The documentation files in this repository (including `README.md`, `MODEL_CREDITS.md`, `DATASET_CREDITS.md`, and other `.md` files) were generated with the assistance of an AI language model.