Edit model card

小雨同學 2x7B

採用聯發科 Breeze 7B Instruct 為基底的國語 MoE (Mixture-of-Experts) 模型,共有兩個 Expert model。

請用 Marcoro14-7B 或是 Breeze-7B-Instruct 所推薦的 Prompt 格式進行操作;以下為模型配置。

  • v0.2 更新了 tokenizer parameters


This is an experimental Mixtral-architecture MoE model with 2 of 7B sized fine-tunes. Breeze and CodeNinja are used on top of Marcoro14-7B-slerp.

Model configuration is as follows:

To use the model, please use either prompt templates suggested by the base models.


Please evaluate before use in any application pipeline. Activation for coding part of the model would be 'code', 'python', 'typescript', 'javascript', 'programming', 'algorithm'.

Downloads last month
Model size
12.9B params
Tensor type
Inference Examples
Inference API (serverless) has been turned off for this model.