Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Quantization made by Richard Erkhov.

Github

Discord

Request more models

Monsoon-7B-exp-1 - GGUF

Original model description:

inference: false language: - zh - en license: unknown model_name: Monsoon-7B-exp-1 pipeline_tag: text-generation prompt_template: ' SYS_PROMPT [INST] QUERY1 [/INST] RESPONSE1 [INST] QUERY2 [/INST]' tags: - nlp - chinese - mistral - mixtral - traditional_chinese - merge - mergekit - MediaTek-Research/Breeze-7B-Instruct-v0_1 - SanjiWatsuki/Silicon-Maid-7B


雨季 7B exp-1

Breeze 7B Instruct 與 Silicon-Maid-7B (角扮用) 的 dare-ties merge 試驗性模型。

請用 Silicon-Maid-7B 或是 Breeze-7B-Instruct 所推薦的 Prompt 格式進行操作;以下為模型配置。

Monsoon 7B exp-1

This is an experimental Mistral-architecture DARE-TIES merge model of 2x 7B sized fine-tunes. Breeze and Silicon Maid are used together.

Model configuration is as follows:

To use the model, please use either prompt templates suggested by the base models.

Merge Method

This model was merged using the DARE TIES merge method.

Downloads last month
256
GGUF
Model size
7.49B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference API
Unable to determine this model's library. Check the docs .