Edit model card

miiqu-105b-v1.0

Developed by Infinimol AI GmbH

Also Available:

8th place on EQ-Bench, beating Qwen1.5-72B-Chat, miqudev/miqu-1-70b, mistral-medium and claude-3-sonnet-20240229. All without fine-tuning or additional training.

Thanks for support from: turboderp, silphendio, sqrkl, and ngxson!

❗ Q4_K_M files are split and require joining

Note: HF does not support uploading files larger than 50GB. The Q4_K_M files are supplied as split files.

Click for instructions regarding Q4_K_M files

Process

Please download:

  • miiqu.gguf-split-aa
  • miiqu.gguf-split-ab
  • miiqu.gguf-split-ac
  • miiqu.gguf-split-ad
  • miiqu.gguf-split-ae
  • miiqu.gguf-split-af

To join the files, do the following:

Linux and macOS:

cat miiqu.gguf-split-a* > miiqu_Q4_K_M.gguf && rm miiqu.gguf-split-a*

Windows command line:

COPY /B miiqu.gguf-split-aa + miiqu.gguf-split-ab + miiqu.gguf-split-ac + miiqu.gguf-split-ad + miiqu.gguf-split-ae + miiqu.gguf-split-af miiqu_Q4_K_M.gguf
DEL miiqu.gguf-split-aa miiqu.gguf-split-ab miiqu.gguf-split-ac miiqu.gguf-split-ad miiqu.gguf-split-ae miiqu.gguf-split-af

Model Details

  • Max Context: 32768 tokens
  • Layers: 105

Prompt template: ChatML or Mistral

chatml:

<|im_start|><|user|>\n<|user-message|><|im_end|>\n<|im_start|><|bot|>\n<|bot-message|><|im_end|>\n

mistral:

[INST] <|user|><|user-message|>[/INST]<|bot|><|bot-message|></s>
Downloads last month
0
Unable to determine this model’s pipeline type. Check the docs .