Merge:

[Xwin (0.66) + ReMM (0.33)] x [Xwin (0.33) + MLewd (0.66)]

The goal was to recreate https://huggingface.co/Undi95/MXLewd-L2-20B-GGUF in 13B without using merge interlacing (will probably be a little less good).

Models used

  • Undi95/MLewd-L2-13B-v2-3
  • Undi95/ReMM-v2.1-L2-13B
  • Xwin-LM/Xwin-LM-13B-V0.1

One part is ReMM (0.33) and Xwin (0.66)

One part is Xwin (0.33) and MLewd (0.66)

Prompt template: Alpaca

Below is an instruction that describes a task. Write a response that completes the request.

### Instruction:
{prompt}

### Response:
Downloads last month
76
GGUF
Model size
13B params
Architecture
llama

4-bit

5-bit

6-bit

8-bit

Inference API
Unable to determine this model's library. Check the docs .