Edit model card

This is an ExLlamaV2 quantized model in 3bpw of jebcarter/Psyfighter-13B using PIPPA as the calibration dataset. (I might reupload this or add a new branch with a different version that uses the default calibration dataset instead)

Alpaca and ChatML prompt templates seem to work fine with this model.

Original Model card

merge_method: task_arithmaetic
base_model: TheBloke/Llama-2-13B-fp16
models:
  - model: TheBloke/Llama-2-13B-fp16
  - model: KoboldAI/LLaMA2-13B-Tiefighter
    parameters:
      weight: 1.0
  - model: chaoyi-wu/MedLLaMA_13B
    parameters:
      weight: 0.01
  - model: Doctor-Shotgun/llama-2-13b-chat-limarp-v2-merged
    parameters:
      weight: 0.02
dtype: float16

This model was made possible thanks to the Compute provided by the KoboldAI community.

Downloads last month
1

Collection including mpasila/Psyfighter-13B-exl2-3bpw