wolfram's picture
Upload folder using huggingface_hub
c39107e verified
|
raw
history blame
4.08 kB
metadata
base_model:
  - 152334H/miqu-1-70b-sf
  - lizpreciatior/lzlv_70b_fp16_hf
language:
  - en
  - de
  - fr
  - es
  - it
library_name: transformers
tags:
  - mergekit
  - merge

miquliz-120b-v2.0

image/jpeg

This is v2.0 of a 120b frankenmerge created by interleaving layers of miqu-1-70b-sf with lzlv_70b_fp16_hf using mergekit.

Inspired by goliath-120b.

Thanks for the support, CopilotKit - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.

Special thanks to a kind soul, who wishes to remain anonymous, for donating the improved recipe for v2.0.

Prompt template: Mistral

<s>[INST] {prompt} [/INST]

See also: 🐺🐦‍⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with 17 different instruct templates : LocalLLaMA

Model Details

  • Max Context: 32768 tokens
  • Layers: 140

Merge Details

Merge Method

This model was merged using the linear merge method.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

merge_method: linear
parameters:
  weight: 1.0
slices:
  - sources:
      - model: 152334H/miqu-1-70b-sf
        layer_range: [0, 1]
      - model: lizpreciatior/lzlv_70b_fp16_hf
        layer_range: [0, 1]
        parameters:
          weight: 0
  - sources:
      - model: 152334H/miqu-1-70b-sf
        layer_range: [1, 20]
  - sources:
      - model: lizpreciatior/lzlv_70b_fp16_hf
        layer_range: [10, 30]
  - sources:
      - model: 152334H/miqu-1-70b-sf
        layer_range: [20, 40]
  - sources:
      - model: lizpreciatior/lzlv_70b_fp16_hf
        layer_range: [30, 50]
  - sources:
      - model: 152334H/miqu-1-70b-sf
        layer_range: [40, 60]
  - sources:
      - model: lizpreciatior/lzlv_70b_fp16_hf
        layer_range: [50, 70]
  - sources:
      - model: 152334H/miqu-1-70b-sf
        layer_range: [60, 79]
  - sources:
      - model: 152334H/miqu-1-70b-sf
        layer_range: [79, 80]
      - model: lizpreciatior/lzlv_70b_fp16_hf
        layer_range: [79, 80]
        parameters:
          weight: 0
dtype: float16
tokenizer_source: model:152334H/miqu-1-70b-sf

Credits & Special Thanks

Support

  • My Ko-fi page if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!

DISCLAIMER: THIS IS BASED ON A LEAKED ASSET AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK.