Edit model card

Hermes-2-Pro-Llama-3-14B

This is the GGUF version of a naive (passthrough) self-merge of two NousResearch/Hermes-2-Pro-Llama-3-8B, using mergekit. The merge config was inpired by mlabonne/Meta-Llama-3-120B-Instruct.

Results

This gives better (and actually usable) results compared to my previous merges. I can clearly see a rise in capability as it responded with a more thought out answer to my function calling prompt, but failed to adhere to the format and instructions fully. Will read more and look at other merge methods. This is also the first GGUF that I converted myself which also might have contributed to the increase in quality.

Conversion Command

Using the convert-hf-to-gguf.py provided in llama.cpp:

python convert-hf-to-gguf.py --outfile hermes2-pro-llama3-14b-gguf --outtype f16 --use-temp-file --model-name Hermes2-Pro-Llama3-14b hermes2-pro-llama3-14b

Merge Details

Merge Method

This model was merged using the passthrough merge method.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

slices:
- sources:
  - layer_range: [0, 8]
    model: NousResearch/Hermes-2-Pro-Llama-3-8B
- sources:
  - layer_range: [4, 12]
    model: NousResearch/Hermes-2-Pro-Llama-3-8B
- sources:
  - layer_range: [8, 16]
    model: NousResearch/Hermes-2-Pro-Llama-3-8B
- sources:
  - layer_range: [12, 20]
    model: NousResearch/Hermes-2-Pro-Llama-3-8B
- sources:
  - layer_range: [16, 24]
    model: NousResearch/Hermes-2-Pro-Llama-3-8B
- sources:
  - layer_range: [20, 28]
    model: NousResearch/Hermes-2-Pro-Llama-3-8B
- sources:
  - layer_range: [24, 32]
    model: NousResearch/Hermes-2-Pro-Llama-3-8B
merge_method: passthrough
dtype: float16
Downloads last month

-

Downloads are not tracked for this model. How to track
Unable to determine this model’s pipeline type. Check the docs .

Finetuned from

Collection including AtakanTekparmak/hermes2-pro-llama3-14b-GGUF