Edit model card

hermes2-pro-llama3-16b

This is a self-merge of NousResearch/Hermes-2-Pro-Llama-3-8B mergekit. For the purposes of experimentation.

Results

Another fail, need to read more about merging and look at more examples, and maybe try more with base models then SFT?

Merge Details

Merge Method

This model was merged using the passthrough merge method.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

slices:
  - sources:
    - model: NousResearch/Hermes-2-Pro-Llama-3-8B
      layer_range: [0, 32]
  - sources:
    - model: NousResearch/Hermes-2-Pro-Llama-3-8B
      layer_range: [0, 32]
merge_method: passthrough
dtype: bfloat16
tokenizer_source: union
Downloads last month
1
Safetensors
Model size
15B params
Tensor type
BF16
·

Finetuned from

Collection including AtakanTekparmak/hermes2-pro-llama3-16b