Edit model card

final_merge_medical

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the DARE TIES merge method using ./evol_merge_storage_medical/input_models/Llama-2-7b-hf_1207016536 as a base.

Models Merged

The following models were included in the merge:

  • ./evol_merge_storage_medical/input_models/meditron-7b_535713315

Configuration

The following YAML configuration was used to produce this model:

base_model: ./evol_merge_storage_medical/input_models/Llama-2-7b-hf_1207016536
dtype: bfloat16
merge_method: dare_ties
parameters:
  int8_mask: 1.0
  normalize: 1.0
slices:
- sources:
  - layer_range: [0, 4]
    model: ./evol_merge_storage_medical/input_models/meditron-7b_535713315
    parameters:
      density: 0.992384363500622
      weight: 0.12828024613059025
  - layer_range: [0, 4]
    model: ./evol_merge_storage_medical/input_models/Llama-2-7b-hf_1207016536
    parameters:
      density: 1.0
      weight: 0.2043801834581346
- sources:
  - layer_range: [4, 8]
    model: ./evol_merge_storage_medical/input_models/meditron-7b_535713315
    parameters:
      density: 0.8166617869585223
      weight: 0.5073401926353303
  - layer_range: [4, 8]
    model: ./evol_merge_storage_medical/input_models/Llama-2-7b-hf_1207016536
    parameters:
      density: 0.9502278365422492
      weight: 0.9538077350045424
- sources:
  - layer_range: [8, 12]
    model: ./evol_merge_storage_medical/input_models/meditron-7b_535713315
    parameters:
      density: 0.9729859873278464
      weight: 0.24983126895812832
  - layer_range: [8, 12]
    model: ./evol_merge_storage_medical/input_models/Llama-2-7b-hf_1207016536
    parameters:
      density: 1.0
      weight: 0.36977171041903323
- sources:
  - layer_range: [12, 16]
    model: ./evol_merge_storage_medical/input_models/meditron-7b_535713315
    parameters:
      density: 0.942566643987019
      weight: 0.25604084632369556
  - layer_range: [12, 16]
    model: ./evol_merge_storage_medical/input_models/Llama-2-7b-hf_1207016536
    parameters:
      density: 1.0
      weight: 0.3479493989699756
- sources:
  - layer_range: [16, 20]
    model: ./evol_merge_storage_medical/input_models/meditron-7b_535713315
    parameters:
      density: 0.9937376494167793
      weight: 0.18395490495846406
  - layer_range: [16, 20]
    model: ./evol_merge_storage_medical/input_models/Llama-2-7b-hf_1207016536
    parameters:
      density: 1.0
      weight: 0.11572624333070619
- sources:
  - layer_range: [20, 24]
    model: ./evol_merge_storage_medical/input_models/meditron-7b_535713315
    parameters:
      density: 0.784005177941743
      weight: 0.5101432868743351
  - layer_range: [20, 24]
    model: ./evol_merge_storage_medical/input_models/Llama-2-7b-hf_1207016536
    parameters:
      density: 0.7983003243622248
      weight: 0.6687460577953303
- sources:
  - layer_range: [24, 28]
    model: ./evol_merge_storage_medical/input_models/meditron-7b_535713315
    parameters:
      density: 0.5794040987288581
      weight: 0.5522697341347537
  - layer_range: [24, 28]
    model: ./evol_merge_storage_medical/input_models/Llama-2-7b-hf_1207016536
    parameters:
      density: 1.0
      weight: 0.8057676781293861
- sources:
  - layer_range: [28, 32]
    model: ./evol_merge_storage_medical/input_models/meditron-7b_535713315
    parameters:
      density: 0.960066282337993
      weight: 0.6270458624800975
  - layer_range: [28, 32]
    model: ./evol_merge_storage_medical/input_models/Llama-2-7b-hf_1207016536
    parameters:
      density: 0.9877495522716813
      weight: 0.15281690912944845
Downloads last month
11
Safetensors
Model size
6.74B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.