Edit model card

Loyal-Silicon-Maid-7B

This is a merge of pre-trained language models created using mergekit.

I liked both Silicon-Maid and Loyal-Macaroni-Maid by SanjiWatsuki, but was looking for a middle ground, so I made my first merge using both. Let me know how it runs for you and what your results or issues are, this is my first attempt so it might be rough.

Merge Details

Merge Method

This model was merged using the DARE TIES merge method using SanjiWatsuki/Loyal-Macaroni-Maid-7B as a base.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

models:
  - model: SanjiWatsuki/Loyal-Macaroni-Maid-7B
  - model: SanjiWatsuki/Silicon-Maid-7B
    parameters:
      weight: 0.4
      density: 0.8
merge_method: dare_ties
base_model: SanjiWatsuki/Loyal-Macaroni-Maid-7B
parameters:
  int8_mask: true
dtype: bfloat16

Prompt Template (Alpaca)

  • Important Note: The limit of the context length is 8192 tokens.
  • Credits: Taken from original card by SanjiWatsuki

I found the best SillyTavern results from using the Noromaid template but please try other templates! Let me know if you find anything good.

SillyTavern config files: Context, Instruct.

Additionally, here is my highly recommended Text Completion preset. You can tweak this by adjusting temperature up or dropping min p to boost creativity or raise min p to increase stability. You shouldn't need to touch anything else!

Below is an instruction that describes a task. Write a response that appropriately completes the request.

### Instruction:
{prompt}

### Response:
Downloads last month
1
Safetensors
Model size
7.24B params
Tensor type
BF16
·

Merge of