Edit model card

Multilingual-SaigaSuzume-8B

Your words are like rain falling from heaven on a tower in a sinful land; can anyone in Babylon understand them?

Multilingual-SaigaSuzume-8B-Logo256.png

This model was created as the basis of multilingual abilities for other models. I think it will be very useful as an integral part of your model. There is some censorship, keep this in mind.

Merge Details

Method

This is a simple, but usefull merge of 7 cool models, created using mergekit.

Models

The following models were included in the merge:

Configuration

The following YAML configurations was used to produce this model:

# Multilingual-SaigaSuzume-8B-BFH
models:
  - model: lightblue/suzume-llama-3-8B-multilingual-orpo-borda-full
  - model: IlyaGusev/saiga_llama3_8b
  - model: lightblue/suzume-llama-3-8B-multilingual-orpo-borda-half
merge_method: model_stock
base_model: huihui-ai/Meta-Llama-3.1-8B-Instruct-abliterated
dtype: bfloat16

# Multilingual-SaigaSuzume-8B-BTP
models:
  - model: lightblue/suzume-llama-3-8B-multilingual-orpo-borda-top75
  - model: IlyaGusev/saiga_llama3_8b
  - model: lightblue/suzume-llama-3-8B-multilingual-orpo-borda-top25
merge_method: model_stock
base_model: huihui-ai/Meta-Llama-3.1-8B-Instruct-abliterated
dtype: bfloat16

# Multilingual-SaigaSuzume-8B-Classic
models:
  - model: IlyaGusev/saiga_llama3_8b
  - model: lightblue/suzume-llama-3-8B-multilingual
merge_method: model_stock
base_model: huihui-ai/Meta-Llama-3.1-8B-Instruct-abliterated
dtype: bfloat16

# Multilingual-SaigaSuzume-8B
models:
  - model: Multilingual-SaigaSuzume-8B-BFH
  - model: Multilingual-SaigaSuzume-8B-BTP
merge_method: model_stock
base_model: Multilingual-SaigaSuzume-8B-Classic
dtype: bfloat16

My thanks to the authors of the original models, your work is incredible. Have a good time 🖤

Downloads last month
0
Safetensors
Model size
8.03B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for Khetterman/Multilingual-SaigaSuzume-8B