Khetterman commited on
Commit
3bbaa8a
1 Parent(s): 4869b3a

Create README.md

Browse files

![Multilingual-SaigaSuzume-8B-Logo256.png](https://cdn-uploads.huggingface.co/production/uploads/673125091920e70ac26c8a2e/aVbK8k3mUMBAOlUSXBK91.png)

Files changed (1) hide show
  1. README.md +98 -0
README.md ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - huihui-ai/Meta-Llama-3.1-8B-Instruct-abliterated
4
+ - IlyaGusev/saiga_llama3_8b
5
+ - lightblue/suzume-llama-3-8B-multilingual
6
+ - lightblue/suzume-llama-3-8B-multilingual-orpo-borda-full
7
+ - lightblue/suzume-llama-3-8B-multilingual-orpo-borda-half
8
+ - lightblue/suzume-llama-3-8B-multilingual-orpo-borda-top25
9
+ - lightblue/suzume-llama-3-8B-multilingual-orpo-borda-top75
10
+ library_name: transformers
11
+ tags:
12
+ - mergekit
13
+ - merge
14
+ - bfloat16
15
+ - safetensors
16
+ - 8b
17
+ - chat
18
+ - conversational
19
+ language:
20
+ - de
21
+ - en
22
+ - es
23
+ - fr
24
+ - hi
25
+ - it
26
+ - ja
27
+ - pt
28
+ - ru
29
+ - th
30
+ - zh
31
+
32
+ ---
33
+ # Multilingual-SaigaSuzume-8B
34
+
35
+ >Your words are like rain falling from heaven on a tower in a sinful land; can anyone in Babylon understand them?
36
+
37
+ ![Multilingual-SaigaSuzume-8B-Logo256.png](https://cdn-uploads.huggingface.co/production/uploads/673125091920e70ac26c8a2e/aVbK8k3mUMBAOlUSXBK91.png)
38
+
39
+ This model was created as the basis of multilingual abilities for other models. I think it will be very useful as an integral part of your model. There is some censorship, keep this in mind.
40
+
41
+ ## Merge Details
42
+ ### Method
43
+
44
+ This is a simple, but usefull merge of **7 cool models**, created using [mergekit](https://github.com/arcee-ai/mergekit).
45
+
46
+ ### Models
47
+
48
+ The following models were included in the merge:
49
+
50
+ * [huihui-ai/Meta-Llama-3.1-8B-Instruct-abliterated](https://huggingface.co/huihui-ai/Meta-Llama-3.1-8B-Instruct-abliterated)
51
+ * [IlyaGusev/saiga_llama3_8b](https://huggingface.co/IlyaGusev/saiga_llama3_8b)
52
+ * [lightblue/suzume-llama-3-8B-multilingual](https://huggingface.co/lightblue/suzume-llama-3-8B-multilingual)
53
+ * [lightblue/suzume-llama-3-8B-multilingual-orpo-borda-full](https://huggingface.co/lightblue/suzume-llama-3-8B-multilingual-orpo-borda-full)
54
+ * [lightblue/suzume-llama-3-8B-multilingual-orpo-borda-half](https://huggingface.co/lightblue/suzume-llama-3-8B-multilingual-orpo-borda-half)
55
+ * [lightblue/suzume-llama-3-8B-multilingual-orpo-borda-top25](https://huggingface.co/lightblue/suzume-llama-3-8B-multilingual-orpo-borda-top25)
56
+ * [lightblue/suzume-llama-3-8B-multilingual-orpo-borda-top75](https://huggingface.co/lightblue/suzume-llama-3-8B-multilingual-orpo-borda-top75)
57
+
58
+ ### Configuration
59
+
60
+ The following YAML configurations was used to produce this model:
61
+
62
+ ```yaml
63
+ # Multilingual-SaigaSuzume-8B-BFH
64
+ models:
65
+ - model: lightblue/suzume-llama-3-8B-multilingual-orpo-borda-full
66
+ - model: IlyaGusev/saiga_llama3_8b
67
+ - model: lightblue/suzume-llama-3-8B-multilingual-orpo-borda-half
68
+ merge_method: model_stock
69
+ base_model: huihui-ai/Meta-Llama-3.1-8B-Instruct-abliterated
70
+ dtype: bfloat16
71
+
72
+ # Multilingual-SaigaSuzume-8B-BTP
73
+ models:
74
+ - model: lightblue/suzume-llama-3-8B-multilingual-orpo-borda-top75
75
+ - model: IlyaGusev/saiga_llama3_8b
76
+ - model: lightblue/suzume-llama-3-8B-multilingual-orpo-borda-top25
77
+ merge_method: model_stock
78
+ base_model: huihui-ai/Meta-Llama-3.1-8B-Instruct-abliterated
79
+ dtype: bfloat16
80
+
81
+ # Multilingual-SaigaSuzume-8B-Classic
82
+ models:
83
+ - model: IlyaGusev/saiga_llama3_8b
84
+ - model: lightblue/suzume-llama-3-8B-multilingual
85
+ merge_method: model_stock
86
+ base_model: huihui-ai/Meta-Llama-3.1-8B-Instruct-abliterated
87
+ dtype: bfloat16
88
+
89
+ # Multilingual-SaigaSuzume-8B
90
+ models:
91
+ - model: Multilingual-SaigaSuzume-8B-BFH
92
+ - model: Multilingual-SaigaSuzume-8B-BTP
93
+ merge_method: model_stock
94
+ base_model: Multilingual-SaigaSuzume-8B-Classic
95
+ dtype: bfloat16
96
+ ```
97
+
98
+ >My thanks to the authors of the original models, your work is incredible. Have a good time 🖤