AndreasThinks
commited on
Commit
•
520314b
1
Parent(s):
ca6faff
Update README.md
Browse files
README.md
CHANGED
@@ -9,8 +9,52 @@ model-index:
|
|
9 |
results: []
|
10 |
---
|
11 |
|
12 |
-
|
13 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
14 |
|
15 |
[<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
|
16 |
<details><summary>See axolotl config</summary>
|
@@ -84,50 +128,6 @@ special_tokens:
|
|
84 |
</details><br>
|
85 |
|
86 |
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/andreasthinks/mistral-nemo-welsh/runs/syq2m3vr)
|
87 |
-
# mistral-7b-english-welsh-translate
|
88 |
-
|
89 |
-
This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.3](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3) on the [Welsh Government Alpaca Welsh-English Instructions](https://huggingface.co/datasets/AndreasThinks/welsh-translation-instruction/blob/main/README.md) dataset.
|
90 |
-
|
91 |
-
This model is trained for English-Welsh translation (in any direction), with a focus on government documents, using Markdown formatting.
|
92 |
-
|
93 |
-
To ensure the highest quality translations, use the Alpaca instruction prompt format with the below structure.
|
94 |
-
|
95 |
-
```
|
96 |
-
### Instruction: {instruction}
|
97 |
-
|
98 |
-
### Input: {input}
|
99 |
-
|
100 |
-
### Response:
|
101 |
-
|
102 |
-
```
|
103 |
-
|
104 |
-
Your instruction should be "Translate the text from English to Welsh." (or vice versa).
|
105 |
-
|
106 |
-
The model is also available [quantized as GGUF](https://huggingface.co/AndreasThinks/mistral-7b-english-welsh-translate-GGUF).
|
107 |
-
|
108 |
-
## Running the model
|
109 |
-
|
110 |
-
The model is intended to be run locally, ideally using [Text generation web UI](https://github.com/oobabooga/text-generation-webui) to ensure correct prompt structure.
|
111 |
-
|
112 |
-
Start the UI as instructed for your system.
|
113 |
-
|
114 |
-
- In the "Model" tab, download either this model or [the quantized version](https://huggingface.co/AndreasThinks/mistral-7b-english-welsh-translate-GGUF). Once the download is complete, load the model.
|
115 |
-
- In the "Parameters" tab, under "Generation", set "auto_max_new_tokens" to maximum, and "Ban the eos_token" to True. In "Custom stopping strings", add "### Input"
|
116 |
-
- In the "Notebook" tab, make sure you are using the "Alpaca-with-input" prompt. Set the instruction as "Translate the text from Welsh to English." (or vice versa).#
|
117 |
-
- Add the text you would like to translate (replacing "Input"), and hit "generate"
|
118 |
-
|
119 |
-
Performance may start to degrade past a certain context window (especially if using the quantized models). Convert in chunks of under 1000 words to avoid these issues.
|
120 |
-
|
121 |
-
## LLM Evals
|
122 |
-
|
123 |
-
Thanks to [YALL - Yet Another LLM Leaderboard](https://huggingface.co/spaces/mlabonne/Yet_Another_LLM_Leaderboard)
|
124 |
-
|
125 |
-
| Model |AGIEval|TruthfulQA|Bigbench|
|
126 |
-
|-------------------------------------------------------------------------------------------------------------|------:|---------:|-------:|
|
127 |
-
|[mistral-7b-english-welsh-translate](https://huggingface.co/AndreasThinks/mistral-7b-english-welsh-translate)| 35.31| 54.5| 38.4|
|
128 |
-
|
129 |
-
|
130 |
-
## Training procedure
|
131 |
|
132 |
### Training hyperparameters
|
133 |
|
|
|
9 |
results: []
|
10 |
---
|
11 |
|
12 |
+
# mistral-7b-english-welsh-translate
|
13 |
+
|
14 |
+
|
15 |
+
|
16 |
+
This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.3](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3) on the [Welsh Government Alpaca Welsh-English Instructions](https://huggingface.co/datasets/AndreasThinks/welsh-translation-instruction/blob/main/README.md) dataset.
|
17 |
+
|
18 |
+
This model is trained for English-Welsh translation (in any direction), with a focus on government documents, using Markdown formatting.
|
19 |
+
|
20 |
+
To ensure the highest quality translations, use the Alpaca instruction prompt format with the below structure.
|
21 |
+
|
22 |
+
```
|
23 |
+
### Instruction: {instruction}
|
24 |
+
|
25 |
+
### Input: {input}
|
26 |
+
|
27 |
+
### Response:
|
28 |
+
|
29 |
+
```
|
30 |
+
|
31 |
+
Your instruction should be "Translate the text from English to Welsh." (or vice versa).
|
32 |
+
|
33 |
+
The model is also available [quantized as GGUF](https://huggingface.co/AndreasThinks/mistral-7b-english-welsh-translate-GGUF).
|
34 |
+
|
35 |
+
## Running the model
|
36 |
+
|
37 |
+
The model is intended to be run locally, ideally using [Text generation web UI](https://github.com/oobabooga/text-generation-webui) to ensure correct prompt structure.
|
38 |
+
|
39 |
+
Start the UI as instructed for your system.
|
40 |
+
|
41 |
+
- In the "Model" tab, download either this model or [the quantized version](https://huggingface.co/AndreasThinks/mistral-7b-english-welsh-translate-GGUF). Once the download is complete, load the model.
|
42 |
+
- In the "Parameters" tab, under "Generation", set "auto_max_new_tokens" to maximum, and "Ban the eos_token" to True. In "Custom stopping strings", add "### Input"
|
43 |
+
- In the "Notebook" tab, make sure you are using the "Alpaca-with-input" prompt. Set the instruction as "Translate the text from Welsh to English." (or vice versa).#
|
44 |
+
- Add the text you would like to translate (replacing "Input"), and hit "generate"
|
45 |
+
|
46 |
+
Performance may start to degrade past a certain context window (especially if using the quantized models). Convert in chunks of under 1000 words to avoid these issues.
|
47 |
+
|
48 |
+
## LLM Evals
|
49 |
+
|
50 |
+
Thanks to [YALL - Yet Another LLM Leaderboard](https://huggingface.co/spaces/mlabonne/Yet_Another_LLM_Leaderboard)
|
51 |
+
|
52 |
+
| Model |AGIEval|TruthfulQA|Bigbench|
|
53 |
+
|-------------------------------------------------------------------------------------------------------------|------:|---------:|-------:|
|
54 |
+
|[mistral-7b-english-welsh-translate](https://huggingface.co/AndreasThinks/mistral-7b-english-welsh-translate)| 35.31| 54.5| 38.4|
|
55 |
+
|
56 |
+
|
57 |
+
## Training procedure
|
58 |
|
59 |
[<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
|
60 |
<details><summary>See axolotl config</summary>
|
|
|
128 |
</details><br>
|
129 |
|
130 |
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/andreasthinks/mistral-nemo-welsh/runs/syq2m3vr)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
131 |
|
132 |
### Training hyperparameters
|
133 |
|