Update README.md
Browse files
README.md
CHANGED
@@ -1,18 +1,49 @@
|
|
1 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
license: llama3.1
|
3 |
base_model: meta-llama/Meta-Llama-3.1-8B-Instruct
|
4 |
-
library_name: peft
|
5 |
tags:
|
6 |
- llama-factory
|
7 |
- lora
|
8 |
datasets:
|
9 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
10 |
language:
|
11 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
12 |
---
|
13 |
|
14 |
|
15 |
-
- Fine-tuning of Llama-3.1-8B on german datasets.
|
16 |
- I've (alway) kept LoRA `QLoRA_German-ORPO` so it can be applied to any *LLaMA-3.1-8B* fine-tuned model but may affect performance.
|
17 |
|
18 |
<details>
|
|
|
1 |
---
|
2 |
+
model_creator: Nekochu
|
3 |
+
quantized_by: Nekochu
|
4 |
+
model_name: Llama-3.1 8B German ORPO
|
5 |
+
pretty_name: Llama-3.1 8B German ORPO
|
6 |
+
model_type: llama3.1
|
7 |
+
prompt_template: >-
|
8 |
+
Below is an instruction that describes a task. Write a response that
|
9 |
+
appropriately completes the request. ### Instruction: {Instruction} {summary} ### input: {category} ### Response: {prompt}
|
10 |
+
library_name: peft
|
11 |
license: llama3.1
|
12 |
base_model: meta-llama/Meta-Llama-3.1-8B-Instruct
|
|
|
13 |
tags:
|
14 |
- llama-factory
|
15 |
- lora
|
16 |
datasets:
|
17 |
+
- mayflowergmbh/intel_orca_dpo_pairs_de
|
18 |
+
- LeoLM/OpenSchnabeltier
|
19 |
+
- LeoLM/German_Songs
|
20 |
+
- LeoLM/German_Poems
|
21 |
+
- bjoernp/ultrachat_de
|
22 |
+
- mayflowergmbh/ultra-chat_de
|
23 |
+
- mayflowergmbh/airoboros-3.0_de
|
24 |
+
- mayflowergmbh/booksum_de
|
25 |
+
- mayflowergmbh/dolphin_de
|
26 |
+
- mayflowergmbh/evol-instruct_de
|
27 |
+
- mayflowergmbh/openschnabeltier_de
|
28 |
+
- mayflowergmbh/alpaca-gpt4_de
|
29 |
+
- mayflowergmbh/dolly-15k_de
|
30 |
+
- mayflowergmbh/oasst_de
|
31 |
language:
|
32 |
+
- de
|
33 |
+
- en
|
34 |
+
pipeline_tag: text-generation
|
35 |
+
task_categories:
|
36 |
+
- question-answering
|
37 |
+
- text2text-generation
|
38 |
+
- conversational
|
39 |
+
inference: True
|
40 |
+
model-index:
|
41 |
+
- name: Llama-3.1-8B-German-ORPO
|
42 |
+
results: []
|
43 |
---
|
44 |
|
45 |
|
46 |
+
- Fine-tuning of Llama-3.1-8B on german datasets. Same datasets used in [Nekochu/Llama-2-13B-German-ORPO](https://huggingface.co/Nekochu/Llama-2-13B-German-ORPO).
|
47 |
- I've (alway) kept LoRA `QLoRA_German-ORPO` so it can be applied to any *LLaMA-3.1-8B* fine-tuned model but may affect performance.
|
48 |
|
49 |
<details>
|