freefallr commited on
Commit
8a2e659
·
1 Parent(s): 5cb70be

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -17,19 +17,19 @@ datasets:
17
  # Llama 2 13b Chat German - GGUF
18
 
19
  This repository contains [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german) in GGUF format.
20
- The original model created by [jphme](https://huggingface.co/jphme) and is a fine-tune of [Llama2 13b Chat](https://huggingface.co/meta-llama/Llama-2-13b-chat) from Meta, focused on German instructions (helpful for RAG).
21
 
22
 
23
  ## Model Sheet
24
 
25
- | **Model Attribute** | **Details** |
26
  |--------------------------|--------------------------------------------------------------------------------------------------------------|
27
  | **Format** | GGUF |
28
  | **Converted with** | llama.cpp (Commit: 9e20231) |
29
- | **Quantization Levels** | - 8 Bit <br> - 5_K_M Bit <br> - 4_K_M Bit |
30
  | **Model** | [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german) |
31
  | **Created by** | [jphme](https://huggingface.co/jphme) |
32
-
33
 
34
  ## Replication Steps
35
  Clone and install llama.cpp *(Commit: 9e20231)* and use the provided `convert.py` file to convert the original model to GGUF with FP16 precision. The converted model will then be used to do further quantization.
 
17
  # Llama 2 13b Chat German - GGUF
18
 
19
  This repository contains [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german) in GGUF format.
20
+ The original model was created by [jphme](https://huggingface.co/jphme) and is a fine-tune of [Llama2 13b Chat](https://huggingface.co/meta-llama/Llama-2-13b-chat) from Meta, trained on German instructions.
21
 
22
 
23
  ## Model Sheet
24
 
25
+ | **Model Attribute** | **Details** |
26
  |--------------------------|--------------------------------------------------------------------------------------------------------------|
27
  | **Format** | GGUF |
28
  | **Converted with** | llama.cpp (Commit: 9e20231) |
29
+ | **Quantization Levels** | 8 Bit<br> 5 Bit K_M <br> 4 Bit K_M |
30
  | **Model** | [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german) |
31
  | **Created by** | [jphme](https://huggingface.co/jphme) |
32
+ | **Training Data** | Prorietary German Conversation Dataset, German SQuAD, German legal SQuAD data, augmented with "wrong" contexts, to improve factual RAG |
33
 
34
  ## Replication Steps
35
  Clone and install llama.cpp *(Commit: 9e20231)* and use the provided `convert.py` file to convert the original model to GGUF with FP16 precision. The converted model will then be used to do further quantization.