BramVanroy commited on
Commit
b960df1
1 Parent(s): af713d1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +34 -12
README.md CHANGED
@@ -1,38 +1,60 @@
1
  ---
2
- license: apache-2.0
3
  base_model: BramVanroy/llama2-13b-ft-mc4_nl_cleaned_tiny
4
  tags:
5
  - generated_from_trainer
 
 
 
6
  datasets:
7
  - BramVanroy/dutch_chat_datasets
8
  model-index:
9
- - name: 2e-4lr+64tbs+32a+4r
10
  results: []
 
 
11
  ---
12
 
13
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
- should probably proofread and complete it, then remove this comment. -->
15
 
16
- # 2e-4lr+64tbs+32a+4r
17
 
18
- This model is a fine-tuned version of [BramVanroy/llama2-13b-ft-mc4_nl_cleaned_tiny](https://huggingface.co/BramVanroy/llama2-13b-ft-mc4_nl_cleaned_tiny) on the BramVanroy/dutch_chat_datasets dataset.
19
- It achieves the following results on the evaluation set:
20
- - Loss: 1.0848
21
 
22
  ## Model description
23
 
24
- More information needed
 
 
 
 
 
 
 
 
25
 
26
  ## Intended uses & limitations
27
 
28
- More information needed
 
 
 
 
29
 
30
- ## Training and evaluation data
31
 
32
- More information needed
33
 
34
  ## Training procedure
35
 
 
 
 
 
 
 
 
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
 
1
  ---
2
+ license: cc-by-nc-sa-4.0
3
  base_model: BramVanroy/llama2-13b-ft-mc4_nl_cleaned_tiny
4
  tags:
5
  - generated_from_trainer
6
+ - llama
7
+ - lora
8
+ - adapters
9
  datasets:
10
  - BramVanroy/dutch_chat_datasets
11
  model-index:
12
+ - name: Llama-2-13b-chat-dutch
13
  results: []
14
+ language:
15
+ - nl
16
  ---
17
 
 
 
18
 
19
+ # Llama-2-13b-chat-dutch
20
 
21
+ This model is a fine-tuned version of [BramVanroy/llama2-13b-ft-mc4_nl_cleaned_tiny](https://huggingface.co/BramVanroy/llama2-13b-ft-mc4_nl_cleaned_tiny)
22
+ on the [BramVanroy/dutch_chat_datasets](https://huggingface.co/datasets/BramVanroy/dutch_chat_datasets) dataset on a context of 4096 tokens.
23
+ See the original [meta-llama/Llama-2-13b-hf](https://huggingface.co/meta-llama/Llama-2-13b-hf) for more information, intended use, and biases.
24
 
25
  ## Model description
26
 
27
+ I could not get Llama 2 13B to produce much Dutch, even though the description paper indicates that it was trained on a (small) portion of Dutch data. I therefore
28
+ continue training the original Llama 2 13B checkpoint on Dutch data [in regular CLM](https://huggingface.co/BramVanroy/llama2-13b-ft-mc4_nl_cleaned_tiny). In a second
29
+ step I finetuned that model on a collection of synthetic (translated) instruction and chat datasets that I have [collected](https://huggingface.co/datasets/BramVanroy/dutch_chat_datasets). See their pages for licensing, usage, creation, and citation information.
30
+
31
+ - https://huggingface.co/datasets/BramVanroy/dolly-15k-dutch
32
+ - https://huggingface.co/datasets/BramVanroy/alpaca-cleaned-dutch-baize
33
+ - https://huggingface.co/datasets/BramVanroy/stackoverflow-chat-dutch
34
+ - https://huggingface.co/datasets/BramVanroy/quora-chat-dutch
35
+
36
 
37
  ## Intended uses & limitations
38
 
39
+ Depending on the prompt, the model can return good results considering that it is only 13B in size and was only marginally pretrained on Dutch. That being said, the
40
+ model was not trained on human feedback and contains no safe-guards so it may produce unexpected and even offensive content depending on the query. The only attempt
41
+ of a safe-guard is the default prompt that it was trained on, which was
42
+
43
+ > Je bent een behulpzame, respectvolle en eerlijke assistent. Antwoord altijd zo behulpzaam mogelijk. Je antwoorden mogen geen schadelijke, onethische, racistische, seksistische, gevaarlijke of illegale inhoud bevatten. Zorg ervoor dat je antwoorden sociaal onbevooroordeeld en positief van aard zijn.\n\nAls een vraag nergens op slaat of feitelijk niet coherent is, leg dan uit waarom in plaats van iets niet correct te antwoorden. Als je het antwoord op een vraag niet weet, deel dan geen onjuiste informatie.\
44
 
45
+ Use with caution and at your own risk!
46
 
47
+ Because the model was trained on synthetic data, translated with OpenAI's API, you cannot use this model to create a competitive product to theirs.
48
 
49
  ## Training procedure
50
 
51
+ Trained with 4096 tokens context length. The dataset was preprocessed so that as many as possible dialogs were put in a single batch, without disrupting
52
+ dialogs. In other words, a dialog was never split up over different sequences or batches. During training, the human prompts were ignored in back propagation.
53
+
54
+ Trained with LoRA targetting ["q_proj", "v_proj"] in 4 bit and merged before upload. Trained with Flash Attention as borrowed from [here](https://github.com/philschmid/deep-learning-pytorch-huggingface/blob/main/training/utils/llama_patch.py).
55
+
56
+ The adapters are in the `adapters` branch.
57
+
58
  ### Training hyperparameters
59
 
60
  The following hyperparameters were used during training: