robinsmits commited on
Commit
ec2ffa3
1 Parent(s): f4e935e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +51 -9
README.md CHANGED
@@ -1,36 +1,66 @@
1
  ---
2
- license: apache-2.0
 
 
 
3
  base_model: DAMO-NLP-MT/polylm-13b
4
  tags:
5
  - generated_from_trainer
 
 
 
 
6
  model-index:
7
  - name: polylm_13b_ft_alpaca_clean_dutch
8
  results: []
 
 
 
 
9
  ---
10
 
11
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
- should probably proofread and complete it, then remove this comment. -->
13
-
14
  # polylm_13b_ft_alpaca_clean_dutch
15
 
16
- This model is a fine-tuned version of [DAMO-NLP-MT/polylm-13b](https://huggingface.co/DAMO-NLP-MT/polylm-13b) on an unknown dataset.
 
 
17
  It achieves the following results on the evaluation set:
18
  - Loss: 1.3355
19
 
20
- ## Model description
21
 
22
- More information needed
23
 
24
  ## Intended uses & limitations
25
 
26
- More information needed
 
 
 
 
 
 
 
 
 
 
 
 
27
 
28
  ## Training and evaluation data
29
 
30
- More information needed
 
 
 
 
31
 
32
  ## Training procedure
33
 
 
 
 
 
34
  ### Training hyperparameters
35
 
36
  The following hyperparameters were used during training:
@@ -45,6 +75,17 @@ The following hyperparameters were used during training:
45
  - lr_scheduler_warmup_steps: 64
46
  - num_epochs: 2
47
 
 
 
 
 
 
 
 
 
 
 
 
48
  ### Training results
49
 
50
  | Training Loss | Epoch | Step | Validation Loss |
@@ -69,3 +110,4 @@ The following hyperparameters were used during training:
69
  - Pytorch 2.0.1+cu118
70
  - Datasets 2.14.0
71
  - Tokenizers 0.13.3
 
 
1
  ---
2
+ license: cc-by-nc-4.0
3
+ inference: false
4
+ datasets:
5
+ - BramVanroy/alpaca-cleaned-dutch
6
  base_model: DAMO-NLP-MT/polylm-13b
7
  tags:
8
  - generated_from_trainer
9
+ - alpaca
10
+ - Transformers
11
+ - PolyLM
12
+ - text-generation-inference
13
  model-index:
14
  - name: polylm_13b_ft_alpaca_clean_dutch
15
  results: []
16
+ language:
17
+ - nl
18
+ library_name: peft
19
+ pipeline_tag: text-generation
20
  ---
21
 
 
 
 
22
  # polylm_13b_ft_alpaca_clean_dutch
23
 
24
+ ## Model description
25
+
26
+ This adapter model is a fine-tuned version of [DAMO-NLP-MT/polylm-13b](https://huggingface.co/DAMO-NLP-MT/polylm-13b).
27
  It achieves the following results on the evaluation set:
28
  - Loss: 1.3355
29
 
30
+ Finetuning was performed on the Dutch [BramVanroy/alpaca-cleaned-dutch](https://www.huggingface.co/datasets/BramVanroy/alpaca-cleaned-dutch) dataset which contains 52K of records with instruction following-data translated from English to Dutch.
31
 
32
+ See [DAMO-NLP-MT/polylm-13b](https://huggingface.co/DAMO-NLP-MT/polylm-13b) for all information about the base model.
33
 
34
  ## Intended uses & limitations
35
 
36
+ The PolyLM-13B model was trained on 18 languages. The primary focus was to create a multi-lingual Open LLM.
37
+ Dutch was one of those 18 languages. For training the model a diverse combination of multi-lingual datasets was used.
38
+
39
+ The generated output and performance of this model for the Dutch language is very likely not always comparable to the various Open-Llama models that have been finetuned on English Alpaca datasets.
40
+
41
+ The primary intention of this finetuned model is to explore and research the use of the Dutch language in combination with an Open LLM model.
42
+
43
+ ## Bias, Risks, and Limitations
44
+
45
+ The information below is copied from the base model's [official model card](https://arxiv.org/pdf/2307.06018.pdf).
46
+ This applies also to the finetuned model.
47
+
48
+ > Our contributions are fully methodological: adding the support of multilingualism to LLM during training and SFT phases. It is unavoidable that PolyLM might exhibit several common deficiencies of language models, e.g. hallucination and toxicity. PolyLM should not be used directly in any application, without a prior assessment of safety and fairness concerns specific to the application.
49
 
50
  ## Training and evaluation data
51
 
52
+ This model was trained on the [BramVanroy/alpaca-cleaned-dutch](https://www.huggingface.co/datasets/BramVanroy/alpaca-cleaned-dutch) dataset.
53
+
54
+ The dataset is the Dutch translation of the English Alpaca Cleaned instruction dataset.
55
+
56
+ Based on the dataset license only Non-Commercial use is allowed. Commercial use is strictly forbidden.
57
 
58
  ## Training procedure
59
 
60
+ This model was finetuned with a QLoRA setup on a Google Colab A100 GPU in about 7.0 hours.
61
+
62
+ The notebook used for training can be found here: [Training Notebook](https://github.com/RobinSmits/Dutch-LLMs/blob/main/PolyLM_13B_Alpaca_Clean_Dutch_Qlora.ipynb)
63
+
64
  ### Training hyperparameters
65
 
66
  The following hyperparameters were used during training:
 
75
  - lr_scheduler_warmup_steps: 64
76
  - num_epochs: 2
77
 
78
+ The following bitsandbytes quantization config was used during training:
79
+ - load_in_8bit: False
80
+ - load_in_4bit: True
81
+ - llm_int8_threshold: 6.0
82
+ - llm_int8_skip_modules: None
83
+ - llm_int8_enable_fp32_cpu_offload: False
84
+ - llm_int8_has_fp16_weight: False
85
+ - bnb_4bit_quant_type: nf4
86
+ - bnb_4bit_use_double_quant: True
87
+ - bnb_4bit_compute_dtype: bfloat16
88
+ -
89
  ### Training results
90
 
91
  | Training Loss | Epoch | Step | Validation Loss |
 
110
  - Pytorch 2.0.1+cu118
111
  - Datasets 2.14.0
112
  - Tokenizers 0.13.3
113
+ - PEFT 0.4.0