RichardErkhov commited on
Commit
90fbeb0
1 Parent(s): fb69854

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +130 -0
README.md ADDED
@@ -0,0 +1,130 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ codellama-13b-oasst-sft-v10 - bnb 4bits
11
+ - Model creator: https://huggingface.co/OpenAssistant/
12
+ - Original model: https://huggingface.co/OpenAssistant/codellama-13b-oasst-sft-v10/
13
+
14
+
15
+
16
+
17
+ Original model description:
18
+ ---
19
+ license: llama2
20
+ datasets:
21
+ - OpenAssistant/oasst1
22
+ - shahules786/orca-best
23
+ language:
24
+ - en
25
+ ---
26
+ # Open-Assistant CodeLlama 13B SFT v10
27
+
28
+ This model is an Open-Assistant fine-tuning of Meta's CodeLlama 13B LLM.
29
+
30
+ **Note**: Due to the new RoPE Theta value (1e6 instead of 1e4), for correct results you must load this model with `trust_remote_code=True` or use the latest main branch of Huggingface transformers (until version 4.33 is released).
31
+
32
+ ## Model Details
33
+
34
+ - **Finetuned from:** [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) via [epfLLM/Megatron-LLM](https://github.com/epfLLM/Megatron-LLM)
35
+ - **Model type:** Causal decoder-only transformer language model
36
+ - **Language:** English
37
+ - **Weights & Biases training logs:** 6123 steps, BS 64 [run56_oa_llamacode](https://wandb.ai/open-assistant/public-sft/runs/run56_oa_llamacode)
38
+ - **Demo:** [Continuations for 250 random prompts (without system message)](https://open-assistant.github.io/oasst-model-eval/?f=https%3A%2F%2Fraw.githubusercontent.com%2FOpen-Assistant%2Foasst-model-eval%2Fmain%2Fsampling_reports%2Foasst-sft%2F2023-08-26_OpenAssistant_codellama-13b-oasst-sft-v10_sampling_noprefix2.json)
39
+ - **License:** [LLAMA 2 COMMUNITY LICENSE AGREEMENT](https://huggingface.co/meta-llama/Llama-2-70b/raw/main/LICENSE.txt)
40
+ - **Contact:** [Open-Assistant Discord](https://ykilcher.com/open-assistant-discord)
41
+
42
+ ## Prompting / Prompt Template
43
+
44
+ Due to public demand (see [survey](https://twitter.com/erhartford/status/1682403597525430272)) we changed the prompt-template for this model from custom prompter/assistant tokens to OpenAI's [chatml](https://github.com/openai/openai-python/blob/main/chatml.md) standard prompt format.
45
+ We hope that this leads to greater compatibility with chat inference/frontend applications.
46
+
47
+ Prompt dialogue template:
48
+
49
+ ```
50
+ """
51
+ <|im_start|>system
52
+ {system_message}<|im_end|>
53
+ <|im_start|>user
54
+ {prompt}<|im_end|>
55
+ <|im_start|>assistant
56
+ """
57
+ ```
58
+
59
+ The model input can contain multiple conversation turns between user and assistant, e.g.
60
+ ```
61
+ <|im_start|>user
62
+ {prompt 1}<|im_end|>
63
+ <|im_start|>assistant
64
+ {reply 1}<|im_end|>
65
+ <|im_start|>user
66
+ {prompt 2}<|im_end|>
67
+ <|im_start|>assistant
68
+ (...)
69
+ ```
70
+
71
+ The model was partly trained with orca system messages.
72
+ For inference we recommend to use the official [Llama2 system message](https://github.com/facebookresearch/llama/blob/ea9f33d6d3ea8ed7d560d270986407fd6c2e52b7/example_chat_completion.py#L57-L61):
73
+ ```
74
+ <|im_start|>system
75
+ You are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature.
76
+
77
+ If a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information.
78
+ <|im_end|>
79
+ ```
80
+
81
+ ### Credits & Special Thanks
82
+
83
+ - Thanks to [Meta AI](https://ai.meta.com/) for training and releasing the CodeLLlama model.
84
+ - Distributed training support was provided by EPFL's [Machine Learning and Optimization Laboratory](https://www.epfl.ch/labs/mlo/), and [Natural Language Processing Lab](https://nlp.epfl.ch/).
85
+ - The open-source [epfLLM/Megatron-LLM](https://github.com/epfLLM/Megatron-LLM) trainer was used for fine-tuning.
86
+ - [rombodawg](https://huggingface.co/rombodawg) curated the [LosslessMegaCodeTrainingV2_1m_Evol_Uncensored](https://huggingface.co/datasets/rombodawg/LosslessMegaCodeTrainingV2_1m_Evol_Uncensored) dataset.
87
+ - [ehartford](https://huggingface.co/ehartford) generated and published the [ehartford/dolphin](https://huggingface.co/datasets/ehartford/dolphin).
88
+ - [shahules786](https://github.com/shahules786) de-duped and filtered the Dolphin and Megacode dataset with a clustering/controid approach and generated orca-best & bestofmegacode.
89
+ - [andreaskoepf](https://github.com/andreaskoepf/) prepared & orchestrated the training.
90
+
91
+ ## Ethical Considerations and Limitations
92
+
93
+ Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios.
94
+ For these reasons, as with all LLMs, the potential outputs of codellama-13b-oasst-sft-v10 cannot be predicted
95
+ in advance, and the model may in some instances produce inaccurate, biased or other objectionable responses
96
+ to user prompts. Therefore, before deploying any applications of codellama-13b-oasst-sft-v10, developers should
97
+ perform safety testing and tuning tailored to their specific applications of the model.
98
+
99
+ Please see Meta's [Responsible Use Guide](https://ai.meta.com/llama/responsible-use-guide/).
100
+
101
+ ## Configuration Details
102
+
103
+ The "pretokenizer" utility used to tokenize the datamix is part of the Open-Assistant github repository and can be found here: [model/pretokenizer](https://github.com/LAION-AI/Open-Assistant/tree/main/model/pretokenizer).
104
+
105
+
106
+ ### Pretokenizer Configuration
107
+
108
+
109
+ ```
110
+ orca_megacode_oasst_best:
111
+ datasets:
112
+ - orca-chat:
113
+ val_split: 0.01
114
+ max_val_set: 1000
115
+ - bestofmegacode:
116
+ val_split: 0.01
117
+ max_val_set: 1000
118
+ - oasst_export:
119
+ lang: "bg,ca,cs,da,de,en,es,fr,hr,hu,it,nl,pl,pt,ro,ru,sl,sr,sv,uk"
120
+ #hf_dataset_name: OpenAssistant/oasst1
121
+ input_file_path: 2023-08-25_oasst_ready.jsonl.gz
122
+ top_k: 1
123
+ val_split: 0.025
124
+ output_dir: "output/orca_megacode_oasst_best"
125
+ filename_prefix: "orca_megacode_oasst_best"
126
+ min_assistant_tokens: 1
127
+ ```
128
+
129
+
130
+