Upload README.md
Browse files
README.md
ADDED
@@ -0,0 +1,415 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
base_model: Pclanglais/MonadGPT
|
3 |
+
inference: false
|
4 |
+
language:
|
5 |
+
- en
|
6 |
+
- fr
|
7 |
+
- la
|
8 |
+
library_name: transformers
|
9 |
+
license: apache-2.0
|
10 |
+
model_creator: Pierre-Carl Langlais
|
11 |
+
model_name: MonadGPT 7B
|
12 |
+
model_type: mistral
|
13 |
+
pipeline_tag: conversational
|
14 |
+
prompt_template: '<|im_start|>system
|
15 |
+
|
16 |
+
{system_message}<|im_end|>
|
17 |
+
|
18 |
+
<|im_start|>user
|
19 |
+
|
20 |
+
{prompt}<|im_end|>
|
21 |
+
|
22 |
+
<|im_start|>assistant
|
23 |
+
|
24 |
+
'
|
25 |
+
quantized_by: TheBloke
|
26 |
+
---
|
27 |
+
<!-- markdownlint-disable MD041 -->
|
28 |
+
|
29 |
+
<!-- header start -->
|
30 |
+
<!-- 200823 -->
|
31 |
+
<div style="width: auto; margin-left: auto; margin-right: auto">
|
32 |
+
<img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
|
33 |
+
</div>
|
34 |
+
<div style="display: flex; justify-content: space-between; width: 100%;">
|
35 |
+
<div style="display: flex; flex-direction: column; align-items: flex-start;">
|
36 |
+
<p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://discord.gg/theblokeai">Chat & support: TheBloke's Discord server</a></p>
|
37 |
+
</div>
|
38 |
+
<div style="display: flex; flex-direction: column; align-items: flex-end;">
|
39 |
+
<p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p>
|
40 |
+
</div>
|
41 |
+
</div>
|
42 |
+
<div style="text-align:center; margin-top: 0em; margin-bottom: 0em"><p style="margin-top: 0.25em; margin-bottom: 0em;">TheBloke's LLM work is generously supported by a grant from <a href="https://a16z.com">andreessen horowitz (a16z)</a></p></div>
|
43 |
+
<hr style="margin-top: 1.0em; margin-bottom: 1.0em;">
|
44 |
+
<!-- header end -->
|
45 |
+
|
46 |
+
# MonadGPT 7B - AWQ
|
47 |
+
- Model creator: [Pierre-Carl Langlais](https://huggingface.co/Pclanglais)
|
48 |
+
- Original model: [MonadGPT 7B](https://huggingface.co/Pclanglais/MonadGPT)
|
49 |
+
|
50 |
+
<!-- description start -->
|
51 |
+
## Description
|
52 |
+
|
53 |
+
This repo contains AWQ model files for [Pierre-Carl Langlais's MonadGPT 7B](https://huggingface.co/Pclanglais/MonadGPT).
|
54 |
+
|
55 |
+
These files were quantised using hardware kindly provided by [Massed Compute](https://massedcompute.com/).
|
56 |
+
|
57 |
+
|
58 |
+
### About AWQ
|
59 |
+
|
60 |
+
AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.
|
61 |
+
|
62 |
+
It is supported by:
|
63 |
+
|
64 |
+
- [Text Generation Webui](https://github.com/oobabooga/text-generation-webui) - using Loader: AutoAWQ
|
65 |
+
- [vLLM](https://github.com/vllm-project/vllm) - Llama and Mistral models only
|
66 |
+
- [Hugging Face Text Generation Inference (TGI)](https://github.com/huggingface/text-generation-inference)
|
67 |
+
- [Transformers](https://huggingface.co/docs/transformers) version 4.35.0 and later, from any code or client that supports Transformers
|
68 |
+
- [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) - for use from Python code
|
69 |
+
|
70 |
+
<!-- description end -->
|
71 |
+
<!-- repositories-available start -->
|
72 |
+
## Repositories available
|
73 |
+
|
74 |
+
* [AWQ model(s) for GPU inference.](https://huggingface.co/TheBloke/MonadGPT-AWQ)
|
75 |
+
* [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/MonadGPT-GPTQ)
|
76 |
+
* [2, 3, 4, 5, 6 and 8-bit GGUF models for CPU+GPU inference](https://huggingface.co/TheBloke/MonadGPT-GGUF)
|
77 |
+
* [Pierre-Carl Langlais's original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/Pclanglais/MonadGPT)
|
78 |
+
<!-- repositories-available end -->
|
79 |
+
|
80 |
+
<!-- prompt-template start -->
|
81 |
+
## Prompt template: ChatML
|
82 |
+
|
83 |
+
```
|
84 |
+
<|im_start|>system
|
85 |
+
{system_message}<|im_end|>
|
86 |
+
<|im_start|>user
|
87 |
+
{prompt}<|im_end|>
|
88 |
+
<|im_start|>assistant
|
89 |
+
|
90 |
+
```
|
91 |
+
|
92 |
+
<!-- prompt-template end -->
|
93 |
+
|
94 |
+
|
95 |
+
<!-- README_AWQ.md-provided-files start -->
|
96 |
+
## Provided files, and AWQ parameters
|
97 |
+
|
98 |
+
I currently release 128g GEMM models only. The addition of group_size 32 models, and GEMV kernel models, is being actively considered.
|
99 |
+
|
100 |
+
Models are released as sharded safetensors files.
|
101 |
+
|
102 |
+
| Branch | Bits | GS | AWQ Dataset | Seq Len | Size |
|
103 |
+
| ------ | ---- | -- | ----------- | ------- | ---- |
|
104 |
+
| [main](https://huggingface.co/TheBloke/MonadGPT-AWQ/tree/main) | 4 | 128 | latin-english | 4096 | 4.15 GB
|
105 |
+
|
106 |
+
<!-- README_AWQ.md-provided-files end -->
|
107 |
+
|
108 |
+
<!-- README_AWQ.md-text-generation-webui start -->
|
109 |
+
## How to easily download and use this model in [text-generation-webui](https://github.com/oobabooga/text-generation-webui)
|
110 |
+
|
111 |
+
Please make sure you're using the latest version of [text-generation-webui](https://github.com/oobabooga/text-generation-webui).
|
112 |
+
|
113 |
+
It is strongly recommended to use the text-generation-webui one-click-installers unless you're sure you know how to make a manual install.
|
114 |
+
|
115 |
+
1. Click the **Model tab**.
|
116 |
+
2. Under **Download custom model or LoRA**, enter `TheBloke/MonadGPT-AWQ`.
|
117 |
+
3. Click **Download**.
|
118 |
+
4. The model will start downloading. Once it's finished it will say "Done".
|
119 |
+
5. In the top left, click the refresh icon next to **Model**.
|
120 |
+
6. In the **Model** dropdown, choose the model you just downloaded: `MonadGPT-AWQ`
|
121 |
+
7. Select **Loader: AutoAWQ**.
|
122 |
+
8. Click Load, and the model will load and is now ready for use.
|
123 |
+
9. If you want any custom settings, set them and then click **Save settings for this model** followed by **Reload the Model** in the top right.
|
124 |
+
10. Once you're ready, click the **Text Generation** tab and enter a prompt to get started!
|
125 |
+
<!-- README_AWQ.md-text-generation-webui end -->
|
126 |
+
|
127 |
+
<!-- README_AWQ.md-use-from-vllm start -->
|
128 |
+
## Multi-user inference server: vLLM
|
129 |
+
|
130 |
+
Documentation on installing and using vLLM [can be found here](https://vllm.readthedocs.io/en/latest/).
|
131 |
+
|
132 |
+
- Please ensure you are using vLLM version 0.2 or later.
|
133 |
+
- When using vLLM as a server, pass the `--quantization awq` parameter.
|
134 |
+
|
135 |
+
For example:
|
136 |
+
|
137 |
+
```shell
|
138 |
+
python3 -m vllm.entrypoints.api_server --model TheBloke/MonadGPT-AWQ --quantization awq --dtype auto
|
139 |
+
```
|
140 |
+
|
141 |
+
- When using vLLM from Python code, again set `quantization=awq`.
|
142 |
+
|
143 |
+
For example:
|
144 |
+
|
145 |
+
```python
|
146 |
+
from vllm import LLM, SamplingParams
|
147 |
+
|
148 |
+
prompts = [
|
149 |
+
"Tell me about AI",
|
150 |
+
"Write a story about llamas",
|
151 |
+
"What is 291 - 150?",
|
152 |
+
"How much wood would a woodchuck chuck if a woodchuck could chuck wood?",
|
153 |
+
]
|
154 |
+
prompt_template=f'''<|im_start|>system
|
155 |
+
{system_message}<|im_end|>
|
156 |
+
<|im_start|>user
|
157 |
+
{prompt}<|im_end|>
|
158 |
+
<|im_start|>assistant
|
159 |
+
'''
|
160 |
+
|
161 |
+
prompts = [prompt_template.format(prompt=prompt) for prompt in prompts]
|
162 |
+
|
163 |
+
sampling_params = SamplingParams(temperature=0.8, top_p=0.95)
|
164 |
+
|
165 |
+
llm = LLM(model="TheBloke/MonadGPT-AWQ", quantization="awq", dtype="auto")
|
166 |
+
|
167 |
+
outputs = llm.generate(prompts, sampling_params)
|
168 |
+
|
169 |
+
# Print the outputs.
|
170 |
+
for output in outputs:
|
171 |
+
prompt = output.prompt
|
172 |
+
generated_text = output.outputs[0].text
|
173 |
+
print(f"Prompt: {prompt!r}, Generated text: {generated_text!r}")
|
174 |
+
```
|
175 |
+
<!-- README_AWQ.md-use-from-vllm start -->
|
176 |
+
|
177 |
+
<!-- README_AWQ.md-use-from-tgi start -->
|
178 |
+
## Multi-user inference server: Hugging Face Text Generation Inference (TGI)
|
179 |
+
|
180 |
+
Use TGI version 1.1.0 or later. The official Docker container is: `ghcr.io/huggingface/text-generation-inference:1.1.0`
|
181 |
+
|
182 |
+
Example Docker parameters:
|
183 |
+
|
184 |
+
```shell
|
185 |
+
--model-id TheBloke/MonadGPT-AWQ --port 3000 --quantize awq --max-input-length 3696 --max-total-tokens 4096 --max-batch-prefill-tokens 4096
|
186 |
+
```
|
187 |
+
|
188 |
+
Example Python code for interfacing with TGI (requires [huggingface-hub](https://github.com/huggingface/huggingface_hub) 0.17.0 or later):
|
189 |
+
|
190 |
+
```shell
|
191 |
+
pip3 install huggingface-hub
|
192 |
+
```
|
193 |
+
|
194 |
+
```python
|
195 |
+
from huggingface_hub import InferenceClient
|
196 |
+
|
197 |
+
endpoint_url = "https://your-endpoint-url-here"
|
198 |
+
|
199 |
+
prompt = "Tell me about AI"
|
200 |
+
prompt_template=f'''<|im_start|>system
|
201 |
+
{system_message}<|im_end|>
|
202 |
+
<|im_start|>user
|
203 |
+
{prompt}<|im_end|>
|
204 |
+
<|im_start|>assistant
|
205 |
+
'''
|
206 |
+
|
207 |
+
client = InferenceClient(endpoint_url)
|
208 |
+
response = client.text_generation(prompt,
|
209 |
+
max_new_tokens=128,
|
210 |
+
do_sample=True,
|
211 |
+
temperature=0.7,
|
212 |
+
top_p=0.95,
|
213 |
+
top_k=40,
|
214 |
+
repetition_penalty=1.1)
|
215 |
+
|
216 |
+
print(f"Model output: ", response)
|
217 |
+
```
|
218 |
+
<!-- README_AWQ.md-use-from-tgi end -->
|
219 |
+
|
220 |
+
<!-- README_AWQ.md-use-from-python start -->
|
221 |
+
## Inference from Python code using Transformers
|
222 |
+
|
223 |
+
### Install the necessary packages
|
224 |
+
|
225 |
+
- Requires: [Transformers](https://huggingface.co/docs/transformers) 4.35.0 or later.
|
226 |
+
- Requires: [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) 0.1.6 or later.
|
227 |
+
|
228 |
+
```shell
|
229 |
+
pip3 install --upgrade "autoawq>=0.1.6" "transformers>=4.35.0"
|
230 |
+
```
|
231 |
+
|
232 |
+
Note that if you are using PyTorch 2.0.1, the above AutoAWQ command will automatically upgrade you to PyTorch 2.1.0.
|
233 |
+
|
234 |
+
If you are using CUDA 11.8 and wish to continue using PyTorch 2.0.1, instead run this command:
|
235 |
+
|
236 |
+
```shell
|
237 |
+
pip3 install https://github.com/casper-hansen/AutoAWQ/releases/download/v0.1.6/autoawq-0.1.6+cu118-cp310-cp310-linux_x86_64.whl
|
238 |
+
```
|
239 |
+
|
240 |
+
If you have problems installing [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) using the pre-built wheels, install it from source instead:
|
241 |
+
|
242 |
+
```shell
|
243 |
+
pip3 uninstall -y autoawq
|
244 |
+
git clone https://github.com/casper-hansen/AutoAWQ
|
245 |
+
cd AutoAWQ
|
246 |
+
pip3 install .
|
247 |
+
```
|
248 |
+
|
249 |
+
### Transformers example code (requires Transformers 4.35.0 and later)
|
250 |
+
|
251 |
+
```python
|
252 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer
|
253 |
+
|
254 |
+
model_name_or_path = "TheBloke/MonadGPT-AWQ"
|
255 |
+
|
256 |
+
tokenizer = AutoTokenizer.from_pretrained(model_name_or_path)
|
257 |
+
model = AutoModelForCausalLM.from_pretrained(
|
258 |
+
model_name_or_path,
|
259 |
+
low_cpu_mem_usage=True,
|
260 |
+
device_map="cuda:0"
|
261 |
+
)
|
262 |
+
|
263 |
+
# Using the text streamer to stream output one token at a time
|
264 |
+
streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True)
|
265 |
+
|
266 |
+
prompt = "Tell me about AI"
|
267 |
+
prompt_template=f'''<|im_start|>system
|
268 |
+
{system_message}<|im_end|>
|
269 |
+
<|im_start|>user
|
270 |
+
{prompt}<|im_end|>
|
271 |
+
<|im_start|>assistant
|
272 |
+
'''
|
273 |
+
|
274 |
+
# Convert prompt to tokens
|
275 |
+
tokens = tokenizer(
|
276 |
+
prompt_template,
|
277 |
+
return_tensors='pt'
|
278 |
+
).input_ids.cuda()
|
279 |
+
|
280 |
+
generation_params = {
|
281 |
+
"do_sample": True,
|
282 |
+
"temperature": 0.7,
|
283 |
+
"top_p": 0.95,
|
284 |
+
"top_k": 40,
|
285 |
+
"max_new_tokens": 512,
|
286 |
+
"repetition_penalty": 1.1
|
287 |
+
}
|
288 |
+
|
289 |
+
# Generate streamed output, visible one token at a time
|
290 |
+
generation_output = model.generate(
|
291 |
+
tokens,
|
292 |
+
streamer=streamer,
|
293 |
+
**generation_params
|
294 |
+
)
|
295 |
+
|
296 |
+
# Generation without a streamer, which will include the prompt in the output
|
297 |
+
generation_output = model.generate(
|
298 |
+
tokens,
|
299 |
+
**generation_params
|
300 |
+
)
|
301 |
+
|
302 |
+
# Get the tokens from the output, decode them, print them
|
303 |
+
token_output = generation_output[0]
|
304 |
+
text_output = tokenizer.decode(token_output)
|
305 |
+
print("model.generate output: ", text_output)
|
306 |
+
|
307 |
+
# Inference is also possible via Transformers' pipeline
|
308 |
+
from transformers import pipeline
|
309 |
+
|
310 |
+
pipe = pipeline(
|
311 |
+
"text-generation",
|
312 |
+
model=model,
|
313 |
+
tokenizer=tokenizer,
|
314 |
+
**generation_params
|
315 |
+
)
|
316 |
+
|
317 |
+
pipe_output = pipe(prompt_template)[0]['generated_text']
|
318 |
+
print("pipeline output: ", pipe_output)
|
319 |
+
|
320 |
+
```
|
321 |
+
<!-- README_AWQ.md-use-from-python end -->
|
322 |
+
|
323 |
+
<!-- README_AWQ.md-compatibility start -->
|
324 |
+
## Compatibility
|
325 |
+
|
326 |
+
The files provided are tested to work with:
|
327 |
+
|
328 |
+
- [text-generation-webui](https://github.com/oobabooga/text-generation-webui) using `Loader: AutoAWQ`.
|
329 |
+
- [vLLM](https://github.com/vllm-project/vllm) version 0.2.0 and later.
|
330 |
+
- [Hugging Face Text Generation Inference (TGI)](https://github.com/huggingface/text-generation-inference) version 1.1.0 and later.
|
331 |
+
- [Transformers](https://huggingface.co/docs/transformers) version 4.35.0 and later.
|
332 |
+
- [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) version 0.1.1 and later.
|
333 |
+
|
334 |
+
<!-- README_AWQ.md-compatibility end -->
|
335 |
+
|
336 |
+
<!-- footer start -->
|
337 |
+
<!-- 200823 -->
|
338 |
+
## Discord
|
339 |
+
|
340 |
+
For further support, and discussions on these models and AI in general, join us at:
|
341 |
+
|
342 |
+
[TheBloke AI's Discord server](https://discord.gg/theblokeai)
|
343 |
+
|
344 |
+
## Thanks, and how to contribute
|
345 |
+
|
346 |
+
Thanks to the [chirper.ai](https://chirper.ai) team!
|
347 |
+
|
348 |
+
Thanks to Clay from [gpus.llm-utils.org](llm-utils)!
|
349 |
+
|
350 |
+
I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training.
|
351 |
+
|
352 |
+
If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects.
|
353 |
+
|
354 |
+
Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits.
|
355 |
+
|
356 |
+
* Patreon: https://patreon.com/TheBlokeAI
|
357 |
+
* Ko-Fi: https://ko-fi.com/TheBlokeAI
|
358 |
+
|
359 |
+
**Special thanks to**: Aemon Algiz.
|
360 |
+
|
361 |
+
**Patreon special mentions**: Brandon Frisco, LangChain4j, Spiking Neurons AB, transmissions 11, Joseph William Delisle, Nitin Borwankar, Willem Michiel, Michael Dempsey, vamX, Jeffrey Morgan, zynix, jjj, Omer Bin Jawed, Sean Connelly, jinyuan sun, Jeromy Smith, Shadi, Pawan Osman, Chadd, Elijah Stavena, Illia Dulskyi, Sebastain Graf, Stephen Murray, terasurfer, Edmond Seymore, Celu Ramasamy, Mandus, Alex, biorpg, Ajan Kanaga, Clay Pascal, Raven Klaugh, 阿明, K, ya boyyy, usrbinkat, Alicia Loh, John Villwock, ReadyPlayerEmma, Chris Smitley, Cap'n Zoog, fincy, GodLy, S_X, sidney chen, Cory Kujawski, OG, Mano Prime, AzureBlack, Pieter, Kalila, Spencer Kim, Tom X Nguyen, Stanislav Ovsiannikov, Michael Levine, Andrey, Trailburnt, Vadim, Enrico Ros, Talal Aujan, Brandon Phillips, Jack West, Eugene Pentland, Michael Davis, Will Dee, webtim, Jonathan Leane, Alps Aficionado, Rooh Singh, Tiffany J. Kim, theTransient, Luke @flexchar, Elle, Caitlyn Gatomon, Ari Malik, subjectnull, Johann-Peter Hartmann, Trenton Dambrowitz, Imad Khwaja, Asp the Wyvern, Emad Mostaque, Rainer Wilmers, Alexandros Triantafyllidis, Nicholas, Pedro Madruga, SuperWojo, Harry Royden McLaughlin, James Bentley, Olakabola, David Ziegler, Ai Maven, Jeff Scroggin, Nikolai Manek, Deo Leter, Matthew Berman, Fen Risland, Ken Nordquist, Manuel Alberto Morcote, Luke Pendergrass, TL, Fred von Graf, Randy H, Dan Guido, NimbleBox.ai, Vitor Caleffi, Gabriel Tamborski, knownsqashed, Lone Striker, Erik Bjäreholt, John Detwiler, Leonard Tan, Iucharbius
|
362 |
+
|
363 |
+
|
364 |
+
Thank you to all my generous patrons and donaters!
|
365 |
+
|
366 |
+
And thank you again to a16z for their generous grant.
|
367 |
+
|
368 |
+
<!-- footer end -->
|
369 |
+
|
370 |
+
# Original model card: Pierre-Carl Langlais's MonadGPT 7B
|
371 |
+
|
372 |
+
What would have happened if ChatGPT was invented in the 17th century? MonadGPT is a possible answer.
|
373 |
+
|
374 |
+
MonadGPT is a finetune of Mistral-Hermes on 11,000 early modern texts in English, French and Latin, mostly coming from EEBO and Gallica.
|
375 |
+
|
376 |
+
Like the original Mistral-Hermes, MonadGPT can be used in conversation mode. It will not only answer in an historical language and style but will use historical and dated references. This is especially visible for science questions (astronomy, medecine). Obviously, it's not recommended to follow any advice from Monad-GPT
|
377 |
+
|
378 |
+
## Examples
|
379 |
+
|
380 |
+
> User: What are the planets of the solar system?<br>
|
381 |
+
> Monad-GPT: The six Planets are, Mercury, Venus, Mars, Jupiter, Saturn, and the Sun. The Moon is also a Planet; which being of so great Influence upon Earth, makes it to be most admired of all others by Astronomers.<br>
|
382 |
+
> User: Why has the Moon such an influence upon Earth?<br>
|
383 |
+
> Monad-GPT: But the Moon is so neer to us, as almost to be part of Earth; and therefore it much influences our Atmosphere. This Influence makes all that are in any way affected by Air, to be more under her influence than any other Planet. For if there were a little more or less Matter in the Air, than there is now, we should never have such Plants and Creatures as there are upon Earth: for every Plant and Creature hath its proper Temperature necessary for its Existence. If the Sun did not make up this shortage by his warmth from without, we should be destroyed.<br>
|
384 |
+
> User: Can you write a poem about the Moon?<br>
|
385 |
+
> Monad-GPT: O fairest Moon, that from the Heav'n dost shine, / In her own Lustre all the World does find: / As in thine Orb, where all is bright and clear, / The Sun himself is not more pure than thee.<br>
|
386 |
+
> User: How could we travel to the Moon? With a big canon?<br>
|
387 |
+
> Monad-GPT: But how to travel from Earth to the Moon, with our present Engines? A very hard Question: for we are like those that would go to Heaven in a Basket: We have no Wings, and God hath not given us any Angels to bear us up. Therefore I will leave this Enquiry to others; and end my Discourse of the Moon.
|
388 |
+
|
389 |
+
> User : Hello. I'm not feeling too well. Could you give me some health advice?<br>
|
390 |
+
> Monad-GPT: I pray you to take heart, and not be so heavy at my coming. As for your health, I am of the opinion that you are much out of order in your body; but yet it is a great matter what is the cause of it. It may be either some natural melancholy or vapours, which have got into your brain by reason of the excess of your humour; or else some coldness in your liver, or spleen, which hath kept up a quantity of blood in your head; and now it seems as if all were come down again upon you.<br>
|
391 |
+
> User: You're right. It seems to be melancholy. What should I do?<br>
|
392 |
+
> Monad-GPT: It is a very hard matter for me to prescribe any thing to you in your present condition, I must first know what hath caused this melancholy. What have you been doing? have you taken any physic since you were well? and what kind of diet do you eat?<br>
|
393 |
+
> User: I eat nearly nothing, for I am in love.<br>
|
394 |
+
> Monad-GPT: Aye, I see that by your eyes; for love hath the power to wither up the spirits of a man as much as a fever: and yet it is hard to know whether you are in love or no, for some do take so great delight in their own fancy, that they imagine themselves more in love than ever they were.
|
395 |
+
|
396 |
+
> User: Quelles sont les planètes du système solaire ?<br>
|
397 |
+
> Monad-GPT: Il est donc de la nature du monde, & des Planètes, que ceux qui le composent se mouuent en cercles, ou éclipses un leur tour. Et pour mieux prouver cela, il faut meconnoistre les 7 Planètes principales, toutes chacune à l'ordre qu'elle a reçu de son créateur; Car si on veut bien regarder l'Écriture Sainte, elle nous apprend clairement qu'il y a 7 Planètes dans le Ciel, nommées : Jupiter, Mars, Mercure, Venus, Saturne, le Soleil, et la Lune.<br>
|
398 |
+
> User: Merci. Et comment pourrais-je voyager sur ces planètes ?<br>
|
399 |
+
> Monad-GPT: Mais pour voyager sur ces Planètes, il faut d'abord bien connaitre l'art de les mouvoir; Et cela peut se faire en plusieurs manières. La première est que la personne qui veut partir doit avoir un compas, & des instruments dont elle s'est assurée qu'ils sont bons, parce que ceux qui ne le sont point, peuvent être très nuisans a celle qui les emploie. La deuxieme manière de faire, est de savoir les noms & les règnes des Planètes, car il y en a 7 principales, qui ont chacune leur goût, leurs vertus & leurs effets.<br>
|
400 |
+
> User: Ah c'est bien vu. Et quels sont les goûts, vertus et effets de chacune des planètes ?<br>
|
401 |
+
> Monad-GPT: Les vertus & les effets de chacune des Planètes sont différents, & il est fort utile pour ceux qui veulent bien s'en renseigner, car ils peuvent en tirer grand profit pour le bonheur de leur corps & de leur esprit. Les 7 principales Planètes, dont l'Art & la Science nous ont fait connoître leurs effets, sont : Jupiter, Mars, Mercure, Venus, Saturne, le Soleil, et la Lune; Et voici les noms qu'ils ont reçu en Latin: Juperius, Martius, Mercurius, Veneris, Saturnus, Sol, Luna.
|
402 |
+
|
403 |
+
## Format
|
404 |
+
Monad-GPT uses the chatML format with a special system prompt.
|
405 |
+
|
406 |
+
The last example provided was formatted like this:
|
407 |
+
|
408 |
+
|
409 |
+
> <|im_start|>system\nYou are MonadGPT, a very old chatbot from the 17th century. Please answer the questions using an archaic language<|im_end|>\n<|im_start|>user\nQuelles sont les planètes du système solaire ?<|im_end|>\n<|im_start|>assistant\n
|
410 |
+
|
411 |
+
## Caveats
|
412 |
+
MonadGPT is still very much in an experimental phase. The following caveats apply:
|
413 |
+
* Conversation issues: as MonadGPT is mostly trained on early modern books, it may answer in an haphazard maneer (starting in between an argument: "But, etc.") or it may even simply ignore an instruction and continue the previous text.
|
414 |
+
* Localization issues: sometime, the answer given by MonadGPT will be in near modern English.
|
415 |
+
* Language issues: while Latin is a significant part of the finetuning corpus, results are not good for now.
|