RichardErkhov commited on
Commit
60f6947
1 Parent(s): b5a72af

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +73 -0
README.md ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ Mistral-7B-v0.1-sharded - bnb 8bits
11
+ - Model creator: https://huggingface.co/alexsherstinsky/
12
+ - Original model: https://huggingface.co/alexsherstinsky/Mistral-7B-v0.1-sharded/
13
+
14
+
15
+
16
+
17
+ Original model description:
18
+ ---
19
+ license: apache-2.0
20
+ pipeline_tag: text-generation
21
+ tags:
22
+ - pretrained
23
+ inference:
24
+ parameters:
25
+ temperature: 0.7
26
+ ---
27
+
28
+ # Note: Sharded Version of the Original "Mistral 7B" Model
29
+
30
+ This is just a version of https://huggingface.co/mistralai/Mistral-7B-v0.1 which is sharded to 2GB maximum parts in order to reduce the RAM required when loading.
31
+
32
+ # Model Card for Mistral-7B-v0.1
33
+
34
+ The Mistral-7B-v0.1 Large Language Model (LLM) is a pretrained generative text model with 7 billion parameters.
35
+ Mistral-7B-v0.1 outperforms Llama 2 13B on all benchmarks we tested.
36
+
37
+ For full details of this model please read our [Release blog post](https://mistral.ai/news/announcing-mistral-7b/)
38
+
39
+ ## Model Architecture
40
+ Mistral-7B-v0.1 is a transformer model, with the following architecture choices:
41
+ - Grouped-Query Attention
42
+ - Sliding-Window Attention
43
+ - Byte-fallback BPE tokenizer
44
+
45
+ ## Troubleshooting
46
+ - If you see the following error:
47
+ ```
48
+ Traceback (most recent call last):
49
+ File "", line 1, in
50
+ File "/transformers/models/auto/auto_factory.py", line 482, in from_pretrained
51
+ config, kwargs = AutoConfig.from_pretrained(
52
+ File "/transformers/models/auto/configuration_auto.py", line 1022, in from_pretrained
53
+ config_class = CONFIG_MAPPING[config_dict["model_type"]]
54
+ File "/transformers/models/auto/configuration_auto.py", line 723, in getitem
55
+ raise KeyError(key)
56
+ KeyError: 'mistral'
57
+ ```
58
+
59
+ Installing transformers from source should solve the issue:
60
+ ```
61
+ pip install git+https://github.com/huggingface/transformers
62
+ ```
63
+ This should not be required after transformers-v4.33.4.
64
+
65
+
66
+ ## Notice
67
+
68
+ Mistral 7B is a pretrained base model and therefore does not have any moderation mechanisms.
69
+
70
+ ## The Mistral AI Team
71
+
72
+ Albert Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lélio Renard Lavaud, Lucile Saulnier, Marie-Anne Lachaux, Pierre Stock, Teven Le Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix, William El Sayed.
73
+