alokabhishek commited on
Commit
ea26a32
1 Parent(s): e76b6f3

Updated Readme

Browse files
Files changed (1) hide show
  1. README.md +129 -6
README.md CHANGED
@@ -1,15 +1,138 @@
1
  ---
 
2
  license: apache-2.0
3
  pipeline_tag: text-generation
4
  tags:
5
- - finetuned
6
- inference: true
7
- widget:
8
- - messages:
9
- - role: user
10
- content: What is your favorite condiment?
 
11
  ---
12
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  # Model Card for Mistral-7B-Instruct-v0.2
14
 
15
  The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an instruct fine-tuned version of the Mistral-7B-v0.2.
 
1
  ---
2
+ library_name: transformers
3
  license: apache-2.0
4
  pipeline_tag: text-generation
5
  tags:
6
+ - ExLlamaV2
7
+ - 4bit
8
+ - Mistral
9
+ - Mistral-7B
10
+ - quantized
11
+ - exl2
12
+ - 4.0-bpw
13
  ---
14
 
15
+ # Model Card for alokabhishek/Mistral-7B-Instruct-v0.2-5.0-bpw-exl2
16
+
17
+ <!-- Provide a quick summary of what the model is/does. -->
18
+ This repo contains 4-bit quantized (using ExLlamaV2) model Mistral AI_'s Mistral-7B-Instruct-v0.2
19
+
20
+
21
+
22
+ ## Model Details
23
+
24
+ - Model creator: [Mistral AI_](https://huggingface.co/mistralai)
25
+ - Original model: [Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2)
26
+
27
+
28
+ ### About 4 bit quantization using ExLlamaV2
29
+
30
+
31
+ - ExLlamaV2 github repo: [ExLlamaV2 github repo](https://github.com/turboderp/exllamav2)
32
+
33
+
34
+
35
+ # How to Get Started with the Model
36
+
37
+ Use the code below to get started with the model.
38
+
39
+
40
+ ## How to run from Python code
41
+
42
+ #### First install the package
43
+ ```shell
44
+ # Install ExLLamaV2
45
+ !git clone https://github.com/turboderp/exllamav2
46
+ !pip install -e exllamav2
47
+ ```
48
+
49
+ #### Import
50
+
51
+ ```python
52
+ from huggingface_hub import login, HfApi, create_repo
53
+ from torch import bfloat16
54
+ import locale
55
+ import torch
56
+ import os
57
+ ```
58
+
59
+ #### set up variables
60
+
61
+ ```python
62
+ # Define the model ID for the desired model
63
+ model_id = "alokabhishek/Mistral-7B-Instruct-v0.2-5.0-bpw-exl2"
64
+ BPW = 5.0
65
+
66
+ # define variables
67
+ model_name = model_id.split("/")[-1]
68
+
69
+ ```
70
+
71
+ #### Download the quantized model
72
+ ```shell
73
+ !git-lfs install
74
+ # download the model to loacl directory
75
+ !git clone https://{username}:{HF_TOKEN}@huggingface.co/{model_id} {model_name}
76
+ ```
77
+
78
+ #### Run Inference on quantized model using
79
+ ```shell
80
+ # Run model
81
+ !python exllamav2/test_inference.py -m {model_name}/ -p "Tell me a funny joke about Large Language Models meeting a Blackhole in an intergalactic Bar."
82
+ ```
83
+
84
+
85
+ ## Uses
86
+
87
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
88
+
89
+ ### Direct Use
90
+
91
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
92
+
93
+ [More Information Needed]
94
+
95
+
96
+ ### Out-of-Scope Use
97
+
98
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
99
+
100
+ [More Information Needed]
101
+
102
+ ## Bias, Risks, and Limitations
103
+
104
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
105
+
106
+ [More Information Needed]
107
+
108
+
109
+ ## Evaluation
110
+
111
+ <!-- This section describes the evaluation protocols and provides the results. -->
112
+
113
+
114
+ #### Metrics
115
+
116
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
117
+
118
+ [More Information Needed]
119
+
120
+ ### Results
121
+
122
+ [More Information Needed]
123
+
124
+
125
+ ## Model Card Authors [optional]
126
+
127
+ [More Information Needed]
128
+
129
+ ## Model Card Contact
130
+
131
+ [More Information Needed]
132
+
133
+
134
+
135
+
136
  # Model Card for Mistral-7B-Instruct-v0.2
137
 
138
  The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an instruct fine-tuned version of the Mistral-7B-v0.2.