Lamimad commited on
Commit
443913f
1 Parent(s): 58cc909

Create Readme.md

Browse files
Files changed (1) hide show
  1. Readme.md +53 -0
Readme.md ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ pipeline_tag: text-generation
4
+ tags:
5
+ - finetuned
6
+ inference:
7
+ parameters:
8
+ temperature: 0.9
9
+ ---
10
+ # Model Card for Luna-standard-0.0.1
11
+
12
+ The Luna-standard-0.0.1 Large Language Model (LLM) is a instruct fine-tuned version of the [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) generative text model using a variety of publicly available conversation datasets.
13
+
14
+ For full details of this model please read our [paper](https://arxiv.org/abs/2310.06825) and [release blog post](https://mistral.ai/news/announcing-mistral-7b/).
15
+
16
+ ## Instruction format
17
+
18
+ In order to leverage instruction fine-tuning, your prompt should be surrounded by `[INST]` and `[/INST]` tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id.
19
+
20
+ E.g.
21
+ ```
22
+ text = "<s>[INST] What is your favourite condiment? [/INST]"
23
+ "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!</s> "
24
+ "[INST] Do you have mayonnaise recipes? [/INST]"
25
+ ```
26
+
27
+ This format is available as a [chat template](https://huggingface.co/docs/transformers/main/chat_templating) via the `apply_chat_template()` method:
28
+
29
+ ```python
30
+ from transformers import AutoModelForCausalLM, AutoTokenizer
31
+ device = "cuda" # the device to load the model onto
32
+ model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-Instruct-v0.1")
33
+ tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.1")
34
+ messages = [
35
+ {"role": "user", "content": "What is your favourite condiment?"},
36
+ {"role": "assistant", "content": "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!"},
37
+ {"role": "user", "content": "Do you have mayonnaise recipes?"}
38
+ ]
39
+ encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt")
40
+ model_inputs = encodeds.to(device)
41
+ model.to(device)
42
+ generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True)
43
+ decoded = tokenizer.batch_decode(generated_ids)
44
+ print(decoded[0])
45
+ ```
46
+
47
+ ## Model Architecture
48
+ This instruction model is based on Mistral-7B-v0.1, a transformer model with the following architecture choices:
49
+ - Grouped-Query Attention
50
+ - Sliding-Window Attention
51
+ - Byte-fallback BPE tokenizer
52
+
53
+