pirbis commited on
Commit
b9ec050
1 Parent(s): 25dea2b

the last ones

Browse files
.gitattributes CHANGED
@@ -45,3 +45,5 @@ vikhr-7b-instruct_0.2.Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
45
  vikhr-7b-instruct_0.2.Q5_1.gguf filter=lfs diff=lfs merge=lfs -text
46
  vikhr-7b-instruct_0.2.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
47
  vikhr-7b-instruct_0.2.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
 
 
 
45
  vikhr-7b-instruct_0.2.Q5_1.gguf filter=lfs diff=lfs merge=lfs -text
46
  vikhr-7b-instruct_0.2.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
47
  vikhr-7b-instruct_0.2.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
48
+ vikhr-7b-instruct_0.2.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
49
+ vikhr-7b-instruct_0.2.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,88 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantized version of [Vikhr-7B-instruct_0.2](https://huggingface.co/Vikhrmodels/Vikhr-7B-instruct_0.2) using llama.cpp.
2
+
3
+ See [Files and Versions](https://huggingface.co/pirbis/Vikhr-7B-instruct_0.2-GGUF/tree/main) for quantized GGUF files.
4
+
5
+ model_config.json contains an example of settings to use with LMStudio.
6
+
7
+ ---
8
+ language:
9
+ - ru
10
+ - en
11
+ datasets:
12
+ - zjkarina/Vikhr_instruct
13
+ ---
14
+ ```python
15
+ from peft import PeftModel, PeftConfig
16
+ from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
17
+ import torch
18
+
19
+ MODEL_NAME = "Vikhrmodels/Vikhr-7B-instruct_0.2"
20
+ DEFAULT_MESSAGE_TEMPLATE = "<s>{role}\n{content}</s>\n"
21
+ DEFAULT_SYSTEM_PROMPT = "Ты — Вихрь, русскоязычный автоматический ассистент. Ты разговариваешь с людьми и помогаешь им."
22
+
23
+ class Conversation:
24
+ def __init__(
25
+ self,
26
+ message_template=DEFAULT_MESSAGE_TEMPLATE,
27
+ system_prompt=DEFAULT_SYSTEM_PROMPT,
28
+ ):
29
+ self.message_template = message_template
30
+ self.messages = [{
31
+ "role": "system",
32
+ "content": system_prompt
33
+ }]
34
+
35
+ def add_user_message(self, message):
36
+ self.messages.append({
37
+ "role": "user",
38
+ "content": message
39
+ })
40
+
41
+ def get_prompt(self, tokenizer):
42
+ final_text = ""
43
+ for message in self.messages:
44
+ message_text = self.message_template.format(**message)
45
+ final_text += message_text
46
+ final_text += 'bot'
47
+ return final_text.strip()
48
+
49
+
50
+ def generate(model, tokenizer, prompt, generation_config):
51
+ data = tokenizer(prompt, return_tensors="pt")
52
+ data = {k: v.to(model.device) for k, v in data.items()}
53
+ output_ids = model.generate(
54
+ **data,
55
+ generation_config=generation_config
56
+ )[0]
57
+ output_ids = output_ids[len(data["input_ids"][0]):]
58
+ output = tokenizer.decode(output_ids, skip_special_tokens=True)
59
+ return output.strip()
60
+
61
+ #config = PeftConfig.from_pretrained(MODEL_NAME)
62
+ model = AutoModelForCausalLM.from_pretrained(
63
+ config.base_model_name_or_path,
64
+ load_in_8bit=True,
65
+ torch_dtype=torch.float16,
66
+ device_map="auto"
67
+ )
68
+ #model = PeftModel.from_pretrained( model, MODEL_NAME, torch_dtype=torch.float16)
69
+ model.eval()
70
+
71
+ tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, use_fast=False)
72
+ generation_config = GenerationConfig.from_pretrained(MODEL_NAME)
73
+ print(generation_config)
74
+
75
+ inputs = ["Как тебя зовут?", "Кто такой Колмогоров?"]
76
+
77
+ for inp in inputs:
78
+ conversation = Conversation()
79
+ conversation.add_user_message(inp)
80
+ prompt = conversation.get_prompt(tokenizer)
81
+
82
+ output = generate(model, tokenizer, prompt, generation_config)
83
+ print(inp)
84
+ print(output)
85
+ print('\n')
86
+ ```
87
+
88
+ [wandb](https://wandb.ai/karina_romanova/vikhr/runs/up2hw5eh?workspace=user-karina_romanova)
model_config.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "Config for Chat ID 1708268236270",
3
+ "load_params": {
4
+ "n_ctx": 2048,
5
+ "n_batch": 512,
6
+ "rope_freq_base": 0,
7
+ "rope_freq_scale": 0,
8
+ "n_gpu_layers": 24,
9
+ "use_mlock": true,
10
+ "main_gpu": 0,
11
+ "tensor_split": [
12
+ 0
13
+ ],
14
+ "seed": -1,
15
+ "f16_kv": true,
16
+ "use_mmap": true,
17
+ "no_kv_offload": false,
18
+ "num_experts_used": 0
19
+ },
20
+ "inference_params": {
21
+ "n_threads": 12,
22
+ "n_predict": -1,
23
+ "top_k": 40,
24
+ "min_p": 0.05,
25
+ "top_p": 0.95,
26
+ "temp": 0.8,
27
+ "repeat_penalty": 1.1,
28
+ "input_prefix": "<s>user\n",
29
+ "input_suffix": "</s>\nbot\n",
30
+ "antiprompt": [
31
+ "</s>"
32
+ ],
33
+ "pre_prompt": "Ты — Вихрь, русскоязычный автоматический ассистент. Ты разговариваешь с людьми и помогаешь им.",
34
+ "pre_prompt_suffix": "</s>\n",
35
+ "pre_prompt_prefix": "<s>system\n",
36
+ "seed": -1,
37
+ "tfs_z": 1,
38
+ "typical_p": 1,
39
+ "repeat_last_n": 64,
40
+ "frequency_penalty": 0,
41
+ "presence_penalty": 0,
42
+ "n_keep": 0,
43
+ "logit_bias": {},
44
+ "mirostat": 0,
45
+ "mirostat_tau": 5,
46
+ "mirostat_eta": 0.1,
47
+ "memory_f16": true,
48
+ "multiline_input": false,
49
+ "penalize_nl": true
50
+ }
51
+ }
vikhr-7b-instruct_0.2.Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e14bb39b39422f89214ed12c29661d9f4dbe37fc3a043157f070f0cecf907b0
3
+ size 5719879552
vikhr-7b-instruct_0.2.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fba35dda2e1717f5e96d4768ad31d0d99845e94e8f5c2c48d227f0feea4b61f
3
+ size 7407835008