AlexWortega commited on
Commit
7b0388d
1 Parent(s): 4cc5d0d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -3
README.md CHANGED
@@ -13,7 +13,8 @@ GGUF версия: https://huggingface.co/pirbis/Vikhr-7B-instruct_0.2-GGUF
13
  from peft import PeftModel, PeftConfig
14
  from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
15
  import torch
16
-
 
17
  MODEL_NAME = "Vikhrmodels/Vikhr-7B-instruct_0.2"
18
  DEFAULT_MESSAGE_TEMPLATE = "<s>{role}\n{content}</s>\n"
19
  DEFAULT_SYSTEM_PROMPT = "Ты — Вихрь, русскоязычный автоматический ассистент. Ты разговариваешь с людьми и помогаешь им."
@@ -55,10 +56,10 @@ def generate(model, tokenizer, prompt, generation_config):
55
  output_ids = output_ids[len(data["input_ids"][0]):]
56
  output = tokenizer.decode(output_ids, skip_special_tokens=True)
57
  return output.strip()
58
-
59
  #config = PeftConfig.from_pretrained(MODEL_NAME)
60
  model = AutoModelForCausalLM.from_pretrained(
61
- config.base_model_name_or_path,
62
  load_in_8bit=True,
63
  torch_dtype=torch.float16,
64
  device_map="auto"
@@ -67,7 +68,12 @@ model = AutoModelForCausalLM.from_pretrained(
67
  model.eval()
68
 
69
  tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, use_fast=False)
 
70
  generation_config = GenerationConfig.from_pretrained(MODEL_NAME)
 
 
 
 
71
  print(generation_config)
72
 
73
  inputs = ["Как тебя зовут?", "Кто такой Колмогоров?"]
 
13
  from peft import PeftModel, PeftConfig
14
  from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
15
  import torch
16
+ import os
17
+ os.environ['HF_HOME']='.'
18
  MODEL_NAME = "Vikhrmodels/Vikhr-7B-instruct_0.2"
19
  DEFAULT_MESSAGE_TEMPLATE = "<s>{role}\n{content}</s>\n"
20
  DEFAULT_SYSTEM_PROMPT = "Ты — Вихрь, русскоязычный автоматический ассистент. Ты разговариваешь с людьми и помогаешь им."
 
56
  output_ids = output_ids[len(data["input_ids"][0]):]
57
  output = tokenizer.decode(output_ids, skip_special_tokens=True)
58
  return output.strip()
59
+
60
  #config = PeftConfig.from_pretrained(MODEL_NAME)
61
  model = AutoModelForCausalLM.from_pretrained(
62
+ MODEL_NAME,
63
  load_in_8bit=True,
64
  torch_dtype=torch.float16,
65
  device_map="auto"
 
68
  model.eval()
69
 
70
  tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, use_fast=False)
71
+
72
  generation_config = GenerationConfig.from_pretrained(MODEL_NAME)
73
+ generation_config.max_length=256
74
+ generation_config.top_p=0.9
75
+ generation_config.top_k=30
76
+ generation_config.do_sample = True
77
  print(generation_config)
78
 
79
  inputs = ["Как тебя зовут?", "Кто такой Колмогоров?"]