PengQu commited on
Commit
8a863c0
1 Parent(s): 59200e7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -28,10 +28,11 @@ open_llama_7b_v2_vicuna_Chinese is a chat model supervised finetuned on vicuna s
28
  from transformers import AutoTokenizer, AutoModelForCausalLM
29
 
30
  tokenizer = AutoTokenizer.from_pretrained("PengQu/open_llama_7b_v2_vicuna_Chinese",use_fast=False)
31
- model = AutoModelForCausalLM.from_pretrained("PengQu/open_llama_7b_v2_vicuna_Chinese").to("cuda:6")
32
 
33
- prompt = '用flask写一个简单的http服务器。'
34
- input_ids = tokenizer(prompt, return_tensors="pt").input_ids.to("cuda:6")
 
35
 
36
  generation_output = model.generate(input_ids=input_ids, max_new_tokens=512)
37
  print(tokenizer.decode(generation_output[0],skip_special_tokens=True))
 
28
  from transformers import AutoTokenizer, AutoModelForCausalLM
29
 
30
  tokenizer = AutoTokenizer.from_pretrained("PengQu/open_llama_7b_v2_vicuna_Chinese",use_fast=False)
31
+ model = AutoModelForCausalLM.from_pretrained("PengQu/open_llama_7b_v2_vicuna_Chinese").to("cuda")
32
 
33
+ instruction = "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: {} ASSISTANT:"
34
+ prompt = instruction.format('用flask写一个简单的http服务器。')
35
+ input_ids = tokenizer(prompt, return_tensors="pt").input_ids.to("cuda")
36
 
37
  generation_output = model.generate(input_ids=input_ids, max_new_tokens=512)
38
  print(tokenizer.decode(generation_output[0],skip_special_tokens=True))