letao670982 commited on
Commit
2659fc4
·
verified ·
1 Parent(s): a19357f

Upload 2 files

Browse files
Files changed (2) hide show
  1. app.py +27 -0
  2. requirement.txt +5 -0
app.py ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ from transformers import AutoTokenizer, AutoModelForCausalLM, TextStreamer
3
+ from threading import Thread
4
+
5
+ tokenizer = AutoTokenizer.from_pretrained("deepseek-ai/deepseek-coder-1.3b-instruct", trust_remote_code=True)
6
+ model = AutoModelForCausalLM.from_pretrained("deepseek-ai/deepseek-coder-1.3b-instruct", trust_remote_code=True, device_map="auto")
7
+
8
+ streamer = TextStreamer(tokenizer, skip_prompt=True)
9
+
10
+ #streaming output
11
+ def respond(message, history):
12
+ messages=[
13
+ { 'role': 'user', 'content': message}]
14
+ inputs = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt").to(model.device)
15
+ # 32021 is the id of <|EOT|> token
16
+ outputs = model.generate(inputs, streamer=streamer, max_new_tokens=1024, do_sample=False, top_k=50, top_p=0.95, num_return_sequences=1, eos_token_id=32021)
17
+ out_answer = tokenizer.decode(outputs[0][len(inputs[0]):], skip_special_tokens=True)
18
+ generated_text = ""
19
+ for new_text in out_answer:
20
+ generated_text += new_text
21
+ yield generated_text
22
+
23
+
24
+ demo_chatbot = gr.ChatInterface(respond, title="Deepseek-Coder", description="Enter text to start chatting.")
25
+
26
+ demo_chatbot.launch()
27
+ #https://huggingface.co/docs/transformers/internal/generation_utils
requirement.txt ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ accelerate
2
+ gradio
3
+ huggingface-hub
4
+ torch
5
+ transformers