Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -11,15 +11,11 @@ checkpoint = "microsoft/phi-2"
|
|
11 |
tokenizer = AutoTokenizer.from_pretrained(checkpoint, trust_remote_code=True)
|
12 |
model = AutoModelForCausalLM.from_pretrained(checkpoint, torch_dtype=torch.float32, device_map="cpu", trust_remote_code=True)
|
13 |
|
14 |
-
# Streamer
|
15 |
-
streamer = TextIteratorStreamer(tokenizer=tokenizer, skip_prompt=True)
|
16 |
-
|
17 |
# Text generation pipeline
|
18 |
phi2 = pipeline(
|
19 |
"text-generation",
|
20 |
tokenizer=tokenizer,
|
21 |
-
model=model,
|
22 |
-
streamer=streamer,
|
23 |
pad_token_id=tokenizer.eos_token_id,
|
24 |
eos_token_id=tokenizer.eos_token_id,
|
25 |
device_map="cpu"
|
@@ -38,7 +34,9 @@ def generate(prompt, chat_history, max_new_tokens):
|
|
38 |
final_prompt += "User: " + prompt + "\n"
|
39 |
final_prompt += "Output:"
|
40 |
|
41 |
-
|
|
|
|
|
42 |
thread.start()
|
43 |
|
44 |
generated_text = ""
|
|
|
11 |
tokenizer = AutoTokenizer.from_pretrained(checkpoint, trust_remote_code=True)
|
12 |
model = AutoModelForCausalLM.from_pretrained(checkpoint, torch_dtype=torch.float32, device_map="cpu", trust_remote_code=True)
|
13 |
|
|
|
|
|
|
|
14 |
# Text generation pipeline
|
15 |
phi2 = pipeline(
|
16 |
"text-generation",
|
17 |
tokenizer=tokenizer,
|
18 |
+
model=model,
|
|
|
19 |
pad_token_id=tokenizer.eos_token_id,
|
20 |
eos_token_id=tokenizer.eos_token_id,
|
21 |
device_map="cpu"
|
|
|
34 |
final_prompt += "User: " + prompt + "\n"
|
35 |
final_prompt += "Output:"
|
36 |
|
37 |
+
# Streamer
|
38 |
+
streamer = TextIteratorStreamer(tokenizer=tokenizer, skip_prompt=True)
|
39 |
+
thread = Thread(target=phi2, kwargs={"text_inputs":final_prompt, "max_new_tokens":max_new_tokens, "streamer":streamer})
|
40 |
thread.start()
|
41 |
|
42 |
generated_text = ""
|