Update app.py
Browse files
app.py
CHANGED
@@ -9,6 +9,7 @@ import torch
|
|
9 |
REPO_NAME = 'schuler/experimental-JP47D21-KPhi-3-micro-4k-instruct'
|
10 |
|
11 |
# How to cache?
|
|
|
12 |
def load_model(repo_name):
|
13 |
tokenizer = AutoTokenizer.from_pretrained(repo_name, trust_remote_code=True)
|
14 |
generator_conf = GenerationConfig.from_pretrained(repo_name)
|
@@ -20,10 +21,12 @@ tokenizer, generator_conf, model = load_model(REPO_NAME)
|
|
20 |
|
21 |
global_error = ''
|
22 |
try:
|
|
|
23 |
generator = pipeline("text-generation", model=model, tokenizer=tokenizer)
|
24 |
except Exception as e:
|
25 |
global_error = f"Failed to load model: {str(e)}"
|
26 |
|
|
|
27 |
def respond(
|
28 |
message,
|
29 |
history: list[tuple[str, str]],
|
|
|
9 |
REPO_NAME = 'schuler/experimental-JP47D21-KPhi-3-micro-4k-instruct'
|
10 |
|
11 |
# How to cache?
|
12 |
+
@spaces.GPU()
|
13 |
def load_model(repo_name):
|
14 |
tokenizer = AutoTokenizer.from_pretrained(repo_name, trust_remote_code=True)
|
15 |
generator_conf = GenerationConfig.from_pretrained(repo_name)
|
|
|
21 |
|
22 |
global_error = ''
|
23 |
try:
|
24 |
+
@spaces.GPU()
|
25 |
generator = pipeline("text-generation", model=model, tokenizer=tokenizer)
|
26 |
except Exception as e:
|
27 |
global_error = f"Failed to load model: {str(e)}"
|
28 |
|
29 |
+
@spaces.GPU()
|
30 |
def respond(
|
31 |
message,
|
32 |
history: list[tuple[str, str]],
|