Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -91,8 +91,8 @@ def progress_bar_html(label: str) -> str:
|
|
| 91 |
return f'''
|
| 92 |
<div style="display: flex; align-items: center;">
|
| 93 |
<span style="margin-right: 10px; font-size: 14px;">{label}</span>
|
| 94 |
-
<div style="width: 110px; height: 5px; background-color: #
|
| 95 |
-
<div style="width: 100%; height: 100%; background-color: #
|
| 96 |
</div>
|
| 97 |
</div>
|
| 98 |
<style>
|
|
@@ -104,7 +104,7 @@ def progress_bar_html(label: str) -> str:
|
|
| 104 |
'''
|
| 105 |
|
| 106 |
|
| 107 |
-
@spaces.GPU
|
| 108 |
def generate(input_dict: dict, chat_history: list[dict],
|
| 109 |
max_new_tokens: int = 1024,
|
| 110 |
temperature: float = 0.6,
|
|
@@ -160,7 +160,7 @@ def generate(input_dict: dict, chat_history: list[dict],
|
|
| 160 |
thread = Thread(target=model_m.generate, kwargs=generation_kwargs)
|
| 161 |
thread.start()
|
| 162 |
buffer = ""
|
| 163 |
-
yield progress_bar_html("Processing video with
|
| 164 |
for new_text in streamer:
|
| 165 |
buffer += new_text
|
| 166 |
buffer = buffer.replace("<|im_end|>", "")
|
|
@@ -190,7 +190,7 @@ def generate(input_dict: dict, chat_history: list[dict],
|
|
| 190 |
thread = Thread(target=model_m.generate, kwargs=generation_kwargs)
|
| 191 |
thread.start()
|
| 192 |
buffer = ""
|
| 193 |
-
yield progress_bar_html("
|
| 194 |
for new_text in streamer:
|
| 195 |
buffer += new_text
|
| 196 |
buffer = buffer.replace("<|im_end|>", "")
|
|
@@ -219,7 +219,7 @@ def generate(input_dict: dict, chat_history: list[dict],
|
|
| 219 |
t = Thread(target=model.generate, kwargs=generation_kwargs)
|
| 220 |
t.start()
|
| 221 |
outputs = []
|
| 222 |
-
yield progress_bar_html("Processing
|
| 223 |
for new_text in streamer:
|
| 224 |
outputs.append(new_text)
|
| 225 |
yield "".join(outputs)
|
|
|
|
| 91 |
return f'''
|
| 92 |
<div style="display: flex; align-items: center;">
|
| 93 |
<span style="margin-right: 10px; font-size: 14px;">{label}</span>
|
| 94 |
+
<div style="width: 110px; height: 5px; background-color: #DC143C; border-radius: 2px; overflow: hidden;">
|
| 95 |
+
<div style="width: 100%; height: 100%; background-color: #FF0000; animation: loading 1.5s linear infinite;"></div>
|
| 96 |
</div>
|
| 97 |
</div>
|
| 98 |
<style>
|
|
|
|
| 104 |
'''
|
| 105 |
|
| 106 |
|
| 107 |
+
@spaces.GPU
|
| 108 |
def generate(input_dict: dict, chat_history: list[dict],
|
| 109 |
max_new_tokens: int = 1024,
|
| 110 |
temperature: float = 0.6,
|
|
|
|
| 160 |
thread = Thread(target=model_m.generate, kwargs=generation_kwargs)
|
| 161 |
thread.start()
|
| 162 |
buffer = ""
|
| 163 |
+
yield progress_bar_html("Processing video with Callisto OCR3")
|
| 164 |
for new_text in streamer:
|
| 165 |
buffer += new_text
|
| 166 |
buffer = buffer.replace("<|im_end|>", "")
|
|
|
|
| 190 |
thread = Thread(target=model_m.generate, kwargs=generation_kwargs)
|
| 191 |
thread.start()
|
| 192 |
buffer = ""
|
| 193 |
+
yield progress_bar_html("Processing image with Callisto OCR3")
|
| 194 |
for new_text in streamer:
|
| 195 |
buffer += new_text
|
| 196 |
buffer = buffer.replace("<|im_end|>", "")
|
|
|
|
| 219 |
t = Thread(target=model.generate, kwargs=generation_kwargs)
|
| 220 |
t.start()
|
| 221 |
outputs = []
|
| 222 |
+
yield progress_bar_html("Processing With Pocket Llama 3B")
|
| 223 |
for new_text in streamer:
|
| 224 |
outputs.append(new_text)
|
| 225 |
yield "".join(outputs)
|