Spaces:
Running
Running
Upload 3 files
Browse files- README.md +14 -14
- app.py +1 -1
- multit2i.py +1 -1
README.md
CHANGED
@@ -1,14 +1,14 @@
|
|
1 |
-
---
|
2 |
-
title: Free Multi Models Text-to-Image Demo V3
|
3 |
-
emoji: ππ
|
4 |
-
colorFrom: blue
|
5 |
-
colorTo: purple
|
6 |
-
sdk: gradio
|
7 |
-
sdk_version: 5.
|
8 |
-
app_file: app.py
|
9 |
-
short_description: Text-to-Image
|
10 |
-
license: mit
|
11 |
-
pinned: true
|
12 |
-
---
|
13 |
-
|
14 |
-
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
|
|
|
1 |
+
---
|
2 |
+
title: Free Multi Models Text-to-Image Demo V3
|
3 |
+
emoji: ππ
|
4 |
+
colorFrom: blue
|
5 |
+
colorTo: purple
|
6 |
+
sdk: gradio
|
7 |
+
sdk_version: 5.9.1
|
8 |
+
app_file: app.py
|
9 |
+
short_description: Text-to-Image
|
10 |
+
license: mit
|
11 |
+
pinned: true
|
12 |
+
---
|
13 |
+
|
14 |
+
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
|
app.py
CHANGED
@@ -131,5 +131,5 @@ with gr.Blocks(theme="NoCrypt/miku@>=1.2.2", fill_width=True, css=css) as demo:
|
|
131 |
.then(translate_to_en, [neg_prompt], [neg_prompt], queue=False, show_api=False)
|
132 |
|
133 |
#demo.queue(default_concurrency_limit=200, max_size=200)
|
134 |
-
demo.launch(max_threads=400)
|
135 |
# https://github.com/gradio-app/gradio/issues/6339
|
|
|
131 |
.then(translate_to_en, [neg_prompt], [neg_prompt], queue=False, show_api=False)
|
132 |
|
133 |
#demo.queue(default_concurrency_limit=200, max_size=200)
|
134 |
+
demo.launch(max_threads=400, ssr_mode=False)
|
135 |
# https://github.com/gradio-app/gradio/issues/6339
|
multit2i.py
CHANGED
@@ -69,7 +69,7 @@ def find_model_list(author: str="", tags: list[str]=[], not_tag="", sort: str="l
|
|
69 |
for model in model_infos:
|
70 |
if not model.private and not model.gated or HF_TOKEN is not None:
|
71 |
loadable = is_loadable(model.id, force_gpu) if check_status else True
|
72 |
-
if not_tag and not_tag in model.tags or not loadable: continue
|
73 |
models.append(model.id)
|
74 |
if len(models) == limit: break
|
75 |
return models
|
|
|
69 |
for model in model_infos:
|
70 |
if not model.private and not model.gated or HF_TOKEN is not None:
|
71 |
loadable = is_loadable(model.id, force_gpu) if check_status else True
|
72 |
+
if not_tag and not_tag in model.tags or not loadable or "not-for-all-audiences" in model.tags: continue
|
73 |
models.append(model.id)
|
74 |
if len(models) == limit: break
|
75 |
return models
|