Spaces:
Running
on
Zero
Running
on
Zero
update template
Browse files- app/src/base_model_template.py +14 -14
- app/src/vlm_template.py +16 -16
app/src/base_model_template.py
CHANGED
@@ -21,20 +21,20 @@ brushnet = BrushNetModel.from_pretrained(brushnet_path, torch_dtype=torch_dtype)
|
|
21 |
|
22 |
|
23 |
base_models_list = [
|
24 |
-
{
|
25 |
-
|
26 |
-
|
27 |
-
|
28 |
-
|
29 |
-
|
30 |
-
},
|
31 |
-
{
|
32 |
-
|
33 |
-
|
34 |
-
|
35 |
-
|
36 |
-
|
37 |
-
},
|
38 |
{
|
39 |
"name": "henmixReal (Preload)",
|
40 |
"local_path": "models/base_model/henmixReal_v5c",
|
|
|
21 |
|
22 |
|
23 |
base_models_list = [
|
24 |
+
# {
|
25 |
+
# "name": "dreamshaper_8 (Preload)",
|
26 |
+
# "local_path": "models/base_model/dreamshaper_8",
|
27 |
+
# "pipe": StableDiffusionBrushNetPipeline.from_pretrained(
|
28 |
+
# "models/base_model/dreamshaper_8", brushnet=brushnet, torch_dtype=torch_dtype, low_cpu_mem_usage=False
|
29 |
+
# ).to(device)
|
30 |
+
# },
|
31 |
+
# {
|
32 |
+
# "name": "epicrealism (Preload)",
|
33 |
+
# "local_path": "models/base_model/epicrealism_naturalSinRC1VAE",
|
34 |
+
# "pipe": StableDiffusionBrushNetPipeline.from_pretrained(
|
35 |
+
# "models/base_model/epicrealism_naturalSinRC1VAE", brushnet=brushnet, torch_dtype=torch_dtype, low_cpu_mem_usage=False
|
36 |
+
# ).to(device)
|
37 |
+
# },
|
38 |
{
|
39 |
"name": "henmixReal (Preload)",
|
40 |
"local_path": "models/base_model/henmixReal_v5c",
|
app/src/vlm_template.py
CHANGED
@@ -28,22 +28,22 @@ vlms_list = [
|
|
28 |
# "llava-hf/llava-v1.6-mistral-7b-hf", torch_dtype=torch_dtype, device_map=device
|
29 |
# ).to("cpu"),
|
30 |
# },
|
31 |
-
{
|
32 |
-
|
33 |
-
|
34 |
-
|
35 |
-
|
36 |
-
|
37 |
-
|
38 |
-
|
39 |
-
|
40 |
-
|
41 |
-
|
42 |
-
|
43 |
-
|
44 |
-
|
45 |
-
|
46 |
-
},
|
47 |
# {
|
48 |
# "type": "llava-next",
|
49 |
# "name": "llava-v1.6-vicuna-13b-hf",
|
|
|
28 |
# "llava-hf/llava-v1.6-mistral-7b-hf", torch_dtype=torch_dtype, device_map=device
|
29 |
# ).to("cpu"),
|
30 |
# },
|
31 |
+
# {
|
32 |
+
# "type": "llava-next",
|
33 |
+
# "name": "llama3-llava-next-8b-hf (Preload)",
|
34 |
+
# "local_path": "models/vlms/llama3-llava-next-8b-hf",
|
35 |
+
# "processor": LlavaNextProcessor.from_pretrained(
|
36 |
+
# "models/vlms/llama3-llava-next-8b-hf"
|
37 |
+
# ) if os.path.exists("models/vlms/llama3-llava-next-8b-hf") else LlavaNextProcessor.from_pretrained(
|
38 |
+
# "llava-hf/llama3-llava-next-8b-hf"
|
39 |
+
# ),
|
40 |
+
# "model": LlavaNextForConditionalGeneration.from_pretrained(
|
41 |
+
# "models/vlms/llama3-llava-next-8b-hf", torch_dtype=torch_dtype, device_map=device
|
42 |
+
# ).to("cpu") if os.path.exists("models/vlms/llama3-llava-next-8b-hf") else
|
43 |
+
# LlavaNextForConditionalGeneration.from_pretrained(
|
44 |
+
# "llava-hf/llama3-llava-next-8b-hf", torch_dtype=torch_dtype, device_map=device
|
45 |
+
# ).to("cpu"),
|
46 |
+
# },
|
47 |
# {
|
48 |
# "type": "llava-next",
|
49 |
# "name": "llava-v1.6-vicuna-13b-hf",
|