Spaces:
Running
on
Zero
Running
on
Zero
Upload 24 files
Browse files- README.md +1 -1
- app.py +12 -9
- constants.py +49 -23
- dc.py +79 -44
- env.py +1 -0
- image_processor.py +2 -2
- llmdolphin.py +44 -46
- llmenv.py +742 -0
- lora_dict.json +14 -0
- modutils.py +6 -6
- packages.txt +1 -1
- requirements.txt +9 -9
- tagger/fl2sd3longcap.py +12 -6
- tagger/tagger.py +1 -1
- tagger/v2.py +4 -5
- utils.py +181 -104
README.md
CHANGED
|
@@ -4,7 +4,7 @@ emoji: 🖼🖼️📦
|
|
| 4 |
colorFrom: purple
|
| 5 |
colorTo: red
|
| 6 |
sdk: gradio
|
| 7 |
-
sdk_version:
|
| 8 |
app_file: app.py
|
| 9 |
license: mit
|
| 10 |
short_description: Text-to-Image
|
|
|
|
| 4 |
colorFrom: purple
|
| 5 |
colorTo: red
|
| 6 |
sdk: gradio
|
| 7 |
+
sdk_version: 5.44.1
|
| 8 |
app_file: app.py
|
| 9 |
license: mit
|
| 10 |
short_description: Text-to-Image
|
app.py
CHANGED
|
@@ -1,6 +1,7 @@
|
|
| 1 |
import spaces
|
| 2 |
import gradio as gr
|
| 3 |
import numpy as np
|
|
|
|
| 4 |
|
| 5 |
# DiffuseCraft
|
| 6 |
from dc import (infer, _infer, pass_result, get_diffusers_model_list, get_samplers, save_image_history,
|
|
@@ -35,7 +36,7 @@ def description_ui():
|
|
| 35 |
"""
|
| 36 |
)
|
| 37 |
|
| 38 |
-
|
| 39 |
MAX_SEED = np.iinfo(np.int32).max
|
| 40 |
MAX_IMAGE_SIZE = 4096
|
| 41 |
MIN_IMAGE_SIZE = 256
|
|
@@ -51,7 +52,7 @@ css = """
|
|
| 51 |
.image { margin: 0px auto; object-fit: contain; }
|
| 52 |
"""
|
| 53 |
|
| 54 |
-
with gr.Blocks(fill_width=True, elem_id="container", css=css
|
| 55 |
gr.Markdown("# Votepurchase Multiple Model", elem_classes="title")
|
| 56 |
state = gr.State(value={})
|
| 57 |
with gr.Tab("Image Generator"):
|
|
@@ -124,7 +125,8 @@ with gr.Blocks(fill_width=True, elem_id="container", css=css, delete_cache=(60,
|
|
| 124 |
return gr.Dropdown(label=label, choices=get_all_lora_tupled_list(), value="", allow_custom_value=True, elem_classes="lora", min_width=320, visible=visible)
|
| 125 |
|
| 126 |
def lora_scale_slider(label, visible=True):
|
| 127 |
-
|
|
|
|
| 128 |
|
| 129 |
def lora_textbox():
|
| 130 |
return gr.Textbox(label="", info="Example of prompt:", value="", show_copy_button=True, interactive=False, visible=False)
|
|
@@ -233,8 +235,8 @@ with gr.Blocks(fill_width=True, elem_id="container", css=css, delete_cache=(60,
|
|
| 233 |
low_threshold = gr.Slider(minimum=1, maximum=255, step=1, value=100, label="'CANNY' low threshold")
|
| 234 |
high_threshold = gr.Slider(minimum=1, maximum=255, step=1, value=200, label="'CANNY' high threshold")
|
| 235 |
with gr.Row():
|
| 236 |
-
value_threshold = gr.Slider(minimum=
|
| 237 |
-
distance_threshold = gr.Slider(minimum=
|
| 238 |
recolor_gamma_correction = gr.Number(minimum=0., maximum=25., value=1., step=0.001, label="'RECOLOR' gamma correction")
|
| 239 |
tile_blur_sigma = gr.Number(minimum=0, maximum=100, value=9, step=1, label="'TILE' blur sigma")
|
| 240 |
|
|
@@ -295,7 +297,7 @@ with gr.Blocks(fill_width=True, elem_id="container", css=css, delete_cache=(60,
|
|
| 295 |
with gr.Row():
|
| 296 |
upscaler_model_path = gr.Dropdown(label="Upscaler", choices=UPSCALER_KEYS, value=UPSCALER_KEYS[0])
|
| 297 |
upscaler_increases_size = gr.Slider(minimum=1.1, maximum=4., step=0.1, value=1.2, label="Upscale by")
|
| 298 |
-
upscaler_tile_size = gr.Slider(minimum=0, maximum=512, step=16, value=0, label="Upscaler Tile Size", info="0 = no tiling")
|
| 299 |
upscaler_tile_overlap = gr.Slider(minimum=0, maximum=48, step=1, value=8, label="Upscaler Tile Overlap")
|
| 300 |
with gr.Row():
|
| 301 |
hires_steps = gr.Slider(minimum=0, value=30, maximum=100, step=1, label="Hires Steps")
|
|
@@ -351,7 +353,7 @@ with gr.Blocks(fill_width=True, elem_id="container", css=css, delete_cache=(60,
|
|
| 351 |
face_restoration_weight = gr.Slider(minimum=0., maximum=1., step=0.001, value=.5, label="Weight", info="(0 = maximum effect, 1 = minimum effect)")
|
| 352 |
|
| 353 |
with gr.Tab("Translation Settings"):
|
| 354 |
-
chatbot = gr.Chatbot(render_markdown=False, visible=False) # component for auto-translation
|
| 355 |
chat_model = gr.Dropdown(choices=get_dolphin_models(), value=get_dolphin_models()[0], allow_custom_value=True, label="Model")
|
| 356 |
chat_model_info = gr.Markdown(value=get_dolphin_model_info(get_dolphin_models()[0]), label="Model info")
|
| 357 |
chat_format = gr.Dropdown(choices=get_llm_formats(), value=get_dolphin_model_format(get_dolphin_models()[0]), label="Message format")
|
|
@@ -682,5 +684,6 @@ with gr.Blocks(fill_width=True, elem_id="container", css=css, delete_cache=(60,
|
|
| 682 |
gr.LoginButton()
|
| 683 |
gr.DuplicateButton(value="Duplicate Space for private use (This demo does not work on CPU. Requires GPU Space)")
|
| 684 |
|
| 685 |
-
|
| 686 |
-
demo.
|
|
|
|
|
|
| 1 |
import spaces
|
| 2 |
import gradio as gr
|
| 3 |
import numpy as np
|
| 4 |
+
import os
|
| 5 |
|
| 6 |
# DiffuseCraft
|
| 7 |
from dc import (infer, _infer, pass_result, get_diffusers_model_list, get_samplers, save_image_history,
|
|
|
|
| 36 |
"""
|
| 37 |
)
|
| 38 |
|
| 39 |
+
IS_ZERO_GPU = bool(os.getenv("SPACES_ZERO_GPU"))
|
| 40 |
MAX_SEED = np.iinfo(np.int32).max
|
| 41 |
MAX_IMAGE_SIZE = 4096
|
| 42 |
MIN_IMAGE_SIZE = 256
|
|
|
|
| 52 |
.image { margin: 0px auto; object-fit: contain; }
|
| 53 |
"""
|
| 54 |
|
| 55 |
+
with gr.Blocks(fill_width=True, elem_id="container", css=css) as demo:
|
| 56 |
gr.Markdown("# Votepurchase Multiple Model", elem_classes="title")
|
| 57 |
state = gr.State(value={})
|
| 58 |
with gr.Tab("Image Generator"):
|
|
|
|
| 125 |
return gr.Dropdown(label=label, choices=get_all_lora_tupled_list(), value="", allow_custom_value=True, elem_classes="lora", min_width=320, visible=visible)
|
| 126 |
|
| 127 |
def lora_scale_slider(label, visible=True):
|
| 128 |
+
val_lora = 8 if IS_ZERO_GPU else 8 #
|
| 129 |
+
return gr.Slider(minimum=-val_lora, maximum=val_lora, step=0.01, value=0.33, label=label, visible=visible)
|
| 130 |
|
| 131 |
def lora_textbox():
|
| 132 |
return gr.Textbox(label="", info="Example of prompt:", value="", show_copy_button=True, interactive=False, visible=False)
|
|
|
|
| 235 |
low_threshold = gr.Slider(minimum=1, maximum=255, step=1, value=100, label="'CANNY' low threshold")
|
| 236 |
high_threshold = gr.Slider(minimum=1, maximum=255, step=1, value=200, label="'CANNY' high threshold")
|
| 237 |
with gr.Row():
|
| 238 |
+
value_threshold = gr.Slider(minimum=0.0, maximum=2.0, step=0.01, value=0.1, label="'MLSD' Hough value threshold")
|
| 239 |
+
distance_threshold = gr.Slider(minimum=0.0, maximum=20.0, step=0.01, value=0.1, label="'MLSD' Hough distance threshold")
|
| 240 |
recolor_gamma_correction = gr.Number(minimum=0., maximum=25., value=1., step=0.001, label="'RECOLOR' gamma correction")
|
| 241 |
tile_blur_sigma = gr.Number(minimum=0, maximum=100, value=9, step=1, label="'TILE' blur sigma")
|
| 242 |
|
|
|
|
| 297 |
with gr.Row():
|
| 298 |
upscaler_model_path = gr.Dropdown(label="Upscaler", choices=UPSCALER_KEYS, value=UPSCALER_KEYS[0])
|
| 299 |
upscaler_increases_size = gr.Slider(minimum=1.1, maximum=4., step=0.1, value=1.2, label="Upscale by")
|
| 300 |
+
upscaler_tile_size = gr.Slider(minimum=0, maximum=512, step=16, value=(0 if IS_ZERO_GPU else 192), label="Upscaler Tile Size", info="0 = no tiling")
|
| 301 |
upscaler_tile_overlap = gr.Slider(minimum=0, maximum=48, step=1, value=8, label="Upscaler Tile Overlap")
|
| 302 |
with gr.Row():
|
| 303 |
hires_steps = gr.Slider(minimum=0, value=30, maximum=100, step=1, label="Hires Steps")
|
|
|
|
| 353 |
face_restoration_weight = gr.Slider(minimum=0., maximum=1., step=0.001, value=.5, label="Weight", info="(0 = maximum effect, 1 = minimum effect)")
|
| 354 |
|
| 355 |
with gr.Tab("Translation Settings"):
|
| 356 |
+
chatbot = gr.Chatbot(type="messages", render_markdown=False, visible=False) # component for auto-translation
|
| 357 |
chat_model = gr.Dropdown(choices=get_dolphin_models(), value=get_dolphin_models()[0], allow_custom_value=True, label="Model")
|
| 358 |
chat_model_info = gr.Markdown(value=get_dolphin_model_info(get_dolphin_models()[0]), label="Model info")
|
| 359 |
chat_format = gr.Dropdown(choices=get_llm_formats(), value=get_dolphin_model_format(get_dolphin_models()[0]), label="Message format")
|
|
|
|
| 684 |
gr.LoginButton()
|
| 685 |
gr.DuplicateButton(value="Duplicate Space for private use (This demo does not work on CPU. Requires GPU Space)")
|
| 686 |
|
| 687 |
+
if __name__ == "__main__":
|
| 688 |
+
demo.queue()
|
| 689 |
+
demo.launch(show_error=True, share=False, debug=True, ssr_mode=False)
|
constants.py
CHANGED
|
@@ -9,6 +9,8 @@ from stablepy import (
|
|
| 9 |
IP_ADAPTERS_SDXL,
|
| 10 |
)
|
| 11 |
|
|
|
|
|
|
|
| 12 |
# - **Download Models**
|
| 13 |
DOWNLOAD_MODEL = "https://huggingface.co/TechnoByte/MilkyWonderland/resolve/main/milkyWonderland_v40.safetensors"
|
| 14 |
|
|
@@ -23,12 +25,12 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
|
|
| 23 |
'Laxhar/noobai-XL-1.1',
|
| 24 |
'Laxhar/noobai-XL-Vpred-1.0',
|
| 25 |
'black-forest-labs/FLUX.1-dev',
|
|
|
|
| 26 |
'John6666/blue-pencil-flux1-v021-fp8-flux',
|
| 27 |
'John6666/wai-ani-flux-v10forfp8-fp8-flux',
|
| 28 |
'John6666/xe-anime-flux-v04-fp8-flux',
|
| 29 |
'John6666/lyh-anime-flux-v2a1-fp8-flux',
|
| 30 |
'John6666/carnival-unchained-v10-fp8-flux',
|
| 31 |
-
'John6666/iniverse-mix-xl-sfwnsfw-fluxdfp16nsfwv11-fp8-flux',
|
| 32 |
'Freepik/flux.1-lite-8B-alpha',
|
| 33 |
'shauray/FluxDev-HyperSD-merged',
|
| 34 |
'mikeyandfriends/PixelWave_FLUX.1-dev_03',
|
|
@@ -37,23 +39,19 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
|
|
| 37 |
# 'ostris/OpenFLUX.1',
|
| 38 |
'shuttleai/shuttle-3-diffusion',
|
| 39 |
'Laxhar/noobai-XL-1.0',
|
| 40 |
-
'John6666/noobai-xl-nai-xl-epsilonpred10version-sdxl',
|
| 41 |
'Laxhar/noobai-XL-0.77',
|
| 42 |
'John6666/noobai-xl-nai-xl-epsilonpred075version-sdxl',
|
| 43 |
'Laxhar/noobai-XL-0.6',
|
| 44 |
'John6666/noobai-xl-nai-xl-epsilonpred05version-sdxl',
|
| 45 |
'John6666/noobai-cyberfix-v10-sdxl',
|
| 46 |
'John6666/noobaiiter-xl-vpred-v075-sdxl',
|
| 47 |
-
'John6666/
|
| 48 |
-
'John6666/
|
| 49 |
-
'John6666/ntr-mix-illustrious-xl-noob-xl-v777-sdxl',
|
| 50 |
-
'John6666/ntr-mix-illustrious-xl-noob-xl-v777forlora-sdxl',
|
| 51 |
'John6666/ntr-mix-illustrious-xl-noob-xl-xi-sdxl',
|
| 52 |
'John6666/ntr-mix-illustrious-xl-noob-xl-xii-sdxl',
|
| 53 |
'John6666/ntr-mix-illustrious-xl-noob-xl-xiii-sdxl',
|
| 54 |
'John6666/mistoon-anime-v10illustrious-sdxl',
|
| 55 |
-
'John6666/hassaku-xl-illustrious-
|
| 56 |
-
'John6666/hassaku-xl-illustrious-v10style-sdxl',
|
| 57 |
'John6666/haruki-mix-illustrious-v10-sdxl',
|
| 58 |
'John6666/noobreal-v10-sdxl',
|
| 59 |
'John6666/complicated-noobai-merge-vprediction-sdxl',
|
|
@@ -64,6 +62,7 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
|
|
| 64 |
'Laxhar/noobai-XL-Vpred-0.65',
|
| 65 |
'Laxhar/noobai-XL-Vpred-0.6',
|
| 66 |
'John6666/cat-tower-noobai-xl-checkpoint-v14vpred-sdxl',
|
|
|
|
| 67 |
'John6666/noobai-xl-nai-xl-vpred05version-sdxl',
|
| 68 |
'John6666/noobai-fusion2-vpred-itercomp-v1-sdxl',
|
| 69 |
'John6666/noobai-xl-nai-xl-vpredtestversion-sdxl',
|
|
@@ -74,19 +73,34 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
|
|
| 74 |
'John6666/obsession-illustriousxl-v21-sdxl',
|
| 75 |
'John6666/obsession-illustriousxl-v30-sdxl',
|
| 76 |
'John6666/obsession-illustriousxl-v31-sdxl',
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 77 |
'John6666/wai-nsfw-illustrious-v70-sdxl',
|
|
|
|
| 78 |
'John6666/illustrious-pony-mix-v3-sdxl',
|
| 79 |
-
'John6666/nova-anime-xl-
|
| 80 |
-
'John6666/nova-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 81 |
'John6666/silvermoon-mix03-illustrious-v10-sdxl',
|
| 82 |
'eienmojiki/Anything-XL',
|
| 83 |
'eienmojiki/Starry-XL-v5.2',
|
|
|
|
| 84 |
'John6666/meinaxl-v2-sdxl',
|
| 85 |
'Eugeoter/artiwaifu-diffusion-2.0',
|
| 86 |
'comin/IterComp',
|
| 87 |
-
'John6666/epicrealism-xl-vxiabeast-sdxl',
|
| 88 |
-
'John6666/epicrealism-xl-v10kiss2-sdxl',
|
| 89 |
'John6666/epicrealism-xl-v8kiss-sdxl',
|
|
|
|
|
|
|
|
|
|
| 90 |
'misri/zavychromaxl_v80',
|
| 91 |
'SG161222/RealVisXL_V4.0',
|
| 92 |
'SG161222/RealVisXL_V5.0',
|
|
@@ -102,8 +116,10 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
|
|
| 102 |
'John6666/ebara-mfcg-pony-mix-v12-sdxl',
|
| 103 |
'John6666/t-ponynai3-v51-sdxl',
|
| 104 |
'John6666/t-ponynai3-v65-sdxl',
|
|
|
|
| 105 |
'John6666/prefect-pony-xl-v3-sdxl',
|
| 106 |
'John6666/prefect-pony-xl-v4-sdxl',
|
|
|
|
| 107 |
'John6666/mala-anime-mix-nsfw-pony-xl-v5-sdxl',
|
| 108 |
'John6666/wai-ani-nsfw-ponyxl-v10-sdxl',
|
| 109 |
'John6666/wai-real-mix-v11-sdxl',
|
|
@@ -111,13 +127,14 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
|
|
| 111 |
'John6666/wai-c-v6-sdxl',
|
| 112 |
'John6666/iniverse-mix-xl-sfwnsfw-pony-guofeng-v43-sdxl',
|
| 113 |
'John6666/sifw-annihilation-xl-v2-sdxl',
|
|
|
|
| 114 |
'John6666/photo-realistic-pony-v5-sdxl',
|
| 115 |
'John6666/pony-realism-v21main-sdxl',
|
| 116 |
'John6666/pony-realism-v22main-sdxl',
|
| 117 |
-
'John6666/
|
| 118 |
-
'John6666/cyberrealistic-pony-v64-sdxl',
|
| 119 |
'John6666/cyberrealistic-pony-v65-sdxl',
|
| 120 |
'John6666/cyberrealistic-pony-v7-sdxl',
|
|
|
|
| 121 |
'GraydientPlatformAPI/realcartoon-pony-diffusion',
|
| 122 |
'John6666/nova-anime-xl-pony-v5-sdxl',
|
| 123 |
'John6666/autismmix-sdxl-autismmix-pony-sdxl',
|
|
@@ -127,13 +144,15 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
|
|
| 127 |
'John6666/duchaiten-pony-real-v11fix-sdxl',
|
| 128 |
'John6666/duchaiten-pony-real-v20-sdxl',
|
| 129 |
'John6666/duchaiten-pony-xl-no-score-v70-sdxl',
|
| 130 |
-
'Spestly/OdysseyXL-3.0',
|
| 131 |
-
'Spestly/OdysseyXL-4.0',
|
| 132 |
'KBlueLeaf/Kohaku-XL-Zeta',
|
| 133 |
'cagliostrolab/animagine-xl-3.1',
|
|
|
|
| 134 |
'yodayo-ai/kivotos-xl-2.0',
|
| 135 |
'yodayo-ai/holodayo-xl-2.1',
|
| 136 |
'yodayo-ai/clandestine-xl-1.0',
|
|
|
|
|
|
|
|
|
|
| 137 |
'digiplay/majicMIX_sombre_v2',
|
| 138 |
'digiplay/majicMIX_realistic_v6',
|
| 139 |
'digiplay/majicMIX_realistic_v7',
|
|
@@ -159,9 +178,9 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
|
|
| 159 |
'GraydientPlatformAPI/realcartoon3d-17',
|
| 160 |
'GraydientPlatformAPI/realcartoon-pixar11',
|
| 161 |
'GraydientPlatformAPI/realcartoon-real17',
|
| 162 |
-
'nitrosocke/Ghibli-Diffusion',
|
| 163 |
]
|
| 164 |
|
|
|
|
| 165 |
DIFFUSERS_FORMAT_LORAS = [
|
| 166 |
"nerijs/animation2k-flux",
|
| 167 |
"XLabs-AI/flux-RealismLora",
|
|
@@ -183,8 +202,11 @@ DIRECTORY_VAES = 'vaes'
|
|
| 183 |
DIRECTORY_EMBEDS = 'embedings'
|
| 184 |
DIRECTORY_UPSCALERS = 'upscalers'
|
| 185 |
|
| 186 |
-
CACHE_HF = "/home/user/.cache/huggingface/hub/"
|
| 187 |
STORAGE_ROOT = "/home/user/"
|
|
|
|
|
|
|
|
|
|
|
|
|
| 188 |
|
| 189 |
TASK_STABLEPY = {
|
| 190 |
'txt2img': 'txt2img',
|
|
@@ -226,6 +248,7 @@ UPSCALER_DICT_GUI = {
|
|
| 226 |
# "realesr-general-x4v3": "https://github.com/xinntao/Real-ESRGAN/releases/download/v0.2.5.0/realesr-general-x4v3.pth",
|
| 227 |
# "realesr-general-wdn-x4v3": "https://github.com/xinntao/Real-ESRGAN/releases/download/v0.2.5.0/realesr-general-wdn-x4v3.pth",
|
| 228 |
"4x-UltraSharp": "https://huggingface.co/Shandypur/ESRGAN-4x-UltraSharp/resolve/main/4x-UltraSharp.pth",
|
|
|
|
| 229 |
"4x_foolhardy_Remacri": "https://huggingface.co/FacehugmanIII/4x_foolhardy_Remacri/resolve/main/4x_foolhardy_Remacri.pth",
|
| 230 |
"Remacri4xExtraSmoother": "https://huggingface.co/hollowstrawberry/upscalers-backup/resolve/main/ESRGAN/Remacri%204x%20ExtraSmoother.pth",
|
| 231 |
"AnimeSharp4x": "https://huggingface.co/hollowstrawberry/upscalers-backup/resolve/main/ESRGAN/AnimeSharp%204x.pth",
|
|
@@ -359,9 +382,11 @@ SUBTITLE_GUI = (
|
|
| 359 |
" to perform different tasks in image generation."
|
| 360 |
)
|
| 361 |
|
|
|
|
|
|
|
| 362 |
HELP_GUI = (
|
| 363 |
-
"""### Help:
|
| 364 |
-
|
| 365 |
- Distorted or strange images often result from high prompt weights, so it's best to use low weights and scales, and consider using Classic variants like 'Classic-original'.
|
| 366 |
- For better results with Pony Diffusion, try using sampler DPM++ 1s or DPM2 with Compel or Classic prompt weights.
|
| 367 |
"""
|
|
@@ -485,7 +510,7 @@ EXAMPLES_GUI = [
|
|
| 485 |
20,
|
| 486 |
4.0,
|
| 487 |
-1,
|
| 488 |
-
"loras/Coloring_book_-_LineArt.safetensors",
|
| 489 |
1.0,
|
| 490 |
"DPM++ 2M SDE",
|
| 491 |
1024,
|
|
@@ -580,6 +605,7 @@ EXAMPLES_GUI = [
|
|
| 580 |
RESOURCES = (
|
| 581 |
"""### Resources
|
| 582 |
- John6666's space has some great features you might find helpful [link](https://huggingface.co/spaces/John6666/DiffuseCraftMod).
|
| 583 |
-
-
|
|
|
|
| 584 |
"""
|
| 585 |
-
)
|
|
|
|
| 9 |
IP_ADAPTERS_SDXL,
|
| 10 |
)
|
| 11 |
|
| 12 |
+
IS_ZERO_GPU = bool(os.getenv("SPACES_ZERO_GPU"))
|
| 13 |
+
|
| 14 |
# - **Download Models**
|
| 15 |
DOWNLOAD_MODEL = "https://huggingface.co/TechnoByte/MilkyWonderland/resolve/main/milkyWonderland_v40.safetensors"
|
| 16 |
|
|
|
|
| 25 |
'Laxhar/noobai-XL-1.1',
|
| 26 |
'Laxhar/noobai-XL-Vpred-1.0',
|
| 27 |
'black-forest-labs/FLUX.1-dev',
|
| 28 |
+
'black-forest-labs/FLUX.1-Krea-dev',
|
| 29 |
'John6666/blue-pencil-flux1-v021-fp8-flux',
|
| 30 |
'John6666/wai-ani-flux-v10forfp8-fp8-flux',
|
| 31 |
'John6666/xe-anime-flux-v04-fp8-flux',
|
| 32 |
'John6666/lyh-anime-flux-v2a1-fp8-flux',
|
| 33 |
'John6666/carnival-unchained-v10-fp8-flux',
|
|
|
|
| 34 |
'Freepik/flux.1-lite-8B-alpha',
|
| 35 |
'shauray/FluxDev-HyperSD-merged',
|
| 36 |
'mikeyandfriends/PixelWave_FLUX.1-dev_03',
|
|
|
|
| 39 |
# 'ostris/OpenFLUX.1',
|
| 40 |
'shuttleai/shuttle-3-diffusion',
|
| 41 |
'Laxhar/noobai-XL-1.0',
|
|
|
|
| 42 |
'Laxhar/noobai-XL-0.77',
|
| 43 |
'John6666/noobai-xl-nai-xl-epsilonpred075version-sdxl',
|
| 44 |
'Laxhar/noobai-XL-0.6',
|
| 45 |
'John6666/noobai-xl-nai-xl-epsilonpred05version-sdxl',
|
| 46 |
'John6666/noobai-cyberfix-v10-sdxl',
|
| 47 |
'John6666/noobaiiter-xl-vpred-v075-sdxl',
|
| 48 |
+
'John6666/ripplemix-noob-vpred10-illustrious01-v14-sdxl',
|
| 49 |
+
'John6666/sigmaih-15-sdxl',
|
|
|
|
|
|
|
| 50 |
'John6666/ntr-mix-illustrious-xl-noob-xl-xi-sdxl',
|
| 51 |
'John6666/ntr-mix-illustrious-xl-noob-xl-xii-sdxl',
|
| 52 |
'John6666/ntr-mix-illustrious-xl-noob-xl-xiii-sdxl',
|
| 53 |
'John6666/mistoon-anime-v10illustrious-sdxl',
|
| 54 |
+
'John6666/hassaku-xl-illustrious-v22-sdxl',
|
|
|
|
| 55 |
'John6666/haruki-mix-illustrious-v10-sdxl',
|
| 56 |
'John6666/noobreal-v10-sdxl',
|
| 57 |
'John6666/complicated-noobai-merge-vprediction-sdxl',
|
|
|
|
| 62 |
'Laxhar/noobai-XL-Vpred-0.65',
|
| 63 |
'Laxhar/noobai-XL-Vpred-0.6',
|
| 64 |
'John6666/cat-tower-noobai-xl-checkpoint-v14vpred-sdxl',
|
| 65 |
+
'John6666/cat-tower-noobai-xl-checkpoint-v15vpred-sdxl',
|
| 66 |
'John6666/noobai-xl-nai-xl-vpred05version-sdxl',
|
| 67 |
'John6666/noobai-fusion2-vpred-itercomp-v1-sdxl',
|
| 68 |
'John6666/noobai-xl-nai-xl-vpredtestversion-sdxl',
|
|
|
|
| 73 |
'John6666/obsession-illustriousxl-v21-sdxl',
|
| 74 |
'John6666/obsession-illustriousxl-v30-sdxl',
|
| 75 |
'John6666/obsession-illustriousxl-v31-sdxl',
|
| 76 |
+
'John6666/one-obsession-13-sdxl',
|
| 77 |
+
'John6666/one-obsession-14-24d-sdxl',
|
| 78 |
+
'John6666/one-obsession-15-noobai-sdxl',
|
| 79 |
+
'John6666/one-obsession-v16-noobai-sdxl',
|
| 80 |
+
'John6666/prefect-illustrious-xl-v3-sdxl',
|
| 81 |
'John6666/wai-nsfw-illustrious-v70-sdxl',
|
| 82 |
+
'John6666/wai-nsfw-illustrious-sdxl-v140-sdxl',
|
| 83 |
'John6666/illustrious-pony-mix-v3-sdxl',
|
| 84 |
+
'John6666/nova-anime-xl-il-v90-sdxl',
|
| 85 |
+
'John6666/nova-anime-xl-il-v110-sdxl',
|
| 86 |
+
'John6666/nova-orange-xl-re-v10-sdxl',
|
| 87 |
+
'John6666/nova-orange-xl-v110-sdxl',
|
| 88 |
+
'John6666/nova-orange-xl-re-v20-sdxl',
|
| 89 |
+
'John6666/nova-unreal-xl-v60-sdxl',
|
| 90 |
+
'John6666/nova-unreal-xl-v70-sdxl',
|
| 91 |
+
'John6666/nova-unreal-xl-v80-sdxl',
|
| 92 |
+
'John6666/nova-cartoon-xl-v40-sdxl',
|
| 93 |
'John6666/silvermoon-mix03-illustrious-v10-sdxl',
|
| 94 |
'eienmojiki/Anything-XL',
|
| 95 |
'eienmojiki/Starry-XL-v5.2',
|
| 96 |
+
'votepurchase/plantMilkModelSuite_walnut',
|
| 97 |
'John6666/meinaxl-v2-sdxl',
|
| 98 |
'Eugeoter/artiwaifu-diffusion-2.0',
|
| 99 |
'comin/IterComp',
|
|
|
|
|
|
|
| 100 |
'John6666/epicrealism-xl-v8kiss-sdxl',
|
| 101 |
+
'John6666/epicrealism-xl-v10kiss2-sdxl',
|
| 102 |
+
'John6666/epicrealism-xl-vxiabeast-sdxl',
|
| 103 |
+
'John6666/epicrealism-xl-vxvii-crystal-clear-realism-sdxl',
|
| 104 |
'misri/zavychromaxl_v80',
|
| 105 |
'SG161222/RealVisXL_V4.0',
|
| 106 |
'SG161222/RealVisXL_V5.0',
|
|
|
|
| 116 |
'John6666/ebara-mfcg-pony-mix-v12-sdxl',
|
| 117 |
'John6666/t-ponynai3-v51-sdxl',
|
| 118 |
'John6666/t-ponynai3-v65-sdxl',
|
| 119 |
+
'John6666/t-ponynai3-v7-sdxl',
|
| 120 |
'John6666/prefect-pony-xl-v3-sdxl',
|
| 121 |
'John6666/prefect-pony-xl-v4-sdxl',
|
| 122 |
+
'John6666/prefect-pony-xl-v50-sdxl',
|
| 123 |
'John6666/mala-anime-mix-nsfw-pony-xl-v5-sdxl',
|
| 124 |
'John6666/wai-ani-nsfw-ponyxl-v10-sdxl',
|
| 125 |
'John6666/wai-real-mix-v11-sdxl',
|
|
|
|
| 127 |
'John6666/wai-c-v6-sdxl',
|
| 128 |
'John6666/iniverse-mix-xl-sfwnsfw-pony-guofeng-v43-sdxl',
|
| 129 |
'John6666/sifw-annihilation-xl-v2-sdxl',
|
| 130 |
+
'John6666/sifw-annihilation-xl-v305illustrious-beta-sdxl',
|
| 131 |
'John6666/photo-realistic-pony-v5-sdxl',
|
| 132 |
'John6666/pony-realism-v21main-sdxl',
|
| 133 |
'John6666/pony-realism-v22main-sdxl',
|
| 134 |
+
'John6666/pony-realism-v23-ultra-sdxl',
|
|
|
|
| 135 |
'John6666/cyberrealistic-pony-v65-sdxl',
|
| 136 |
'John6666/cyberrealistic-pony-v7-sdxl',
|
| 137 |
+
'John6666/cyberrealistic-pony-v127-alternative-sdxl',
|
| 138 |
'GraydientPlatformAPI/realcartoon-pony-diffusion',
|
| 139 |
'John6666/nova-anime-xl-pony-v5-sdxl',
|
| 140 |
'John6666/autismmix-sdxl-autismmix-pony-sdxl',
|
|
|
|
| 144 |
'John6666/duchaiten-pony-real-v11fix-sdxl',
|
| 145 |
'John6666/duchaiten-pony-real-v20-sdxl',
|
| 146 |
'John6666/duchaiten-pony-xl-no-score-v70-sdxl',
|
|
|
|
|
|
|
| 147 |
'KBlueLeaf/Kohaku-XL-Zeta',
|
| 148 |
'cagliostrolab/animagine-xl-3.1',
|
| 149 |
+
'cagliostrolab/animagine-xl-4.0',
|
| 150 |
'yodayo-ai/kivotos-xl-2.0',
|
| 151 |
'yodayo-ai/holodayo-xl-2.1',
|
| 152 |
'yodayo-ai/clandestine-xl-1.0',
|
| 153 |
+
'https://huggingface.co/chemwolf/Karmix-XL-v0/resolve/main/Karmix-XL-v0.safetensors?download=true',
|
| 154 |
+
'https://civitai.com/api/download/models/128713?type=Model&format=SafeTensor&size=pruned&fp=fp16',
|
| 155 |
+
'https://civitai.com/models/30240?modelVersionId=125771',
|
| 156 |
'digiplay/majicMIX_sombre_v2',
|
| 157 |
'digiplay/majicMIX_realistic_v6',
|
| 158 |
'digiplay/majicMIX_realistic_v7',
|
|
|
|
| 178 |
'GraydientPlatformAPI/realcartoon3d-17',
|
| 179 |
'GraydientPlatformAPI/realcartoon-pixar11',
|
| 180 |
'GraydientPlatformAPI/realcartoon-real17',
|
|
|
|
| 181 |
]
|
| 182 |
|
| 183 |
+
|
| 184 |
DIFFUSERS_FORMAT_LORAS = [
|
| 185 |
"nerijs/animation2k-flux",
|
| 186 |
"XLabs-AI/flux-RealismLora",
|
|
|
|
| 202 |
DIRECTORY_EMBEDS = 'embedings'
|
| 203 |
DIRECTORY_UPSCALERS = 'upscalers'
|
| 204 |
|
|
|
|
| 205 |
STORAGE_ROOT = "/home/user/"
|
| 206 |
+
CACHE_HF_ROOT = os.path.expanduser(os.getenv("HF_HOME")) if os.getenv("HF_HOME") else os.path.expanduser("~/cache/huggingface")
|
| 207 |
+
CACHE_HF = os.path.join(CACHE_HF_ROOT, "hub")
|
| 208 |
+
if IS_ZERO_GPU:
|
| 209 |
+
os.environ["HF_HOME"] = CACHE_HF
|
| 210 |
|
| 211 |
TASK_STABLEPY = {
|
| 212 |
'txt2img': 'txt2img',
|
|
|
|
| 248 |
# "realesr-general-x4v3": "https://github.com/xinntao/Real-ESRGAN/releases/download/v0.2.5.0/realesr-general-x4v3.pth",
|
| 249 |
# "realesr-general-wdn-x4v3": "https://github.com/xinntao/Real-ESRGAN/releases/download/v0.2.5.0/realesr-general-wdn-x4v3.pth",
|
| 250 |
"4x-UltraSharp": "https://huggingface.co/Shandypur/ESRGAN-4x-UltraSharp/resolve/main/4x-UltraSharp.pth",
|
| 251 |
+
"Real-ESRGAN-Anime-finetuning": "https://huggingface.co/danhtran2mind/Real-ESRGAN-Anime-finetuning/resolve/main/Real-ESRGAN-Anime-finetuning.pth",
|
| 252 |
"4x_foolhardy_Remacri": "https://huggingface.co/FacehugmanIII/4x_foolhardy_Remacri/resolve/main/4x_foolhardy_Remacri.pth",
|
| 253 |
"Remacri4xExtraSmoother": "https://huggingface.co/hollowstrawberry/upscalers-backup/resolve/main/ESRGAN/Remacri%204x%20ExtraSmoother.pth",
|
| 254 |
"AnimeSharp4x": "https://huggingface.co/hollowstrawberry/upscalers-backup/resolve/main/ESRGAN/AnimeSharp%204x.pth",
|
|
|
|
| 382 |
" to perform different tasks in image generation."
|
| 383 |
)
|
| 384 |
|
| 385 |
+
msg_zero = "" if not IS_ZERO_GPU else "- The current space runs on a ZERO GPU which is assigned for approximately 60 seconds; Therefore, if you submit expensive tasks, the operation may be canceled upon reaching the maximum allowed time with 'GPU TASK ABORTED'."
|
| 386 |
+
|
| 387 |
HELP_GUI = (
|
| 388 |
+
f"""### Help:
|
| 389 |
+
{msg_zero}
|
| 390 |
- Distorted or strange images often result from high prompt weights, so it's best to use low weights and scales, and consider using Classic variants like 'Classic-original'.
|
| 391 |
- For better results with Pony Diffusion, try using sampler DPM++ 1s or DPM2 with Compel or Classic prompt weights.
|
| 392 |
"""
|
|
|
|
| 510 |
20,
|
| 511 |
4.0,
|
| 512 |
-1,
|
| 513 |
+
("loras/Coloring_book_-_LineArt.safetensors" if os.path.exists("loras/Coloring_book_-_LineArt.safetensors") else "None"),
|
| 514 |
1.0,
|
| 515 |
"DPM++ 2M SDE",
|
| 516 |
1024,
|
|
|
|
| 605 |
RESOURCES = (
|
| 606 |
"""### Resources
|
| 607 |
- John6666's space has some great features you might find helpful [link](https://huggingface.co/spaces/John6666/DiffuseCraftMod).
|
| 608 |
+
- Try the image generator in Colab’s free tier, which provides free GPU [link](https://github.com/R3gm/SD_diffusers_interactive).
|
| 609 |
+
- `DiffuseCraft` in Colab:[link](https://github.com/R3gm/DiffuseCraft?tab=readme-ov-file#diffusecraft).
|
| 610 |
"""
|
| 611 |
+
)
|
dc.py
CHANGED
|
@@ -1,5 +1,6 @@
|
|
| 1 |
import spaces
|
| 2 |
import os
|
|
|
|
| 3 |
from stablepy import (
|
| 4 |
Model_Diffusers,
|
| 5 |
SCHEDULE_TYPE_OPTIONS,
|
|
@@ -23,6 +24,7 @@ from constants import (
|
|
| 23 |
DIFFUSERS_CONTROLNET_MODEL,
|
| 24 |
IP_MODELS,
|
| 25 |
MODE_IP_OPTIONS,
|
|
|
|
| 26 |
)
|
| 27 |
from stablepy.diffusers_vanilla.style_prompt_config import STYLE_NAMES
|
| 28 |
import torch
|
|
@@ -41,6 +43,7 @@ from utils import (
|
|
| 41 |
progress_step_bar,
|
| 42 |
html_template_message,
|
| 43 |
escape_html,
|
|
|
|
| 44 |
)
|
| 45 |
from image_processor import preprocessor_tab
|
| 46 |
from datetime import datetime
|
|
@@ -53,15 +56,17 @@ from diffusers import FluxPipeline
|
|
| 53 |
# import urllib.parse
|
| 54 |
import subprocess
|
| 55 |
|
| 56 |
-
|
| 57 |
-
if
|
| 58 |
subprocess.run("rm -rf /data-nvme/zerogpu-offload/*", env={}, shell=True)
|
| 59 |
-
|
|
|
|
|
|
|
|
|
|
| 60 |
|
| 61 |
ImageFile.LOAD_TRUNCATED_IMAGES = True
|
| 62 |
-
|
| 63 |
# os.environ["PYTORCH_NO_CUDA_MEMORY_CACHING"] = "1"
|
| 64 |
-
print(os.getenv("SPACES_ZERO_GPU"))
|
| 65 |
|
| 66 |
## BEGIN MOD
|
| 67 |
logging.getLogger("diffusers").setLevel(logging.ERROR)
|
|
@@ -103,14 +108,11 @@ for directory in directories:
|
|
| 103 |
|
| 104 |
# Download stuffs
|
| 105 |
for url in [url.strip() for url in download_model.split(',')]:
|
| 106 |
-
|
| 107 |
-
download_things(DIRECTORY_MODELS, url, HF_TOKEN, CIVITAI_API_KEY)
|
| 108 |
for url in [url.strip() for url in download_vae.split(',')]:
|
| 109 |
-
|
| 110 |
-
download_things(DIRECTORY_VAES, url, HF_TOKEN, CIVITAI_API_KEY)
|
| 111 |
for url in [url.strip() for url in download_lora.split(',')]:
|
| 112 |
-
|
| 113 |
-
download_things(DIRECTORY_LORAS, url, HF_TOKEN, CIVITAI_API_KEY)
|
| 114 |
|
| 115 |
# Download Embeddings
|
| 116 |
for url_embed in DOWNLOAD_EMBEDS:
|
|
@@ -138,15 +140,32 @@ def get_embed_list(pipeline_name):
|
|
| 138 |
|
| 139 |
print('\033[33m🏁 Download and listing of valid models completed.\033[0m')
|
| 140 |
|
| 141 |
-
|
| 142 |
-
|
| 143 |
-
flux_repo
|
| 144 |
-
|
| 145 |
-
|
| 146 |
-
|
| 147 |
-
|
| 148 |
-
|
| 149 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 150 |
|
| 151 |
## BEGIN MOD
|
| 152 |
class GuiSD:
|
|
@@ -164,6 +183,15 @@ class GuiSD:
|
|
| 164 |
removal_candidate = self.inventory.pop(0)
|
| 165 |
delete_model(removal_candidate)
|
| 166 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 167 |
def update_inventory(self, model_name):
|
| 168 |
if model_name not in single_file_model_list:
|
| 169 |
self.inventory = [
|
|
@@ -174,19 +202,24 @@ class GuiSD:
|
|
| 174 |
def load_new_model(self, model_name, vae_model, task, controlnet_model, progress=gr.Progress(track_tqdm=True)):
|
| 175 |
|
| 176 |
# download link model > model_name
|
| 177 |
-
if "http"
|
|
|
|
|
|
|
| 178 |
model_name, model_type = download_link_model(model_name, DIRECTORY_MODELS) #
|
|
|
|
|
|
|
| 179 |
is_link_model = True #
|
| 180 |
else: is_link_model = False #
|
| 181 |
|
| 182 |
-
|
|
|
|
| 183 |
|
| 184 |
vae_model = vae_model if vae_model != "None" else None
|
| 185 |
model_type = get_model_type(model_name) if not is_link_model else model_type #
|
| 186 |
dtype_model = torch.bfloat16 if model_type == "FLUX" else torch.float16
|
| 187 |
|
| 188 |
if not os.path.exists(model_name):
|
| 189 |
-
|
| 190 |
_ = download_diffuser_repo(
|
| 191 |
repo_name=model_name,
|
| 192 |
model_type=model_type,
|
|
@@ -231,10 +264,10 @@ class GuiSD:
|
|
| 231 |
type_model_precision=dtype_model,
|
| 232 |
retain_task_model_in_cache=False,
|
| 233 |
controlnet_model=controlnet_model,
|
| 234 |
-
device="cpu",
|
| 235 |
env_components=components,
|
| 236 |
)
|
| 237 |
-
self.model.advanced_params(image_preprocessor_cuda_active=
|
| 238 |
else:
|
| 239 |
if self.model.base_model_id != model_name:
|
| 240 |
load_now_time = datetime.now()
|
|
@@ -244,7 +277,8 @@ class GuiSD:
|
|
| 244 |
print("Waiting for the previous model's time ops...")
|
| 245 |
time.sleep(9 - elapsed_time)
|
| 246 |
|
| 247 |
-
|
|
|
|
| 248 |
self.model.load_pipe(
|
| 249 |
model_name,
|
| 250 |
task_name=TASK_STABLEPY[task],
|
|
@@ -404,7 +438,7 @@ class GuiSD:
|
|
| 404 |
lora_model_list = get_lora_model_list()
|
| 405 |
## END MOD
|
| 406 |
|
| 407 |
-
|
| 408 |
|
| 409 |
task = TASK_STABLEPY[task]
|
| 410 |
|
|
@@ -502,19 +536,19 @@ class GuiSD:
|
|
| 502 |
"distance_threshold": distance_threshold,
|
| 503 |
"recolor_gamma_correction": float(recolor_gamma_correction),
|
| 504 |
"tile_blur_sigma": int(tile_blur_sigma),
|
| 505 |
-
"lora_A": lora1
|
| 506 |
"lora_scale_A": lora_scale1,
|
| 507 |
-
"lora_B": lora2
|
| 508 |
"lora_scale_B": lora_scale2,
|
| 509 |
-
"lora_C": lora3
|
| 510 |
"lora_scale_C": lora_scale3,
|
| 511 |
-
"lora_D": lora4
|
| 512 |
"lora_scale_D": lora_scale4,
|
| 513 |
-
"lora_E": lora5
|
| 514 |
"lora_scale_E": lora_scale5,
|
| 515 |
-
"lora_F": lora6
|
| 516 |
"lora_scale_F": lora_scale6,
|
| 517 |
-
"lora_G": lora7
|
| 518 |
"lora_scale_G": lora_scale7,
|
| 519 |
## BEGIN MOD
|
| 520 |
"textual_inversion": get_embed_list(self.model.class_name) if textual_inversion else [],
|
|
@@ -574,11 +608,11 @@ class GuiSD:
|
|
| 574 |
# kwargs for diffusers pipeline
|
| 575 |
if guidance_rescale:
|
| 576 |
pipe_params["guidance_rescale"] = guidance_rescale
|
| 577 |
-
|
| 578 |
-
|
| 579 |
-
|
| 580 |
-
|
| 581 |
-
|
| 582 |
|
| 583 |
actual_progress = 0
|
| 584 |
info_images = gr.update()
|
|
@@ -608,7 +642,7 @@ class GuiSD:
|
|
| 608 |
|
| 609 |
download_links = "<br>".join(
|
| 610 |
[
|
| 611 |
-
f'<a href="{path.replace("/images/", "/file
|
| 612 |
for i, path in enumerate(image_path)
|
| 613 |
]
|
| 614 |
)
|
|
@@ -720,7 +754,8 @@ def sd_gen_generate_pipeline(*args):
|
|
| 720 |
|
| 721 |
@spaces.GPU(duration=15)
|
| 722 |
def process_upscale(image, upscaler_name, upscaler_size):
|
| 723 |
-
if image is None:
|
|
|
|
| 724 |
|
| 725 |
from stablepy.diffusers_vanilla.utils import save_pil_image_with_metadata
|
| 726 |
from stablepy import load_upscaler_model
|
|
@@ -737,7 +772,7 @@ def process_upscale(image, upscaler_name, upscaler_size):
|
|
| 737 |
|
| 738 |
name_upscaler = f"./{DIRECTORY_UPSCALERS}/{name_upscaler.split('/')[-1]}"
|
| 739 |
|
| 740 |
-
scaler_beta = load_upscaler_model(model=name_upscaler, tile=0, tile_overlap=8, device="cuda", half=
|
| 741 |
image_up = scaler_beta.upscale(image, upscaler_size, True)
|
| 742 |
|
| 743 |
image_path = save_pil_image_with_metadata(image_up, f'{os.getcwd()}/up_images', exif_image)
|
|
@@ -746,8 +781,8 @@ def process_upscale(image, upscaler_name, upscaler_size):
|
|
| 746 |
|
| 747 |
|
| 748 |
# https://huggingface.co/spaces/BestWishYsh/ConsisID-preview-Space/discussions/1#674969a022b99c122af5d407
|
| 749 |
-
dynamic_gpu_duration.zerogpu = True
|
| 750 |
-
sd_gen_generate_pipeline.zerogpu = True
|
| 751 |
sd_gen = GuiSD()
|
| 752 |
|
| 753 |
|
|
|
|
| 1 |
import spaces
|
| 2 |
import os
|
| 3 |
+
from argparse import ArgumentParser
|
| 4 |
from stablepy import (
|
| 5 |
Model_Diffusers,
|
| 6 |
SCHEDULE_TYPE_OPTIONS,
|
|
|
|
| 24 |
DIFFUSERS_CONTROLNET_MODEL,
|
| 25 |
IP_MODELS,
|
| 26 |
MODE_IP_OPTIONS,
|
| 27 |
+
CACHE_HF_ROOT,
|
| 28 |
)
|
| 29 |
from stablepy.diffusers_vanilla.style_prompt_config import STYLE_NAMES
|
| 30 |
import torch
|
|
|
|
| 43 |
progress_step_bar,
|
| 44 |
html_template_message,
|
| 45 |
escape_html,
|
| 46 |
+
clear_hf_cache,
|
| 47 |
)
|
| 48 |
from image_processor import preprocessor_tab
|
| 49 |
from datetime import datetime
|
|
|
|
| 56 |
# import urllib.parse
|
| 57 |
import subprocess
|
| 58 |
|
| 59 |
+
IS_ZERO_GPU = bool(os.getenv("SPACES_ZERO_GPU"))
|
| 60 |
+
if IS_ZERO_GPU:
|
| 61 |
subprocess.run("rm -rf /data-nvme/zerogpu-offload/*", env={}, shell=True)
|
| 62 |
+
IS_GPU_MODE = True if IS_ZERO_GPU else (True if torch.cuda.is_available() else False)
|
| 63 |
+
img_path = "./images/"
|
| 64 |
+
allowed_path = os.path.abspath(img_path)
|
| 65 |
+
delete_cache_time = (9600, 9600) if IS_ZERO_GPU else (86400, 86400)
|
| 66 |
|
| 67 |
ImageFile.LOAD_TRUNCATED_IMAGES = True
|
| 68 |
+
torch.backends.cuda.matmul.allow_tf32 = True
|
| 69 |
# os.environ["PYTORCH_NO_CUDA_MEMORY_CACHING"] = "1"
|
|
|
|
| 70 |
|
| 71 |
## BEGIN MOD
|
| 72 |
logging.getLogger("diffusers").setLevel(logging.ERROR)
|
|
|
|
| 108 |
|
| 109 |
# Download stuffs
|
| 110 |
for url in [url.strip() for url in download_model.split(',')]:
|
| 111 |
+
download_things(DIRECTORY_MODELS, url, HF_TOKEN, CIVITAI_API_KEY)
|
|
|
|
| 112 |
for url in [url.strip() for url in download_vae.split(',')]:
|
| 113 |
+
download_things(DIRECTORY_VAES, url, HF_TOKEN, CIVITAI_API_KEY)
|
|
|
|
| 114 |
for url in [url.strip() for url in download_lora.split(',')]:
|
| 115 |
+
download_things(DIRECTORY_LORAS, url, HF_TOKEN, CIVITAI_API_KEY)
|
|
|
|
| 116 |
|
| 117 |
# Download Embeddings
|
| 118 |
for url_embed in DOWNLOAD_EMBEDS:
|
|
|
|
| 140 |
|
| 141 |
print('\033[33m🏁 Download and listing of valid models completed.\033[0m')
|
| 142 |
|
| 143 |
+
components = None
|
| 144 |
+
if IS_ZERO_GPU:
|
| 145 |
+
flux_repo = "camenduru/FLUX.1-dev-diffusers"
|
| 146 |
+
flux_pipe = FluxPipeline.from_pretrained(
|
| 147 |
+
flux_repo,
|
| 148 |
+
transformer=None,
|
| 149 |
+
torch_dtype=torch.bfloat16,
|
| 150 |
+
)#.to("cuda")
|
| 151 |
+
components = flux_pipe.components
|
| 152 |
+
delete_model(flux_repo)
|
| 153 |
+
|
| 154 |
+
parser = ArgumentParser(description='DiffuseCraft: Create images from text prompts.', add_help=True)
|
| 155 |
+
parser.add_argument("--share", action="store_true", dest="share_enabled", default=False, help="Enable sharing")
|
| 156 |
+
parser.add_argument('--theme', type=str, default="NoCrypt/miku", help='Set the theme (default: NoCrypt/miku)')
|
| 157 |
+
parser.add_argument("--ssr", action="store_true", help="Enable SSR (Server-Side Rendering)")
|
| 158 |
+
parser.add_argument("--log-level", type=str, default="INFO", choices=["DEBUG", "INFO", "WARNING", "ERROR", "CRITICAL"], help="Set logging level (default: INFO)")
|
| 159 |
+
args = parser.parse_args()
|
| 160 |
+
|
| 161 |
+
logger.setLevel(
|
| 162 |
+
"INFO" if IS_ZERO_GPU else getattr(logging, args.log_level.upper())
|
| 163 |
+
)
|
| 164 |
+
|
| 165 |
+
def lora_chk(lora_):
|
| 166 |
+
if isinstance(lora_, str) and lora_.strip() not in ["", "None"]:
|
| 167 |
+
return lora_
|
| 168 |
+
return None
|
| 169 |
|
| 170 |
## BEGIN MOD
|
| 171 |
class GuiSD:
|
|
|
|
| 183 |
removal_candidate = self.inventory.pop(0)
|
| 184 |
delete_model(removal_candidate)
|
| 185 |
|
| 186 |
+
# Cleanup after 60 seconds of inactivity
|
| 187 |
+
lowPrioCleanup = max((datetime.now() - self.last_load).total_seconds(), 0) > 60
|
| 188 |
+
if lowPrioCleanup and not self.status_loading and get_used_storage_gb(CACHE_HF_ROOT) > (storage_floor_gb * 2):
|
| 189 |
+
print("Cleaning up Hugging Face cache...")
|
| 190 |
+
clear_hf_cache()
|
| 191 |
+
self.inventory = [
|
| 192 |
+
m for m in self.inventory if os.path.exists(m)
|
| 193 |
+
]
|
| 194 |
+
|
| 195 |
def update_inventory(self, model_name):
|
| 196 |
if model_name not in single_file_model_list:
|
| 197 |
self.inventory = [
|
|
|
|
| 202 |
def load_new_model(self, model_name, vae_model, task, controlnet_model, progress=gr.Progress(track_tqdm=True)):
|
| 203 |
|
| 204 |
# download link model > model_name
|
| 205 |
+
if model_name.startswith("http"): #
|
| 206 |
+
yield f"Downloading model: {model_name}"
|
| 207 |
+
#model_name = download_things(DIRECTORY_MODELS, model_name, HF_TOKEN, CIVITAI_API_KEY)
|
| 208 |
model_name, model_type = download_link_model(model_name, DIRECTORY_MODELS) #
|
| 209 |
+
if not model_name:
|
| 210 |
+
raise ValueError("Error retrieving model information from URL")
|
| 211 |
is_link_model = True #
|
| 212 |
else: is_link_model = False #
|
| 213 |
|
| 214 |
+
if IS_ZERO_GPU:
|
| 215 |
+
self.update_storage_models()
|
| 216 |
|
| 217 |
vae_model = vae_model if vae_model != "None" else None
|
| 218 |
model_type = get_model_type(model_name) if not is_link_model else model_type #
|
| 219 |
dtype_model = torch.bfloat16 if model_type == "FLUX" else torch.float16
|
| 220 |
|
| 221 |
if not os.path.exists(model_name):
|
| 222 |
+
logger.debug(f"model_name={model_name}, vae_model={vae_model}, task={task}, controlnet_model={controlnet_model}")
|
| 223 |
_ = download_diffuser_repo(
|
| 224 |
repo_name=model_name,
|
| 225 |
model_type=model_type,
|
|
|
|
| 264 |
type_model_precision=dtype_model,
|
| 265 |
retain_task_model_in_cache=False,
|
| 266 |
controlnet_model=controlnet_model,
|
| 267 |
+
device="cpu" if IS_ZERO_GPU else None,
|
| 268 |
env_components=components,
|
| 269 |
)
|
| 270 |
+
self.model.advanced_params(image_preprocessor_cuda_active=IS_GPU_MODE)
|
| 271 |
else:
|
| 272 |
if self.model.base_model_id != model_name:
|
| 273 |
load_now_time = datetime.now()
|
|
|
|
| 277 |
print("Waiting for the previous model's time ops...")
|
| 278 |
time.sleep(9 - elapsed_time)
|
| 279 |
|
| 280 |
+
if IS_ZERO_GPU:
|
| 281 |
+
self.model.device = torch.device("cpu")
|
| 282 |
self.model.load_pipe(
|
| 283 |
model_name,
|
| 284 |
task_name=TASK_STABLEPY[task],
|
|
|
|
| 438 |
lora_model_list = get_lora_model_list()
|
| 439 |
## END MOD
|
| 440 |
|
| 441 |
+
logger.debug(f"Config model: {model_name}, {vae_model}, {loras_list}")
|
| 442 |
|
| 443 |
task = TASK_STABLEPY[task]
|
| 444 |
|
|
|
|
| 536 |
"distance_threshold": distance_threshold,
|
| 537 |
"recolor_gamma_correction": float(recolor_gamma_correction),
|
| 538 |
"tile_blur_sigma": int(tile_blur_sigma),
|
| 539 |
+
"lora_A": lora_chk(lora1),
|
| 540 |
"lora_scale_A": lora_scale1,
|
| 541 |
+
"lora_B": lora_chk(lora2),
|
| 542 |
"lora_scale_B": lora_scale2,
|
| 543 |
+
"lora_C": lora_chk(lora3),
|
| 544 |
"lora_scale_C": lora_scale3,
|
| 545 |
+
"lora_D": lora_chk(lora4),
|
| 546 |
"lora_scale_D": lora_scale4,
|
| 547 |
+
"lora_E": lora_chk(lora5),
|
| 548 |
"lora_scale_E": lora_scale5,
|
| 549 |
+
"lora_F": lora_chk(lora6),
|
| 550 |
"lora_scale_F": lora_scale6,
|
| 551 |
+
"lora_G": lora_chk(lora7),
|
| 552 |
"lora_scale_G": lora_scale7,
|
| 553 |
## BEGIN MOD
|
| 554 |
"textual_inversion": get_embed_list(self.model.class_name) if textual_inversion else [],
|
|
|
|
| 608 |
# kwargs for diffusers pipeline
|
| 609 |
if guidance_rescale:
|
| 610 |
pipe_params["guidance_rescale"] = guidance_rescale
|
| 611 |
+
if IS_ZERO_GPU:
|
| 612 |
+
self.model.device = torch.device("cuda:0")
|
| 613 |
+
if hasattr(self.model.pipe, "transformer") and loras_list != ["None"] * self.model.num_loras:
|
| 614 |
+
self.model.pipe.transformer.to(self.model.device)
|
| 615 |
+
logger.debug("transformer to cuda")
|
| 616 |
|
| 617 |
actual_progress = 0
|
| 618 |
info_images = gr.update()
|
|
|
|
| 642 |
|
| 643 |
download_links = "<br>".join(
|
| 644 |
[
|
| 645 |
+
f'<a href="{path.replace("/images/", f"/gradio_api/file={allowed_path}/")}" download="{os.path.basename(path)}">Download Image {i + 1}</a>'
|
| 646 |
for i, path in enumerate(image_path)
|
| 647 |
]
|
| 648 |
)
|
|
|
|
| 754 |
|
| 755 |
@spaces.GPU(duration=15)
|
| 756 |
def process_upscale(image, upscaler_name, upscaler_size):
|
| 757 |
+
if image is None:
|
| 758 |
+
return None
|
| 759 |
|
| 760 |
from stablepy.diffusers_vanilla.utils import save_pil_image_with_metadata
|
| 761 |
from stablepy import load_upscaler_model
|
|
|
|
| 772 |
|
| 773 |
name_upscaler = f"./{DIRECTORY_UPSCALERS}/{name_upscaler.split('/')[-1]}"
|
| 774 |
|
| 775 |
+
scaler_beta = load_upscaler_model(model=name_upscaler, tile=(0 if IS_ZERO_GPU else 192), tile_overlap=8, device=("cuda" if IS_GPU_MODE else "cpu"), half=IS_GPU_MODE)
|
| 776 |
image_up = scaler_beta.upscale(image, upscaler_size, True)
|
| 777 |
|
| 778 |
image_path = save_pil_image_with_metadata(image_up, f'{os.getcwd()}/up_images', exif_image)
|
|
|
|
| 781 |
|
| 782 |
|
| 783 |
# https://huggingface.co/spaces/BestWishYsh/ConsisID-preview-Space/discussions/1#674969a022b99c122af5d407
|
| 784 |
+
# dynamic_gpu_duration.zerogpu = True
|
| 785 |
+
# sd_gen_generate_pipeline.zerogpu = True
|
| 786 |
sd_gen = GuiSD()
|
| 787 |
|
| 788 |
|
env.py
CHANGED
|
@@ -71,6 +71,7 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
|
|
| 71 |
'Raelina/Raehoshi-illust-XL-4',
|
| 72 |
'Raelina/Raehoshi-illust-XL-5',
|
| 73 |
'Raelina/Raehoshi-illust-XL-5.1',
|
|
|
|
| 74 |
'camenduru/FLUX.1-dev-diffusers',
|
| 75 |
'black-forest-labs/FLUX.1-schnell',
|
| 76 |
'sayakpaul/FLUX.1-merged',
|
|
|
|
| 71 |
'Raelina/Raehoshi-illust-XL-4',
|
| 72 |
'Raelina/Raehoshi-illust-XL-5',
|
| 73 |
'Raelina/Raehoshi-illust-XL-5.1',
|
| 74 |
+
'Raelina/Raehoshi-illust-XL-6',
|
| 75 |
'camenduru/FLUX.1-dev-diffusers',
|
| 76 |
'black-forest-labs/FLUX.1-schnell',
|
| 77 |
'sayakpaul/FLUX.1-merged',
|
image_processor.py
CHANGED
|
@@ -92,8 +92,8 @@ def preprocessor_tab():
|
|
| 92 |
pre_processor_resolution = gr.Slider(minimum=64, maximum=2048, step=64, value=512, label="Preprocessor Resolution")
|
| 93 |
pre_low_threshold = gr.Slider(minimum=1, maximum=255, step=1, value=100, label="'CANNY' low threshold")
|
| 94 |
pre_high_threshold = gr.Slider(minimum=1, maximum=255, step=1, value=200, label="'CANNY' high threshold")
|
| 95 |
-
pre_value_threshold = gr.Slider(minimum=
|
| 96 |
-
pre_distance_threshold = gr.Slider(minimum=
|
| 97 |
pre_recolor_mode = gr.Dropdown(label="'RECOLOR' mode", choices=["luminance", "intensity"], value="luminance")
|
| 98 |
pre_recolor_gamma_correction = gr.Number(minimum=0., maximum=25., value=1., step=0.001, label="'RECOLOR' gamma correction")
|
| 99 |
pre_blur_k_size = gr.Number(minimum=0, maximum=100, value=9, step=1, label="'BLUR' sigma")
|
|
|
|
| 92 |
pre_processor_resolution = gr.Slider(minimum=64, maximum=2048, step=64, value=512, label="Preprocessor Resolution")
|
| 93 |
pre_low_threshold = gr.Slider(minimum=1, maximum=255, step=1, value=100, label="'CANNY' low threshold")
|
| 94 |
pre_high_threshold = gr.Slider(minimum=1, maximum=255, step=1, value=200, label="'CANNY' high threshold")
|
| 95 |
+
pre_value_threshold = gr.Slider(minimum=0., maximum=2.0, step=0.01, value=0.1, label="'MLSD' Hough value threshold")
|
| 96 |
+
pre_distance_threshold = gr.Slider(minimum=0., maximum=20.0, step=0.01, value=0.1, label="'MLSD' Hough distance threshold")
|
| 97 |
pre_recolor_mode = gr.Dropdown(label="'RECOLOR' mode", choices=["luminance", "intensity"], value="luminance")
|
| 98 |
pre_recolor_gamma_correction = gr.Number(minimum=0., maximum=25., value=1., step=0.001, label="'RECOLOR' gamma correction")
|
| 99 |
pre_blur_k_size = gr.Number(minimum=0, maximum=100, value=9, step=1, label="'BLUR' sigma")
|
llmdolphin.py
CHANGED
|
@@ -7,6 +7,7 @@ import gc
|
|
| 7 |
import os
|
| 8 |
import urllib
|
| 9 |
from typing import Any
|
|
|
|
| 10 |
from huggingface_hub import hf_hub_download, HfApi
|
| 11 |
from llama_cpp import Llama
|
| 12 |
from llama_cpp_agent import LlamaCppAgent, MessagesFormatterType
|
|
@@ -376,10 +377,10 @@ def get_raw_prompt(msg: str):
|
|
| 376 |
|
| 377 |
# https://llama-cpp-python.readthedocs.io/en/latest/api-reference/
|
| 378 |
@torch.inference_mode()
|
| 379 |
-
@spaces.GPU(duration=
|
| 380 |
def dolphin_respond(
|
| 381 |
message: str,
|
| 382 |
-
history: list[
|
| 383 |
model: str = default_llm_model_filename,
|
| 384 |
system_message: str = get_dolphin_sysprompt(),
|
| 385 |
max_tokens: int = 1024,
|
|
@@ -434,16 +435,12 @@ def dolphin_respond(
|
|
| 434 |
messages = BasicChatHistory()
|
| 435 |
|
| 436 |
for msn in history:
|
| 437 |
-
|
| 438 |
-
'role': Roles.user,
|
| 439 |
-
|
| 440 |
-
|
| 441 |
-
|
| 442 |
-
|
| 443 |
-
'content': msn[1]
|
| 444 |
-
}
|
| 445 |
-
messages.add_message(user)
|
| 446 |
-
messages.add_message(assistant)
|
| 447 |
|
| 448 |
stream = agent.get_chat_response(
|
| 449 |
message,
|
|
@@ -455,28 +452,28 @@ def dolphin_respond(
|
|
| 455 |
|
| 456 |
progress(0.5, desc="Processing...")
|
| 457 |
|
| 458 |
-
|
|
|
|
| 459 |
for output in stream:
|
| 460 |
-
|
| 461 |
-
yield
|
| 462 |
except Exception as e:
|
| 463 |
print(e)
|
| 464 |
raise gr.Error(f"Error: {e}")
|
| 465 |
-
#yield [("", None)]
|
| 466 |
finally:
|
| 467 |
torch.cuda.empty_cache()
|
| 468 |
gc.collect()
|
| 469 |
|
| 470 |
|
| 471 |
def dolphin_parse(
|
| 472 |
-
history: list[
|
| 473 |
state: dict,
|
| 474 |
):
|
| 475 |
try:
|
| 476 |
dolphin_sysprompt_mode = get_state(state, "dolphin_sysprompt_mode")
|
| 477 |
if dolphin_sysprompt_mode == "Chat with LLM" or not history or len(history) < 1:
|
| 478 |
return "", gr.update(), gr.update()
|
| 479 |
-
msg = history[-1][
|
| 480 |
raw_prompt = get_raw_prompt(msg)
|
| 481 |
prompts = []
|
| 482 |
if dolphin_sysprompt_mode == "Japanese to Danbooru Dictionary" and is_japanese(raw_prompt):
|
|
@@ -490,10 +487,10 @@ def dolphin_parse(
|
|
| 490 |
|
| 491 |
|
| 492 |
@torch.inference_mode()
|
| 493 |
-
@spaces.GPU(duration=
|
| 494 |
def dolphin_respond_auto(
|
| 495 |
message: str,
|
| 496 |
-
history: list[
|
| 497 |
model: str = default_llm_model_filename,
|
| 498 |
system_message: str = get_dolphin_sysprompt(),
|
| 499 |
max_tokens: int = 1024,
|
|
@@ -549,16 +546,12 @@ def dolphin_respond_auto(
|
|
| 549 |
messages = BasicChatHistory()
|
| 550 |
|
| 551 |
for msn in history:
|
| 552 |
-
|
| 553 |
-
'role': Roles.user,
|
| 554 |
-
|
| 555 |
-
|
| 556 |
-
|
| 557 |
-
|
| 558 |
-
'content': msn[1]
|
| 559 |
-
}
|
| 560 |
-
messages.add_message(user)
|
| 561 |
-
messages.add_message(assistant)
|
| 562 |
|
| 563 |
progress(0, desc="Translating...")
|
| 564 |
stream = agent.get_chat_response(
|
|
@@ -571,13 +564,16 @@ def dolphin_respond_auto(
|
|
| 571 |
|
| 572 |
progress(0.5, desc="Processing...")
|
| 573 |
|
| 574 |
-
|
|
|
|
| 575 |
for output in stream:
|
| 576 |
-
|
| 577 |
-
yield
|
| 578 |
except Exception as e:
|
| 579 |
print(e)
|
| 580 |
-
|
|
|
|
|
|
|
| 581 |
finally:
|
| 582 |
torch.cuda.empty_cache()
|
| 583 |
gc.collect()
|
|
@@ -585,14 +581,14 @@ def dolphin_respond_auto(
|
|
| 585 |
|
| 586 |
def dolphin_parse_simple(
|
| 587 |
message: str,
|
| 588 |
-
history: list[
|
| 589 |
state: dict,
|
| 590 |
):
|
| 591 |
try:
|
| 592 |
#if not is_japanese(message): return message
|
| 593 |
dolphin_sysprompt_mode = get_state(state, "dolphin_sysprompt_mode")
|
| 594 |
if dolphin_sysprompt_mode == "Chat with LLM" or not history or len(history) < 1: return message
|
| 595 |
-
msg = history[-1][
|
| 596 |
raw_prompt = get_raw_prompt(msg)
|
| 597 |
prompts = []
|
| 598 |
if dolphin_sysprompt_mode == "Japanese to Danbooru Dictionary" and is_japanese(raw_prompt):
|
|
@@ -611,10 +607,10 @@ cv2.setNumThreads(1)
|
|
| 611 |
|
| 612 |
|
| 613 |
@torch.inference_mode()
|
| 614 |
-
@spaces.GPU(duration=
|
| 615 |
def respond_playground(
|
| 616 |
message: str,
|
| 617 |
-
history: list[
|
| 618 |
model: str = default_llm_model_filename,
|
| 619 |
system_message: str = get_dolphin_sysprompt(),
|
| 620 |
max_tokens: int = 1024,
|
|
@@ -669,10 +665,12 @@ def respond_playground(
|
|
| 669 |
|
| 670 |
# Add user and assistant messages to the history
|
| 671 |
for msn in history:
|
| 672 |
-
|
| 673 |
-
|
| 674 |
-
|
| 675 |
-
|
|
|
|
|
|
|
| 676 |
|
| 677 |
# Stream the response
|
| 678 |
stream = agent.get_chat_response(
|
|
@@ -683,14 +681,14 @@ def respond_playground(
|
|
| 683 |
print_output=False
|
| 684 |
)
|
| 685 |
|
| 686 |
-
|
|
|
|
| 687 |
for output in stream:
|
| 688 |
-
|
| 689 |
-
yield
|
| 690 |
except Exception as e:
|
| 691 |
print(e)
|
| 692 |
raise gr.Error(f"Error: {e}")
|
| 693 |
-
#yield ""
|
| 694 |
finally:
|
| 695 |
torch.cuda.empty_cache()
|
| 696 |
gc.collect()
|
|
|
|
| 7 |
import os
|
| 8 |
import urllib
|
| 9 |
from typing import Any
|
| 10 |
+
from gradio import MessageDict
|
| 11 |
from huggingface_hub import hf_hub_download, HfApi
|
| 12 |
from llama_cpp import Llama
|
| 13 |
from llama_cpp_agent import LlamaCppAgent, MessagesFormatterType
|
|
|
|
| 377 |
|
| 378 |
# https://llama-cpp-python.readthedocs.io/en/latest/api-reference/
|
| 379 |
@torch.inference_mode()
|
| 380 |
+
@spaces.GPU(duration=30)
|
| 381 |
def dolphin_respond(
|
| 382 |
message: str,
|
| 383 |
+
history: list[MessageDict],
|
| 384 |
model: str = default_llm_model_filename,
|
| 385 |
system_message: str = get_dolphin_sysprompt(),
|
| 386 |
max_tokens: int = 1024,
|
|
|
|
| 435 |
messages = BasicChatHistory()
|
| 436 |
|
| 437 |
for msn in history:
|
| 438 |
+
if msn["role"] == "user":
|
| 439 |
+
user = {'role': Roles.user, 'content': msn["content"]}
|
| 440 |
+
messages.add_message(user)
|
| 441 |
+
elif msn["role"] == "assistant":
|
| 442 |
+
assistant = {'role': Roles.assistant, 'content': msn["content"]}
|
| 443 |
+
messages.add_message(assistant)
|
|
|
|
|
|
|
|
|
|
|
|
|
| 444 |
|
| 445 |
stream = agent.get_chat_response(
|
| 446 |
message,
|
|
|
|
| 452 |
|
| 453 |
progress(0.5, desc="Processing...")
|
| 454 |
|
| 455 |
+
history.append({"role": "user", "content": message})
|
| 456 |
+
history.append({"role": "assistant", "content": ""})
|
| 457 |
for output in stream:
|
| 458 |
+
history[-1]['content'] += output
|
| 459 |
+
yield history
|
| 460 |
except Exception as e:
|
| 461 |
print(e)
|
| 462 |
raise gr.Error(f"Error: {e}")
|
|
|
|
| 463 |
finally:
|
| 464 |
torch.cuda.empty_cache()
|
| 465 |
gc.collect()
|
| 466 |
|
| 467 |
|
| 468 |
def dolphin_parse(
|
| 469 |
+
history: list[MessageDict],
|
| 470 |
state: dict,
|
| 471 |
):
|
| 472 |
try:
|
| 473 |
dolphin_sysprompt_mode = get_state(state, "dolphin_sysprompt_mode")
|
| 474 |
if dolphin_sysprompt_mode == "Chat with LLM" or not history or len(history) < 1:
|
| 475 |
return "", gr.update(), gr.update()
|
| 476 |
+
msg = history[-1]["content"]
|
| 477 |
raw_prompt = get_raw_prompt(msg)
|
| 478 |
prompts = []
|
| 479 |
if dolphin_sysprompt_mode == "Japanese to Danbooru Dictionary" and is_japanese(raw_prompt):
|
|
|
|
| 487 |
|
| 488 |
|
| 489 |
@torch.inference_mode()
|
| 490 |
+
@spaces.GPU(duration=30)
|
| 491 |
def dolphin_respond_auto(
|
| 492 |
message: str,
|
| 493 |
+
history: list[MessageDict],
|
| 494 |
model: str = default_llm_model_filename,
|
| 495 |
system_message: str = get_dolphin_sysprompt(),
|
| 496 |
max_tokens: int = 1024,
|
|
|
|
| 546 |
messages = BasicChatHistory()
|
| 547 |
|
| 548 |
for msn in history:
|
| 549 |
+
if msn["role"] == "user":
|
| 550 |
+
user = {'role': Roles.user, 'content': msn["content"]}
|
| 551 |
+
messages.add_message(user)
|
| 552 |
+
elif msn["role"] == "assistant":
|
| 553 |
+
assistant = {'role': Roles.assistant, 'content': msn["content"]}
|
| 554 |
+
messages.add_message(assistant)
|
|
|
|
|
|
|
|
|
|
|
|
|
| 555 |
|
| 556 |
progress(0, desc="Translating...")
|
| 557 |
stream = agent.get_chat_response(
|
|
|
|
| 564 |
|
| 565 |
progress(0.5, desc="Processing...")
|
| 566 |
|
| 567 |
+
history.append({"role": "user", "content": message})
|
| 568 |
+
history.append({"role": "assistant", "content": ""})
|
| 569 |
for output in stream:
|
| 570 |
+
history[-1]['content'] += output
|
| 571 |
+
yield history, gr.update(), gr.update()
|
| 572 |
except Exception as e:
|
| 573 |
print(e)
|
| 574 |
+
history.append({"role": "user", "content": message})
|
| 575 |
+
history.append({"role": "assistant", "content": message})
|
| 576 |
+
yield history, gr.update(), gr.update()
|
| 577 |
finally:
|
| 578 |
torch.cuda.empty_cache()
|
| 579 |
gc.collect()
|
|
|
|
| 581 |
|
| 582 |
def dolphin_parse_simple(
|
| 583 |
message: str,
|
| 584 |
+
history: list[MessageDict],
|
| 585 |
state: dict,
|
| 586 |
):
|
| 587 |
try:
|
| 588 |
#if not is_japanese(message): return message
|
| 589 |
dolphin_sysprompt_mode = get_state(state, "dolphin_sysprompt_mode")
|
| 590 |
if dolphin_sysprompt_mode == "Chat with LLM" or not history or len(history) < 1: return message
|
| 591 |
+
msg = history[-1]["content"]
|
| 592 |
raw_prompt = get_raw_prompt(msg)
|
| 593 |
prompts = []
|
| 594 |
if dolphin_sysprompt_mode == "Japanese to Danbooru Dictionary" and is_japanese(raw_prompt):
|
|
|
|
| 607 |
|
| 608 |
|
| 609 |
@torch.inference_mode()
|
| 610 |
+
@spaces.GPU(duration=30)
|
| 611 |
def respond_playground(
|
| 612 |
message: str,
|
| 613 |
+
history: list[MessageDict],
|
| 614 |
model: str = default_llm_model_filename,
|
| 615 |
system_message: str = get_dolphin_sysprompt(),
|
| 616 |
max_tokens: int = 1024,
|
|
|
|
| 665 |
|
| 666 |
# Add user and assistant messages to the history
|
| 667 |
for msn in history:
|
| 668 |
+
if msn["role"] == "user":
|
| 669 |
+
user = {'role': Roles.user, 'content': msn["content"]}
|
| 670 |
+
messages.add_message(user)
|
| 671 |
+
elif msn["role"] == "assistant":
|
| 672 |
+
assistant = {'role': Roles.assistant, 'content': msn["content"]}
|
| 673 |
+
messages.add_message(assistant)
|
| 674 |
|
| 675 |
# Stream the response
|
| 676 |
stream = agent.get_chat_response(
|
|
|
|
| 681 |
print_output=False
|
| 682 |
)
|
| 683 |
|
| 684 |
+
history.append({"role": "user", "content": message})
|
| 685 |
+
history.append({"role": "assistant", "content": ""})
|
| 686 |
for output in stream:
|
| 687 |
+
history[-1]['content'] += output
|
| 688 |
+
yield history
|
| 689 |
except Exception as e:
|
| 690 |
print(e)
|
| 691 |
raise gr.Error(f"Error: {e}")
|
|
|
|
| 692 |
finally:
|
| 693 |
torch.cuda.empty_cache()
|
| 694 |
gc.collect()
|
llmenv.py
CHANGED
|
@@ -14,6 +14,7 @@ llm_models = {
|
|
| 14 |
"mn-12b-lyra-v2a1-q5_k_m.gguf": ["HalleyStarbun/MN-12B-Lyra-v2a1-Q5_K_M-GGUF", MessagesFormatterType.CHATML],
|
| 15 |
"L3-8B-Tamamo-v1.i1-Q5_K_M.gguf": ["mradermacher/L3-8B-Tamamo-v1-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 16 |
"MN-Chinofun-12B-2.i1-Q4_K_M.gguf": ["mradermacher/MN-Chinofun-12B-2-i1-GGUF", MessagesFormatterType.MISTRAL],
|
|
|
|
| 17 |
"Dolphin-5.1-7b.Q5_K_M.gguf": ["mradermacher/Dolphin-5.1-7b-GGUF", MessagesFormatterType.MISTRAL],
|
| 18 |
"Captain-Eris-Diogenes_Twighlight-V0.420-12B.i1-Q4_K_M.gguf": ["mradermacher/Captain-Eris-Diogenes_Twighlight-V0.420-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 19 |
"Captain-Eris-Diogenes_Twilight-V0.420-12B.i1-Q4_K_M.gguf": ["mradermacher/Captain-Eris-Diogenes_Twilight-V0.420-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
|
@@ -129,9 +130,41 @@ llm_models = {
|
|
| 129 |
"Aurora-SCE-12B.Q4_K_M.gguf": ["mradermacher/Aurora-SCE-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 130 |
"Rune-14b.Q4_K_M.gguf": ["mradermacher/Rune-14b-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 131 |
"L3.1-Athena-e-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Athena-e-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
|
|
|
| 132 |
"PR2-14B-Instruct.Q4_K_M.gguf": ["mradermacher/PR2-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
|
|
|
|
|
|
|
|
|
| 133 |
"Nemo-12b-Humanize-KTO-Experimental-Latest.Q4_K_M.gguf": ["mradermacher/Nemo-12b-Humanize-KTO-Experimental-Latest-GGUF", MessagesFormatterType.CHATML],
|
| 134 |
"Baptist-Christian-Bible-Expert-v1.1-12B.Q4_K_M.gguf": ["mradermacher/Baptist-Christian-Bible-Expert-v1.1-12B-GGUF", MessagesFormatterType.MISTRAL],
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 135 |
"openbuddy-qwen2.5llamaify-14b-v23.1-200k.i1-Q4_K_M.gguf": ["mradermacher/openbuddy-qwen2.5llamaify-14b-v23.1-200k-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 136 |
"Trinas_Nectar-8B-model_stock.i1-Q4_K_M.gguf": ["mradermacher/Trinas_Nectar-8B-model_stock-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 137 |
"ChatWaifu_Magnum_V0.2.Q4_K_M.gguf": ["mradermacher/ChatWaifu_Magnum_V0.2-GGUF", MessagesFormatterType.MISTRAL],
|
|
@@ -145,6 +178,715 @@ llm_models = {
|
|
| 145 |
#"": ["", MessagesFormatterType.OPEN_CHAT],
|
| 146 |
#"": ["", MessagesFormatterType.CHATML],
|
| 147 |
#"": ["", MessagesFormatterType.PHI_3],
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 148 |
"Hush-Qwen2.5-7B-MST-v1.1.Q5_K_M.gguf": ["mradermacher/Hush-Qwen2.5-7B-MST-v1.1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 149 |
"MFANNv0.22.1.Q5_K_M.gguf": ["mradermacher/MFANNv0.22.1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 150 |
"gemma2-9b-cpt-sea-lionv3-instruct.Q4_K_M.gguf": ["mradermacher/gemma2-9b-cpt-sea-lionv3-instruct-GGUF", MessagesFormatterType.ALPACA],
|
|
|
|
| 14 |
"mn-12b-lyra-v2a1-q5_k_m.gguf": ["HalleyStarbun/MN-12B-Lyra-v2a1-Q5_K_M-GGUF", MessagesFormatterType.CHATML],
|
| 15 |
"L3-8B-Tamamo-v1.i1-Q5_K_M.gguf": ["mradermacher/L3-8B-Tamamo-v1-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 16 |
"MN-Chinofun-12B-2.i1-Q4_K_M.gguf": ["mradermacher/MN-Chinofun-12B-2-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 17 |
+
"Shirayukihime-12B.Q4_K_M.gguf": ["mradermacher/Shirayukihime-12B-GGUF", MessagesFormatterType.CHATML],
|
| 18 |
"Dolphin-5.1-7b.Q5_K_M.gguf": ["mradermacher/Dolphin-5.1-7b-GGUF", MessagesFormatterType.MISTRAL],
|
| 19 |
"Captain-Eris-Diogenes_Twighlight-V0.420-12B.i1-Q4_K_M.gguf": ["mradermacher/Captain-Eris-Diogenes_Twighlight-V0.420-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 20 |
"Captain-Eris-Diogenes_Twilight-V0.420-12B.i1-Q4_K_M.gguf": ["mradermacher/Captain-Eris-Diogenes_Twilight-V0.420-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
|
|
|
| 130 |
"Aurora-SCE-12B.Q4_K_M.gguf": ["mradermacher/Aurora-SCE-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 131 |
"Rune-14b.Q4_K_M.gguf": ["mradermacher/Rune-14b-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 132 |
"L3.1-Athena-e-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Athena-e-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 133 |
+
"ElitePhi.i1-Q4_K_M.gguf": ["mradermacher/ElitePhi-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 134 |
"PR2-14B-Instruct.Q4_K_M.gguf": ["mradermacher/PR2-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 135 |
+
"Nemo_Pony_2-merged.Q4_K_M.gguf": ["mradermacher/Nemo_Pony_2-merged-GGUF", MessagesFormatterType.MISTRAL],
|
| 136 |
+
"MN-12B-Mimicore-Nocturne.Q4_K_M.gguf": ["mradermacher/MN-12B-Mimicore-Nocturne-GGUF", MessagesFormatterType.MISTRAL],
|
| 137 |
+
"Orca-2-13b-abliterated.Q4_K_M.gguf": ["mradermacher/Orca-2-13b-abliterated-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 138 |
"Nemo-12b-Humanize-KTO-Experimental-Latest.Q4_K_M.gguf": ["mradermacher/Nemo-12b-Humanize-KTO-Experimental-Latest-GGUF", MessagesFormatterType.CHATML],
|
| 139 |
"Baptist-Christian-Bible-Expert-v1.1-12B.Q4_K_M.gguf": ["mradermacher/Baptist-Christian-Bible-Expert-v1.1-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 140 |
+
"MN-Hekate-Enodia-12B.Q4_K_M.gguf": ["mradermacher/MN-Hekate-Enodia-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 141 |
+
"MN-Hekate-Ekklesia-12B.Q4_K_M.gguf": ["mradermacher/MN-Hekate-Ekklesia-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 142 |
+
"KANN-I-0.1-7b.Q5_K_M.gguf": ["mradermacher/KANN-I-0.1-7b-GGUF", MessagesFormatterType.MISTRAL],
|
| 143 |
+
"UnslopNemo-12B-v1.Q4_K_M.gguf": ["mradermacher/UnslopNemo-12B-v1-GGUF", MessagesFormatterType.MISTRAL],
|
| 144 |
+
"Irix-12B-Model_Stock.Q4_K_M.gguf": ["mradermacher/Irix-12B-Model_Stock-GGUF", MessagesFormatterType.CHATML],
|
| 145 |
+
"MagnaMellRei-v1-12B.Q4_K_M.gguf": ["mradermacher/MagnaMellRei-v1-12B-GGUF", MessagesFormatterType.CHATML],
|
| 146 |
+
"HMS-Slerp-12B.Q4_K_M.gguf": ["mradermacher/HMS-Slerp-12B-GGUF", MessagesFormatterType.CHATML],
|
| 147 |
+
"SnowElf-12B-v2.Q4_K_M.gguf": ["mradermacher/SnowElf-12B-v2-GGUF", MessagesFormatterType.CHATML],
|
| 148 |
+
"Queen-2.5-14B-aka.Q4_K_M.gguf": ["mradermacher/Queen-2.5-14B-aka-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 149 |
+
"KnowledgeCore-12B.Q4_K_M.gguf": ["mradermacher/KnowledgeCore-12B-GGUF", MessagesFormatterType.CHATML],
|
| 150 |
+
"Mistral-Nemo-2407-Role-Playing-Final-4data_ga16_lr6e7.Q4_K_M.gguf": ["mradermacher/Mistral-Nemo-2407-Role-Playing-Final-4data_ga16_lr6e7-GGUF", MessagesFormatterType.CHATML],
|
| 151 |
+
"PatriSlush-DarkRPReign-12B.Q4_K_M.gguf": ["mradermacher/PatriSlush-DarkRPReign-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 152 |
+
"BianCang-Qwen2.5-14B-Instruct.Q4_K_M.gguf": ["mradermacher/BianCang-Qwen2.5-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 153 |
+
"Simulation_LLM_wiki_14B_V2.Q4_K_M.gguf": ["mradermacher/Simulation_LLM_wiki_14B_V2-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 154 |
+
"Neona-12B.i1-Q4_K_M.gguf": ["mradermacher/Neona-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 155 |
+
"NeoSage-12B.Q4_K_M.gguf": ["mradermacher/NeoSage-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 156 |
+
"Patricide-12B-Forgottenslop-Mell.i1-Q4_K_M.gguf": ["mradermacher/Patricide-12B-Forgottenslop-Mell-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 157 |
+
"Pinecone-Rune-12b.Q4_K_M.gguf": ["mradermacher/Pinecone-Rune-12b-GGUF", MessagesFormatterType.MISTRAL],
|
| 158 |
+
"mn-12b-rp-without-dumb.Q4_K_M.gguf": ["mradermacher/mn-12b-rp-without-dumb-GGUF", MessagesFormatterType.MISTRAL],
|
| 159 |
+
"Denker-mistral-nemo-12B.Q4_K_M.gguf": ["mradermacher/Denker-mistral-nemo-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 160 |
+
"Goldcide-12B-Forgottenslop-Mell.Q4_K_M.gguf": ["mradermacher/Goldcide-12B-Forgottenslop-Mell-GGUF", MessagesFormatterType.MISTRAL],
|
| 161 |
+
"claude-3.7-sonnet-reasoning-gemma3-12B.Q4_K_M.gguf": ["mradermacher/claude-3.7-sonnet-reasoning-gemma3-12B-GGUF", MessagesFormatterType.ALPACA],
|
| 162 |
+
"allura-org_MN-Lyrebird-12B-Q4_K_M.gguf": ["bartowski/allura-org_MN-Lyrebird-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 163 |
+
"ape-fiction-2-mistral-nemo.Q4_K_M.gguf": ["mradermacher/ape-fiction-2-mistral-nemo-GGUF", MessagesFormatterType.MISTRAL],
|
| 164 |
+
"Irixxed_Homunculus-12B-Q3T-v.0.3-Reasoner.Q4_K_M.gguf": ["mradermacher/Irixxed_Homunculus-12B-Q3T-v.0.3-Reasoner-GGUF", MessagesFormatterType.MISTRAL],
|
| 165 |
+
"Gemma-2-Llama-Swallow-9b-pt-v0.1.Q4_K_M.gguf": ["mradermacher/Gemma-2-Llama-Swallow-9b-pt-v0.1-GGUF", MessagesFormatterType.ALPACA],
|
| 166 |
+
"Qwen2.5-7B-base-french-bespoke-stratos-full-sft.i1-Q5_K_S.gguf": ["mradermacher/Qwen2.5-7B-base-french-bespoke-stratos-full-sft-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 167 |
+
"Protestant-Christian-Bible-Expert-v2.0-12B.Q4_K_M.gguf": ["mradermacher/Protestant-Christian-Bible-Expert-v2.0-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 168 |
"openbuddy-qwen2.5llamaify-14b-v23.1-200k.i1-Q4_K_M.gguf": ["mradermacher/openbuddy-qwen2.5llamaify-14b-v23.1-200k-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 169 |
"Trinas_Nectar-8B-model_stock.i1-Q4_K_M.gguf": ["mradermacher/Trinas_Nectar-8B-model_stock-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 170 |
"ChatWaifu_Magnum_V0.2.Q4_K_M.gguf": ["mradermacher/ChatWaifu_Magnum_V0.2-GGUF", MessagesFormatterType.MISTRAL],
|
|
|
|
| 178 |
#"": ["", MessagesFormatterType.OPEN_CHAT],
|
| 179 |
#"": ["", MessagesFormatterType.CHATML],
|
| 180 |
#"": ["", MessagesFormatterType.PHI_3],
|
| 181 |
+
"SauerHuatuoSkyworkDeepWatt-o1-Llama-3.1-8B.Q5_K_M.gguf": ["mradermacher/SauerHuatuoSkyworkDeepWatt-o1-Llama-3.1-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 182 |
+
"care-japanese-llama3.1-8b.Q5_K_M.gguf": ["mradermacher/care-japanese-llama3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 183 |
+
"UltraPatriMerge-12B.Q4_K_M.gguf": ["mradermacher/UltraPatriMerge-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 184 |
+
"Llama-3.1-Amelia-MTFT-8B-v1.Q5_K_M.gguf": ["mradermacher/Llama-3.1-Amelia-MTFT-8B-v1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 185 |
+
"llama3-archimate-merged.Q5_K_M.gguf": ["mradermacher/llama3-archimate-merged-GGUF", MessagesFormatterType.LLAMA_3],
|
| 186 |
+
"Mistral-Nemo-Base-2407-RP-Merge.Q4_K_M.gguf": ["mradermacher/Mistral-Nemo-Base-2407-RP-Merge-GGUF", MessagesFormatterType.CHATML],
|
| 187 |
+
"PatriMaid-12B-Forgottenslop-NeonMell.Q4_K_M.gguf": ["mradermacher/PatriMaid-12B-Forgottenslop-NeonMell-GGUF", MessagesFormatterType.MISTRAL],
|
| 188 |
+
"Magnolia-v3-medis-dilute-12B.Q4_K_M.gguf": ["mradermacher/Magnolia-v3-medis-dilute-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 189 |
+
"Magnolia-v3b-12B.Q4_K_M.gguf": ["mradermacher/Magnolia-v3b-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 190 |
+
"MN-Mystic-Rune-12B.Q4_K_S.gguf": ["mradermacher/MN-Mystic-Rune-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 191 |
+
"MarinaraSpaghetti-NemoMix-Unleashed-12B-chat.Q4_K_M.gguf": ["mradermacher/MarinaraSpaghetti-NemoMix-Unleashed-12B-chat-GGUF", MessagesFormatterType.MISTRAL],
|
| 192 |
+
"GoldFox-12B-Forgottenslop-Mell.i1-Q4_K_M.gguf": ["mradermacher/GoldFox-12B-Forgottenslop-Mell-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 193 |
+
"Magnolia-Mell-v1-12B.Q4_K_M.gguf": ["mradermacher/Magnolia-Mell-v1-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 194 |
+
"Shisa-DellaTest-12B.Q4_K_M.gguf": ["mradermacher/Shisa-DellaTest-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 195 |
+
"Q2.5-Coldbrew14B-FusionMix.i1-Q4_K_M.gguf": ["mradermacher/Q2.5-Coldbrew14B-FusionMix-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 196 |
+
"nemo-instruct-books-model-stock.Q4_K_M.gguf": ["mradermacher/nemo-instruct-books-model-stock-GGUF", MessagesFormatterType.MISTRAL],
|
| 197 |
+
"FoxCide-12B-Forgottenslop-Mell.Q4_K_M.gguf": ["mradermacher/FoxCide-12B-Forgottenslop-Mell-GGUF", MessagesFormatterType.MISTRAL],
|
| 198 |
+
"BMO-CaptianMaid-12B.i1-Q4_K_M.gguf": ["mradermacher/BMO-CaptianMaid-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 199 |
+
"Aurore-Reveil_Koto-Small-7B-IT-Q5_K_M.gguf": ["bartowski/Aurore-Reveil_Koto-Small-7B-IT-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 200 |
+
"Pinecone-Rune-12b-Token-Surgery-Chatml-v0.1a.i1-Q4_K_M.gguf": ["mradermacher/Pinecone-Rune-12b-Token-Surgery-Chatml-v0.1a-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 201 |
+
"Kitsune-Symphony-V0.0-12B.Q4_K_M.gguf": ["mradermacher/Kitsune-Symphony-V0.0-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 202 |
+
"Anora-12b.i1-Q4_K_M.gguf": ["mradermacher/Anora-12b-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 203 |
+
"Minor-Repo-12B-omg.Q4_K_M.gguf": ["mradermacher/Minor-Repo-12B-omg-GGUF", MessagesFormatterType.MISTRAL],
|
| 204 |
+
"Luna.i1-Q5_K_M.gguf": ["mradermacher/Luna-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 205 |
+
"OmegaMaid-DarkWay-FinalNoctis-12B.Q4_K_M.gguf": ["mradermacher/OmegaMaid-DarkWay-FinalNoctis-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 206 |
+
"CaptainMaid-12B-VioletMell-V0.420.Q4_K_M.gguf": ["mradermacher/CaptainMaid-12B-VioletMell-V0.420-GGUF", MessagesFormatterType.MISTRAL],
|
| 207 |
+
"LatentSoup-modelstock-8b.Q5_K_M.gguf": ["mradermacher/LatentSoup-modelstock-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 208 |
+
"Erotic-Model.v1.Q5_K_M.gguf": ["mradermacher/Erotic-Model.v1-GGUF", MessagesFormatterType.MISTRAL],
|
| 209 |
+
"Llama-3.1-Amelia-CD-8B-v1.Q5_K_M.gguf": ["mradermacher/Llama-3.1-Amelia-CD-8B-v1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 210 |
+
"funny-nemo-embedding-merged.Q4_K_M.gguf": ["mradermacher/funny-nemo-embedding-merged-GGUF", MessagesFormatterType.MISTRAL],
|
| 211 |
+
"EviOmni-nq_train-7B.Q5_K_M.gguf": ["mradermacher/EviOmni-nq_train-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 212 |
+
"NuMarkdown-8B-Thinking-fork.Q5_K_M.gguf": ["mradermacher/NuMarkdown-8B-Thinking-fork-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 213 |
+
"Comet_12B_V.7.Q4_K_M.gguf": ["mradermacher/Comet_12B_V.7-GGUF", MessagesFormatterType.ALPACA],
|
| 214 |
+
"Llama-3.1-8B-Instruct-wildfeedback-seed-RPO-0.1.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-Instruct-wildfeedback-seed-RPO-0.1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 215 |
+
"NeuralMerge-9B-Dare.Q5_K_M.gguf": ["mradermacher/NeuralMerge-9B-Dare-GGUF", MessagesFormatterType.MISTRAL],
|
| 216 |
+
"CaptainErisNebula-12B-Chimera-v0.420.Q4_K_M.gguf": ["mradermacher/CaptainErisNebula-12B-Chimera-v0.420-GGUF", MessagesFormatterType.MISTRAL],
|
| 217 |
+
"Nemo-12B-OldSpice.Q4_K_M.gguf": ["mradermacher/Nemo-12B-OldSpice-GGUF", MessagesFormatterType.MISTRAL],
|
| 218 |
+
"funny-nemo-embedding-testing.Q4_K_M.gguf": ["mradermacher/funny-nemo-embedding-testing-GGUF", MessagesFormatterType.MISTRAL],
|
| 219 |
+
"francois-v3.Q4_K_M.gguf": ["mradermacher/francois-v3-GGUF", MessagesFormatterType.CHATML],
|
| 220 |
+
"KansenSakura-Eclipse-RP-12b.Q4_K_M.gguf": ["mradermacher/KansenSakura-Eclipse-RP-12b-GGUF", MessagesFormatterType.CHATML],
|
| 221 |
+
"Sugoi-14B-Ultra-HF.Q4_K_M.gguf": ["mradermacher/Sugoi-14B-Ultra-HF-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 222 |
+
"CaptainErisNebula-12B-Chimera-v1.1.i1-Q4_K_M.gguf": ["mradermacher/CaptainErisNebula-12B-Chimera-v1.1-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 223 |
+
"NOVER1-Qwen2.5-7B.i1-Q5_K_M.gguf": ["mradermacher/NOVER1-Qwen2.5-7B-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 224 |
+
"MT-Gen6-U-gemma-2-Tv3Tv2-9B.Q4_K_M.gguf": ["mradermacher/MT-Gen6-U-gemma-2-Tv3Tv2-9B-GGUF", MessagesFormatterType.ALPACA],
|
| 225 |
+
"SimpleChat-4B-V1.i1-Q5_K_M.gguf": ["mradermacher/SimpleChat-4B-V1-i1-GGUF", MessagesFormatterType.ALPACA],
|
| 226 |
+
"MathTutor-7B-MDPO_v0.1.Q5_K_M.gguf": ["mradermacher/MathTutor-7B-MDPO_v0.1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 227 |
+
"MathTutor-7B-H_v0.0.1.Q5_K_M.gguf": ["mradermacher/MathTutor-7B-H_v0.0.1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 228 |
+
"Wingless-Alice-RP-NSFW-8B.Q4_K_S.gguf": ["mradermacher/Wingless-Alice-RP-NSFW-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 229 |
+
"Shisa-K-12B.Q4_K_M.gguf": ["mradermacher/Shisa-K-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 230 |
+
"UI-Venus-Ground-7B.Q5_K_M.gguf": ["mradermacher/UI-Venus-Ground-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 231 |
+
"Smilodon-9B-v0.5.i1-Q4_K_M.gguf": ["mradermacher/Smilodon-9B-v0.5-i1-GGUF", MessagesFormatterType.ALPACA],
|
| 232 |
+
"Qwen2.5-VL-7B-NSFW-Caption-V3.Q5_K_M.gguf": ["mradermacher/Qwen2.5-VL-7B-NSFW-Caption-V3-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 233 |
+
"Arrow-R1-7B.Q5_K_M.gguf": ["mradermacher/Arrow-R1-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 234 |
+
"Moondark-12B.i1-Q4_K_M.gguf": ["mradermacher/Moondark-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 235 |
+
"Moonviolet-12B.Q4_K_M.gguf": ["mradermacher/Moonviolet-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 236 |
+
"MoonMega-12B.Q4_K_M.gguf": ["mradermacher/MoonMega-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 237 |
+
"Qwen2.5-7B-Instruct-wildfeedback-iterDPO-iter2-4k.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-wildfeedback-iterDPO-iter2-4k-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 238 |
+
"Josiefied-Qwen3-4B-Instruct-2507-gabliterated-v1.i1-Q5_K_M.gguf": ["mradermacher/Josiefied-Qwen3-4B-Instruct-2507-gabliterated-v1-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 239 |
+
"Qwen2.5-14B-Instruct-wildfeedback-RPO-DRIFT-iter1-4k.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Instruct-wildfeedback-RPO-DRIFT-iter1-4k-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 240 |
+
"gemma3-12b-tolkien_v3.Q4_K_M.gguf": ["mradermacher/gemma3-12b-tolkien_v3-GGUF", MessagesFormatterType.ALPACA],
|
| 241 |
+
"DellaMix-12B.i1-Q4_K_M.gguf": ["mradermacher/DellaMix-12B-i1-GGUF", MessagesFormatterType.CHATML],
|
| 242 |
+
"EtherealAurora-12B-Lorablated.Q4_K_M.gguf": ["mradermacher/EtherealAurora-12B-Lorablated-GGUF", MessagesFormatterType.MISTRAL],
|
| 243 |
+
"Not-WizardLM-2-The-Omega-Directive-7b-Unslop-v2.1.i1-Q5_K_M.gguf": ["mradermacher/Not-WizardLM-2-The-Omega-Directive-7b-Unslop-v2.1-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 244 |
+
"gemma3-12b-tolkien_v2.i1-Q4_K_M.gguf": ["mradermacher/gemma3-12b-tolkien_v2-i1-GGUF", MessagesFormatterType.ALPACA],
|
| 245 |
+
"Qwen-2.5-3b-Evol-CoT.Q5_K_M.gguf": ["mradermacher/Qwen-2.5-3b-Evol-CoT-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 246 |
+
"VersaVid-R1.Q5_K_M.gguf": ["mradermacher/VersaVid-R1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 247 |
+
"sphinxnautics-7b-kl-distilled.Q5_K_M.gguf": ["mradermacher/sphinxnautics-7b-kl-distilled-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 248 |
+
"gemma3-12b-tolkien.Q5_K_M.gguf": ["mradermacher/gemma3-12b-tolkien-GGUF", MessagesFormatterType.ALPACA],
|
| 249 |
+
"ThinkLite-Critic-7b.Q5_K_M.gguf": ["mradermacher/ThinkLite-Critic-7b-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 250 |
+
"Qwen2.5-nsfw.Q5_K_M.gguf": ["mradermacher/Qwen2.5-nsfw-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 251 |
+
"Rewiz-Tom-7B.Q5_K_M.gguf": ["mradermacher/Rewiz-Tom-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 252 |
+
"NextCoder-Mirage-sol-7B.i1-Q5_K_M.gguf": ["mradermacher/NextCoder-Mirage-sol-7B-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 253 |
+
"qwen2.5-7b-v1.Q5_K_M.gguf": ["mradermacher/qwen2.5-7b-v1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 254 |
+
"CriticLeanGPT-Qwen2.5-7B-Instruct-SFT.i1-Q5_K_M.gguf": ["mradermacher/CriticLeanGPT-Qwen2.5-7B-Instruct-SFT-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 255 |
+
"OpenR1-Qwen-7B-Instruct-SFT-lora-ctx16k.Q5_K_M.gguf": ["mradermacher/OpenR1-Qwen-7B-Instruct-SFT-lora-ctx16k-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 256 |
+
"CriticLeanGPT-Qwen2.5-14B-Instruct-SFT-RL.Q4_K_M.gguf": ["mradermacher/CriticLeanGPT-Qwen2.5-14B-Instruct-SFT-RL-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 257 |
+
"Qwen3-14B-SFT.Q5_K_M.gguf": ["mradermacher/Qwen3-14B-SFT-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 258 |
+
"Blossom-V6.1-8B.Q5_K_M.gguf": ["mradermacher/Blossom-V6.1-8B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 259 |
+
"Homunculus-abliterated.Q4_K_M.gguf": ["mradermacher/Homunculus-abliterated-GGUF", MessagesFormatterType.CHATML],
|
| 260 |
+
"GThinker-7B.i1-Q5_K_M.gguf": ["mradermacher/GThinker-7B-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 261 |
+
"TAMA-QWen2.5.i1-Q5_K_M.gguf": ["mradermacher/TAMA-QWen2.5-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 262 |
+
"Qwen3-psychological-reasoning-1.7B.Q5_K_M.gguf": ["mradermacher/Qwen3-psychological-reasoning-1.7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 263 |
+
"QuasiStarSynth-12B.Q4_K_M.gguf": ["mradermacher/QuasiStarSynth-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 264 |
+
"Solution-back-7B.Q5_K_M.gguf": ["mradermacher/Solution-back-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 265 |
+
"Qwen2.5-3B-UFO-hotpotqa-1turn.Q5_K_M.gguf": ["mradermacher/Qwen2.5-3B-UFO-hotpotqa-1turn-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 266 |
+
"AbyssSynth-12B.Q4_K_M.gguf": ["mradermacher/AbyssSynth-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 267 |
+
"E-Star-12B-v0.1.Q4_K_M.gguf": ["mradermacher/E-Star-12B-v0.1-GGUF", MessagesFormatterType.ALPACA],
|
| 268 |
+
"SingularitySynth-12B.i1-Q4_K_M.gguf": ["mradermacher/SingularitySynth-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 269 |
+
"MT2-Gen2_gemma-3-12B.Q4_K_M.gguf": ["mradermacher/MT2-Gen2_gemma-3-12B-GGUF", MessagesFormatterType.ALPACA],
|
| 270 |
+
"LilithCore-v1-12B.Q4_K_M.gguf": ["mradermacher/LilithCore-v1-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 271 |
+
"WeThink-Qwen2.5VL-7B.i1-Q5_K_M.gguf": ["mradermacher/WeThink-Qwen2.5VL-7B-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 272 |
+
"RuleReasoner-8B.Q5_K_M.gguf": ["mradermacher/RuleReasoner-8B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 273 |
+
"Vision-Matters-7B.Q5_K_M.gguf": ["mradermacher/Vision-Matters-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 274 |
+
"gemma-3-cn-novel-4b.Q5_K_M.gguf": ["mradermacher/gemma-3-cn-novel-4b-GGUF", MessagesFormatterType.ALPACA],
|
| 275 |
+
"Qwen2.5-7B-Instruct-ultrafeedback-10k.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-ultrafeedback-10k-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 276 |
+
"RACRO-7B-CRO.i1-Q5_K_M.gguf": ["mradermacher/RACRO-7B-CRO-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 277 |
+
"QVikhr-3-4B-Instruction.Q6_K.gguf": ["mradermacher/QVikhr-3-4B-Instruction-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 278 |
+
"care-japanese-gemma2-9b.Q4_K_M.gguf": ["mradermacher/care-japanese-gemma2-9b-GGUF", MessagesFormatterType.ALPACA],
|
| 279 |
+
"Geo-SFT.i1-Q5_K_M.gguf": ["mradermacher/Geo-SFT-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 280 |
+
"Geo-R1-CI-0620.Q5_K_M.gguf": ["mradermacher/Geo-R1-CI-0620-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 281 |
+
"SpaceQwen2.5-VL-3B-Instruct.Q5_K_M.gguf": ["mradermacher/SpaceQwen2.5-VL-3B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 282 |
+
"Metis-RISE-7B.i1-Q5_K_M.gguf": ["mradermacher/Metis-RISE-7B-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 283 |
+
"OpenBuddy-R10528DistillQwen-14B-v27.1.i1-Q4_K_M.gguf": ["mradermacher/OpenBuddy-R10528DistillQwen-14B-v27.1-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 284 |
+
"sft_subtitles_fixer_v1_merged.Q4_K_M.gguf": ["mradermacher/sft_subtitles_fixer_v1_merged-GGUF", MessagesFormatterType.ALPACA],
|
| 285 |
+
"Qwen3-8B-Josiefied-iSMART.Q6_K.gguf": ["mradermacher/Qwen3-8B-Josiefied-iSMART-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 286 |
+
"Polar-G3-DPO-12b.i1-Q4_K_M.gguf": ["mradermacher/Polar-G3-DPO-12b-i1-GGUF", MessagesFormatterType.ALPACA],
|
| 287 |
+
"DREX-062225-exp.Q5_K_M.gguf": ["mradermacher/DREX-062225-exp-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 288 |
+
"Japanese-Qwen2.5-14B-Instruct-V1.Q4_K_M.gguf": ["mradermacher/Japanese-Qwen2.5-14B-Instruct-V1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 289 |
+
"Qwen2.5-7B-Instruct-userfeedback-on-policy-iter2.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-userfeedback-on-policy-iter2-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 290 |
+
"Configurable-Hermes-3-Llama-3.1-8B.Q5_K_M.gguf": ["mradermacher/Configurable-Hermes-3-Llama-3.1-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 291 |
+
"KiloNovaSynth-12B.Q4_K_M.gguf": ["mradermacher/KiloNovaSynth-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 292 |
+
"Qwen2.5-7B-Instruct-userfeedback-on-policy-iter1.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-userfeedback-on-policy-iter1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 293 |
+
"Llama-3.1-Swallow-8B-Instruct-v0.5-Q5_K_M.gguf": ["mmnga/Llama-3.1-Swallow-8B-Instruct-v0.5-gguf", MessagesFormatterType.LLAMA_3],
|
| 294 |
+
"Bohdi-Qwen2.5-7B-Instruct.i1-Q5_K_M.gguf": ["mradermacher/Bohdi-Qwen2.5-7B-Instruct-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 295 |
+
"Q2.5-ColdBrew-RxV.Q4_K_M.gguf": ["mradermacher/Q2.5-ColdBrew-RxV-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 296 |
+
"SuperNovaSynth-12B.Q4_K_M.gguf": ["mradermacher/SuperNovaSynth-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 297 |
+
"HyperNovaSynth-12B.Q4_K_M.gguf": ["mradermacher/HyperNovaSynth-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 298 |
+
"Simulation_LLM_wiki_7B_V2.Q5_K_M.gguf": ["mradermacher/Simulation_LLM_wiki_7B_V2-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 299 |
+
"Qwen2.5-7B-Instruct-userfeedback-SPIN-iter1.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-userfeedback-SPIN-iter1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 300 |
+
"Qwen2.5-7B-Instruct-userfeedback-SFT-SPIN-iter1.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-userfeedback-SFT-SPIN-iter1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 301 |
+
"PatriSlush-DarkReign-12B.Q4_K_M.gguf": ["mradermacher/PatriSlush-DarkReign-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 302 |
+
"Hermes-2-Theta-Llama-3-8B-32k.i1-Q5_K_M.gguf": ["mradermacher/Hermes-2-Theta-Llama-3-8B-32k-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 303 |
+
"QFFT-S1-7B.Q5_K_M.gguf": ["mradermacher/QFFT-S1-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 304 |
+
"QFFT-LIMO-7B.Q5_K_M.gguf": ["mradermacher/QFFT-LIMO-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 305 |
+
"Omega-Darker_The-Final-Directive-Longform-Stage2-ERP-12B-v0.2.Q4_K_M.gguf": ["mradermacher/Omega-Darker_The-Final-Directive-Longform-Stage2-ERP-12B-v0.2-GGUF", MessagesFormatterType.MISTRAL],
|
| 306 |
+
"Qwen2.5-7B-Instruct-userfeedback-sentiment-seed.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-userfeedback-sentiment-seed-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 307 |
+
"Irix-12B_Slush_V2.i1-Q4_K_M.gguf": ["mradermacher/Irix-12B_Slush_V2-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 308 |
+
"NeonViolin.Q4_K_M.gguf": ["mradermacher/NeonViolin-GGUF", MessagesFormatterType.MISTRAL],
|
| 309 |
+
"Qwen2.5-3B-rat-polished.Q5_K_M.gguf": ["mradermacher/Qwen2.5-3B-rat-polished-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 310 |
+
"Mistral-qwq-12b-merge.Q4_K_M.gguf": ["mradermacher/Mistral-qwq-12b-merge-GGUF", MessagesFormatterType.MISTRAL],
|
| 311 |
+
"Situations-Model-1.1.Q5_K_M.gguf": ["mradermacher/Situations-Model-1.1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 312 |
+
"Ego-R1-SFT-3B.Q5_K_M.gguf": ["mradermacher/Ego-R1-SFT-3B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 313 |
+
"Ego-R1-Agent-3B.Q5_K_M.gguf": ["mradermacher/Ego-R1-Agent-3B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 314 |
+
"JudgeLRM-14B.Q4_K_M.gguf": ["mradermacher/JudgeLRM-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 315 |
+
"Irix_MemecapSlerp-12B.Q4_K_M.gguf": ["mradermacher/Irix_MemecapSlerp-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 316 |
+
"Irixxed-Magcap-12B-Slerp.Q4_K_M.gguf": ["mradermacher/Irixxed-Magcap-12B-Slerp-GGUF", MessagesFormatterType.MISTRAL],
|
| 317 |
+
"The-Omega-Directive-M-12B-Unslop-v2.0.Q4_K_M.gguf": ["mradermacher/The-Omega-Directive-M-12B-Unslop-v2.0-GGUF", MessagesFormatterType.MISTRAL],
|
| 318 |
+
"ZYH-LLM-Qwen2.5-14B-V5.Q4_K_M.gguf": ["mradermacher/ZYH-LLM-Qwen2.5-14B-V5-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 319 |
+
"LorablatedStock-12B.Q4_K_M.gguf": ["mradermacher/LorablatedStock-12B-GGUF", MessagesFormatterType.CHATML],
|
| 320 |
+
"Qwen2.5-7B-Chess-MMXL.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Chess-MMXL-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 321 |
+
"Mistral-Nemo-Gutenberg-Encore-12B.Q4_K_M.gguf": ["mradermacher/Mistral-Nemo-Gutenberg-Encore-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 322 |
+
"Qwen2.5-7B-Instruct-wildfeedback.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-wildfeedback-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 323 |
+
"Shisa-v2-Mistral-Nemo-12B-Lorablated.Q4_K_M.gguf": ["mradermacher/Shisa-v2-Mistral-Nemo-12B-Lorablated-GGUF", MessagesFormatterType.MISTRAL],
|
| 324 |
+
"L3-Dark-Planet-8B-wordstorm2.Q5_K_M.gguf": ["mradermacher/L3-Dark-Planet-8B-wordstorm2-GGUF", MessagesFormatterType.LLAMA_3],
|
| 325 |
+
"Qwen2.5-7B-Instruct-userfeedback.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-userfeedback-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 326 |
+
"L3-Dark-Planet-8B-wordstorm-cr2.Q5_K_M.gguf": ["mradermacher/L3-Dark-Planet-8B-wordstorm-cr2-GGUF", MessagesFormatterType.LLAMA_3],
|
| 327 |
+
"L3-Dark-Planet-8B-wordstorm-b6.i1-Q5_K_M.gguf": ["mradermacher/L3-Dark-Planet-8B-wordstorm-b6-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 328 |
+
"Gilded-Tempest-12B.Q4_K_S.gguf": ["mradermacher/Gilded-Tempest-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 329 |
+
"Qwen2.5-7B-Instruct-GRPO-Nei-Tokens.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-GRPO-Nei-Tokens-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 330 |
+
"FusionEngine-12B.Q4_K_M.gguf": ["mradermacher/FusionEngine-12B-GGUF", MessagesFormatterType.CHATML],
|
| 331 |
+
"ShiZhen-V0.1.Q5_K_M.gguf": ["mradermacher/ShiZhen-V0.1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 332 |
+
"HMS-Fusion-12B.Q4_K_M.gguf": ["mradermacher/HMS-Fusion-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 333 |
+
"TutorRL-7B.Q5_K_M.gguf": ["mradermacher/TutorRL-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 334 |
+
"Omega-Darker_The-Final-Directive-14B.Q4_K_M.gguf": ["mradermacher/Omega-Darker_The-Final-Directive-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 335 |
+
"Student-rl_Qwen2.5-7B-Instruct_v0.0.1.Q5_K_M.gguf": ["mradermacher/Student-rl_Qwen2.5-7B-Instruct_v0.0.1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 336 |
+
"Contrl-Stheno-v1-8B.Q5_K_M.gguf": ["mradermacher/Contrl-Stheno-v1-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 337 |
+
"fluxiia_14b.Q4_K_M.gguf": ["mradermacher/fluxiia_14b-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 338 |
+
"JwenJwen1-1.5b.Q5_K_M.gguf": ["mradermacher/JwenJwen1-1.5b-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 339 |
+
"SecGPT-14B-Reasoning.Q4_K_M.gguf": ["mradermacher/SecGPT-14B-Reasoning-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 340 |
+
"Dirty-Muse-Writer-v01-Uncensored-Erotica-NSFW.Q4_K_M.gguf": ["mradermacher/Dirty-Muse-Writer-v01-Uncensored-Erotica-NSFW-GGUF", MessagesFormatterType.ALPACA],
|
| 341 |
+
"SearchR1-nq_hotpotqa_train-qwen2.5-7b-em-ppo.Q4_K_M.gguf": ["mradermacher/SearchR1-nq_hotpotqa_train-qwen2.5-7b-em-ppo-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 342 |
+
"ForgottenMaid-12B-v2.Q4_K_M.gguf": ["mradermacher/ForgottenMaid-12B-v2-GGUF", MessagesFormatterType.MISTRAL],
|
| 343 |
+
"Socratic-Qwen2.5-7B-v2.Q5_K_M.gguf": ["mradermacher/Socratic-Qwen2.5-7B-v2-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 344 |
+
"EsotericSage-12B.i1-Q4_K_M.gguf": ["mradermacher/EsotericSage-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 345 |
+
"praxis-bookwriter-llama3.1-8b-sft.Q5_K_M.gguf": ["mradermacher/praxis-bookwriter-llama3.1-8b-sft-GGUF", MessagesFormatterType.LLAMA_3],
|
| 346 |
+
"LinearWriter-12B.Q4_K_M.gguf": ["mradermacher/LinearWriter-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 347 |
+
"Qwen2.5-CoderX-14B-v0.5.Q4_K_M.gguf": ["mradermacher/Qwen2.5-CoderX-14B-v0.5-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 348 |
+
"Violet_MagCap-Rebase-12B.Q4_K_M.gguf": ["mradermacher/Violet_MagCap-Rebase-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 349 |
+
"VeriReason-Qwen2.5-7b-SFT-Reasoning.i1-Q5_K_M.gguf": ["mradermacher/VeriReason-Qwen2.5-7b-SFT-Reasoning-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 350 |
+
"Emilia-Multislerp-12B.Q4_K_M.gguf": ["mradermacher/Emilia-Multislerp-12B-GGUF", MessagesFormatterType.CHATML],
|
| 351 |
+
"TaskRouter-1.5B.Q5_K_M.gguf": ["mradermacher/TaskRouter-1.5B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 352 |
+
"NeonMaid-12B-v2.Q4_K_M.gguf": ["mradermacher/NeonMaid-12B-v2-GGUF", MessagesFormatterType.CHATML],
|
| 353 |
+
"Orihime-Gutenberg-12B.Q4_K_M.gguf": ["mradermacher/Orihime-Gutenberg-12B-GGUF", MessagesFormatterType.CHATML],
|
| 354 |
+
"SearchSimulation_3B.Q5_K_M.gguf": ["mradermacher/SearchSimulation_3B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 355 |
+
"SearchSimulation_14B.Q4_K_M.gguf": ["mradermacher/SearchSimulation_14B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 356 |
+
"NeonMaid-12B.Q4_K_M.gguf": ["mradermacher/NeonMaid-12B-GGUF", MessagesFormatterType.CHATML],
|
| 357 |
+
"Chief_chain_generator-Qwen2.5-7B-Instruct.Q5_K_M.gguf": ["mradermacher/Chief_chain_generator-Qwen2.5-7B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 358 |
+
"Emotion_inferencer-Qwen2.5-7B-Instruct.Q5_K_M.gguf": ["mradermacher/Emotion_inferencer-Qwen2.5-7B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 359 |
+
"ForgottenMaid-12B.Q4_K_M.gguf": ["mradermacher/ForgottenMaid-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 360 |
+
"AceCoder-Qwen2.5-Coder-7B-Ins-V1.1.Q5_K_M.gguf": ["mradermacher/AceCoder-Qwen2.5-Coder-7B-Ins-V1.1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 361 |
+
"TheDrummer_Rivermind-Lux-12B-v1-Q4_K_M.gguf": ["bartowski/TheDrummer_Rivermind-Lux-12B-v1-GGUF", MessagesFormatterType.MISTRAL],
|
| 362 |
+
"Qwen2.5-GenX-14B.Q4_K_M.gguf": ["mradermacher/Qwen2.5-GenX-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 363 |
+
"vanilla-cn-roleplay-0.1.Q5_K_M.gguf": ["mradermacher/vanilla-cn-roleplay-0.1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 364 |
+
"EsotericLight-12B.Q4_K_M.gguf": ["mradermacher/EsotericLight-12B-GGUF", MessagesFormatterType.CHATML],
|
| 365 |
+
"llama-3-Daredevil-Mahou-8B.Q5_K_M.gguf": ["mradermacher/llama-3-Daredevil-Mahou-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 366 |
+
"Orihime-12B.Q4_K_M.gguf": ["mradermacher/Orihime-12B-GGUF", MessagesFormatterType.CHATML],
|
| 367 |
+
"YM-12B-Model_Stock.Q4_K_M.gguf": ["mradermacher/YM-12B-Model_Stock-GGUF", MessagesFormatterType.MISTRAL],
|
| 368 |
+
"Qwen2.5-7B-Instruct-HardLambda0.1-220.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-HardLambda0.1-220-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 369 |
+
"gemma-2-abliterated-Ifable-9B-untied.Q4_K_M.gguf": ["mradermacher/gemma-2-abliterated-Ifable-9B-untied-GGUF", MessagesFormatterType.ALPACA],
|
| 370 |
+
"Smoothie-Qwen2.5-3B-Instruct.i1-Q5_K_M.gguf": ["mradermacher/Smoothie-Qwen2.5-3B-Instruct-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 371 |
+
"Qwen2.5-14B-Aletheia-cogito.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Aletheia-cogito-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 372 |
+
"Qwen2.5-14B-EMPO-Natural-Reasoning_simple_full.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-EMPO-Natural-Reasoning_simple_full-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 373 |
+
"exurbia-advance-q4_k_m.gguf": ["ClaudioItaly/Exurbia-Advance-Q4_K_M-GGUF", MessagesFormatterType.ALPACA],
|
| 374 |
+
"Levelled-12B-Model_Stock.Q4_K_M.gguf": ["mradermacher/Levelled-12B-Model_Stock-GGUF", MessagesFormatterType.MISTRAL],
|
| 375 |
+
"sapie-gemma2-9B-IT.i1-Q4_K_M.gguf": ["mradermacher/sapie-gemma2-9B-IT-i1-GGUF", MessagesFormatterType.ALPACA],
|
| 376 |
+
"Persona-Qwen2.5-3B-Instruct.Q5_K_M.gguf": ["mradermacher/Persona-Qwen2.5-3B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 377 |
+
"Socratic-Qwen2.5-7B.Q5_K_M.gguf": ["mradermacher/Socratic-Qwen2.5-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 378 |
+
"qwen2.5-1.5B-base-abliterated.Q5_K_M.gguf": ["mradermacher/qwen2.5-1.5B-base-abliterated-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 379 |
+
"Rei-V3-KTO-12B.Q4_K_M.gguf": ["mradermacher/Rei-V3-KTO-12B-GGUF", MessagesFormatterType.CHATML],
|
| 380 |
+
"Qwen2.5-14B-Instinct-Talk.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Instinct-Talk-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 381 |
+
"Llama3.1-Aloe-Beta-8B.Q5_K_M.gguf": ["mradermacher/Llama3.1-Aloe-Beta-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 382 |
+
"gemma-2-9B-it-blend.Q4_K_S.gguf": ["mradermacher/gemma-2-9B-it-blend-GGUF", MessagesFormatterType.ALPACA],
|
| 383 |
+
"recoilme-gemma-2-Ataraxy-9B-v0.1-t0.75.Q4_K_M.gguf": ["mradermacher/recoilme-gemma-2-Ataraxy-9B-v0.1-t0.75-GGUF", MessagesFormatterType.ALPACA],
|
| 384 |
+
"Smoothie-Qwen2.5-1.5B-Instruct.Q5_K_M.gguf": ["mradermacher/Smoothie-Qwen2.5-1.5B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 385 |
+
"ELYZA-Shortcut-1.0-Qwen-7B-Q5_K_M.gguf": ["mmnga/ELYZA-Shortcut-1.0-Qwen-7B-gguf", MessagesFormatterType.OPEN_CHAT],
|
| 386 |
+
"Qwen2.5-7B-Instruct-abliterated.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-abliterated-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 387 |
+
"Crispers-14B-v1.Q4_K_M.gguf": ["mradermacher/Crispers-14B-v1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 388 |
+
"amoral-cogito-Zara-14B-ties.i1-Q4_K_M.gguf": ["mradermacher/amoral-cogito-Zara-14B-ties-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 389 |
+
"Qwen2.5-Kunoulise-D.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Kunoulise-D-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 390 |
+
"Qwen2.5-Kunoulise-C.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Kunoulise-C-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 391 |
+
"Qwen2.5-Kunoulise-B.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Kunoulise-B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 392 |
+
"Qwen2.5-Kunoulise-A.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Kunoulise-A-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 393 |
+
"Crispers-7B-v1.i1-Q5_K_M.gguf": ["mradermacher/Crispers-7B-v1-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 394 |
+
"SnowElf-12B.Q4_K_M.gguf": ["mradermacher/SnowElf-12B-GGUF", MessagesFormatterType.CHATML],
|
| 395 |
+
"Fusion4-14B-Instruct.Q4_K_M.gguf": ["mradermacher/Fusion4-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 396 |
+
"VeriCoder_Qwen14B.Q4_K_M.gguf": ["mradermacher/VeriCoder_Qwen14B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 397 |
+
"hms-slerp-12b-v2-q4_k_m.gguf": ["yamatazen/HMS-Slerp-12B-v2-Q4_K_M-GGUF", MessagesFormatterType.MISTRAL],
|
| 398 |
+
"grpo-qwen7b-triton-10ep.Q5_K_M.gguf": ["mradermacher/grpo-qwen7b-triton-10ep-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 399 |
+
"grpo-qwen7b-triton-5ep.Q5_K_M.gguf": ["mradermacher/grpo-qwen7b-triton-5ep-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 400 |
+
"Kuugo-LawLLM-14B.Q4_K_M.gguf": ["mradermacher/Kuugo-LawLLM-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 401 |
+
"T3Q-qwen2.5-14b-v1.2-e2.Q4_K_M.gguf": ["mradermacher/T3Q-qwen2.5-14b-v1.2-e2-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 402 |
+
"ablation-200-a163.finaldpo2.1e7-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-200-a163.finaldpo2.1e7-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 403 |
+
"ablation-204-a163.finaldpo2.2e7-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-204-a163.finaldpo2.2e7-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 404 |
+
"Vera-v0.1.Q5_K_M.gguf": ["mradermacher/Vera-v0.1-GGUF", MessagesFormatterType.MISTRAL],
|
| 405 |
+
"ablation-192-a175.finaldpo2.8e7-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-192-a175.finaldpo2.8e7-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 406 |
+
"MagSoup-v1-12B.Q4_K_M.gguf": ["mradermacher/MagSoup-v1-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 407 |
+
"LUFFY-Qwen-Instruct-7B.Q5_K_M.gguf": ["mradermacher/LUFFY-Qwen-Instruct-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 408 |
+
"VeriPrefer-Qwen2.5-Coder-7B.Q5_K_M.gguf": ["mradermacher/VeriPrefer-Qwen2.5-Coder-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 409 |
+
"ToolRL-Qwen2.5-3B.Q5_K_M.gguf": ["mradermacher/ToolRL-Qwen2.5-3B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 410 |
+
"TaxoQwen2.5-7B.Q5_K_M.gguf": ["mradermacher/TaxoQwen2.5-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 411 |
+
"Magnolia-v8-12B.Q4_K_M.gguf": ["mradermacher/Magnolia-v8-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 412 |
+
"Llama3.2-3B-Reasoning-SFT-fTuned-v1.Q5_K_M.gguf": ["mradermacher/Llama3.2-3B-Reasoning-SFT-fTuned-v1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 413 |
+
"Qwen2.5-Freyalise-A.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Freyalise-A-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 414 |
+
"Magnolia-v9-12B.Q4_K_M.gguf": ["mradermacher/Magnolia-v9-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 415 |
+
"Magnolia-v10-12B.Q4_K_M.gguf": ["mradermacher/Magnolia-v10-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 416 |
+
"sft_3b_wopolicy.Q5_K_M.gguf": ["mradermacher/sft_3b_wopolicy-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 417 |
+
"Shisa-v2-Mistral-Nemo-12B-Abliterated.Q4_K_M.gguf": ["mradermacher/Shisa-v2-Mistral-Nemo-12B-Abliterated-GGUF", MessagesFormatterType.MISTRAL],
|
| 418 |
+
"Qwen2.5-14B-ZaraTimpist.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-ZaraTimpist-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 419 |
+
"Chronicle-Qwen2.5-7B.Q5_K_M.gguf": ["mradermacher/Chronicle-Qwen2.5-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 420 |
+
"Llama-3.1-1million-ctx-Dark-Planet-v1.01-8B.Q5_K_M.gguf": ["mradermacher/Llama-3.1-1million-ctx-Dark-Planet-v1.01-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 421 |
+
"MN-Hekate-Pyrtania-12B.Q4_K_M.gguf": ["mradermacher/MN-Hekate-Pyrtania-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 422 |
+
"Qwen2.5-14B-Instruct-LIMO-new.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Instruct-LIMO-new-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 423 |
+
"Qwen2.5-14B-Instruct-CognitiveSFT-v0.1.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Instruct-CognitiveSFT-v0.1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 424 |
+
"Qwen2.5-14B-Instruct-o4.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Instruct-o4-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 425 |
+
"Twilight-SCE-12B-v2.Q4_K_S.gguf": ["mradermacher/Twilight-SCE-12B-v2-GGUF", MessagesFormatterType.CHATML],
|
| 426 |
+
"Twilight-SCE-12B.Q4_K_M.gguf": ["mradermacher/Twilight-SCE-12B-GGUF", MessagesFormatterType.CHATML],
|
| 427 |
+
"MN-Hekate-Limenoskopos-12B.Q4_K_M.gguf": ["mradermacher/MN-Hekate-Limenoskopos-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 428 |
+
"Reformed-Christian-Bible-Expert-v2.1-12B.Q4_K_M.gguf": ["mradermacher/Reformed-Christian-Bible-Expert-v2.1-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 429 |
+
"AnotherOne-Unslop-Mell-12B.Q4_K_M.gguf": ["mradermacher/AnotherOne-Unslop-Mell-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 430 |
+
"Golden-Curry-12B.Q4_K_M.gguf": ["mradermacher/Golden-Curry-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 431 |
+
"Violet_Magcap-12B.i1-Q4_K_M.gguf": ["mradermacher/Violet_Magcap-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 432 |
+
"StarrySky-12B.Q4_K_M.gguf": ["mradermacher/StarrySky-12B-GGUF", MessagesFormatterType.CHATML],
|
| 433 |
+
"kunoichi-lemon-royale-v2experiment2-32K-7B.i1-Q5_K_M.gguf": ["mradermacher/kunoichi-lemon-royale-v2experiment2-32K-7B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 434 |
+
"BlueLight-12B.Q4_K_M.gguf": ["mradermacher/BlueLight-12B-GGUF", MessagesFormatterType.CHATML],
|
| 435 |
+
"kunoichi-lemon-royale-hamansu-v1-32k-7B.Q5_K_M.gguf": ["mradermacher/kunoichi-lemon-royale-hamansu-v1-32k-7B-GGUF", MessagesFormatterType.MISTRAL],
|
| 436 |
+
"Llama-3.1-128k-Dark-Planet-Uncensored-8B.Q5_K_M.gguf": ["mradermacher/Llama-3.1-128k-Dark-Planet-Uncensored-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 437 |
+
"Llama-3.1-1-million-ctx-DeepHermes-Deep-Reasoning-8B.Q5_K_M.gguf": ["mradermacher/Llama-3.1-1-million-ctx-DeepHermes-Deep-Reasoning-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 438 |
+
"Llama-3.1-1million-ctx-Dark-Planet-8B.Q5_K_M.gguf": ["mradermacher/Llama-3.1-1million-ctx-Dark-Planet-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 439 |
+
"kunoichi-lemon-royale-v2experiment1-32K-7B.i1-Q5_K_M.gguf": ["mradermacher/kunoichi-lemon-royale-v2experiment1-32K-7B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 440 |
+
"ABEJA-Qwen2.5-7b-Japanese-v0.1-Q5_K_M.gguf": ["mmnga/ABEJA-Qwen2.5-7b-Japanese-v0.1-gguf", MessagesFormatterType.OPEN_CHAT],
|
| 441 |
+
"Txt-Polisher-Douyin-Style.Q4_K_M.gguf": ["mradermacher/Txt-Polisher-Douyin-Style-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 442 |
+
"Llama-3.1-Nemotron-8B-UltraLong-1M-Instruct.Q4_K_S.gguf": ["mradermacher/Llama-3.1-Nemotron-8B-UltraLong-1M-Instruct-GGUF", MessagesFormatterType.LLAMA_3],
|
| 443 |
+
"dreamgen_lucid-v1-nemo-Q4_K_M.gguf": ["bartowski/dreamgen_lucid-v1-nemo-GGUF", MessagesFormatterType.MISTRAL],
|
| 444 |
+
"TheDrummer_Rivermind-12B-v1-Q4_K_M.gguf": ["bartowski/TheDrummer_Rivermind-12B-v1-GGUF", MessagesFormatterType.MISTRAL],
|
| 445 |
+
"Josiefied-Qwen2.5-Coder-14B-Instruct-abliterated-v1.Q4_K_M.gguf": ["mradermacher/Josiefied-Qwen2.5-Coder-14B-Instruct-abliterated-v1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 446 |
+
"KunoRZN-Llama-3-3B.Q5_K_M.gguf": ["mradermacher/KunoRZN-Llama-3-3B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 447 |
+
"L3-8B-Niitamamo-v0.1-della.Q5_K_M.gguf": ["mradermacher/L3-8B-Niitamamo-v0.1-della-GGUF", MessagesFormatterType.LLAMA_3],
|
| 448 |
+
"Qwen2.5-14B-Imstinct.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Imstinct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 449 |
+
"Pearl-14B.Q4_K_M.gguf": ["mradermacher/Pearl-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 450 |
+
"Qwen2.5-14B-Instinct-RP.i1-Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Instinct-RP-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 451 |
+
"Qwen2.5-Recursive-Coder-14B-Instruct.i1-Q4_K_M.gguf": ["mradermacher/Qwen2.5-Recursive-Coder-14B-Instruct-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 452 |
+
"nemo-12b-instruct-creative.Q4_K_M.gguf": ["mradermacher/nemo-12b-instruct-creative-GGUF", MessagesFormatterType.MISTRAL],
|
| 453 |
+
"Qwen2.5-Slick-Coder-14B-Instruct.Q4_K_S.gguf": ["mradermacher/Qwen2.5-Slick-Coder-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 454 |
+
"Gemma2-Evelyn-9B.Q4_K_M.gguf": ["mradermacher/Gemma2-Evelyn-9B-GGUF", MessagesFormatterType.ALPACA],
|
| 455 |
+
"The-Omega-Abomination-M-12B-v1.0.Q4_K_M.gguf": ["mradermacher/The-Omega-Abomination-M-12B-v1.0-GGUF", MessagesFormatterType.MISTRAL],
|
| 456 |
+
"Qwen2.5-Wicked-Coder-14B-Instruct.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Wicked-Coder-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 457 |
+
"ElvenMaid-12B-Stock.Q4_K_M.gguf": ["mradermacher/ElvenMaid-12B-Stock-GGUF", MessagesFormatterType.CHATML],
|
| 458 |
+
"shisa-v2-unphi4-14b.Q4_K_M.gguf": ["mradermacher/shisa-v2-unphi4-14b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 459 |
+
"L3-Urvashi-v3.5.Q5_K_M.gguf": ["mradermacher/L3-Urvashi-v3.5-GGUF", MessagesFormatterType.LLAMA_3],
|
| 460 |
+
"shisa-v2-qwen2.5-7b.Q5_K_M.gguf": ["mradermacher/shisa-v2-qwen2.5-7b-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 461 |
+
"Qwen2.5-14B-YOYO-V1.5.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-YOYO-V1.5-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 462 |
+
"ElvenMaid-12B-v3.Q4_K_M.gguf": ["mradermacher/ElvenMaid-12B-v3-GGUF", MessagesFormatterType.CHATML],
|
| 463 |
+
"Qwen2.5-14B-Instruct-abliterated-SFT.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Instruct-abliterated-SFT-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 464 |
+
"Kaolinite-Kitara-12B.Q4_K_M.gguf": ["mradermacher/Kaolinite-Kitara-12B-GGUF", MessagesFormatterType.CHATML],
|
| 465 |
+
"deneb-v1-7b.Q5_K_M.gguf": ["mradermacher/deneb-v1-7b-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 466 |
+
"Ice0.105-13.04-RP.Q4_K_S.gguf": ["mradermacher/Ice0.105-13.04-RP-GGUF", MessagesFormatterType.MISTRAL],
|
| 467 |
+
"Gemma2-BlueMoon-9B.i1-Q4_K_M.gguf": ["mradermacher/Gemma2-BlueMoon-9B-i1-GGUF", MessagesFormatterType.ALPACA],
|
| 468 |
+
"Himeyuri-Magnum-12B-v2.i1-Q4_K_M.gguf": ["mradermacher/Himeyuri-Magnum-12B-v2-i1-GGUF", MessagesFormatterType.CHATML],
|
| 469 |
+
"ElvenMaid-12B-v2.Q4_K_M.gguf": ["mradermacher/ElvenMaid-12B-v2-GGUF", MessagesFormatterType.CHATML],
|
| 470 |
+
"ToolMaestro-Distil-T1-7B.Q5_K_M.gguf": ["mradermacher/ToolMaestro-Distil-T1-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 471 |
+
"OpenThoughts-114k-Rewrite-Thoughts-Llama-3.2-3B-Instruct.Q5_K_M.gguf": ["mradermacher/OpenThoughts-114k-Rewrite-Thoughts-Llama-3.2-3B-Instruct-GGUF", MessagesFormatterType.LLAMA_3],
|
| 472 |
+
"La-Superba-14B-Y.2.Q4_K_M.gguf": ["mradermacher/La-Superba-14B-Y.2-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 473 |
+
"Qwen2.5-3B-Instruct-Gate.i1-Q5_K_M.gguf": ["mradermacher/Qwen2.5-3B-Instruct-Gate-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 474 |
+
"cogito-v1-preview-qwen-14B.Q4_K_M.gguf": ["mradermacher/cogito-v1-preview-qwen-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 475 |
+
"Magnolia-v3a-12B.Q4_K_M.gguf": ["mradermacher/Magnolia-v3a-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 476 |
+
"Qwen2.5-14B-YOYO-stock.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-YOYO-stock-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 477 |
+
"ablation-164-shisav2.if100-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-164-shisav2.if100-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 478 |
+
"ablation-190-a175.finaldpo2.7.2e7-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-190-a175.finaldpo2.7.2e7-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 479 |
+
"ablation-183-a177.finaldpo2.4.5e7-shisa-v2-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/ablation-183-a177.finaldpo2.4.5e7-shisa-v2-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
|
| 480 |
+
"Qwen2.5-Casual-Coder-14B-Instruct.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Casual-Coder-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 481 |
+
"Ice0.102-10.04-RP.Q4_K_S.gguf": ["mradermacher/Ice0.102-10.04-RP-GGUF", MessagesFormatterType.MISTRAL],
|
| 482 |
+
"Qwen2.5-Flow-Coder-14B-Instruct.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Flow-Coder-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 483 |
+
"Qwen2.5-CompositeFlow-Coder-14B-Instruct.Q4_K_M.gguf": ["mradermacher/Qwen2.5-CompositeFlow-Coder-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 484 |
+
"Gemma-Muesli-2-9B-it.Q4_K_M.gguf": ["mradermacher/Gemma-Muesli-2-9B-it-GGUF", MessagesFormatterType.ALPACA],
|
| 485 |
+
"Chat-KTO-4B.Q5_K_M.gguf": ["mradermacher/Chat-KTO-4B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 486 |
+
"Gliese-B1257.Q4_K_M.gguf": ["mradermacher/Gliese-B1257-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 487 |
+
"Qwen2.5-Sharp-Coder-14B-Instruct.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Sharp-Coder-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 488 |
+
"Llama-3.1-8B-UltraLong-1M-Instruct.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-UltraLong-1M-Instruct-GGUF", MessagesFormatterType.LLAMA_3],
|
| 489 |
+
"Final-Llama-CMD-1L.Q5_K_M.gguf": ["mradermacher/Final-Llama-CMD-1L-GGUF", MessagesFormatterType.LLAMA_3],
|
| 490 |
+
"Final-Llama-Cyber-1L.Q5_K_M.gguf": ["mradermacher/Final-Llama-Cyber-1L-GGUF", MessagesFormatterType.LLAMA_3],
|
| 491 |
+
"MagnaRei-v2-12B.Q4_K_M.gguf": ["mradermacher/MagnaRei-v2-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 492 |
+
"ThinkEdit-deepseek-llama3-8b.i1-Q5_K_M.gguf": ["mradermacher/ThinkEdit-deepseek-llama3-8b-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 493 |
+
"M-Prometheus-14B.Q4_K_M.gguf": ["mradermacher/M-Prometheus-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 494 |
+
"Llama-Breeze2-8B-Instruct-text-only.Q5_K_M.gguf": ["mradermacher/Llama-Breeze2-8B-Instruct-text-only-GGUF", MessagesFormatterType.LLAMA_3],
|
| 495 |
+
"Magnolia-v7-12B.Q4_K_M.gguf": ["mradermacher/Magnolia-v7-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 496 |
+
"Magnolia-v6-12B.Q4_K_M.gguf": ["mradermacher/Magnolia-v6-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 497 |
+
"L3.1-Smthv1-12B.Q4_K_M.gguf": ["mradermacher/L3.1-Smthv1-12B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 498 |
+
"The-Omega-Directive-M-12B-v1.0.Q4_K_M.gguf": ["mradermacher/The-Omega-Directive-M-12B-v1.0-GGUF", MessagesFormatterType.MISTRAL],
|
| 499 |
+
"The-Omega-Directive-M-8B-v1.0.Q5_K_M.gguf": ["mradermacher/The-Omega-Directive-M-8B-v1.0-GGUF", MessagesFormatterType.MISTRAL],
|
| 500 |
+
"MagnaRei-v1-12B.Q4_K_M.gguf": ["mradermacher/MagnaRei-v1-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 501 |
+
"Hamanasu-Magnum-4B.Q5_K_M.gguf": ["mradermacher/Hamanasu-Magnum-4B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 502 |
+
"kd1-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/kd1-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
|
| 503 |
+
"kd2-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/kd2-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
|
| 504 |
+
"ablation-146-a128.dpo.armorm.rp.tl.1e6-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-146-a128.dpo.armorm.rp.tl.1e6-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 505 |
+
"ablation-151-finalsft-shisa-v2-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/ablation-151-finalsft-shisa-v2-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
|
| 506 |
+
"ablation-147-a128.dpo.armorm.rp.tl.1.5e6-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-147-a128.dpo.armorm.rp.tl.1.5e6-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 507 |
+
"ablation-158-a128.dpo.finaldpo.if50-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-158-a128.dpo.finaldpo.if50-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 508 |
+
"ablation-167-finaldpo.3e7-shisa-v2-unphi-4-14b.Q4_K_M.gguf": ["mradermacher/ablation-167-finaldpo.3e7-shisa-v2-unphi-4-14b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 509 |
+
"ablation-155-a151.dpofinal-shisa-v2-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/ablation-155-a151.dpofinal-shisa-v2-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
|
| 510 |
+
"ablation-156-a152.dpofinal-shisa-v2-mistral-nemo-japanese-12b.Q4_K_M.gguf": ["mradermacher/ablation-156-a152.dpofinal-shisa-v2-mistral-nemo-japanese-12b-GGUF", MessagesFormatterType.MISTRAL],
|
| 511 |
+
"Qwen2.5-7B-YOYO-super.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-YOYO-super-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 512 |
+
"ChaiML-Nemo-DPO-V17.Q4_K_M.gguf": ["mradermacher/ChaiML-Nemo-DPO-V17-GGUF", MessagesFormatterType.MISTRAL],
|
| 513 |
+
"dippy-roleplay-llama-v2.8.Q5_K_M.gguf": ["mradermacher/dippy-roleplay-llama-v2.8-GGUF", MessagesFormatterType.LLAMA_3],
|
| 514 |
+
"Proto-Harpy-Burst-v0.1-7B.Q5_K_M.gguf": ["mradermacher/Proto-Harpy-Burst-v0.1-7B-GGUF", MessagesFormatterType.MISTRAL],
|
| 515 |
+
"OncoLlama-3.1-8B-Instruct-0.6.Q5_K_M.gguf": ["mradermacher/OncoLlama-3.1-8B-Instruct-0.6-GGUF", MessagesFormatterType.LLAMA_3],
|
| 516 |
+
"Mistral-nemo-ja-rp-v0.2.Q4_K_M.gguf": ["mradermacher/Mistral-nemo-ja-rp-v0.2-GGUF", MessagesFormatterType.CHATML],
|
| 517 |
+
"L3.1-Dark-Reasoning-LewdPlay-evo-Hermes-R1-Uncensored-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Dark-Reasoning-LewdPlay-evo-Hermes-R1-Uncensored-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 518 |
+
"L3.1-Dark-Reasoning-Unholy-Hermes-R1-Uncensored-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Dark-Reasoning-Unholy-Hermes-R1-Uncensored-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 519 |
+
"Llama-3.1-8B-Fusion-8020.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-Fusion-8020-GGUF", MessagesFormatterType.LLAMA_3],
|
| 520 |
+
"Llama-3.1-8B-Fusion-9010.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-Fusion-9010-GGUF", MessagesFormatterType.LLAMA_3],
|
| 521 |
+
"Proto-Harpy-Avian-v0.1-7B.Q5_K_M.gguf": ["mradermacher/Proto-Harpy-Avian-v0.1-7B-GGUF", MessagesFormatterType.MISTRAL],
|
| 522 |
+
"Qwen2.5-14B-ties.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-ties-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 523 |
+
"Qwen2.5-14B-ties-1M.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-ties-1M-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 524 |
+
"Qwen2.5-14B-della.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-della-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 525 |
+
"RoLlama3-8b-Instruct-2024-10-09.Q5_K_M.gguf": ["mradermacher/RoLlama3-8b-Instruct-2024-10-09-GGUF", MessagesFormatterType.LLAMA_3],
|
| 526 |
+
"VietCoMath-01-14B.Q4_K_M.gguf": ["mradermacher/VietCoMath-01-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 527 |
+
"t0-s1-14B.Q4_K_M.gguf": ["mradermacher/t0-s1-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 528 |
+
"Gemma2-9b-it-Girl-v1.Q4_K_M.gguf": ["mradermacher/Gemma2-9b-it-Girl-v1-GGUF", MessagesFormatterType.ALPACA],
|
| 529 |
+
"Nemo-DPO-V22.Q4_K_S.gguf": ["mradermacher/Nemo-DPO-V22-GGUF", MessagesFormatterType.MISTRAL],
|
| 530 |
+
"LoyalMaid-12B.Q4_K_S.gguf": ["mradermacher/LoyalMaid-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 531 |
+
"t0-s1.1-3B.Q5_K_M.gguf": ["mradermacher/t0-s1.1-3B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 532 |
+
"NQLSG-Qwen2.5-14B-MegaFusion-v9.4.Q4_K_M.gguf": ["mradermacher/NQLSG-Qwen2.5-14B-MegaFusion-v9.4-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 533 |
+
"Ultron-Summarizer-8B.Q5_K_M.gguf": ["mradermacher/Ultron-Summarizer-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 534 |
+
"NQLSG-Qwen2.5-14B-MegaFusion-v9.4-Coder.Q4_K_M.gguf": ["mradermacher/NQLSG-Qwen2.5-14B-MegaFusion-v9.4-Coder-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 535 |
+
"ElvenMaid-12B.Q4_K_M.gguf": ["mradermacher/ElvenMaid-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 536 |
+
"Llama-3.1-8B-Fusion-7030.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-Fusion-7030-GGUF", MessagesFormatterType.LLAMA_3],
|
| 537 |
+
"MedicalEDI-14b-EDI-Reasoning-5-bf16.Q4_K_M.gguf": ["mradermacher/MedicalEDI-14b-EDI-Reasoning-5-bf16-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 538 |
+
"Llama-3.1-8B-Athena-Apollo-exp.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-Athena-Apollo-exp-GGUF", MessagesFormatterType.LLAMA_3],
|
| 539 |
+
"Llama-3.1-8B-Fusion-5050.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-Fusion-5050-GGUF", MessagesFormatterType.LLAMA_3],
|
| 540 |
+
"Nemo-DPO-V21.Q4_K_M.gguf": ["mradermacher/Nemo-DPO-V21-GGUF", MessagesFormatterType.MISTRAL],
|
| 541 |
+
"ChatWaifu_32B_reasoning.Q4_K_M.gguf": ["mradermacher/ChatWaifu_32B_reasoning-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 542 |
+
"Llama3_Storyteller.Q5_K_M.gguf": ["mradermacher/Llama3_Storyteller-GGUF", MessagesFormatterType.LLAMA_3],
|
| 543 |
+
"ablation-110-openrlhf.dpo.a106.5e7ax-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-110-openrlhf.dpo.a106.5e7ax-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 544 |
+
"DeepResearcher-7b.Q5_K_M.gguf": ["mradermacher/DeepResearcher-7b-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 545 |
+
"ablation-141-a128.dpo.armorm.rp-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-141-a128.dpo.armorm.rp-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 546 |
+
"Bootes-Xuange-Qwen-14B.Q4_K_M.gguf": ["mradermacher/Bootes-Xuange-Qwen-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 547 |
+
"ablation-123-a114.dpo.armorm.rp-shisa-v2-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/ablation-123-a114.dpo.armorm.rp-shisa-v2-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
|
| 548 |
+
"ablation-138-shisav2.gbs128.1.6e5-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-138-shisav2.gbs128.1.6e5-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 549 |
+
"ablation-122-a114.dpo.armorm.rp-shisa-v2-unphi-4-14b.Q4_K_M.gguf": ["mradermacher/ablation-122-a114.dpo.armorm.rp-shisa-v2-unphi-4-14b-GGUF", MessagesFormatterType.MISTRAL],
|
| 550 |
+
"Qwen2.5-14B-stock-v3.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-stock-v3-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 551 |
+
"Humanize-Rei-Slerp.Q4_K_M.gguf": ["mradermacher/Humanize-Rei-Slerp-GGUF", MessagesFormatterType.MISTRAL],
|
| 552 |
+
"ablation-142-a128.dpo.armorm.rp.tl-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-142-a128.dpo.armorm.rp.tl-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 553 |
+
"ablation-139-shisav2.ds.gbs128.1e5-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-139-shisav2.ds.gbs128.1e5-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 554 |
+
"L3.1-Dark-Reasoning-Halu-Blackroot-Hermes-R1-Uncensored-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Dark-Reasoning-Halu-Blackroot-Hermes-R1-Uncensored-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 555 |
+
"L3.1-Dark-Reasoning-Anjir-Hermes-R1-Uncensored-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Dark-Reasoning-Anjir-Hermes-R1-Uncensored-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 556 |
+
"ablation-115-bestofn.atlo.rp.tlx.newmix2-shisa-v2-mistral-nemo-12b.Q4_K_S.gguf": ["mradermacher/ablation-115-bestofn.atlo.rp.tlx.newmix2-shisa-v2-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
|
| 557 |
+
"L3.1-Dark-Reasoning-Super-Nova-RP-Hermes-R1-Uncensored-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Dark-Reasoning-Super-Nova-RP-Hermes-R1-Uncensored-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 558 |
+
"L3.1-Dark-Reasoning-Celeste-V1.2-Hermes-R1-Uncensored-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Dark-Reasoning-Celeste-V1.2-Hermes-R1-Uncensored-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 559 |
+
"DeeperHermes3_R1_D_L3_8b.Q5_K_M.gguf": ["mradermacher/DeeperHermes3_R1_D_L3_8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 560 |
+
"Nemo-DPO-V20.Q4_K_S.gguf": ["mradermacher/Nemo-DPO-V20-GGUF", MessagesFormatterType.MISTRAL],
|
| 561 |
+
"Badudu-8B-instruct-v1.Q4_K_M.gguf": ["mradermacher/Badudu-8B-instruct-v1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 562 |
+
"Himeyuri-Magnum-12B.Q4_K_M.gguf": ["mradermacher/Himeyuri-Magnum-12B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 563 |
+
"Aura-12B-FT.i1-Q4_K_M.gguf": ["mradermacher/Aura-12B-FT-i1-GGUF", MessagesFormatterType.CHATML],
|
| 564 |
+
"NarrowMaid-8B.Q5_K_M.gguf": ["mradermacher/NarrowMaid-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 565 |
+
"L3.1-Dark-Reasoning-Dark-Planet-Hermes-R1-Uncensored-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Dark-Reasoning-Dark-Planet-Hermes-R1-Uncensored-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 566 |
+
"Qwen2.5-Vibe-Coder-14B-Instruct.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Vibe-Coder-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 567 |
+
"josie-7b-v6.0.Q5_K_M.gguf": ["mradermacher/josie-7b-v6.0-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 568 |
+
"NeuDistRo-a1-laser.Q5_K_M.gguf": ["mradermacher/NeuDistRo-a1-laser-GGUF", MessagesFormatterType.MISTRAL],
|
| 569 |
+
"Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v1.Q5_K_M.gguf": ["mradermacher/Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 570 |
+
"DreadMix.Q5_K_M.gguf": ["mradermacher/DreadMix-GGUF", MessagesFormatterType.LLAMA_3],
|
| 571 |
+
"EvolCodeLlama-3.1-8B-Instruct.Q5_K_M.gguf": ["mradermacher/EvolCodeLlama-3.1-8B-Instruct-GGUF", MessagesFormatterType.LLAMA_3],
|
| 572 |
+
"Soaring-8b.Q5_K_M.gguf": ["mradermacher/Soaring-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 573 |
+
"RoLlama3.1-8b-Instruct-DPO.Q5_K_M.gguf": ["mradermacher/RoLlama3.1-8b-Instruct-DPO-GGUF", MessagesFormatterType.LLAMA_3],
|
| 574 |
+
"qwen-14b-chisel.Q4_K_M.gguf": ["mradermacher/qwen-14b-chisel-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 575 |
+
"HarmReporter.Q5_K_M.gguf": ["mradermacher/HarmReporter-GGUF", MessagesFormatterType.LLAMA_3],
|
| 576 |
+
"Gemma2-Alicia-9B.Q4_K_M.gguf": ["mradermacher/Gemma2-Alicia-9B-GGUF", MessagesFormatterType.ALPACA],
|
| 577 |
+
"bad-apple-8b-v1.Q5_K_M.gguf": ["mradermacher/bad-apple-8b-v1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 578 |
+
"Eris-Daturamix-Neural-7b-slerp.Q5_K_M.gguf": ["mradermacher/Eris-Daturamix-Neural-7b-slerp-GGUF", MessagesFormatterType.MISTRAL],
|
| 579 |
+
"miscii-1023-001-200.Q4_K_M.gguf": ["mradermacher/miscii-1023-001-200-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 580 |
+
"Faheng-14B.Q4_K_M.gguf": ["mradermacher/Faheng-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 581 |
+
"sft-math-llama3-8b-inst.Q5_K_M.gguf": ["mradermacher/sft-math-llama3-8b-inst-GGUF", MessagesFormatterType.LLAMA_3],
|
| 582 |
+
"ablation-104-openrlhf.dpo.a84.5e7ring-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-104-openrlhf.dpo.a84.5e7ring-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 583 |
+
"ablation-102-openrlhf.dpo.a84.5e7-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-102-openrlhf.dpo.a84.5e7-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 584 |
+
"ablation-105-bestofn.atlo.rp.tlx25.newmixall-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-105-bestofn.atlo.rp.tlx25.newmixall-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 585 |
+
"ablation-103-openrlhf.dpo.a84.5e7ax-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-103-openrlhf.dpo.a84.5e7ax-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 586 |
+
"Qwen2.5-1.5B-Instruct-Multiple-Choice-Maker.Q5_K_M.gguf": ["mradermacher/Qwen2.5-1.5B-Instruct-Multiple-Choice-Maker-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 587 |
+
"Nemo-DPO-V18.Q4_K_M.gguf": ["mradermacher/Nemo-DPO-V18-GGUF", MessagesFormatterType.MISTRAL],
|
| 588 |
+
"L3-8B-sunfall-v0.3.i1-Q5_K_M.gguf": ["mradermacher/L3-8B-sunfall-v0.3-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 589 |
+
"bad-apple-8b-v2.Q5_K_M.gguf": ["mradermacher/bad-apple-8b-v2-GGUF", MessagesFormatterType.LLAMA_3],
|
| 590 |
+
"dippy-roleplay-llama-v3.1-0-10.Q5_K_M.gguf": ["mradermacher/dippy-roleplay-llama-v3.1-0-10-GGUF", MessagesFormatterType.LLAMA_3],
|
| 591 |
+
"CatunaMayo.Q5_K_M.gguf": ["mradermacher/CatunaMayo-GGUF", MessagesFormatterType.MISTRAL],
|
| 592 |
+
"DeepRetrieval-PubMed-3B.Q5_K_M.gguf": ["mradermacher/DeepRetrieval-PubMed-3B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 593 |
+
"Gemma_Writer-9b.Q4_K_M.gguf": ["mradermacher/Gemma_Writer-9b-GGUF", MessagesFormatterType.ALPACA],
|
| 594 |
+
"Nemo-DPO-V19.Q4_K_M.gguf": ["mradermacher/Nemo-DPO-V19-GGUF", MessagesFormatterType.MISTRAL],
|
| 595 |
+
"supermario-slerp-v3.Q5_K_M.gguf": ["mradermacher/supermario-slerp-v3-GGUF", MessagesFormatterType.MISTRAL],
|
| 596 |
+
"RoLlama3-8b-Instruct-DPO-2024-10-09.Q5_K_M.gguf": ["mradermacher/RoLlama3-8b-Instruct-DPO-2024-10-09-GGUF", MessagesFormatterType.LLAMA_3],
|
| 597 |
+
"Atlantum-12B-v1.Q4_K_M.gguf": ["mradermacher/Atlantum-12B-v1-GGUF", MessagesFormatterType.MISTRAL],
|
| 598 |
+
"Neural-Eris-Daturamix-Chat-7b-slerp.Q5_K_M.gguf": ["mradermacher/Neural-Eris-Daturamix-Chat-7b-slerp-GGUF", MessagesFormatterType.MISTRAL],
|
| 599 |
+
"supermario-v2.Q5_K_M.gguf": ["mradermacher/supermario-v2-GGUF", MessagesFormatterType.MISTRAL],
|
| 600 |
+
"rainbowfish-7B-v9.Q5_K_M.gguf": ["mradermacher/rainbowfish-7B-v9-GGUF", MessagesFormatterType.MISTRAL],
|
| 601 |
+
"Con-J-Qwen2-7B.Q5_K_M.gguf": ["mradermacher/Con-J-Qwen2-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 602 |
+
"Gemma2-9b-it-Youki-v1.Q4_K_M.gguf": ["mradermacher/Gemma2-9b-it-Youki-v1-GGUF", MessagesFormatterType.ALPACA],
|
| 603 |
+
"Reasoning-Mistral-Nemo.Q4_K_M.gguf": ["mradermacher/Reasoning-Mistral-Nemo-GGUF", MessagesFormatterType.MISTRAL],
|
| 604 |
+
"DeepRetrieval-SQL-7B.Q5_K_M.gguf": ["mradermacher/DeepRetrieval-SQL-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 605 |
+
"Qwen2.5-14B-dpo-it-della.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-dpo-it-della-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 606 |
+
"Aura-12B.Q4_K_M.gguf": ["mradermacher/Aura-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 607 |
+
"JudgeLRM-7B.Q5_K_M.gguf": ["mradermacher/JudgeLRM-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 608 |
+
"LLaMAntino-3-SLIMER-IT-PARALLEL.Q5_K_M.gguf": ["mradermacher/LLaMAntino-3-SLIMER-IT-PARALLEL-GGUF", MessagesFormatterType.LLAMA_3],
|
| 609 |
+
"Gemma-Spoon-2-9B-it.i1-Q4_K_M.gguf": ["mradermacher/Gemma-Spoon-2-9B-it-i1-GGUF", MessagesFormatterType.ALPACA],
|
| 610 |
+
"L3-GothicMaid-8B.Q5_K_M.gguf": ["mradermacher/L3-GothicMaid-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 611 |
+
"Neuronal-7b-FT.Q5_K_M.gguf": ["mradermacher/Neuronal-7b-FT-GGUF", MessagesFormatterType.MISTRAL],
|
| 612 |
+
"L3.1-SemiEvol-MMLUPro-8B.Q5_K_M.gguf": ["mradermacher/L3.1-SemiEvol-MMLUPro-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 613 |
+
"meme-12b-e2.Q4_K_M.gguf": ["mradermacher/meme-12b-e2-GGUF", MessagesFormatterType.MISTRAL],
|
| 614 |
+
"ablation-100-bestofn.atlo.rp.tlx50-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-100-bestofn.atlo.rp.tlx50-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 615 |
+
"ablation-101-bestofn.atlo.rp.tlx25-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-101-bestofn.atlo.rp.tlx25-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 616 |
+
"TaCIE-Qwen2-7B.Q5_K_M.gguf": ["mradermacher/TaCIE-Qwen2-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 617 |
+
"Qwen2.5-Coder-7B-Instruct-Uncensored.i1-Q5_K_M.gguf": ["mradermacher/Qwen2.5-Coder-7B-Instruct-Uncensored-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 618 |
+
"Korean-llama-3.1-8B.Q5_K_M.gguf": ["mradermacher/Korean-llama-3.1-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 619 |
+
"Gr3.1-8B-Desert-Rose.Q5_K_M.gguf": ["mradermacher/Gr3.1-8B-Desert-Rose-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 620 |
+
"L3-Tyche-8B-v1.0.Q5_K_M.gguf": ["mradermacher/L3-Tyche-8B-v1.0-GGUF", MessagesFormatterType.LLAMA_3],
|
| 621 |
+
"Eris-LelantaclesV2-7b.Q5_K_M.gguf": ["mradermacher/Eris-LelantaclesV2-7b-GGUF", MessagesFormatterType.MISTRAL],
|
| 622 |
+
"Eris-Lelantacles-7b.Q5_K_M.gguf": ["mradermacher/Eris-Lelantacles-7b-GGUF", MessagesFormatterType.MISTRAL],
|
| 623 |
+
"jessi-v0.1-qwen2.5-7b-instruct.Q5_K_M.gguf": ["mradermacher/jessi-v0.1-qwen2.5-7b-instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 624 |
+
"Qwen2.5-7B-Latent_Verification.i1-Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Latent_Verification-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 625 |
+
"Agentic-Qwen7b-GRPO.Q5_K_M.gguf": ["mradermacher/Agentic-Qwen7b-GRPO-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 626 |
+
"ablation-95-a85.dpo.armorm.rp-shisa-v2-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/ablation-95-a85.dpo.armorm.rp-shisa-v2-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
|
| 627 |
+
"ablation-89-rp.dpo.test-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-89-rp.dpo.test-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 628 |
+
"Gemma-Noon-2-9B-it.Q4_K_M.gguf": ["mradermacher/Gemma-Noon-2-9B-it-GGUF", MessagesFormatterType.ALPACA],
|
| 629 |
+
"ablation-99-bestofn.atlo.rp.tlx-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-99-bestofn.atlo.rp.tlx-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 630 |
+
"ablation-86-bestofn.atlo.rp.tl-shisa-v2-mistral-nemo-12b-5e6.Q4_K_M.gguf": ["mradermacher/ablation-86-bestofn.atlo.rp.tl-shisa-v2-mistral-nemo-12b-5e6-GGUF", MessagesFormatterType.MISTRAL],
|
| 631 |
+
"ablation-85-bestofn.atlo.rp.tl-shisa-v2-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/ablation-85-bestofn.atlo.rp.tl-shisa-v2-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
|
| 632 |
+
"MistThena7B.Q5_K_M.gguf": ["mradermacher/MistThena7B-GGUF", MessagesFormatterType.MISTRAL],
|
| 633 |
+
"Pervepsion-12B-x1.Q4_K_M.gguf": ["mradermacher/Pervepsion-12B-x1-GGUF", MessagesFormatterType.MISTRAL],
|
| 634 |
+
"ablation-98-bestofn.atlo.rp.tl.geniac-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-98-bestofn.atlo.rp.tl.geniac-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 635 |
+
"Dark-Triad-Adam-Llama3.2-3B-v2.Q5_K_M.gguf": ["mradermacher/Dark-Triad-Adam-Llama3.2-3B-v2-GGUF", MessagesFormatterType.LLAMA_3],
|
| 636 |
+
"Qandora-2.5-7B.Q5_K_M.gguf": ["mradermacher/Qandora-2.5-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 637 |
+
"Tesslate_Tessa-T1-14B-Q4_K_M.gguf": ["bartowski/Tesslate_Tessa-T1-14B-GGUF", MessagesFormatterType.CHATML],
|
| 638 |
+
"Tesslate_Tessa-T1-7B-Q5_K_M.gguf": ["bartowski/Tesslate_Tessa-T1-7B-GGUF", MessagesFormatterType.CHATML],
|
| 639 |
+
"tabletop_game_design.Q5_K_M.gguf": ["mradermacher/tabletop_game_design-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 640 |
+
"mergekit-model_stock-tiwlqms.Q4_K_M.gguf": ["mradermacher/mergekit-model_stock-tiwlqms-GGUF", MessagesFormatterType.MISTRAL],
|
| 641 |
+
"Buttocks-7B-v2.1.Q5_K_M.gguf": ["mradermacher/Buttocks-7B-v2.1-GGUF", MessagesFormatterType.MISTRAL],
|
| 642 |
+
"MN-Hekate-Geneteira-12B.Q4_K_M.gguf": ["mradermacher/MN-Hekate-Geneteira-12B-GGUF", MessagesFormatterType.CHATML],
|
| 643 |
+
"Agentic-Qwen7b.Q5_K_M.gguf": ["mradermacher/Agentic-Qwen7b-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 644 |
+
"Mistral-Nemo-12B-Instruct.Q4_K_M.gguf": ["mradermacher/Mistral-Nemo-12B-Instruct-GGUF", MessagesFormatterType.MISTRAL],
|
| 645 |
+
"EvaMedius-14B-Instruct.Q4_K_M.gguf": ["mradermacher/EvaMedius-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 646 |
+
"Captain_Eris_Twighlight-12B-v2.Q5_K_M.gguf": ["mradermacher/Captain_Eris_Twighlight-12B-v2-GGUF", MessagesFormatterType.MISTRAL],
|
| 647 |
+
"Llama-3-8B-WizardLM-196K.Q5_K_M.gguf": ["mradermacher/Llama-3-8B-WizardLM-196K-GGUF", MessagesFormatterType.LLAMA_3],
|
| 648 |
+
"zephyr-qwen2-7b-dpo.Q5_K_M.gguf": ["mradermacher/zephyr-qwen2-7b-dpo-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 649 |
+
"L3-Umbral-Mind-RP-v0.6.2-8B.i1-Q5_K_M.gguf": ["mradermacher/L3-Umbral-Mind-RP-v0.6.2-8B-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 650 |
+
"Final-Llama-IoI-1E.Q5_K_M.gguf": ["mradermacher/Final-Llama-IoI-1E-GGUF", MessagesFormatterType.LLAMA_3],
|
| 651 |
+
"NQLSG-Qwen2.5-14B-MegaFusion-v9.3-Coder.Q4_K_M.gguf": ["mradermacher/NQLSG-Qwen2.5-14B-MegaFusion-v9.3-Coder-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 652 |
+
"ReSearch-Qwen-7B-Instruct.Q5_K_M.gguf": ["mradermacher/ReSearch-Qwen-7B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 653 |
+
"MN-Hekate-Panopaia-12B.Q4_K_S.gguf": ["mradermacher/MN-Hekate-Panopaia-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 654 |
+
"Hathor_Ultra-Mix-Final-0.420-L3-8B.Q5_K_M.gguf": ["mradermacher/Hathor_Ultra-Mix-Final-0.420-L3-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 655 |
+
"Nova-1.5-Conversational-7B.Q5_K_M.gguf": ["mradermacher/Nova-1.5-Conversational-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 656 |
+
"Raptor-X1.Q4_K_M.gguf": ["mradermacher/Raptor-X1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 657 |
+
"Noxis-8B-LINEAR.Q5_K_M.gguf": ["mradermacher/Noxis-8B-LINEAR-GGUF", MessagesFormatterType.LLAMA_3],
|
| 658 |
+
"Q2.5-Instruct-1M_Harmony.Q5_K_M.gguf": ["mradermacher/Q2.5-Instruct-1M_Harmony-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 659 |
+
"Qwen2.5-14B-Instruct-1M-Unalign.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Instruct-1M-Unalign-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 660 |
+
"Llama-3.2-3B-Instruct-Explainable-Propaganda-Detection.Q5_K_M.gguf": ["mradermacher/Llama-3.2-3B-Instruct-Explainable-Propaganda-Detection-GGUF", MessagesFormatterType.LLAMA_3],
|
| 661 |
+
"Qwen-2.5-7B-GRA-WizardLM.Q5_K_M.gguf": ["mradermacher/Qwen-2.5-7B-GRA-WizardLM-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 662 |
+
"dpo_plroleplay.Q4_K_M.gguf": ["mradermacher/dpo_plroleplay-GGUF", MessagesFormatterType.MISTRAL],
|
| 663 |
+
"YinkaMath-14B.Q4_K_M.gguf": ["mradermacher/YinkaMath-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 664 |
+
"CosmoCoder.Q4_K_M.gguf": ["mradermacher/CosmoCoder-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 665 |
+
"Sinner_12B_V.1.Q4_K_M.gguf": ["mradermacher/Sinner_12B_V.1-GGUF", MessagesFormatterType.MISTRAL],
|
| 666 |
+
"Nova-0.5-e3-7B.i1-Q5_K_M.gguf": ["mradermacher/Nova-0.5-e3-7B-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 667 |
+
"ChatingDeveloper-7B-slerp.Q5_K_M.gguf": ["mradermacher/ChatingDeveloper-7B-slerp-GGUF", MessagesFormatterType.MISTRAL],
|
| 668 |
+
"Neo-1-16B.Q5_K_M.gguf": ["mradermacher/Neo-1-16B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 669 |
+
"S5-0014-Plus-81-14B-Exp.Q4_K_M.gguf": ["mradermacher/S5-0014-Plus-81-14B-Exp-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 670 |
+
"Ganymede-Llama-3.3-3B-Preview.Q5_K_M.gguf": ["mradermacher/Ganymede-Llama-3.3-3B-Preview-GGUF", MessagesFormatterType.LLAMA_3],
|
| 671 |
+
"Alpha-UMa-Qwen-14B.Q4_K_M.gguf": ["mradermacher/Alpha-UMa-Qwen-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 672 |
+
"Ice0.101-20.03-RP-GRPO-2.Q5_K_M.gguf": ["mradermacher/Ice0.101-20.03-RP-GRPO-2-GGUF", MessagesFormatterType.MISTRAL],
|
| 673 |
+
"Dirty-Shirley-Writer-v01-Uncensored.i1-Q4_K_M.gguf": ["mradermacher/Dirty-Shirley-Writer-v01-Uncensored-i1-GGUF", MessagesFormatterType.ALPACA],
|
| 674 |
+
"Dirty-Shirley-Writer-v0-abliterated.Q4_K_M.gguf": ["mradermacher/Dirty-Shirley-Writer-v0-abliterated-GGUF", MessagesFormatterType.ALPACA],
|
| 675 |
+
"Wicked-Love-12B-v2.Q4_K_M.gguf": ["mradermacher/Wicked-Love-12B-v2-GGUF", MessagesFormatterType.MISTRAL],
|
| 676 |
+
"Qwen2.5-14B-YOYO-V5.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-YOYO-V5-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 677 |
+
"gemma-2-9b-slimorca20k.Q4_K_M.gguf": ["mradermacher/gemma-2-9b-slimorca20k-GGUF", MessagesFormatterType.ALPACA],
|
| 678 |
+
"oh-dcft-v3.1-gemini-1.5-pro.i1-Q5_K_M.gguf": ["mradermacher/oh-dcft-v3.1-gemini-1.5-pro-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 679 |
+
"Community_Request-03-12B.Q4_K_M.gguf": ["mradermacher/Community_Request-03-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 680 |
+
"Community_Request-01-12B.i1-Q4_K_M.gguf": ["mradermacher/Community_Request-01-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 681 |
+
"Community_Request-02-12B.i1-Q4_K_M.gguf": ["mradermacher/Community_Request-02-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 682 |
+
"mt-merge9-gemma-2-9b-q4_k_m.gguf": ["Otakadelic/MT-Merge9-gemma-2-9B-Q4_K_M-GGUF", MessagesFormatterType.ALPACA],
|
| 683 |
+
"Qwen2.5-14B-YOYO-V2.i1-Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-YOYO-V2-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 684 |
+
"AV-FI-Qwen2.5-1.5B-PT-BR-Instruct.Q5_K_M.gguf": ["mradermacher/AV-FI-Qwen2.5-1.5B-PT-BR-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 685 |
+
"Qwen2.5-7B-YOYO.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-YOYO-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 686 |
+
"Ice0.101-20.03-RP.Q5_K_M.gguf": ["mradermacher/Ice0.101-20.03-RP-GGUF", MessagesFormatterType.MISTRAL],
|
| 687 |
+
"Qwen2.5-Osiris-7B-Instruct.Q5_K_M.gguf": ["mradermacher/Qwen2.5-Osiris-7B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 688 |
+
"Forgotten-Abomination-8B-v4.0.Q5_K_M.gguf": ["mradermacher/Forgotten-Abomination-8B-v4.0-GGUF", MessagesFormatterType.MISTRAL],
|
| 689 |
+
"Qwen2.5-7B-sft-ultrachat-all-pool.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-sft-ultrachat-all-pool-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 690 |
+
"Qwen2.5-7B-sft-ultrachat-SPIN-Qwen2.5-72B-Instruct.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-sft-ultrachat-SPIN-Qwen2.5-72B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 691 |
+
"Qwen2.5-1.5B-YOYO.Q5_K_M.gguf": ["mradermacher/Qwen2.5-1.5B-YOYO-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 692 |
+
"Qwen2.5-3B-YOYO.Q5_K_M.gguf": ["mradermacher/Qwen2.5-3B-YOYO-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 693 |
+
"NeMo-NeMo-12B.Q4_K_M.gguf": ["mradermacher/NeMo-NeMo-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 694 |
+
"Qwen2.5-Coder-14B-YOYO.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Coder-14B-YOYO-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 695 |
+
"Qwen2.5-14B-YOYO.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-YOYO-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 696 |
+
"Qwen2.5-14B-YOYO-V3.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-YOYO-V3-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 697 |
+
"Saanvi-C0-12B.Q4_K_M.gguf": ["mradermacher/Saanvi-C0-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 698 |
+
"Celestial-Harmony-14b-v1.0-Experimental-1015.Q4_K_M.gguf": ["mradermacher/Celestial-Harmony-14b-v1.0-Experimental-1015-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 699 |
+
"Qwen2.5-3B-EMPO-TQA.Q5_K_M.gguf": ["mradermacher/Qwen2.5-3B-EMPO-TQA-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 700 |
+
"Wicked-Love-12B.Q4_K_M.gguf": ["mradermacher/Wicked-Love-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 701 |
+
"UnslopNemo-Mag-Mell_T-2.Q4_K_M.gguf": ["mradermacher/UnslopNemo-Mag-Mell_T-2-GGUF", MessagesFormatterType.MISTRAL],
|
| 702 |
+
"Neporalito-12B-x2.Q4_K_M.gguf": ["mradermacher/Neporalito-12B-x2-GGUF", MessagesFormatterType.MISTRAL],
|
| 703 |
+
"Qwen2.5-7B-sft-ultrachat-peers-pool.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-sft-ultrachat-peers-pool-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 704 |
+
"Qwen2-7B-sft-ultrachat-peers-pool.Q5_K_M.gguf": ["mradermacher/Qwen2-7B-sft-ultrachat-peers-pool-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 705 |
+
"Fiendish_LLAMA_3B.Q5_K_M.gguf": ["mradermacher/Fiendish_LLAMA_3B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 706 |
+
"Qwen3b-bs128-GRPO.Q5_K_M.gguf": ["mradermacher/Qwen3b-bs128-GRPO-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 707 |
+
"UnslopNemo-Mag-Mell_T-1.Q4_K_M.gguf": ["mradermacher/UnslopNemo-Mag-Mell_T-1-GGUF", MessagesFormatterType.MISTRAL],
|
| 708 |
+
"llama3.1-8b-instruct-hasanpiker-abliterated.Q5_K_M.gguf": ["mradermacher/llama3.1-8b-instruct-hasanpiker-abliterated-GGUF", MessagesFormatterType.LLAMA_3],
|
| 709 |
+
"qwen-writerdemo-7b-s500.Q5_K_M.gguf": ["mradermacher/qwen-writerdemo-7b-s500-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 710 |
+
"AlignXpert-PBA-Subset.Q5_K_M.gguf": ["mradermacher/AlignXpert-PBA-Subset-GGUF", MessagesFormatterType.LLAMA_3],
|
| 711 |
+
"Glowing-Forest-2-12B.Q4_K_M.gguf": ["mradermacher/Glowing-Forest-2-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 712 |
+
"Kosmos-EVAA-gamma-v16-8B.Q5_K_M.gguf": ["mradermacher/Kosmos-EVAA-gamma-v16-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 713 |
+
"gemma-2-aeria-infinity-9b.Q4_K_M.gguf": ["mradermacher/gemma-2-aeria-infinity-9b-GGUF", MessagesFormatterType.ALPACA],
|
| 714 |
+
"Kosmos-EVAA-gamma-v13-8B.Q5_K_M.gguf": ["mradermacher/Kosmos-EVAA-gamma-v13-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 715 |
+
"DeepHermes-3-Llama-3-3B-Preview-abliterated.i1-Q5_K_M.gguf": ["mradermacher/DeepHermes-3-Llama-3-3B-Preview-abliterated-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 716 |
+
"GeneralThought-Qwen-7B-instruct-SFT.Q5_K_M.gguf": ["mradermacher/GeneralThought-Qwen-7B-instruct-SFT-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 717 |
+
"MN-Hekate-Kleidoukhos-12B.Q4_K_M.gguf": ["mradermacher/MN-Hekate-Kleidoukhos-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 718 |
+
"Forgotten-Abomination-12B-v4.0.Q4_K_M.gguf": ["mradermacher/Forgotten-Abomination-12B-v4.0-GGUF", MessagesFormatterType.MISTRAL],
|
| 719 |
+
"Kosmos-EVAA-gamma-v18-8B.Q5_K_M.gguf": ["Kosmos-EVAA-gamma-v18-8B.Q5_K_M.gguf", MessagesFormatterType.LLAMA_3],
|
| 720 |
+
"Llama-3.2-3B-Instruct-abliterated-finetuned.Q5_K_M.gguf": ["mradermacher/Llama-3.2-3B-Instruct-abliterated-finetuned-GGUF", MessagesFormatterType.LLAMA_3],
|
| 721 |
+
"Neo-1-7B-Flash.Q5_K_M.gguf": ["mradermacher/Neo-1-7B-Flash-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 722 |
+
"NightWind-12B.Q4_K_M.gguf": ["mradermacher/NightWind-12B-GGUF", MessagesFormatterType.CHATML],
|
| 723 |
+
"ablation-81-bestofn.athenev2-shisa-v2-llama-3.1-8b-lr8e6.Q5_K_M.gguf": ["mradermacher/ablation-81-bestofn.athenev2-shisa-v2-llama-3.1-8b-lr8e6-GGUF", MessagesFormatterType.LLAMA_3],
|
| 724 |
+
"formula-bs-mix-7B-level-1-0318.Q5_K_M.gguf": ["mradermacher/formula-bs-mix-7B-level-1-0318-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 725 |
+
"nano-r1-model.Q5_K_M.gguf": ["mradermacher/nano-r1-model-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 726 |
+
"NM-12B-Lyris-dev-2.Q4_K_M.gguf": ["mradermacher/NM-12B-Lyris-dev-2-GGUF", MessagesFormatterType.MISTRAL],
|
| 727 |
+
"RAt0.25-gemma-2-RI-9B.Q4_K_M.gguf": ["mradermacher/RAt0.25-gemma-2-RI-9B-GGUF", MessagesFormatterType.ALPACA],
|
| 728 |
+
"MN-Hecate-Chthonia-12B.Q4_K_M.gguf": ["mradermacher/MN-Hecate-Chthonia-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 729 |
+
"Dark-Science-Forgotten-Safeword-12B.Q4_K_M.gguf": ["mradermacher/Dark-Science-Forgotten-Safeword-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 730 |
+
"BBmath-Qwen2.5-3B-GRPO.Q5_K_M.gguf": ["mradermacher/BBmath-Qwen2.5-3B-GRPO-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 731 |
+
"Delta-Pavonis-Qwen-14B.Q4_K_S.gguf": ["mradermacher/Delta-Pavonis-Qwen-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 732 |
+
"BeaverAI_MN-2407-DSK-QwQify-v0.1-12B-Q4_K_M.gguf": ["bartowski/BeaverAI_MN-2407-DSK-QwQify-v0.1-12B-GGUF", MessagesFormatterType.MISTRAL],
|
| 733 |
+
"EtherealNight-12B.Q4_K_S.gguf": ["mradermacher/EtherealNight-12B-GGUF", MessagesFormatterType.CHATML],
|
| 734 |
+
"MN-Nyx-Chthonia-12B.i1-IQ3_M.gguf": ["mradermacher/MN-Nyx-Chthonia-12B-i1-GGUF", MessagesFormatterType.CHATML],
|
| 735 |
+
"Nu2-Lupi-Qwen-14B.i1-Q4_K_S.gguf": ["mradermacher/Nu2-Lupi-Qwen-14B-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 736 |
+
"KPT-7b-v0.2_stock.Q5_K_M.gguf": ["mradermacher/KPT-7b-v0.2_stock-GGUF", MessagesFormatterType.MISTRAL],
|
| 737 |
+
"Violet-Lyra-Gutenberg-v4.Q4_K_M.gguf": ["mradermacher/Violet-Lyra-Gutenberg-v4-GGUF", MessagesFormatterType.MISTRAL],
|
| 738 |
+
"Kosmos-EVAA-v12-8B.Q5_K_M.gguf": ["mradermacher/Kosmos-EVAA-v12-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 739 |
+
"mistral-nemo-instruct-2407-Reasoning.Q4_K_M.gguf": ["mradermacher/mistral-nemo-instruct-2407-Reasoning-GGUF", MessagesFormatterType.MISTRAL],
|
| 740 |
+
"MT-Merge-gemma-2-9B.Q4_K_M.gguf": ["mradermacher/MT-Merge-gemma-2-9B-GGUF", MessagesFormatterType.ALPACA],
|
| 741 |
+
"Blagoveshchensk_14B_V4.Q4_K_M.gguf": ["mradermacher/Blagoveshchensk_14B_V4-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 742 |
+
"ablation-69-a67.dpo.armorm-ablation-67-nemo-12b.Q4_K_S.gguf": ["mradermacher/ablation-69-a67.dpo.armorm-ablation-67-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
|
| 743 |
+
"ablation-07-raftulu70b-shisa-v2-llama3.1-8b-lr8e6.Q5_K_M.gguf": ["mradermacher/ablation-07-raftulu70b-shisa-v2-llama3.1-8b-lr8e6-GGUF", MessagesFormatterType.LLAMA_3],
|
| 744 |
+
"ablation-60-a55.dpo.enja-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-60-a55.dpo.enja-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 745 |
+
"ablation-53-rafathenev2.rp-shisa-v2-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/ablation-53-rafathenev2.rp-shisa-v2-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
|
| 746 |
+
"ablation-61-a55.dpo.enjanot-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-61-a55.dpo.enjanot-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 747 |
+
"ablation-32-rafbestseq.reasoning-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-32-rafbestseq.reasoning-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
|
| 748 |
+
"ablation-42-rafathenev2.cmrmixen.masked-shisa-v2-llama-3.1-8b-lr8e6.Q5_K_M.gguf": ["mradermacher/ablation-42-rafathenev2.cmrmixen.masked-shisa-v2-llama-3.1-8b-lr8e6-GGUF", MessagesFormatterType.LLAMA_3],
|
| 749 |
+
"ablation-23-rafbestseqshuffle.2epoch-shisa-v2-llama-3.1-8b-lr8e6.Q5_K_M.gguf": ["mradermacher/ablation-23-rafbestseqshuffle.2epoch-shisa-v2-llama-3.1-8b-lr8e6-GGUF", MessagesFormatterType.LLAMA_3],
|
| 750 |
+
"ablation-67-rafathenev2.rp.tl.yahoo-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/ablation-67-rafathenev2.rp.tl.yahoo-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
|
| 751 |
+
"EtherealMoon-12B.Q4_K_M.gguf": ["mradermacher/EtherealMoon-12B-GGUF", MessagesFormatterType.CHATML],
|
| 752 |
+
"ablation-56-rafathenev2.rp.tl.gamma-shisa-v2-gamma-7b.Q5_K_M.gguf": ["mradermacher/ablation-56-rafathenev2.rp.tl.gamma-shisa-v2-gamma-7b-GGUF", MessagesFormatterType.MISTRAL],
|
| 753 |
+
"ablation-11-rafathenev2-shisa-v2-llama-3.1-8b-lr8e6.Q5_K_M.gguf": ["mradermacher/ablation-11-rafathenev2-shisa-v2-llama-3.1-8b-lr8e6-GGUF", MessagesFormatterType.LLAMA_3],
|
| 754 |
+
"ablation-36-cmrmix-shisa-v2-llama-3.1-8b-lr8e6.Q5_K_M.gguf": ["mradermacher/ablation-36-cmrmix-shisa-v2-llama-3.1-8b-lr8e6-GGUF", MessagesFormatterType.LLAMA_3],
|
| 755 |
+
"L3.1-Athena-n-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Athena-n-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 756 |
+
"NQLSG-Qwen2.5-14B-MegaFusion-v9.3.Q4_K_M.gguf": ["mradermacher/NQLSG-Qwen2.5-14B-MegaFusion-v9.3-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 757 |
+
"NQLSG-Qwen2.5-14B-MegaFusion-v9.2-fusechat-sft.Q4_K_M.gguf": ["mradermacher/NQLSG-Qwen2.5-14B-MegaFusion-v9.2-fusechat-sft-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 758 |
+
"TopicalChat_gemma.Q4_K_M.gguf": ["mradermacher/TopicalChat_gemma-GGUF", MessagesFormatterType.ALPACA],
|
| 759 |
+
"general7Bv2-ECE-PRYMMAL-Martial.Q5_K_M.gguf": ["mradermacher/general7Bv2-ECE-PRYMMAL-Martial-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 760 |
+
"NM-12B-Lyris-dev-3.Q4_K_M.gguf": ["mradermacher/NM-12B-Lyris-dev-3-GGUF", MessagesFormatterType.CHATML],
|
| 761 |
+
"DeepHermes-3-Llama-3-3B-Preview.i1-Q5_K_M.gguf": ["mradermacher/DeepHermes-3-Llama-3-3B-Preview-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 762 |
+
"Kosmos-EVAA-v10-8B.Q5_K_M.gguf": ["mradermacher/Kosmos-EVAA-v10-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 763 |
+
"Scrollplay.Q5_K_M.gguf": ["mradermacher/Scrollplay-GGUF", MessagesFormatterType.MISTRAL],
|
| 764 |
+
"Kosmos-EVAA-gamma-v17-8B.Q5_K_M.gguf": ["mradermacher/Kosmos-EVAA-gamma-v17-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 765 |
+
"Cheng-2-Base.i1-Q4_K_M.gguf": ["mradermacher/Cheng-2-Base-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 766 |
+
"Cheng-2.i1-Q4_K_M.gguf": ["mradermacher/Cheng-2-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 767 |
+
"Perpicon-Experimental-3.1.Q5_K_M.gguf": ["mradermacher/Perpicon-Experimental-3.1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 768 |
+
"Gemma-2-Ataraxy-v3b-9B.Q4_K_M.gguf": ["mradermacher/Gemma-2-Ataraxy-v3b-9B-GGUF", MessagesFormatterType.ALPACA],
|
| 769 |
+
"llama-3-1-8b-answer-generator-ft-2.Q5_K_M.gguf": ["mradermacher/llama-3-1-8b-answer-generator-ft-2-GGUF", MessagesFormatterType.LLAMA_3],
|
| 770 |
+
"nsfw-back-to-model-stock.Q5_K_M.gguf": ["mradermacher/nsfw-back-to-model-stock-GGUF", MessagesFormatterType.LLAMA_3],
|
| 771 |
+
"MN-Chthonia-12B.i1-Q4_K_M.gguf": ["mradermacher/MN-Chthonia-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
|
| 772 |
+
"exurbia-enhanced-q4_k_m.gguf": ["ClaudioItaly/Exurbia-Enhanced-Q4_K_M-GGUF", MessagesFormatterType.ALPACA],
|
| 773 |
+
"Gemma-2-Ataraxy-v3j-9B.Q4_K_M.gguf": ["mradermacher/Gemma-2-Ataraxy-v3j-9B-GGUF", MessagesFormatterType.ALPACA],
|
| 774 |
+
"Abus-14B-Instruct.Q4_K_M.gguf": ["mradermacher/Abus-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 775 |
+
"Qwen2.5-7B-task8.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-task8-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 776 |
+
"Cheng-2-v1.1.Q4_K_M.gguf": ["mradermacher/Cheng-2-v1.1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 777 |
+
"Qwen2.5-7B-Fuse-Exp.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Fuse-Exp-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 778 |
+
"Llama-3.1-8B-sft-ultrachat-gen-dpo-10k.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-sft-ultrachat-gen-dpo-10k-GGUF", MessagesFormatterType.LLAMA_3],
|
| 779 |
+
"NousResearch_DeepHermes-3-Llama-3-3B-Preview-Q5_K_M.gguf": ["bartowski/NousResearch_DeepHermes-3-Llama-3-3B-Preview-GGUF", MessagesFormatterType.LLAMA_3],
|
| 780 |
+
"Romulus-cpt-Llama-3.1-8B-v0.1-Instruct.Q5_K_M.gguf": ["mradermacher/Romulus-cpt-Llama-3.1-8B-v0.1-Instruct-GGUF", MessagesFormatterType.LLAMA_3],
|
| 781 |
+
"caliburn-v2-12b.Q4_K_M.gguf": ["mradermacher/caliburn-v2-12b-GGUF", MessagesFormatterType.MISTRAL],
|
| 782 |
+
"hermes-llama3-roleplay-4000-v2.Q5_K_M.gguf": ["mradermacher/hermes-llama3-roleplay-4000-v2-GGUF", MessagesFormatterType.LLAMA_3],
|
| 783 |
+
"hercule-bn.Q5_K_M.gguf": ["mradermacher/hercule-bn-GGUF", MessagesFormatterType.LLAMA_3],
|
| 784 |
+
"G2-GSHT-32K.Q4_K_M.gguf": ["mradermacher/G2-GSHT-32K-GGUF", MessagesFormatterType.ALPACA],
|
| 785 |
+
"krx-qwen2.5-7B-pt.Q5_K_M.gguf": ["mradermacher/krx-qwen2.5-7B-pt-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 786 |
+
"etherealaurora-12b-v3-q4_k_m.gguf": ["yamatazen/EtherealAurora-12B-v3-Q4_K_M-GGUF", MessagesFormatterType.MISTRAL],
|
| 787 |
+
"4PRYMMAL-GEMMA2-9B-SLERP.Q4_K_M.gguf": ["mradermacher/4PRYMMAL-GEMMA2-9B-SLERP-GGUF", MessagesFormatterType.ALPACA],
|
| 788 |
+
"gemma-2-aeria-9b.Q4_K_M.gguf": ["mradermacher/gemma-2-aeria-9b-GGUF", MessagesFormatterType.ALPACA],
|
| 789 |
+
"gemma-2-inko-9b.Q4_K_M.gguf": ["mradermacher/gemma-2-inko-9b-GGUF", MessagesFormatterType.ALPACA],
|
| 790 |
+
"krx-qwen2.5-7b-it-prompt.Q5_K_M.gguf": ["mradermacher/krx-qwen2.5-7b-it-prompt-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 791 |
+
"EmojiLlama-3.1-8B.Q5_K_M.gguf": ["mradermacher/EmojiLlama-3.1-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 792 |
+
"Kosmos-EVAA-gamma-v15-8B.Q5_K_M.gguf": ["mradermacher/Kosmos-EVAA-gamma-v15-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 793 |
+
"Qwen2.5-7B-Instruct-preference.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-preference-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 794 |
+
"ZYH-LLM-Qwen2.5-14B-V4.Q4_K_M.gguf": ["mradermacher/ZYH-LLM-Qwen2.5-14B-V4-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 795 |
+
"Agente-GPT-Qwen2.5-3B-Instruct-Asistant-16bit-v2.Q5_K_M.gguf": ["mradermacher/Agente-GPT-Qwen2.5-3B-Instruct-Asistant-16bit-v2-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 796 |
+
"TopicalChat_mistral_nemo.Q4_K_M.gguf": ["mradermacher/TopicalChat_mistral_nemo-GGUF", MessagesFormatterType.MISTRAL],
|
| 797 |
+
"Llama-3.2-3B-Fluxed-uncensored.Q5_K_M.gguf": ["mradermacher/Llama-3.2-3B-Fluxed-uncensored-GGUF", MessagesFormatterType.LLAMA_3],
|
| 798 |
+
"Qwen2.5-Coder-3B-few-shot-model.Q5_K_M.gguf": ["mradermacher/Qwen2.5-Coder-3B-few-shot-model-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 799 |
+
"Gigantes-v2-gemma2-9b-it.Q4_K_M.gguf": ["mradermacher/Gigantes-v2-gemma2-9b-it-GGUF", MessagesFormatterType.ALPACA],
|
| 800 |
+
"Unichat-llama3.1-Chinese-8B.Q5_K_M.gguf": ["mradermacher/Unichat-llama3.1-Chinese-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 801 |
+
"HollandV2R1.Q5_K_M.gguf": ["mradermacher/HollandV2R1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 802 |
+
"Qwen-2.5-3b-Text_to_SQL.Q5_K_M.gguf": ["mradermacher/Qwen-2.5-3b-Text_to_SQL-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 803 |
+
"TableGPT2-7B.Q5_K_M.gguf": ["mradermacher/TableGPT2-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 804 |
+
"Kosmos-EVAA-gamma-v14-8B.Q5_K_M.gguf": ["mradermacher/Kosmos-EVAA-gamma-v14-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 805 |
+
"KPT-7b-v0.3.Q5_K_M.gguf": ["mradermacher/KPT-7b-v0.3-GGUF", MessagesFormatterType.MISTRAL],
|
| 806 |
+
"qwen2.5-coder-7b-instruct-fma.Q5_K_M.gguf": ["mradermacher/qwen2.5-coder-7b-instruct-fma-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 807 |
+
"oh_v1.3_unnatural_instructions_x8.Q5_K_M.gguf": ["mradermacher/oh_v1.3_unnatural_instructions_x8-GGUF", MessagesFormatterType.LLAMA_3],
|
| 808 |
+
"Qwen2.5-Taiwan-7B-Instruct.Q5_K_M.gguf": ["mradermacher/Qwen2.5-Taiwan-7B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 809 |
+
"NQLSG-Qwen2.5-14B-MegaFusion-v9.2.Q4_K_M.gguf": ["mradermacher/NQLSG-Qwen2.5-14B-MegaFusion-v9.2-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 810 |
+
"Ext2Gen-8B-R2.Q5_K_M.gguf": ["mradermacher/Ext2Gen-8B-R2-GGUF", MessagesFormatterType.LLAMA_3],
|
| 811 |
+
"Meta-Llama-3-8B-Instruct-abliterated.i1-Q5_K_M.gguf": ["mradermacher/Meta-Llama-3-8B-Instruct-abliterated-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 812 |
+
"Dirty-Shirley-Writer-v3-9B.i1-Q4_K_M.gguf": ["mradermacher/Dirty-Shirley-Writer-v3-9B-i1-GGUF", MessagesFormatterType.ALPACA],
|
| 813 |
+
"EtherealAurora-12B-v2.Q4_K_M.gguf": ["mradermacher/EtherealAurora-12B-v2-GGUF", MessagesFormatterType.MISTRAL],
|
| 814 |
+
"Galactic-Qwen-14B-Exp2.Q4_K_M.gguf": ["mradermacher/Galactic-Qwen-14B-Exp2-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 815 |
+
"Qwen2.5-14B-Verged.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Verged-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 816 |
+
"Galactic-Qwen-14B-Exp1.Q4_K_M.gguf": ["mradermacher/Galactic-Qwen-14B-Exp1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 817 |
+
"Volans-Opus-14B-Exp.Q4_K_M.gguf": ["mradermacher/Volans-Opus-14B-Exp-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 818 |
+
"MN-Anathema-12B.Q4_K_M.gguf": ["mradermacher/MN-Anathema-12B-GGUF", MessagesFormatterType.CHATML],
|
| 819 |
+
"Gaea-Opus-14B-Exp.Q4_K_M.gguf": ["mradermacher/Gaea-Opus-14B-Exp-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 820 |
+
"OnlyQ.Q5_K_M.gguf": ["mradermacher/OnlyQ-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 821 |
+
"asteroid-14b-v0.1.Q4_K_M.gguf": ["mradermacher/asteroid-14b-v0.1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 822 |
+
"Forgotten-Safeword-8B-V3.0.Q5_K_M.gguf": ["mradermacher/Forgotten-Safeword-8B-V3.0-GGUF", MessagesFormatterType.MISTRAL],
|
| 823 |
+
"Qwen2.5-7B-PubMeQA-Distill.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-PubMeQA-Distill-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 824 |
+
"Llama-3.2-3b-RP-Toxic-Fuse.Q5_K_M.gguf": ["mradermacher/Llama-3.2-3b-RP-Toxic-Fuse-GGUF", MessagesFormatterType.LLAMA_3],
|
| 825 |
+
"v1-deepseek-instruct.Q5_K_M.gguf": ["mradermacher/v1-deepseek-instruct-GGUF", MessagesFormatterType.LLAMA_3],
|
| 826 |
+
"L3.1-Athena-l2-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Athena-l2-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 827 |
+
"L3.1-Athena-l4-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Athena-l4-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 828 |
+
"L3.1-Athena-l3-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Athena-l3-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 829 |
+
"Llama-3.1-8B-AthenaSky-MegaMix.i1-Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-AthenaSky-MegaMix-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 830 |
+
"Writing-Model-Qwen-7B.Q5_K_M.gguf": ["mradermacher/Writing-Model-Qwen-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 831 |
+
"open-r1_OlympicCoder-7B-Q5_K_M.gguf": ["bartowski/open-r1_OlympicCoder-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 832 |
+
"Qwen-2.5-7B-deepscale-RL.Q5_K_M.gguf": ["mradermacher/Qwen-2.5-7B-deepscale-RL-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 833 |
+
"L3.1-Athena-l-8B.i1-Q5_K_M.gguf": ["mradermacher/L3.1-Athena-l-8B-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 834 |
+
"Llama-3.2-3B-Instruct-lyrics.Q5_K_M.gguf": ["mradermacher/Llama-3.2-3B-Instruct-lyrics-GGUF", MessagesFormatterType.LLAMA_3],
|
| 835 |
+
"Llama-3-Base-8B-DICE-Iter2.Q5_K_M.gguf": ["mradermacher/Llama-3-Base-8B-DICE-Iter2-GGUF", MessagesFormatterType.LLAMA_3],
|
| 836 |
+
"Llama-3-Base-8B-DICE-Iter1.Q5_K_M.gguf": ["mradermacher/Llama-3-Base-8B-DICE-Iter1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 837 |
+
"Llama-3-8B-Instruct-Coding-Expert.Q5_K_M.gguf": ["mradermacher/Llama-3-8B-Instruct-Coding-Expert-GGUF", MessagesFormatterType.LLAMA_3],
|
| 838 |
+
"Llama-O-3.1-8B-levent.Q5_K_M.gguf": ["mradermacher/Llama-O-3.1-8B-levent-GGUF", MessagesFormatterType.LLAMA_3],
|
| 839 |
+
"Qwen2.5-7B-task6.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-task6-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 840 |
+
"finance-Llama3-8B.Q5_K_M.gguf": ["mradermacher/finance-Llama3-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 841 |
+
"OpenPipe_Deductive-Reasoning-Qwen-14B-Q4_K_M.gguf": ["bartowski/OpenPipe_Deductive-Reasoning-Qwen-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 842 |
+
"Dirty-Shirley-Writer-v2-9B-Uncensored.i1-Q4_K_M.gguf": ["mradermacher/Dirty-Shirley-Writer-v2-9B-Uncensored-i1-GGUF", MessagesFormatterType.ALPACA],
|
| 843 |
+
"ichor_1.1-8B-Model_Stock.Q5_K_M.gguf": ["mradermacher/ichor_1.1-8B-Model_Stock-GGUF", MessagesFormatterType.LLAMA_3],
|
| 844 |
+
"gemma-2-2b-fa-v2.Q5_K_M.gguf": ["mradermacher/gemma-2-2b-fa-v2-GGUF", MessagesFormatterType.ALPACA],
|
| 845 |
+
"H_the_eighth-8B-LINEAR.Q5_K_M.gguf": ["mradermacher/H_the_eighth-8B-LINEAR-GGUF", MessagesFormatterType.LLAMA_3],
|
| 846 |
+
"granite-3.2-8b-instruct.Q5_K_M.gguf": ["mradermacher/granite-3.2-8b-instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 847 |
+
"tempestissimo-14b-0309.Q4_K_M.gguf": ["mradermacher/tempestissimo-14b-0309-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 848 |
+
"Deepseek-R1-Distill-NSFW-RP-vRedux-Proper.Q5_K_M.gguf": ["mradermacher/Deepseek-R1-Distill-NSFW-RP-vRedux-Proper-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 849 |
+
"inexpertus_1.1-8B-LINEAR.Q5_K_M.gguf": ["mradermacher/inexpertus_1.1-8B-LINEAR-GGUF", MessagesFormatterType.LLAMA_3],
|
| 850 |
+
"Llama-3.1-8B-RainbowLight-EtherealMix.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-RainbowLight-EtherealMix-GGUF", MessagesFormatterType.LLAMA_3],
|
| 851 |
+
"Llama-3.1-8B-SuperNova-EtherealHermes.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-SuperNova-EtherealHermes-GGUF", MessagesFormatterType.LLAMA_3],
|
| 852 |
+
"Dirty-Shirley-Quill-v1-gemma-2-Ifable-9B-Uncensored-slerp.Q4_K_M.gguf": ["mradermacher/Dirty-Shirley-Quill-v1-gemma-2-Ifable-9B-Uncensored-slerp-GGUF", MessagesFormatterType.ALPACA],
|
| 853 |
+
"Llama3.1-SuperHawk-8B.Q5_K_M.gguf": ["mradermacher/Llama3.1-SuperHawk-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 854 |
+
"Qwen-2.5-7B-Reasoning.Q5_K_M.gguf": ["mradermacher/Qwen-2.5-7B-Reasoning-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 855 |
+
"nsfw-sce-test-2-redux.Q5_K_M.gguf": ["mradermacher/nsfw-sce-test-2-redux-GGUF", MessagesFormatterType.LLAMA_3],
|
| 856 |
+
"DCFT-camel_physics_seed_science-etash.Q5_K_M.gguf": ["mradermacher/DCFT-camel_physics_seed_science-etash-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 857 |
+
"Qwen2.5-Coder-7B-Instruct-Ties.Q5_K_M.gguf": ["mradermacher/Qwen2.5-Coder-7B-Instruct-Ties-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 858 |
+
"Meta-Llama-3-8B-Instruct-abliterated.Q5_K_M.gguf": ["mradermacher/Meta-Llama-3-8B-Instruct-abliterated-GGUF", MessagesFormatterType.LLAMA_3],
|
| 859 |
+
"MFANN-llama3.1-abliterated-v2.Q5_K_M.gguf": ["mradermacher/MFANN-llama3.1-abliterated-v2-GGUF", MessagesFormatterType.LLAMA_3],
|
| 860 |
+
"lamarckvergence-14b-tensopolis-v1.Q4_K_M.gguf": ["mradermacher/lamarckvergence-14b-tensopolis-v1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 861 |
+
"NQLSG-Qwen2.5-14B-MegaFusion-v9.1.i1-Q4_K_M.gguf": ["mradermacher/NQLSG-Qwen2.5-14B-MegaFusion-v9.1-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 862 |
+
"s1-Qwen2.5-7B-Instruct.Q5_K_M.gguf": ["mradermacher/s1-Qwen2.5-7B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 863 |
+
"s1-Qwen2.5-7B.Q5_K_M.gguf": ["mradermacher/s1-Qwen2.5-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 864 |
+
"Llama3.1-IgneousIguana-8B.Q5_K_M.gguf": ["mradermacher/Llama3.1-IgneousIguana-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 865 |
+
"Llama-3-Instruct-8B-SPPO-Iter3-SimPO-merge.Q5_K_M.gguf": ["mradermacher/Llama-3-Instruct-8B-SPPO-Iter3-SimPO-merge-GGUF", MessagesFormatterType.LLAMA_3],
|
| 866 |
+
"recoilme-gemma-2-Ataraxy-9B-v0.1.i1-Q4_K_M.gguf": ["mradermacher/recoilme-gemma-2-Ataraxy-9B-v0.1-i1-GGUF", MessagesFormatterType.ALPACA],
|
| 867 |
+
"Gemma-Moon-2-9B-it.Q4_K_M.gguf": ["mradermacher/Gemma-Moon-2-9B-it-GGUF", MessagesFormatterType.ALPACA],
|
| 868 |
+
"CardProjector-7B-v2.Q5_K_M.gguf": ["mradermacher/CardProjector-7B-v2-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 869 |
+
"PerpetualNight-12B.Q4_K_M.gguf": ["mradermacher/PerpetualNight-12B-GGUF", MessagesFormatterType.CHATML],
|
| 870 |
+
"Llama3.2-3B-Instruct-KAI.Q5_K_M.gguf": ["mradermacher/Llama3.2-3B-Instruct-KAI-GGUF", MessagesFormatterType.LLAMA_3],
|
| 871 |
+
"Llama-3.1-8B-SuperTulu-LexiNova.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-SuperTulu-LexiNova-GGUF", MessagesFormatterType.LLAMA_3],
|
| 872 |
+
"Qwen2-1.5B-Ita_v5.Q5_K_M.gguf": ["mradermacher/Qwen2-1.5B-Ita_v5-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 873 |
+
"Llama-VARCO-8B-Instruct-LLaMAfied.i1-Q5_K_M.gguf": ["mradermacher/Llama-VARCO-8B-Instruct-LLaMAfied-i1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 874 |
+
"llama3-1_8b_mlfoundations-dev-stackexchange_scicomp.Q5_K_M.gguf": ["mradermacher/llama3-1_8b_mlfoundations-dev-stackexchange_scicomp-GGUF", MessagesFormatterType.LLAMA_3],
|
| 875 |
+
"Gemma-Ko-Merge.i1-Q4_K_M.gguf": ["mradermacher/Gemma-Ko-Merge-i1-GGUF", MessagesFormatterType.ALPACA],
|
| 876 |
+
"gemma-2-9b-it-russian-strict-function-calling-DPO.Q4_K_M.gguf": ["mradermacher/gemma-2-9b-it-russian-strict-function-calling-DPO-GGUF", MessagesFormatterType.ALPACA],
|
| 877 |
+
"Llama-3.1-Supernova-Lite-Instruct-merge-abliterated-8B.Q5_K_M.gguf": ["mradermacher/Llama-3.1-Supernova-Lite-Instruct-merge-abliterated-8B-GGUF", MessagesFormatterType.LLAMA_3],
|
| 878 |
+
"Qwenslerp2-14B-F32.Q4_K_M.gguf": ["mradermacher/Qwenslerp2-14B-F32-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 879 |
+
"Llama-3-Instruct-8B-SimPO-SPPO-Iter3-merge.Q5_K_M.gguf": ["mradermacher/Llama-3-Instruct-8B-SimPO-SPPO-Iter3-merge-GGUF", MessagesFormatterType.LLAMA_3],
|
| 880 |
+
"gemma-2-baku-2b-it.i1-Q5_K_M.gguf": ["mradermacher/gemma-2-baku-2b-it-i1-GGUF", MessagesFormatterType.ALPACA],
|
| 881 |
+
"Nemo-Picaro-12B.i1-Q4_K_M.gguf": ["mradermacher/Nemo-Picaro-12B-i1-GGUF", MessagesFormatterType.CHATML],
|
| 882 |
+
"OpenCrystal-L3-15B-v2.1.Q4_K_M.gguf": ["mradermacher/OpenCrystal-L3-15B-v2.1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 883 |
+
"LLMEvoLLaMA-3.1-8B-v0.1.Q5_K_M.gguf": ["mradermacher/LLMEvoLLaMA-3.1-8B-v0.1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 884 |
+
"Peach-2.0-9B-8k-Roleplay.Q4_K_M.gguf": ["mradermacher/Peach-2.0-9B-8k-Roleplay-GGUF", MessagesFormatterType.CHATML],
|
| 885 |
+
"llama-3.1-8b-sarcasm.Q5_K_M.gguf": ["mradermacher/llama-3.1-8b-sarcasm-GGUF", MessagesFormatterType.LLAMA_3],
|
| 886 |
+
"walrus-gemma-2-9b-it-v01.Q4_K_M.gguf": ["mradermacher/walrus-gemma-2-9b-it-v01-GGUF", MessagesFormatterType.ALPACA],
|
| 887 |
+
"L3.1-Diffbot-Small-2412.Q5_K_M.gguf": ["mradermacher/L3.1-Diffbot-Small-2412-GGUF", MessagesFormatterType.LLAMA_3],
|
| 888 |
+
"s1.1-14B.Q4_K_M.gguf": ["mradermacher/s1.1-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 889 |
+
"Lix-14B-v0.1-Uncensored-DeLMAT.Q4_K_M.gguf": ["mradermacher/Lix-14B-v0.1-Uncensored-DeLMAT-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 890 |
"Hush-Qwen2.5-7B-MST-v1.1.Q5_K_M.gguf": ["mradermacher/Hush-Qwen2.5-7B-MST-v1.1-GGUF", MessagesFormatterType.OPEN_CHAT],
|
| 891 |
"MFANNv0.22.1.Q5_K_M.gguf": ["mradermacher/MFANNv0.22.1-GGUF", MessagesFormatterType.LLAMA_3],
|
| 892 |
"gemma2-9b-cpt-sea-lionv3-instruct.Q4_K_M.gguf": ["mradermacher/gemma2-9b-cpt-sea-lionv3-instruct-GGUF", MessagesFormatterType.ALPACA],
|
lora_dict.json
CHANGED
|
@@ -8882,6 +8882,13 @@
|
|
| 8882 |
"https://civitai.com/models/491760",
|
| 8883 |
"https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/36267afb-7093-4c71-ae47-516c7219fabd/width=450/14403049.jpeg"
|
| 8884 |
],
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 8885 |
"triple_oral_-_triple_barrel_fellatio_-_triple_fellatio-000014": [
|
| 8886 |
"tripplebj / read description",
|
| 8887 |
"Pony",
|
|
@@ -9470,6 +9477,13 @@
|
|
| 9470 |
"https://civitai.com/models/850497",
|
| 9471 |
"https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/0a48013a-809a-41a5-823d-e9183eae0dd0/width=450/34362448.jpeg"
|
| 9472 |
],
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9473 |
"\u7523\u5375\u65ad\u9762\u56f3": [
|
| 9474 |
"girl, egg laying, eggs in uterus",
|
| 9475 |
"Pony",
|
|
|
|
| 8882 |
"https://civitai.com/models/491760",
|
| 8883 |
"https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/36267afb-7093-4c71-ae47-516c7219fabd/width=450/14403049.jpeg"
|
| 8884 |
],
|
| 8885 |
+
"tribbie-hsrIL-v1": [
|
| 8886 |
+
"tribbie, wings, detached collar, puffy long sleeves, black gloves, black and white boots",
|
| 8887 |
+
"Illustrious",
|
| 8888 |
+
"[IL & PONY & SD1.5] Honkai Star Rail - Tribbie | \u7f07\u5b9d 3+1",
|
| 8889 |
+
"https://civitai.com/models/1340805",
|
| 8890 |
+
"https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/f21f1b54-0b5a-4e92-ae08-6bf81eb2f0c2/width=1688/62634838.jpeg"
|
| 8891 |
+
],
|
| 8892 |
"triple_oral_-_triple_barrel_fellatio_-_triple_fellatio-000014": [
|
| 8893 |
"tripplebj / read description",
|
| 8894 |
"Pony",
|
|
|
|
| 9477 |
"https://civitai.com/models/850497",
|
| 9478 |
"https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/0a48013a-809a-41a5-823d-e9183eae0dd0/width=450/34362448.jpeg"
|
| 9479 |
],
|
| 9480 |
+
"\u7231\u8bfa-Illustrious": [
|
| 9481 |
+
"aino (genshin impact), pink hair, grey eyes, bangs pinned back, hair ornament, green dress, white coat, oversized clothes, mechanical tail, sleeves past fingers, mismatched legwear, asymmetrical legwear, open coat, sleeves past wrists",
|
| 9482 |
+
"Illustrious",
|
| 9483 |
+
"\u3010Genshin Impact\u3011\u7231\u8bfa\u4e28Aino\u4e28\u30a2\u30a4\u30ce",
|
| 9484 |
+
"https://civitai.com/models/1792295",
|
| 9485 |
+
"https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/ced12386-d16d-42b2-ad02-2ef70e736f5e/width=1536/94782692.jpeg"
|
| 9486 |
+
],
|
| 9487 |
"\u7523\u5375\u65ad\u9762\u56f3": [
|
| 9488 |
"girl, egg laying, eggs in uterus",
|
| 9489 |
"Pony",
|
modutils.py
CHANGED
|
@@ -1730,7 +1730,7 @@ EXAMPLES_GUI = [
|
|
| 1730 |
"votepurchase/ponyDiffusionV6XL",
|
| 1731 |
],
|
| 1732 |
[
|
| 1733 |
-
"1girl, oomuro sakurako, yuru yuri, official art, school uniform, anime artwork, anime style,
|
| 1734 |
"photo, deformed, black and white, realism, disfigured, low contrast, lowres, (bad), text, error, fewer, extra, missing, worst quality, jpeg artifacts, low quality, watermark, unfinished, displeasing, oldest, early, chromatic aberration, signature, extra digits, artistic error, username, scan, [abstract]",
|
| 1735 |
1,
|
| 1736 |
40,
|
|
@@ -1743,17 +1743,17 @@ EXAMPLES_GUI = [
|
|
| 1743 |
"Raelina/Rae-Diffusion-XL-V2",
|
| 1744 |
],
|
| 1745 |
[
|
| 1746 |
-
"1girl, akaza akari, yuru yuri, official art, anime
|
| 1747 |
-
"
|
| 1748 |
1,
|
| 1749 |
-
|
| 1750 |
-
|
| 1751 |
True,
|
| 1752 |
-1,
|
| 1753 |
"Euler",
|
| 1754 |
1024,
|
| 1755 |
1024,
|
| 1756 |
-
"Raelina/
|
| 1757 |
],
|
| 1758 |
[
|
| 1759 |
"yoshida yuuko, machikado mazoku, 1girl, solo, demon horns,horns, school uniform, long hair, open mouth, skirt, demon girl, ahoge, shiny, shiny hair, anime artwork",
|
|
|
|
| 1730 |
"votepurchase/ponyDiffusionV6XL",
|
| 1731 |
],
|
| 1732 |
[
|
| 1733 |
+
"1girl, oomuro sakurako, yuru yuri, official art, school uniform, anime artwork, anime style, studio anime, highly detailed, masterpiece, best quality, very aesthetic, absurdres",
|
| 1734 |
"photo, deformed, black and white, realism, disfigured, low contrast, lowres, (bad), text, error, fewer, extra, missing, worst quality, jpeg artifacts, low quality, watermark, unfinished, displeasing, oldest, early, chromatic aberration, signature, extra digits, artistic error, username, scan, [abstract]",
|
| 1735 |
1,
|
| 1736 |
40,
|
|
|
|
| 1743 |
"Raelina/Rae-Diffusion-XL-V2",
|
| 1744 |
],
|
| 1745 |
[
|
| 1746 |
+
"1girl, akaza akari, yuru yuri, official art, anime screencap, anime coloring, masterpiece, best quality, absurdres",
|
| 1747 |
+
"bad quality, worst quality, poorly drawn, sketch, multiple views, bad anatomy, bad hands, missing fingers, extra fingers, extra digits, fewer digits, signature, watermark, username",
|
| 1748 |
1,
|
| 1749 |
+
28,
|
| 1750 |
+
5.5,
|
| 1751 |
True,
|
| 1752 |
-1,
|
| 1753 |
"Euler",
|
| 1754 |
1024,
|
| 1755 |
1024,
|
| 1756 |
+
"Raelina/Raehoshi-illust-XL-6",
|
| 1757 |
],
|
| 1758 |
[
|
| 1759 |
"yoshida yuuko, machikado mazoku, 1girl, solo, demon horns,horns, school uniform, long hair, open mouth, skirt, demon girl, ahoge, shiny, shiny hair, anime artwork",
|
packages.txt
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
git-lfs aria2
|
|
|
|
| 1 |
+
git-lfs aria2 ffmpeg
|
requirements.txt
CHANGED
|
@@ -1,10 +1,8 @@
|
|
| 1 |
stablepy==0.6.2
|
| 2 |
-
|
| 3 |
-
|
| 4 |
accelerate
|
| 5 |
-
|
| 6 |
-
datasets
|
| 7 |
-
torch==2.4.0
|
| 8 |
numpy<2
|
| 9 |
gdown
|
| 10 |
opencv-python
|
|
@@ -12,17 +10,19 @@ huggingface_hub
|
|
| 12 |
hf_xet
|
| 13 |
hf_transfer
|
| 14 |
scikit-build-core
|
| 15 |
-
https://github.com/abetlen/llama-cpp-python/releases/download/v0.3.
|
| 16 |
-
git+https://github.com/
|
| 17 |
pybind11>=2.12
|
| 18 |
rapidfuzz
|
| 19 |
torchvision
|
| 20 |
optimum[onnxruntime]
|
| 21 |
-
dartrs
|
|
|
|
| 22 |
translatepy
|
| 23 |
timm
|
| 24 |
wrapt-timeout-decorator
|
| 25 |
sentencepiece
|
| 26 |
unidecode
|
| 27 |
-
|
|
|
|
| 28 |
pydantic==2.10.6
|
|
|
|
| 1 |
stablepy==0.6.2
|
| 2 |
+
diffusers
|
| 3 |
+
transformers<=4.48.3
|
| 4 |
accelerate
|
| 5 |
+
torch==2.5.1
|
|
|
|
|
|
|
| 6 |
numpy<2
|
| 7 |
gdown
|
| 8 |
opencv-python
|
|
|
|
| 10 |
hf_xet
|
| 11 |
hf_transfer
|
| 12 |
scikit-build-core
|
| 13 |
+
https://github.com/abetlen/llama-cpp-python/releases/download/v0.3.16-cu124/llama_cpp_python-0.3.16-cp310-cp310-linux_x86_64.whl
|
| 14 |
+
git+https://github.com/John6666cat/llama-cpp-agent
|
| 15 |
pybind11>=2.12
|
| 16 |
rapidfuzz
|
| 17 |
torchvision
|
| 18 |
optimum[onnxruntime]
|
| 19 |
+
#dartrs
|
| 20 |
+
git+https://github.com/John6666cat/dartrs
|
| 21 |
translatepy
|
| 22 |
timm
|
| 23 |
wrapt-timeout-decorator
|
| 24 |
sentencepiece
|
| 25 |
unidecode
|
| 26 |
+
matplotlib-inline
|
| 27 |
+
https://github.com/mjun0812/flash-attention-prebuild-wheels/releases/download/v0.4.11/flash_attn-2.8.3+cu124torch2.5-cp310-cp310-linux_x86_64.whl
|
| 28 |
pydantic==2.10.6
|
tagger/fl2sd3longcap.py
CHANGED
|
@@ -1,16 +1,22 @@
|
|
| 1 |
-
from transformers import AutoProcessor, AutoModelForCausalLM
|
| 2 |
import spaces
|
|
|
|
| 3 |
import re
|
| 4 |
from PIL import Image
|
| 5 |
import torch
|
| 6 |
|
| 7 |
-
import
|
| 8 |
-
|
|
|
|
|
|
|
| 9 |
|
| 10 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
| 11 |
-
fl_model = AutoModelForCausalLM.from_pretrained('gokaygokay/Florence-2-SD3-Captioner', trust_remote_code=True).to("cpu").eval()
|
| 12 |
-
fl_processor = AutoProcessor.from_pretrained('gokaygokay/Florence-2-SD3-Captioner', trust_remote_code=True)
|
| 13 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 14 |
|
| 15 |
def fl_modify_caption(caption: str) -> str:
|
| 16 |
"""
|
|
@@ -41,7 +47,7 @@ def fl_modify_caption(caption: str) -> str:
|
|
| 41 |
return modified_caption if modified_caption != caption else caption
|
| 42 |
|
| 43 |
|
| 44 |
-
@spaces.GPU(duration=
|
| 45 |
def fl_run_example(image):
|
| 46 |
task_prompt = "<DESCRIPTION>"
|
| 47 |
prompt = task_prompt + "Describe this image in great detail."
|
|
|
|
|
|
|
| 1 |
import spaces
|
| 2 |
+
from transformers import AutoProcessor, AutoModelForCausalLM
|
| 3 |
import re
|
| 4 |
from PIL import Image
|
| 5 |
import torch
|
| 6 |
|
| 7 |
+
from transformers.utils import is_flash_attn_2_available
|
| 8 |
+
if not is_flash_attn_2_available():
|
| 9 |
+
import subprocess
|
| 10 |
+
subprocess.run('pip install flash-attn --no-build-isolation', env={'FLASH_ATTENTION_SKIP_CUDA_BUILD': "TRUE"}, shell=True)
|
| 11 |
|
| 12 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
|
|
|
|
|
|
| 13 |
|
| 14 |
+
try:
|
| 15 |
+
fl_model = AutoModelForCausalLM.from_pretrained('gokaygokay/Florence-2-SD3-Captioner', trust_remote_code=True).to("cpu").eval()
|
| 16 |
+
fl_processor = AutoProcessor.from_pretrained('gokaygokay/Florence-2-SD3-Captioner', trust_remote_code=True)
|
| 17 |
+
except Exception as e:
|
| 18 |
+
print(e)
|
| 19 |
+
fl_model = fl_processor = None
|
| 20 |
|
| 21 |
def fl_modify_caption(caption: str) -> str:
|
| 22 |
"""
|
|
|
|
| 47 |
return modified_caption if modified_caption != caption else caption
|
| 48 |
|
| 49 |
|
| 50 |
+
@spaces.GPU(duration=10)
|
| 51 |
def fl_run_example(image):
|
| 52 |
task_prompt = "<DESCRIPTION>"
|
| 53 |
prompt = task_prompt + "Describe this image in great detail."
|
tagger/tagger.py
CHANGED
|
@@ -504,7 +504,7 @@ def gen_prompt(rating: list[str], character: list[str], general: list[str]):
|
|
| 504 |
return ", ".join(all_tags)
|
| 505 |
|
| 506 |
|
| 507 |
-
@spaces.GPU(duration=
|
| 508 |
def predict_tags(image: Image.Image, general_threshold: float = 0.3, character_threshold: float = 0.8):
|
| 509 |
inputs = wd_processor.preprocess(image, return_tensors="pt")
|
| 510 |
|
|
|
|
| 504 |
return ", ".join(all_tags)
|
| 505 |
|
| 506 |
|
| 507 |
+
@spaces.GPU(duration=10)
|
| 508 |
def predict_tags(image: Image.Image, general_threshold: float = 0.3, character_threshold: float = 0.8):
|
| 509 |
inputs = wd_processor.preprocess(image, return_tensors="pt")
|
| 510 |
|
tagger/v2.py
CHANGED
|
@@ -1,4 +1,4 @@
|
|
| 1 |
-
import time
|
| 2 |
import torch
|
| 3 |
from typing import Callable
|
| 4 |
from pathlib import Path
|
|
@@ -16,16 +16,15 @@ from dartrs.v2 import (
|
|
| 16 |
from dartrs.dartrs import DartTokenizer
|
| 17 |
from dartrs.utils import get_generation_config
|
| 18 |
|
| 19 |
-
|
| 20 |
import gradio as gr
|
| 21 |
from gradio.components import Component
|
| 22 |
|
| 23 |
-
|
| 24 |
try:
|
| 25 |
from output import UpsamplingOutput
|
| 26 |
except:
|
| 27 |
from .output import UpsamplingOutput
|
| 28 |
|
|
|
|
| 29 |
|
| 30 |
V2_ALL_MODELS = {
|
| 31 |
"dart-v2-moe-sft": {
|
|
@@ -43,8 +42,8 @@ V2_ALL_MODELS = {
|
|
| 43 |
|
| 44 |
def prepare_models(model_config: dict):
|
| 45 |
model_name = model_config["repo"]
|
| 46 |
-
tokenizer = DartTokenizer.from_pretrained(model_name)
|
| 47 |
-
model = model_config["class"].from_pretrained(model_name)
|
| 48 |
|
| 49 |
return {
|
| 50 |
"tokenizer": tokenizer,
|
|
|
|
| 1 |
+
import time, os
|
| 2 |
import torch
|
| 3 |
from typing import Callable
|
| 4 |
from pathlib import Path
|
|
|
|
| 16 |
from dartrs.dartrs import DartTokenizer
|
| 17 |
from dartrs.utils import get_generation_config
|
| 18 |
|
|
|
|
| 19 |
import gradio as gr
|
| 20 |
from gradio.components import Component
|
| 21 |
|
|
|
|
| 22 |
try:
|
| 23 |
from output import UpsamplingOutput
|
| 24 |
except:
|
| 25 |
from .output import UpsamplingOutput
|
| 26 |
|
| 27 |
+
HF_TOKEN = os.getenv("HF_TOKEN", None)
|
| 28 |
|
| 29 |
V2_ALL_MODELS = {
|
| 30 |
"dart-v2-moe-sft": {
|
|
|
|
| 42 |
|
| 43 |
def prepare_models(model_config: dict):
|
| 44 |
model_name = model_config["repo"]
|
| 45 |
+
tokenizer = DartTokenizer.from_pretrained(model_name, auth_token=HF_TOKEN)
|
| 46 |
+
model = model_config["class"].from_pretrained(model_name, auth_token=HF_TOKEN)
|
| 47 |
|
| 48 |
return {
|
| 49 |
"tokenizer": tokenizer,
|
utils.py
CHANGED
|
@@ -9,6 +9,7 @@ from constants import (
|
|
| 9 |
DIRECTORY_LORAS,
|
| 10 |
DIRECTORY_MODELS,
|
| 11 |
DIFFUSECRAFT_CHECKPOINT_NAME,
|
|
|
|
| 12 |
CACHE_HF,
|
| 13 |
STORAGE_ROOT,
|
| 14 |
)
|
|
@@ -28,6 +29,7 @@ from urllib3.util import Retry
|
|
| 28 |
import shutil
|
| 29 |
import subprocess
|
| 30 |
|
|
|
|
| 31 |
USER_AGENT = 'Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:127.0) Gecko/20100101 Firefox/127.0'
|
| 32 |
|
| 33 |
|
|
@@ -66,7 +68,8 @@ class ModelInformation:
|
|
| 66 |
)
|
| 67 |
self.filename_url = self.filename_url if self.filename_url else ""
|
| 68 |
self.description = json_data.get("description", "")
|
| 69 |
-
if self.description is None:
|
|
|
|
| 70 |
self.model_name = json_data.get("model", {}).get("name", "")
|
| 71 |
self.model_type = json_data.get("model", {}).get("type", "")
|
| 72 |
self.nsfw = json_data.get("model", {}).get("nsfw", False)
|
|
@@ -76,118 +79,175 @@ class ModelInformation:
|
|
| 76 |
self.original_json = copy.deepcopy(json_data)
|
| 77 |
|
| 78 |
|
| 79 |
-
def
|
| 80 |
-
|
| 81 |
-
|
| 82 |
-
|
| 83 |
-
|
| 84 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 85 |
|
| 86 |
|
| 87 |
-
def
|
| 88 |
-
|
| 89 |
-
downloaded_file_path = None
|
| 90 |
|
| 91 |
-
if
|
| 92 |
-
|
| 93 |
-
os.
|
| 94 |
-
|
| 95 |
-
os.chdir(original_dir)
|
| 96 |
-
elif "huggingface.co" in url:
|
| 97 |
-
url = url.replace("?download=true", "")
|
| 98 |
-
# url = urllib.parse.quote(url, safe=':/') # fix encoding
|
| 99 |
-
if "/blob/" in url:
|
| 100 |
-
url = url.replace("/blob/", "/resolve/")
|
| 101 |
-
user_header = f'"Authorization: Bearer {hf_token}"'
|
| 102 |
|
| 103 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 104 |
|
| 105 |
-
|
| 106 |
-
os.system(f"aria2c --console-log-level=error --summary-interval=10 --header={user_header} -c -x 16 -k 1M -s 16 {url} -d {directory} -o {filename}")
|
| 107 |
-
else:
|
| 108 |
-
os.system(f"aria2c --optimize-concurrent-downloads --console-log-level=error --summary-interval=10 -c -x 16 -k 1M -s 16 {url} -d {directory} -o {filename}")
|
| 109 |
|
| 110 |
-
|
|
|
|
| 111 |
|
| 112 |
-
|
|
|
|
| 113 |
|
| 114 |
-
|
| 115 |
-
|
| 116 |
-
|
| 117 |
-
|
| 118 |
-
|
| 119 |
-
|
| 120 |
-
and model_profile.download_url
|
| 121 |
-
and model_profile.filename_url
|
| 122 |
-
):
|
| 123 |
-
url = model_profile.download_url
|
| 124 |
-
filename = unidecode(model_profile.filename_url) if romanize else model_profile.filename_url
|
| 125 |
-
else:
|
| 126 |
-
if "?" in url:
|
| 127 |
-
url = url.split("?")[0]
|
| 128 |
-
filename = ""
|
| 129 |
|
| 130 |
-
|
| 131 |
-
|
| 132 |
|
| 133 |
-
|
| 134 |
-
|
| 135 |
-
|
|
|
|
|
|
|
| 136 |
|
| 137 |
-
|
| 138 |
-
|
| 139 |
-
|
| 140 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 141 |
os.system(aria2_command)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 142 |
|
| 143 |
-
|
| 144 |
-
|
| 145 |
-
|
| 146 |
-
# # PLAN B
|
| 147 |
-
# # Follow the redirect to get the actual download URL
|
| 148 |
-
# curl_command = (
|
| 149 |
-
# f'curl -L -sI --connect-timeout 5 --max-time 5 '
|
| 150 |
-
# f'-H "Content-Type: application/json" '
|
| 151 |
-
# f'-H "Authorization: Bearer {civitai_api_key}" "{url}"'
|
| 152 |
-
# )
|
| 153 |
-
|
| 154 |
-
# headers = os.popen(curl_command).read()
|
| 155 |
-
|
| 156 |
-
# # Look for the redirected "Location" URL
|
| 157 |
-
# location_match = re.search(r'location: (.+)', headers, re.IGNORECASE)
|
| 158 |
-
|
| 159 |
-
# if location_match:
|
| 160 |
-
# redirect_url = location_match.group(1).strip()
|
| 161 |
-
|
| 162 |
-
# # Extract the filename from the redirect URL's "Content-Disposition"
|
| 163 |
-
# filename_match = re.search(r'filename%3D%22(.+?)%22', redirect_url)
|
| 164 |
-
# if filename_match:
|
| 165 |
-
# encoded_filename = filename_match.group(1)
|
| 166 |
-
# # Decode the URL-encoded filename
|
| 167 |
-
# decoded_filename = urllib.parse.unquote(encoded_filename)
|
| 168 |
-
|
| 169 |
-
# filename = unidecode(decoded_filename) if romanize else decoded_filename
|
| 170 |
-
# print(f"Filename: {filename}")
|
| 171 |
-
|
| 172 |
-
# aria2_command = (
|
| 173 |
-
# f'aria2c --console-log-level=error --summary-interval=10 -c -x 16 '
|
| 174 |
-
# f'-k 1M -s 16 -d "{directory}" -o "{filename}" "{redirect_url}"'
|
| 175 |
-
# )
|
| 176 |
-
# return_code = os.system(aria2_command)
|
| 177 |
-
|
| 178 |
-
# # if return_code != 0:
|
| 179 |
-
# # raise RuntimeError(f"Failed to download file: {filename}. Error code: {return_code}")
|
| 180 |
-
# downloaded_file_path = os.path.join(directory, filename)
|
| 181 |
-
# if not os.path.exists(downloaded_file_path):
|
| 182 |
-
# downloaded_file_path = None
|
| 183 |
-
|
| 184 |
-
# if not downloaded_file_path:
|
| 185 |
-
# # Old method
|
| 186 |
-
# if "?" in url:
|
| 187 |
-
# url = url.split("?")[0]
|
| 188 |
-
# url = url + f"?token={civitai_api_key}"
|
| 189 |
-
# os.system(f"aria2c --console-log-level=error --summary-interval=10 -c -x 16 -k 1M -s 16 -d {directory} {url}")
|
| 190 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 191 |
else:
|
| 192 |
os.system(f"aria2c --console-log-level=error --summary-interval=10 -c -x 16 -k 1M -s 16 -d {directory} {url}")
|
| 193 |
|
|
@@ -216,14 +276,15 @@ def extract_parameters(input_string):
|
|
| 216 |
if "Steps:" in input_string:
|
| 217 |
input_string = input_string.replace("Steps:", "Negative prompt: Steps:")
|
| 218 |
else:
|
| 219 |
-
|
|
|
|
|
|
|
| 220 |
parameters["prompt"] = input_string
|
| 221 |
return parameters
|
| 222 |
|
| 223 |
parm = input_string.split("Negative prompt:")
|
| 224 |
parameters["prompt"] = parm[0].strip()
|
| 225 |
if "Steps:" not in parm[1]:
|
| 226 |
-
print("Steps not detected")
|
| 227 |
parameters["neg_prompt"] = parm[1].strip()
|
| 228 |
return parameters
|
| 229 |
parm = parm[1].split("Steps:")
|
|
@@ -306,7 +367,8 @@ def get_model_type(repo_id: str):
|
|
| 306 |
model = api.model_info(repo_id=repo_id, timeout=5.0)
|
| 307 |
tags = model.tags
|
| 308 |
for tag in tags:
|
| 309 |
-
if tag in MODEL_TYPE_CLASS.keys():
|
|
|
|
| 310 |
|
| 311 |
except Exception:
|
| 312 |
return default
|
|
@@ -433,9 +495,9 @@ def get_folder_size_gb(folder_path):
|
|
| 433 |
return total_size_gb
|
| 434 |
|
| 435 |
|
| 436 |
-
def get_used_storage_gb():
|
| 437 |
try:
|
| 438 |
-
used_gb = get_folder_size_gb(
|
| 439 |
print(f"Used Storage: {used_gb:.2f} GB")
|
| 440 |
except Exception as e:
|
| 441 |
used_gb = 999
|
|
@@ -455,6 +517,21 @@ def delete_model(removal_candidate):
|
|
| 455 |
shutil.rmtree(diffusers_model)
|
| 456 |
|
| 457 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 458 |
def progress_step_bar(step, total):
|
| 459 |
# Calculate the percentage for the progress bar width
|
| 460 |
percentage = min(100, ((step / total) * 100))
|
|
|
|
| 9 |
DIRECTORY_LORAS,
|
| 10 |
DIRECTORY_MODELS,
|
| 11 |
DIFFUSECRAFT_CHECKPOINT_NAME,
|
| 12 |
+
CACHE_HF_ROOT,
|
| 13 |
CACHE_HF,
|
| 14 |
STORAGE_ROOT,
|
| 15 |
)
|
|
|
|
| 29 |
import shutil
|
| 30 |
import subprocess
|
| 31 |
|
| 32 |
+
IS_ZERO_GPU = bool(os.getenv("SPACES_ZERO_GPU"))
|
| 33 |
USER_AGENT = 'Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:127.0) Gecko/20100101 Firefox/127.0'
|
| 34 |
|
| 35 |
|
|
|
|
| 68 |
)
|
| 69 |
self.filename_url = self.filename_url if self.filename_url else ""
|
| 70 |
self.description = json_data.get("description", "")
|
| 71 |
+
if self.description is None:
|
| 72 |
+
self.description = ""
|
| 73 |
self.model_name = json_data.get("model", {}).get("name", "")
|
| 74 |
self.model_type = json_data.get("model", {}).get("type", "")
|
| 75 |
self.nsfw = json_data.get("model", {}).get("nsfw", False)
|
|
|
|
| 79 |
self.original_json = copy.deepcopy(json_data)
|
| 80 |
|
| 81 |
|
| 82 |
+
def get_civit_params(url):
|
| 83 |
+
try:
|
| 84 |
+
json_data = request_json_data(url)
|
| 85 |
+
mdc = ModelInformation(json_data)
|
| 86 |
+
if mdc.download_url and mdc.filename_url:
|
| 87 |
+
return mdc.download_url, mdc.filename_url, mdc.model_url
|
| 88 |
+
else:
|
| 89 |
+
ValueError("Invalid Civitai model URL")
|
| 90 |
+
except Exception as e:
|
| 91 |
+
print(f"Error retrieving Civitai metadata: {e} — fallback to direct download")
|
| 92 |
+
return url, None, None
|
| 93 |
|
| 94 |
|
| 95 |
+
def civ_redirect_down(url, dir_, civitai_api_key, romanize, alternative_name):
|
| 96 |
+
filename_base = filename = None
|
|
|
|
| 97 |
|
| 98 |
+
if alternative_name:
|
| 99 |
+
output_path = os.path.join(dir_, alternative_name)
|
| 100 |
+
if os.path.exists(output_path):
|
| 101 |
+
return output_path, alternative_name
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 102 |
|
| 103 |
+
# Follow the redirect to get the actual download URL
|
| 104 |
+
curl_command = (
|
| 105 |
+
f'curl -L -sI --connect-timeout 5 --max-time 5 '
|
| 106 |
+
f'-H "Content-Type: application/json" '
|
| 107 |
+
f'-H "Authorization: Bearer {civitai_api_key}" "{url}"'
|
| 108 |
+
)
|
| 109 |
|
| 110 |
+
headers = os.popen(curl_command).read()
|
|
|
|
|
|
|
|
|
|
| 111 |
|
| 112 |
+
# Look for the redirected "Location" URL
|
| 113 |
+
location_match = re.search(r'location: (.+)', headers, re.IGNORECASE)
|
| 114 |
|
| 115 |
+
if location_match:
|
| 116 |
+
redirect_url = location_match.group(1).strip()
|
| 117 |
|
| 118 |
+
# Extract the filename from the redirect URL's "Content-Disposition"
|
| 119 |
+
filename_match = re.search(r'filename%3D%22(.+?)%22', redirect_url)
|
| 120 |
+
if filename_match:
|
| 121 |
+
encoded_filename = filename_match.group(1)
|
| 122 |
+
# Decode the URL-encoded filename
|
| 123 |
+
decoded_filename = urllib.parse.unquote(encoded_filename)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 124 |
|
| 125 |
+
filename = unidecode(decoded_filename) if romanize else decoded_filename
|
| 126 |
+
# print(f"Filename redirect: {filename}")
|
| 127 |
|
| 128 |
+
filename_base = alternative_name if alternative_name else filename
|
| 129 |
+
if not filename_base:
|
| 130 |
+
return None, None
|
| 131 |
+
elif os.path.exists(os.path.join(dir_, filename_base)):
|
| 132 |
+
return os.path.join(dir_, filename_base), filename_base
|
| 133 |
|
| 134 |
+
aria2_command = (
|
| 135 |
+
f'aria2c --console-log-level=error --summary-interval=10 -c -x 16 '
|
| 136 |
+
f'-k 1M -s 16 -d "{dir_}" -o "{filename_base}" "{redirect_url}"'
|
| 137 |
+
)
|
| 138 |
+
r_code = os.system(aria2_command) # noqa
|
| 139 |
+
|
| 140 |
+
# if r_code != 0:
|
| 141 |
+
# raise RuntimeError(f"Failed to download file: {filename_base}. Error code: {r_code}")
|
| 142 |
+
|
| 143 |
+
output_path = os.path.join(dir_, filename_base)
|
| 144 |
+
if not os.path.exists(output_path):
|
| 145 |
+
return None, filename_base
|
| 146 |
+
|
| 147 |
+
return output_path, filename_base
|
| 148 |
+
|
| 149 |
+
|
| 150 |
+
def civ_api_down(url, dir_, civitai_api_key, civ_filename):
|
| 151 |
+
"""
|
| 152 |
+
This method is susceptible to being blocked because it generates a lot of temp redirect links with aria2c.
|
| 153 |
+
If an API key limit is reached, generating a new API key and using it can fix the issue.
|
| 154 |
+
"""
|
| 155 |
+
output_path = None
|
| 156 |
+
|
| 157 |
+
url_dl = url + f"?token={civitai_api_key}"
|
| 158 |
+
if not civ_filename:
|
| 159 |
+
aria2_command = f'aria2c -c -x 1 -s 1 -d "{dir_}" "{url_dl}"'
|
| 160 |
os.system(aria2_command)
|
| 161 |
+
else:
|
| 162 |
+
output_path = os.path.join(dir_, civ_filename)
|
| 163 |
+
if not os.path.exists(output_path):
|
| 164 |
+
aria2_command = (
|
| 165 |
+
f'aria2c --console-log-level=error --summary-interval=10 -c -x 16 '
|
| 166 |
+
f'-k 1M -s 16 -d "{dir_}" -o "{civ_filename}" "{url_dl}"'
|
| 167 |
+
)
|
| 168 |
+
os.system(aria2_command)
|
| 169 |
+
|
| 170 |
+
return output_path
|
| 171 |
+
|
| 172 |
+
|
| 173 |
+
def drive_down(url, dir_):
|
| 174 |
+
import gdown
|
| 175 |
+
|
| 176 |
+
output_path = None
|
| 177 |
|
| 178 |
+
drive_id, _ = gdown.parse_url.parse_url(url, warning=False)
|
| 179 |
+
dir_files = os.listdir(dir_)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 180 |
|
| 181 |
+
for dfile in dir_files:
|
| 182 |
+
if drive_id in dfile:
|
| 183 |
+
output_path = os.path.join(dir_, dfile)
|
| 184 |
+
break
|
| 185 |
+
|
| 186 |
+
if not output_path:
|
| 187 |
+
original_path = gdown.download(url, f"{dir_}/", fuzzy=True)
|
| 188 |
+
|
| 189 |
+
dir_name, base_name = os.path.split(original_path)
|
| 190 |
+
name, ext = base_name.rsplit(".", 1)
|
| 191 |
+
new_name = f"{name}_{drive_id}.{ext}"
|
| 192 |
+
output_path = os.path.join(dir_name, new_name)
|
| 193 |
+
|
| 194 |
+
os.rename(original_path, output_path)
|
| 195 |
+
|
| 196 |
+
return output_path
|
| 197 |
+
|
| 198 |
+
|
| 199 |
+
def hf_down(url, dir_, hf_token, romanize):
|
| 200 |
+
url = url.replace("?download=true", "")
|
| 201 |
+
# url = urllib.parse.quote(url, safe=':/') # fix encoding
|
| 202 |
+
|
| 203 |
+
filename = unidecode(url.split('/')[-1]) if romanize else url.split('/')[-1]
|
| 204 |
+
output_path = os.path.join(dir_, filename)
|
| 205 |
+
|
| 206 |
+
if os.path.exists(output_path):
|
| 207 |
+
return output_path
|
| 208 |
+
|
| 209 |
+
if "/blob/" in url:
|
| 210 |
+
url = url.replace("/blob/", "/resolve/")
|
| 211 |
+
|
| 212 |
+
if hf_token:
|
| 213 |
+
user_header = f'"Authorization: Bearer {hf_token}"'
|
| 214 |
+
os.system(f"aria2c --console-log-level=error --summary-interval=10 --header={user_header} -c -x 16 -k 1M -s 16 {url} -d {dir_} -o {filename}")
|
| 215 |
+
else:
|
| 216 |
+
os.system(f"aria2c --optimize-concurrent-downloads --console-log-level=error --summary-interval=10 -c -x 16 -k 1M -s 16 {url} -d {dir_} -o {filename}")
|
| 217 |
+
|
| 218 |
+
return output_path
|
| 219 |
+
|
| 220 |
+
|
| 221 |
+
def download_things(directory, url, hf_token="", civitai_api_key="", romanize=False):
|
| 222 |
+
url = url.strip()
|
| 223 |
+
downloaded_file_path = None
|
| 224 |
+
|
| 225 |
+
if "drive.google.com" in url:
|
| 226 |
+
downloaded_file_path = drive_down(url, directory)
|
| 227 |
+
elif "huggingface.co" in url:
|
| 228 |
+
downloaded_file_path = hf_down(url, directory, hf_token, romanize)
|
| 229 |
+
elif "civitai.com" in url:
|
| 230 |
+
if not civitai_api_key:
|
| 231 |
+
msg = "You need an API key to download Civitai models."
|
| 232 |
+
print(f"\033[91m{msg}\033[0m")
|
| 233 |
+
gr.Warning(msg)
|
| 234 |
+
return None
|
| 235 |
+
|
| 236 |
+
url, civ_filename, civ_page = get_civit_params(url)
|
| 237 |
+
if civ_page and not IS_ZERO_GPU:
|
| 238 |
+
print(f"\033[92mCivitai model: {civ_filename} [page: {civ_page}]\033[0m")
|
| 239 |
+
|
| 240 |
+
downloaded_file_path, civ_filename = civ_redirect_down(url, directory, civitai_api_key, romanize, civ_filename)
|
| 241 |
+
|
| 242 |
+
if not downloaded_file_path:
|
| 243 |
+
msg = (
|
| 244 |
+
"Download failed.\n"
|
| 245 |
+
"If this is due to an API limit, generating a new API key may resolve the issue.\n"
|
| 246 |
+
"Attempting to download using the old method..."
|
| 247 |
+
)
|
| 248 |
+
print(msg)
|
| 249 |
+
gr.Warning(msg)
|
| 250 |
+
downloaded_file_path = civ_api_down(url, directory, civitai_api_key, civ_filename)
|
| 251 |
else:
|
| 252 |
os.system(f"aria2c --console-log-level=error --summary-interval=10 -c -x 16 -k 1M -s 16 -d {directory} {url}")
|
| 253 |
|
|
|
|
| 276 |
if "Steps:" in input_string:
|
| 277 |
input_string = input_string.replace("Steps:", "Negative prompt: Steps:")
|
| 278 |
else:
|
| 279 |
+
msg = "Generation data is invalid."
|
| 280 |
+
gr.Warning(msg)
|
| 281 |
+
print(msg)
|
| 282 |
parameters["prompt"] = input_string
|
| 283 |
return parameters
|
| 284 |
|
| 285 |
parm = input_string.split("Negative prompt:")
|
| 286 |
parameters["prompt"] = parm[0].strip()
|
| 287 |
if "Steps:" not in parm[1]:
|
|
|
|
| 288 |
parameters["neg_prompt"] = parm[1].strip()
|
| 289 |
return parameters
|
| 290 |
parm = parm[1].split("Steps:")
|
|
|
|
| 367 |
model = api.model_info(repo_id=repo_id, timeout=5.0)
|
| 368 |
tags = model.tags
|
| 369 |
for tag in tags:
|
| 370 |
+
if tag in MODEL_TYPE_CLASS.keys():
|
| 371 |
+
return MODEL_TYPE_CLASS.get(tag, default)
|
| 372 |
|
| 373 |
except Exception:
|
| 374 |
return default
|
|
|
|
| 495 |
return total_size_gb
|
| 496 |
|
| 497 |
|
| 498 |
+
def get_used_storage_gb(path_storage=STORAGE_ROOT):
|
| 499 |
try:
|
| 500 |
+
used_gb = get_folder_size_gb(path_storage)
|
| 501 |
print(f"Used Storage: {used_gb:.2f} GB")
|
| 502 |
except Exception as e:
|
| 503 |
used_gb = 999
|
|
|
|
| 517 |
shutil.rmtree(diffusers_model)
|
| 518 |
|
| 519 |
|
| 520 |
+
def clear_hf_cache():
|
| 521 |
+
"""
|
| 522 |
+
Clears the entire Hugging Face cache at ~/.cache/huggingface.
|
| 523 |
+
Hugging Face will re-download models as needed later.
|
| 524 |
+
"""
|
| 525 |
+
try:
|
| 526 |
+
if os.path.exists(CACHE_HF_ROOT):
|
| 527 |
+
shutil.rmtree(CACHE_HF_ROOT, ignore_errors=True)
|
| 528 |
+
print(f"Hugging Face cache cleared: {CACHE_HF_ROOT}")
|
| 529 |
+
else:
|
| 530 |
+
print(f"No Hugging Face cache found at: {CACHE_HF_ROOT}")
|
| 531 |
+
except Exception as e:
|
| 532 |
+
print(f"Error clearing Hugging Face cache: {e}")
|
| 533 |
+
|
| 534 |
+
|
| 535 |
def progress_step_bar(step, total):
|
| 536 |
# Calculate the percentage for the progress bar width
|
| 537 |
percentage = min(100, ((step / total) * 100))
|