John6666 commited on
Commit
3b2dce0
·
verified ·
1 Parent(s): 8771e29

Upload 24 files

Browse files
Files changed (16) hide show
  1. README.md +1 -1
  2. app.py +12 -9
  3. constants.py +49 -23
  4. dc.py +79 -44
  5. env.py +1 -0
  6. image_processor.py +2 -2
  7. llmdolphin.py +44 -46
  8. llmenv.py +742 -0
  9. lora_dict.json +14 -0
  10. modutils.py +6 -6
  11. packages.txt +1 -1
  12. requirements.txt +9 -9
  13. tagger/fl2sd3longcap.py +12 -6
  14. tagger/tagger.py +1 -1
  15. tagger/v2.py +4 -5
  16. utils.py +181 -104
README.md CHANGED
@@ -4,7 +4,7 @@ emoji: 🖼🖼️📦
4
  colorFrom: purple
5
  colorTo: red
6
  sdk: gradio
7
- sdk_version: 4.41.0
8
  app_file: app.py
9
  license: mit
10
  short_description: Text-to-Image
 
4
  colorFrom: purple
5
  colorTo: red
6
  sdk: gradio
7
+ sdk_version: 5.44.1
8
  app_file: app.py
9
  license: mit
10
  short_description: Text-to-Image
app.py CHANGED
@@ -1,6 +1,7 @@
1
  import spaces
2
  import gradio as gr
3
  import numpy as np
 
4
 
5
  # DiffuseCraft
6
  from dc import (infer, _infer, pass_result, get_diffusers_model_list, get_samplers, save_image_history,
@@ -35,7 +36,7 @@ def description_ui():
35
  """
36
  )
37
 
38
-
39
  MAX_SEED = np.iinfo(np.int32).max
40
  MAX_IMAGE_SIZE = 4096
41
  MIN_IMAGE_SIZE = 256
@@ -51,7 +52,7 @@ css = """
51
  .image { margin: 0px auto; object-fit: contain; }
52
  """
53
 
54
- with gr.Blocks(fill_width=True, elem_id="container", css=css, delete_cache=(60, 3600)) as demo:
55
  gr.Markdown("# Votepurchase Multiple Model", elem_classes="title")
56
  state = gr.State(value={})
57
  with gr.Tab("Image Generator"):
@@ -124,7 +125,8 @@ with gr.Blocks(fill_width=True, elem_id="container", css=css, delete_cache=(60,
124
  return gr.Dropdown(label=label, choices=get_all_lora_tupled_list(), value="", allow_custom_value=True, elem_classes="lora", min_width=320, visible=visible)
125
 
126
  def lora_scale_slider(label, visible=True):
127
- return gr.Slider(minimum=-2, maximum=2, step=0.01, value=1.00, label=label, visible=visible)
 
128
 
129
  def lora_textbox():
130
  return gr.Textbox(label="", info="Example of prompt:", value="", show_copy_button=True, interactive=False, visible=False)
@@ -233,8 +235,8 @@ with gr.Blocks(fill_width=True, elem_id="container", css=css, delete_cache=(60,
233
  low_threshold = gr.Slider(minimum=1, maximum=255, step=1, value=100, label="'CANNY' low threshold")
234
  high_threshold = gr.Slider(minimum=1, maximum=255, step=1, value=200, label="'CANNY' high threshold")
235
  with gr.Row():
236
- value_threshold = gr.Slider(minimum=1, maximum=2.0, step=0.01, value=0.1, label="'MLSD' Hough value threshold")
237
- distance_threshold = gr.Slider(minimum=1, maximum=20.0, step=0.01, value=0.1, label="'MLSD' Hough distance threshold")
238
  recolor_gamma_correction = gr.Number(minimum=0., maximum=25., value=1., step=0.001, label="'RECOLOR' gamma correction")
239
  tile_blur_sigma = gr.Number(minimum=0, maximum=100, value=9, step=1, label="'TILE' blur sigma")
240
 
@@ -295,7 +297,7 @@ with gr.Blocks(fill_width=True, elem_id="container", css=css, delete_cache=(60,
295
  with gr.Row():
296
  upscaler_model_path = gr.Dropdown(label="Upscaler", choices=UPSCALER_KEYS, value=UPSCALER_KEYS[0])
297
  upscaler_increases_size = gr.Slider(minimum=1.1, maximum=4., step=0.1, value=1.2, label="Upscale by")
298
- upscaler_tile_size = gr.Slider(minimum=0, maximum=512, step=16, value=0, label="Upscaler Tile Size", info="0 = no tiling")
299
  upscaler_tile_overlap = gr.Slider(minimum=0, maximum=48, step=1, value=8, label="Upscaler Tile Overlap")
300
  with gr.Row():
301
  hires_steps = gr.Slider(minimum=0, value=30, maximum=100, step=1, label="Hires Steps")
@@ -351,7 +353,7 @@ with gr.Blocks(fill_width=True, elem_id="container", css=css, delete_cache=(60,
351
  face_restoration_weight = gr.Slider(minimum=0., maximum=1., step=0.001, value=.5, label="Weight", info="(0 = maximum effect, 1 = minimum effect)")
352
 
353
  with gr.Tab("Translation Settings"):
354
- chatbot = gr.Chatbot(render_markdown=False, visible=False) # component for auto-translation
355
  chat_model = gr.Dropdown(choices=get_dolphin_models(), value=get_dolphin_models()[0], allow_custom_value=True, label="Model")
356
  chat_model_info = gr.Markdown(value=get_dolphin_model_info(get_dolphin_models()[0]), label="Model info")
357
  chat_format = gr.Dropdown(choices=get_llm_formats(), value=get_dolphin_model_format(get_dolphin_models()[0]), label="Message format")
@@ -682,5 +684,6 @@ with gr.Blocks(fill_width=True, elem_id="container", css=css, delete_cache=(60,
682
  gr.LoginButton()
683
  gr.DuplicateButton(value="Duplicate Space for private use (This demo does not work on CPU. Requires GPU Space)")
684
 
685
- demo.queue()
686
- demo.launch(show_error=True, debug=True)
 
 
1
  import spaces
2
  import gradio as gr
3
  import numpy as np
4
+ import os
5
 
6
  # DiffuseCraft
7
  from dc import (infer, _infer, pass_result, get_diffusers_model_list, get_samplers, save_image_history,
 
36
  """
37
  )
38
 
39
+ IS_ZERO_GPU = bool(os.getenv("SPACES_ZERO_GPU"))
40
  MAX_SEED = np.iinfo(np.int32).max
41
  MAX_IMAGE_SIZE = 4096
42
  MIN_IMAGE_SIZE = 256
 
52
  .image { margin: 0px auto; object-fit: contain; }
53
  """
54
 
55
+ with gr.Blocks(fill_width=True, elem_id="container", css=css) as demo:
56
  gr.Markdown("# Votepurchase Multiple Model", elem_classes="title")
57
  state = gr.State(value={})
58
  with gr.Tab("Image Generator"):
 
125
  return gr.Dropdown(label=label, choices=get_all_lora_tupled_list(), value="", allow_custom_value=True, elem_classes="lora", min_width=320, visible=visible)
126
 
127
  def lora_scale_slider(label, visible=True):
128
+ val_lora = 8 if IS_ZERO_GPU else 8 #
129
+ return gr.Slider(minimum=-val_lora, maximum=val_lora, step=0.01, value=0.33, label=label, visible=visible)
130
 
131
  def lora_textbox():
132
  return gr.Textbox(label="", info="Example of prompt:", value="", show_copy_button=True, interactive=False, visible=False)
 
235
  low_threshold = gr.Slider(minimum=1, maximum=255, step=1, value=100, label="'CANNY' low threshold")
236
  high_threshold = gr.Slider(minimum=1, maximum=255, step=1, value=200, label="'CANNY' high threshold")
237
  with gr.Row():
238
+ value_threshold = gr.Slider(minimum=0.0, maximum=2.0, step=0.01, value=0.1, label="'MLSD' Hough value threshold")
239
+ distance_threshold = gr.Slider(minimum=0.0, maximum=20.0, step=0.01, value=0.1, label="'MLSD' Hough distance threshold")
240
  recolor_gamma_correction = gr.Number(minimum=0., maximum=25., value=1., step=0.001, label="'RECOLOR' gamma correction")
241
  tile_blur_sigma = gr.Number(minimum=0, maximum=100, value=9, step=1, label="'TILE' blur sigma")
242
 
 
297
  with gr.Row():
298
  upscaler_model_path = gr.Dropdown(label="Upscaler", choices=UPSCALER_KEYS, value=UPSCALER_KEYS[0])
299
  upscaler_increases_size = gr.Slider(minimum=1.1, maximum=4., step=0.1, value=1.2, label="Upscale by")
300
+ upscaler_tile_size = gr.Slider(minimum=0, maximum=512, step=16, value=(0 if IS_ZERO_GPU else 192), label="Upscaler Tile Size", info="0 = no tiling")
301
  upscaler_tile_overlap = gr.Slider(minimum=0, maximum=48, step=1, value=8, label="Upscaler Tile Overlap")
302
  with gr.Row():
303
  hires_steps = gr.Slider(minimum=0, value=30, maximum=100, step=1, label="Hires Steps")
 
353
  face_restoration_weight = gr.Slider(minimum=0., maximum=1., step=0.001, value=.5, label="Weight", info="(0 = maximum effect, 1 = minimum effect)")
354
 
355
  with gr.Tab("Translation Settings"):
356
+ chatbot = gr.Chatbot(type="messages", render_markdown=False, visible=False) # component for auto-translation
357
  chat_model = gr.Dropdown(choices=get_dolphin_models(), value=get_dolphin_models()[0], allow_custom_value=True, label="Model")
358
  chat_model_info = gr.Markdown(value=get_dolphin_model_info(get_dolphin_models()[0]), label="Model info")
359
  chat_format = gr.Dropdown(choices=get_llm_formats(), value=get_dolphin_model_format(get_dolphin_models()[0]), label="Message format")
 
684
  gr.LoginButton()
685
  gr.DuplicateButton(value="Duplicate Space for private use (This demo does not work on CPU. Requires GPU Space)")
686
 
687
+ if __name__ == "__main__":
688
+ demo.queue()
689
+ demo.launch(show_error=True, share=False, debug=True, ssr_mode=False)
constants.py CHANGED
@@ -9,6 +9,8 @@ from stablepy import (
9
  IP_ADAPTERS_SDXL,
10
  )
11
 
 
 
12
  # - **Download Models**
13
  DOWNLOAD_MODEL = "https://huggingface.co/TechnoByte/MilkyWonderland/resolve/main/milkyWonderland_v40.safetensors"
14
 
@@ -23,12 +25,12 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
23
  'Laxhar/noobai-XL-1.1',
24
  'Laxhar/noobai-XL-Vpred-1.0',
25
  'black-forest-labs/FLUX.1-dev',
 
26
  'John6666/blue-pencil-flux1-v021-fp8-flux',
27
  'John6666/wai-ani-flux-v10forfp8-fp8-flux',
28
  'John6666/xe-anime-flux-v04-fp8-flux',
29
  'John6666/lyh-anime-flux-v2a1-fp8-flux',
30
  'John6666/carnival-unchained-v10-fp8-flux',
31
- 'John6666/iniverse-mix-xl-sfwnsfw-fluxdfp16nsfwv11-fp8-flux',
32
  'Freepik/flux.1-lite-8B-alpha',
33
  'shauray/FluxDev-HyperSD-merged',
34
  'mikeyandfriends/PixelWave_FLUX.1-dev_03',
@@ -37,23 +39,19 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
37
  # 'ostris/OpenFLUX.1',
38
  'shuttleai/shuttle-3-diffusion',
39
  'Laxhar/noobai-XL-1.0',
40
- 'John6666/noobai-xl-nai-xl-epsilonpred10version-sdxl',
41
  'Laxhar/noobai-XL-0.77',
42
  'John6666/noobai-xl-nai-xl-epsilonpred075version-sdxl',
43
  'Laxhar/noobai-XL-0.6',
44
  'John6666/noobai-xl-nai-xl-epsilonpred05version-sdxl',
45
  'John6666/noobai-cyberfix-v10-sdxl',
46
  'John6666/noobaiiter-xl-vpred-v075-sdxl',
47
- 'John6666/ntr-mix-illustrious-xl-noob-xl-v40-sdxl',
48
- 'John6666/ntr-mix-illustrious-xl-noob-xl-ntrmix35-sdxl',
49
- 'John6666/ntr-mix-illustrious-xl-noob-xl-v777-sdxl',
50
- 'John6666/ntr-mix-illustrious-xl-noob-xl-v777forlora-sdxl',
51
  'John6666/ntr-mix-illustrious-xl-noob-xl-xi-sdxl',
52
  'John6666/ntr-mix-illustrious-xl-noob-xl-xii-sdxl',
53
  'John6666/ntr-mix-illustrious-xl-noob-xl-xiii-sdxl',
54
  'John6666/mistoon-anime-v10illustrious-sdxl',
55
- 'John6666/hassaku-xl-illustrious-v10-sdxl',
56
- 'John6666/hassaku-xl-illustrious-v10style-sdxl',
57
  'John6666/haruki-mix-illustrious-v10-sdxl',
58
  'John6666/noobreal-v10-sdxl',
59
  'John6666/complicated-noobai-merge-vprediction-sdxl',
@@ -64,6 +62,7 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
64
  'Laxhar/noobai-XL-Vpred-0.65',
65
  'Laxhar/noobai-XL-Vpred-0.6',
66
  'John6666/cat-tower-noobai-xl-checkpoint-v14vpred-sdxl',
 
67
  'John6666/noobai-xl-nai-xl-vpred05version-sdxl',
68
  'John6666/noobai-fusion2-vpred-itercomp-v1-sdxl',
69
  'John6666/noobai-xl-nai-xl-vpredtestversion-sdxl',
@@ -74,19 +73,34 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
74
  'John6666/obsession-illustriousxl-v21-sdxl',
75
  'John6666/obsession-illustriousxl-v30-sdxl',
76
  'John6666/obsession-illustriousxl-v31-sdxl',
 
 
 
 
 
77
  'John6666/wai-nsfw-illustrious-v70-sdxl',
 
78
  'John6666/illustrious-pony-mix-v3-sdxl',
79
- 'John6666/nova-anime-xl-illustriousv10-sdxl',
80
- 'John6666/nova-orange-xl-v30-sdxl',
 
 
 
 
 
 
 
81
  'John6666/silvermoon-mix03-illustrious-v10-sdxl',
82
  'eienmojiki/Anything-XL',
83
  'eienmojiki/Starry-XL-v5.2',
 
84
  'John6666/meinaxl-v2-sdxl',
85
  'Eugeoter/artiwaifu-diffusion-2.0',
86
  'comin/IterComp',
87
- 'John6666/epicrealism-xl-vxiabeast-sdxl',
88
- 'John6666/epicrealism-xl-v10kiss2-sdxl',
89
  'John6666/epicrealism-xl-v8kiss-sdxl',
 
 
 
90
  'misri/zavychromaxl_v80',
91
  'SG161222/RealVisXL_V4.0',
92
  'SG161222/RealVisXL_V5.0',
@@ -102,8 +116,10 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
102
  'John6666/ebara-mfcg-pony-mix-v12-sdxl',
103
  'John6666/t-ponynai3-v51-sdxl',
104
  'John6666/t-ponynai3-v65-sdxl',
 
105
  'John6666/prefect-pony-xl-v3-sdxl',
106
  'John6666/prefect-pony-xl-v4-sdxl',
 
107
  'John6666/mala-anime-mix-nsfw-pony-xl-v5-sdxl',
108
  'John6666/wai-ani-nsfw-ponyxl-v10-sdxl',
109
  'John6666/wai-real-mix-v11-sdxl',
@@ -111,13 +127,14 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
111
  'John6666/wai-c-v6-sdxl',
112
  'John6666/iniverse-mix-xl-sfwnsfw-pony-guofeng-v43-sdxl',
113
  'John6666/sifw-annihilation-xl-v2-sdxl',
 
114
  'John6666/photo-realistic-pony-v5-sdxl',
115
  'John6666/pony-realism-v21main-sdxl',
116
  'John6666/pony-realism-v22main-sdxl',
117
- 'John6666/cyberrealistic-pony-v63-sdxl',
118
- 'John6666/cyberrealistic-pony-v64-sdxl',
119
  'John6666/cyberrealistic-pony-v65-sdxl',
120
  'John6666/cyberrealistic-pony-v7-sdxl',
 
121
  'GraydientPlatformAPI/realcartoon-pony-diffusion',
122
  'John6666/nova-anime-xl-pony-v5-sdxl',
123
  'John6666/autismmix-sdxl-autismmix-pony-sdxl',
@@ -127,13 +144,15 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
127
  'John6666/duchaiten-pony-real-v11fix-sdxl',
128
  'John6666/duchaiten-pony-real-v20-sdxl',
129
  'John6666/duchaiten-pony-xl-no-score-v70-sdxl',
130
- 'Spestly/OdysseyXL-3.0',
131
- 'Spestly/OdysseyXL-4.0',
132
  'KBlueLeaf/Kohaku-XL-Zeta',
133
  'cagliostrolab/animagine-xl-3.1',
 
134
  'yodayo-ai/kivotos-xl-2.0',
135
  'yodayo-ai/holodayo-xl-2.1',
136
  'yodayo-ai/clandestine-xl-1.0',
 
 
 
137
  'digiplay/majicMIX_sombre_v2',
138
  'digiplay/majicMIX_realistic_v6',
139
  'digiplay/majicMIX_realistic_v7',
@@ -159,9 +178,9 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
159
  'GraydientPlatformAPI/realcartoon3d-17',
160
  'GraydientPlatformAPI/realcartoon-pixar11',
161
  'GraydientPlatformAPI/realcartoon-real17',
162
- 'nitrosocke/Ghibli-Diffusion',
163
  ]
164
 
 
165
  DIFFUSERS_FORMAT_LORAS = [
166
  "nerijs/animation2k-flux",
167
  "XLabs-AI/flux-RealismLora",
@@ -183,8 +202,11 @@ DIRECTORY_VAES = 'vaes'
183
  DIRECTORY_EMBEDS = 'embedings'
184
  DIRECTORY_UPSCALERS = 'upscalers'
185
 
186
- CACHE_HF = "/home/user/.cache/huggingface/hub/"
187
  STORAGE_ROOT = "/home/user/"
 
 
 
 
188
 
189
  TASK_STABLEPY = {
190
  'txt2img': 'txt2img',
@@ -226,6 +248,7 @@ UPSCALER_DICT_GUI = {
226
  # "realesr-general-x4v3": "https://github.com/xinntao/Real-ESRGAN/releases/download/v0.2.5.0/realesr-general-x4v3.pth",
227
  # "realesr-general-wdn-x4v3": "https://github.com/xinntao/Real-ESRGAN/releases/download/v0.2.5.0/realesr-general-wdn-x4v3.pth",
228
  "4x-UltraSharp": "https://huggingface.co/Shandypur/ESRGAN-4x-UltraSharp/resolve/main/4x-UltraSharp.pth",
 
229
  "4x_foolhardy_Remacri": "https://huggingface.co/FacehugmanIII/4x_foolhardy_Remacri/resolve/main/4x_foolhardy_Remacri.pth",
230
  "Remacri4xExtraSmoother": "https://huggingface.co/hollowstrawberry/upscalers-backup/resolve/main/ESRGAN/Remacri%204x%20ExtraSmoother.pth",
231
  "AnimeSharp4x": "https://huggingface.co/hollowstrawberry/upscalers-backup/resolve/main/ESRGAN/AnimeSharp%204x.pth",
@@ -359,9 +382,11 @@ SUBTITLE_GUI = (
359
  " to perform different tasks in image generation."
360
  )
361
 
 
 
362
  HELP_GUI = (
363
- """### Help:
364
- - The current space runs on a ZERO GPU which is assigned for approximately 60 seconds; Therefore, if you submit expensive tasks, the operation may be canceled upon reaching the maximum allowed time with 'GPU TASK ABORTED'.
365
  - Distorted or strange images often result from high prompt weights, so it's best to use low weights and scales, and consider using Classic variants like 'Classic-original'.
366
  - For better results with Pony Diffusion, try using sampler DPM++ 1s or DPM2 with Compel or Classic prompt weights.
367
  """
@@ -485,7 +510,7 @@ EXAMPLES_GUI = [
485
  20,
486
  4.0,
487
  -1,
488
- "loras/Coloring_book_-_LineArt.safetensors",
489
  1.0,
490
  "DPM++ 2M SDE",
491
  1024,
@@ -580,6 +605,7 @@ EXAMPLES_GUI = [
580
  RESOURCES = (
581
  """### Resources
582
  - John6666's space has some great features you might find helpful [link](https://huggingface.co/spaces/John6666/DiffuseCraftMod).
583
- - You can also try the image generator in Colab’s free tier, which provides free GPU [link](https://github.com/R3gm/SD_diffusers_interactive).
 
584
  """
585
- )
 
9
  IP_ADAPTERS_SDXL,
10
  )
11
 
12
+ IS_ZERO_GPU = bool(os.getenv("SPACES_ZERO_GPU"))
13
+
14
  # - **Download Models**
15
  DOWNLOAD_MODEL = "https://huggingface.co/TechnoByte/MilkyWonderland/resolve/main/milkyWonderland_v40.safetensors"
16
 
 
25
  'Laxhar/noobai-XL-1.1',
26
  'Laxhar/noobai-XL-Vpred-1.0',
27
  'black-forest-labs/FLUX.1-dev',
28
+ 'black-forest-labs/FLUX.1-Krea-dev',
29
  'John6666/blue-pencil-flux1-v021-fp8-flux',
30
  'John6666/wai-ani-flux-v10forfp8-fp8-flux',
31
  'John6666/xe-anime-flux-v04-fp8-flux',
32
  'John6666/lyh-anime-flux-v2a1-fp8-flux',
33
  'John6666/carnival-unchained-v10-fp8-flux',
 
34
  'Freepik/flux.1-lite-8B-alpha',
35
  'shauray/FluxDev-HyperSD-merged',
36
  'mikeyandfriends/PixelWave_FLUX.1-dev_03',
 
39
  # 'ostris/OpenFLUX.1',
40
  'shuttleai/shuttle-3-diffusion',
41
  'Laxhar/noobai-XL-1.0',
 
42
  'Laxhar/noobai-XL-0.77',
43
  'John6666/noobai-xl-nai-xl-epsilonpred075version-sdxl',
44
  'Laxhar/noobai-XL-0.6',
45
  'John6666/noobai-xl-nai-xl-epsilonpred05version-sdxl',
46
  'John6666/noobai-cyberfix-v10-sdxl',
47
  'John6666/noobaiiter-xl-vpred-v075-sdxl',
48
+ 'John6666/ripplemix-noob-vpred10-illustrious01-v14-sdxl',
49
+ 'John6666/sigmaih-15-sdxl',
 
 
50
  'John6666/ntr-mix-illustrious-xl-noob-xl-xi-sdxl',
51
  'John6666/ntr-mix-illustrious-xl-noob-xl-xii-sdxl',
52
  'John6666/ntr-mix-illustrious-xl-noob-xl-xiii-sdxl',
53
  'John6666/mistoon-anime-v10illustrious-sdxl',
54
+ 'John6666/hassaku-xl-illustrious-v22-sdxl',
 
55
  'John6666/haruki-mix-illustrious-v10-sdxl',
56
  'John6666/noobreal-v10-sdxl',
57
  'John6666/complicated-noobai-merge-vprediction-sdxl',
 
62
  'Laxhar/noobai-XL-Vpred-0.65',
63
  'Laxhar/noobai-XL-Vpred-0.6',
64
  'John6666/cat-tower-noobai-xl-checkpoint-v14vpred-sdxl',
65
+ 'John6666/cat-tower-noobai-xl-checkpoint-v15vpred-sdxl',
66
  'John6666/noobai-xl-nai-xl-vpred05version-sdxl',
67
  'John6666/noobai-fusion2-vpred-itercomp-v1-sdxl',
68
  'John6666/noobai-xl-nai-xl-vpredtestversion-sdxl',
 
73
  'John6666/obsession-illustriousxl-v21-sdxl',
74
  'John6666/obsession-illustriousxl-v30-sdxl',
75
  'John6666/obsession-illustriousxl-v31-sdxl',
76
+ 'John6666/one-obsession-13-sdxl',
77
+ 'John6666/one-obsession-14-24d-sdxl',
78
+ 'John6666/one-obsession-15-noobai-sdxl',
79
+ 'John6666/one-obsession-v16-noobai-sdxl',
80
+ 'John6666/prefect-illustrious-xl-v3-sdxl',
81
  'John6666/wai-nsfw-illustrious-v70-sdxl',
82
+ 'John6666/wai-nsfw-illustrious-sdxl-v140-sdxl',
83
  'John6666/illustrious-pony-mix-v3-sdxl',
84
+ 'John6666/nova-anime-xl-il-v90-sdxl',
85
+ 'John6666/nova-anime-xl-il-v110-sdxl',
86
+ 'John6666/nova-orange-xl-re-v10-sdxl',
87
+ 'John6666/nova-orange-xl-v110-sdxl',
88
+ 'John6666/nova-orange-xl-re-v20-sdxl',
89
+ 'John6666/nova-unreal-xl-v60-sdxl',
90
+ 'John6666/nova-unreal-xl-v70-sdxl',
91
+ 'John6666/nova-unreal-xl-v80-sdxl',
92
+ 'John6666/nova-cartoon-xl-v40-sdxl',
93
  'John6666/silvermoon-mix03-illustrious-v10-sdxl',
94
  'eienmojiki/Anything-XL',
95
  'eienmojiki/Starry-XL-v5.2',
96
+ 'votepurchase/plantMilkModelSuite_walnut',
97
  'John6666/meinaxl-v2-sdxl',
98
  'Eugeoter/artiwaifu-diffusion-2.0',
99
  'comin/IterComp',
 
 
100
  'John6666/epicrealism-xl-v8kiss-sdxl',
101
+ 'John6666/epicrealism-xl-v10kiss2-sdxl',
102
+ 'John6666/epicrealism-xl-vxiabeast-sdxl',
103
+ 'John6666/epicrealism-xl-vxvii-crystal-clear-realism-sdxl',
104
  'misri/zavychromaxl_v80',
105
  'SG161222/RealVisXL_V4.0',
106
  'SG161222/RealVisXL_V5.0',
 
116
  'John6666/ebara-mfcg-pony-mix-v12-sdxl',
117
  'John6666/t-ponynai3-v51-sdxl',
118
  'John6666/t-ponynai3-v65-sdxl',
119
+ 'John6666/t-ponynai3-v7-sdxl',
120
  'John6666/prefect-pony-xl-v3-sdxl',
121
  'John6666/prefect-pony-xl-v4-sdxl',
122
+ 'John6666/prefect-pony-xl-v50-sdxl',
123
  'John6666/mala-anime-mix-nsfw-pony-xl-v5-sdxl',
124
  'John6666/wai-ani-nsfw-ponyxl-v10-sdxl',
125
  'John6666/wai-real-mix-v11-sdxl',
 
127
  'John6666/wai-c-v6-sdxl',
128
  'John6666/iniverse-mix-xl-sfwnsfw-pony-guofeng-v43-sdxl',
129
  'John6666/sifw-annihilation-xl-v2-sdxl',
130
+ 'John6666/sifw-annihilation-xl-v305illustrious-beta-sdxl',
131
  'John6666/photo-realistic-pony-v5-sdxl',
132
  'John6666/pony-realism-v21main-sdxl',
133
  'John6666/pony-realism-v22main-sdxl',
134
+ 'John6666/pony-realism-v23-ultra-sdxl',
 
135
  'John6666/cyberrealistic-pony-v65-sdxl',
136
  'John6666/cyberrealistic-pony-v7-sdxl',
137
+ 'John6666/cyberrealistic-pony-v127-alternative-sdxl',
138
  'GraydientPlatformAPI/realcartoon-pony-diffusion',
139
  'John6666/nova-anime-xl-pony-v5-sdxl',
140
  'John6666/autismmix-sdxl-autismmix-pony-sdxl',
 
144
  'John6666/duchaiten-pony-real-v11fix-sdxl',
145
  'John6666/duchaiten-pony-real-v20-sdxl',
146
  'John6666/duchaiten-pony-xl-no-score-v70-sdxl',
 
 
147
  'KBlueLeaf/Kohaku-XL-Zeta',
148
  'cagliostrolab/animagine-xl-3.1',
149
+ 'cagliostrolab/animagine-xl-4.0',
150
  'yodayo-ai/kivotos-xl-2.0',
151
  'yodayo-ai/holodayo-xl-2.1',
152
  'yodayo-ai/clandestine-xl-1.0',
153
+ 'https://huggingface.co/chemwolf/Karmix-XL-v0/resolve/main/Karmix-XL-v0.safetensors?download=true',
154
+ 'https://civitai.com/api/download/models/128713?type=Model&format=SafeTensor&size=pruned&fp=fp16',
155
+ 'https://civitai.com/models/30240?modelVersionId=125771',
156
  'digiplay/majicMIX_sombre_v2',
157
  'digiplay/majicMIX_realistic_v6',
158
  'digiplay/majicMIX_realistic_v7',
 
178
  'GraydientPlatformAPI/realcartoon3d-17',
179
  'GraydientPlatformAPI/realcartoon-pixar11',
180
  'GraydientPlatformAPI/realcartoon-real17',
 
181
  ]
182
 
183
+
184
  DIFFUSERS_FORMAT_LORAS = [
185
  "nerijs/animation2k-flux",
186
  "XLabs-AI/flux-RealismLora",
 
202
  DIRECTORY_EMBEDS = 'embedings'
203
  DIRECTORY_UPSCALERS = 'upscalers'
204
 
 
205
  STORAGE_ROOT = "/home/user/"
206
+ CACHE_HF_ROOT = os.path.expanduser(os.getenv("HF_HOME")) if os.getenv("HF_HOME") else os.path.expanduser("~/cache/huggingface")
207
+ CACHE_HF = os.path.join(CACHE_HF_ROOT, "hub")
208
+ if IS_ZERO_GPU:
209
+ os.environ["HF_HOME"] = CACHE_HF
210
 
211
  TASK_STABLEPY = {
212
  'txt2img': 'txt2img',
 
248
  # "realesr-general-x4v3": "https://github.com/xinntao/Real-ESRGAN/releases/download/v0.2.5.0/realesr-general-x4v3.pth",
249
  # "realesr-general-wdn-x4v3": "https://github.com/xinntao/Real-ESRGAN/releases/download/v0.2.5.0/realesr-general-wdn-x4v3.pth",
250
  "4x-UltraSharp": "https://huggingface.co/Shandypur/ESRGAN-4x-UltraSharp/resolve/main/4x-UltraSharp.pth",
251
+ "Real-ESRGAN-Anime-finetuning": "https://huggingface.co/danhtran2mind/Real-ESRGAN-Anime-finetuning/resolve/main/Real-ESRGAN-Anime-finetuning.pth",
252
  "4x_foolhardy_Remacri": "https://huggingface.co/FacehugmanIII/4x_foolhardy_Remacri/resolve/main/4x_foolhardy_Remacri.pth",
253
  "Remacri4xExtraSmoother": "https://huggingface.co/hollowstrawberry/upscalers-backup/resolve/main/ESRGAN/Remacri%204x%20ExtraSmoother.pth",
254
  "AnimeSharp4x": "https://huggingface.co/hollowstrawberry/upscalers-backup/resolve/main/ESRGAN/AnimeSharp%204x.pth",
 
382
  " to perform different tasks in image generation."
383
  )
384
 
385
+ msg_zero = "" if not IS_ZERO_GPU else "- The current space runs on a ZERO GPU which is assigned for approximately 60 seconds; Therefore, if you submit expensive tasks, the operation may be canceled upon reaching the maximum allowed time with 'GPU TASK ABORTED'."
386
+
387
  HELP_GUI = (
388
+ f"""### Help:
389
+ {msg_zero}
390
  - Distorted or strange images often result from high prompt weights, so it's best to use low weights and scales, and consider using Classic variants like 'Classic-original'.
391
  - For better results with Pony Diffusion, try using sampler DPM++ 1s or DPM2 with Compel or Classic prompt weights.
392
  """
 
510
  20,
511
  4.0,
512
  -1,
513
+ ("loras/Coloring_book_-_LineArt.safetensors" if os.path.exists("loras/Coloring_book_-_LineArt.safetensors") else "None"),
514
  1.0,
515
  "DPM++ 2M SDE",
516
  1024,
 
605
  RESOURCES = (
606
  """### Resources
607
  - John6666's space has some great features you might find helpful [link](https://huggingface.co/spaces/John6666/DiffuseCraftMod).
608
+ - Try the image generator in Colab’s free tier, which provides free GPU [link](https://github.com/R3gm/SD_diffusers_interactive).
609
+ - `DiffuseCraft` in Colab:[link](https://github.com/R3gm/DiffuseCraft?tab=readme-ov-file#diffusecraft).
610
  """
611
+ )
dc.py CHANGED
@@ -1,5 +1,6 @@
1
  import spaces
2
  import os
 
3
  from stablepy import (
4
  Model_Diffusers,
5
  SCHEDULE_TYPE_OPTIONS,
@@ -23,6 +24,7 @@ from constants import (
23
  DIFFUSERS_CONTROLNET_MODEL,
24
  IP_MODELS,
25
  MODE_IP_OPTIONS,
 
26
  )
27
  from stablepy.diffusers_vanilla.style_prompt_config import STYLE_NAMES
28
  import torch
@@ -41,6 +43,7 @@ from utils import (
41
  progress_step_bar,
42
  html_template_message,
43
  escape_html,
 
44
  )
45
  from image_processor import preprocessor_tab
46
  from datetime import datetime
@@ -53,15 +56,17 @@ from diffusers import FluxPipeline
53
  # import urllib.parse
54
  import subprocess
55
 
56
- IS_ZERO = True if os.getenv("SPACES_ZERO_GPU", None) else False
57
- if IS_ZERO:
58
  subprocess.run("rm -rf /data-nvme/zerogpu-offload/*", env={}, shell=True)
59
- torch.backends.cuda.matmul.allow_tf32 = True
 
 
 
60
 
61
  ImageFile.LOAD_TRUNCATED_IMAGES = True
62
-
63
  # os.environ["PYTORCH_NO_CUDA_MEMORY_CACHING"] = "1"
64
- print(os.getenv("SPACES_ZERO_GPU"))
65
 
66
  ## BEGIN MOD
67
  logging.getLogger("diffusers").setLevel(logging.ERROR)
@@ -103,14 +108,11 @@ for directory in directories:
103
 
104
  # Download stuffs
105
  for url in [url.strip() for url in download_model.split(',')]:
106
- if not os.path.exists(f"./models/{url.split('/')[-1]}"):
107
- download_things(DIRECTORY_MODELS, url, HF_TOKEN, CIVITAI_API_KEY)
108
  for url in [url.strip() for url in download_vae.split(',')]:
109
- if not os.path.exists(f"./vaes/{url.split('/')[-1]}"):
110
- download_things(DIRECTORY_VAES, url, HF_TOKEN, CIVITAI_API_KEY)
111
  for url in [url.strip() for url in download_lora.split(',')]:
112
- if not os.path.exists(f"./loras/{url.split('/')[-1]}"):
113
- download_things(DIRECTORY_LORAS, url, HF_TOKEN, CIVITAI_API_KEY)
114
 
115
  # Download Embeddings
116
  for url_embed in DOWNLOAD_EMBEDS:
@@ -138,15 +140,32 @@ def get_embed_list(pipeline_name):
138
 
139
  print('\033[33m🏁 Download and listing of valid models completed.\033[0m')
140
 
141
- flux_repo = "camenduru/FLUX.1-dev-diffusers"
142
- flux_pipe = FluxPipeline.from_pretrained(
143
- flux_repo,
144
- transformer=None,
145
- torch_dtype=torch.bfloat16,
146
- )#.to("cuda")
147
- components = flux_pipe.components
148
- delete_model(flux_repo)
149
- #components = None
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
150
 
151
  ## BEGIN MOD
152
  class GuiSD:
@@ -164,6 +183,15 @@ class GuiSD:
164
  removal_candidate = self.inventory.pop(0)
165
  delete_model(removal_candidate)
166
 
 
 
 
 
 
 
 
 
 
167
  def update_inventory(self, model_name):
168
  if model_name not in single_file_model_list:
169
  self.inventory = [
@@ -174,19 +202,24 @@ class GuiSD:
174
  def load_new_model(self, model_name, vae_model, task, controlnet_model, progress=gr.Progress(track_tqdm=True)):
175
 
176
  # download link model > model_name
177
- if "http" in model_name: #
 
 
178
  model_name, model_type = download_link_model(model_name, DIRECTORY_MODELS) #
 
 
179
  is_link_model = True #
180
  else: is_link_model = False #
181
 
182
- self.update_storage_models()
 
183
 
184
  vae_model = vae_model if vae_model != "None" else None
185
  model_type = get_model_type(model_name) if not is_link_model else model_type #
186
  dtype_model = torch.bfloat16 if model_type == "FLUX" else torch.float16
187
 
188
  if not os.path.exists(model_name):
189
- print("debug", model_name, vae_model, task, controlnet_model)
190
  _ = download_diffuser_repo(
191
  repo_name=model_name,
192
  model_type=model_type,
@@ -231,10 +264,10 @@ class GuiSD:
231
  type_model_precision=dtype_model,
232
  retain_task_model_in_cache=False,
233
  controlnet_model=controlnet_model,
234
- device="cpu",
235
  env_components=components,
236
  )
237
- self.model.advanced_params(image_preprocessor_cuda_active=True)
238
  else:
239
  if self.model.base_model_id != model_name:
240
  load_now_time = datetime.now()
@@ -244,7 +277,8 @@ class GuiSD:
244
  print("Waiting for the previous model's time ops...")
245
  time.sleep(9 - elapsed_time)
246
 
247
- self.model.device = torch.device("cpu")
 
248
  self.model.load_pipe(
249
  model_name,
250
  task_name=TASK_STABLEPY[task],
@@ -404,7 +438,7 @@ class GuiSD:
404
  lora_model_list = get_lora_model_list()
405
  ## END MOD
406
 
407
- print("Config model:", model_name, vae_model, loras_list)
408
 
409
  task = TASK_STABLEPY[task]
410
 
@@ -502,19 +536,19 @@ class GuiSD:
502
  "distance_threshold": distance_threshold,
503
  "recolor_gamma_correction": float(recolor_gamma_correction),
504
  "tile_blur_sigma": int(tile_blur_sigma),
505
- "lora_A": lora1 if lora1 != "None" else None,
506
  "lora_scale_A": lora_scale1,
507
- "lora_B": lora2 if lora2 != "None" else None,
508
  "lora_scale_B": lora_scale2,
509
- "lora_C": lora3 if lora3 != "None" else None,
510
  "lora_scale_C": lora_scale3,
511
- "lora_D": lora4 if lora4 != "None" else None,
512
  "lora_scale_D": lora_scale4,
513
- "lora_E": lora5 if lora5 != "None" else None,
514
  "lora_scale_E": lora_scale5,
515
- "lora_F": lora6 if lora6 != "None" else None,
516
  "lora_scale_F": lora_scale6,
517
- "lora_G": lora7 if lora7 != "None" else None,
518
  "lora_scale_G": lora_scale7,
519
  ## BEGIN MOD
520
  "textual_inversion": get_embed_list(self.model.class_name) if textual_inversion else [],
@@ -574,11 +608,11 @@ class GuiSD:
574
  # kwargs for diffusers pipeline
575
  if guidance_rescale:
576
  pipe_params["guidance_rescale"] = guidance_rescale
577
-
578
- self.model.device = torch.device("cuda:0")
579
- if hasattr(self.model.pipe, "transformer") and loras_list != ["None"] * self.model.num_loras:
580
- self.model.pipe.transformer.to(self.model.device)
581
- print("transformer to cuda")
582
 
583
  actual_progress = 0
584
  info_images = gr.update()
@@ -608,7 +642,7 @@ class GuiSD:
608
 
609
  download_links = "<br>".join(
610
  [
611
- f'<a href="{path.replace("/images/", "/file=/home/user/app/images/")}" download="{os.path.basename(path)}">Download Image {i + 1}</a>'
612
  for i, path in enumerate(image_path)
613
  ]
614
  )
@@ -720,7 +754,8 @@ def sd_gen_generate_pipeline(*args):
720
 
721
  @spaces.GPU(duration=15)
722
  def process_upscale(image, upscaler_name, upscaler_size):
723
- if image is None: return None
 
724
 
725
  from stablepy.diffusers_vanilla.utils import save_pil_image_with_metadata
726
  from stablepy import load_upscaler_model
@@ -737,7 +772,7 @@ def process_upscale(image, upscaler_name, upscaler_size):
737
 
738
  name_upscaler = f"./{DIRECTORY_UPSCALERS}/{name_upscaler.split('/')[-1]}"
739
 
740
- scaler_beta = load_upscaler_model(model=name_upscaler, tile=0, tile_overlap=8, device="cuda", half=True)
741
  image_up = scaler_beta.upscale(image, upscaler_size, True)
742
 
743
  image_path = save_pil_image_with_metadata(image_up, f'{os.getcwd()}/up_images', exif_image)
@@ -746,8 +781,8 @@ def process_upscale(image, upscaler_name, upscaler_size):
746
 
747
 
748
  # https://huggingface.co/spaces/BestWishYsh/ConsisID-preview-Space/discussions/1#674969a022b99c122af5d407
749
- dynamic_gpu_duration.zerogpu = True
750
- sd_gen_generate_pipeline.zerogpu = True
751
  sd_gen = GuiSD()
752
 
753
 
 
1
  import spaces
2
  import os
3
+ from argparse import ArgumentParser
4
  from stablepy import (
5
  Model_Diffusers,
6
  SCHEDULE_TYPE_OPTIONS,
 
24
  DIFFUSERS_CONTROLNET_MODEL,
25
  IP_MODELS,
26
  MODE_IP_OPTIONS,
27
+ CACHE_HF_ROOT,
28
  )
29
  from stablepy.diffusers_vanilla.style_prompt_config import STYLE_NAMES
30
  import torch
 
43
  progress_step_bar,
44
  html_template_message,
45
  escape_html,
46
+ clear_hf_cache,
47
  )
48
  from image_processor import preprocessor_tab
49
  from datetime import datetime
 
56
  # import urllib.parse
57
  import subprocess
58
 
59
+ IS_ZERO_GPU = bool(os.getenv("SPACES_ZERO_GPU"))
60
+ if IS_ZERO_GPU:
61
  subprocess.run("rm -rf /data-nvme/zerogpu-offload/*", env={}, shell=True)
62
+ IS_GPU_MODE = True if IS_ZERO_GPU else (True if torch.cuda.is_available() else False)
63
+ img_path = "./images/"
64
+ allowed_path = os.path.abspath(img_path)
65
+ delete_cache_time = (9600, 9600) if IS_ZERO_GPU else (86400, 86400)
66
 
67
  ImageFile.LOAD_TRUNCATED_IMAGES = True
68
+ torch.backends.cuda.matmul.allow_tf32 = True
69
  # os.environ["PYTORCH_NO_CUDA_MEMORY_CACHING"] = "1"
 
70
 
71
  ## BEGIN MOD
72
  logging.getLogger("diffusers").setLevel(logging.ERROR)
 
108
 
109
  # Download stuffs
110
  for url in [url.strip() for url in download_model.split(',')]:
111
+ download_things(DIRECTORY_MODELS, url, HF_TOKEN, CIVITAI_API_KEY)
 
112
  for url in [url.strip() for url in download_vae.split(',')]:
113
+ download_things(DIRECTORY_VAES, url, HF_TOKEN, CIVITAI_API_KEY)
 
114
  for url in [url.strip() for url in download_lora.split(',')]:
115
+ download_things(DIRECTORY_LORAS, url, HF_TOKEN, CIVITAI_API_KEY)
 
116
 
117
  # Download Embeddings
118
  for url_embed in DOWNLOAD_EMBEDS:
 
140
 
141
  print('\033[33m🏁 Download and listing of valid models completed.\033[0m')
142
 
143
+ components = None
144
+ if IS_ZERO_GPU:
145
+ flux_repo = "camenduru/FLUX.1-dev-diffusers"
146
+ flux_pipe = FluxPipeline.from_pretrained(
147
+ flux_repo,
148
+ transformer=None,
149
+ torch_dtype=torch.bfloat16,
150
+ )#.to("cuda")
151
+ components = flux_pipe.components
152
+ delete_model(flux_repo)
153
+
154
+ parser = ArgumentParser(description='DiffuseCraft: Create images from text prompts.', add_help=True)
155
+ parser.add_argument("--share", action="store_true", dest="share_enabled", default=False, help="Enable sharing")
156
+ parser.add_argument('--theme', type=str, default="NoCrypt/miku", help='Set the theme (default: NoCrypt/miku)')
157
+ parser.add_argument("--ssr", action="store_true", help="Enable SSR (Server-Side Rendering)")
158
+ parser.add_argument("--log-level", type=str, default="INFO", choices=["DEBUG", "INFO", "WARNING", "ERROR", "CRITICAL"], help="Set logging level (default: INFO)")
159
+ args = parser.parse_args()
160
+
161
+ logger.setLevel(
162
+ "INFO" if IS_ZERO_GPU else getattr(logging, args.log_level.upper())
163
+ )
164
+
165
+ def lora_chk(lora_):
166
+ if isinstance(lora_, str) and lora_.strip() not in ["", "None"]:
167
+ return lora_
168
+ return None
169
 
170
  ## BEGIN MOD
171
  class GuiSD:
 
183
  removal_candidate = self.inventory.pop(0)
184
  delete_model(removal_candidate)
185
 
186
+ # Cleanup after 60 seconds of inactivity
187
+ lowPrioCleanup = max((datetime.now() - self.last_load).total_seconds(), 0) > 60
188
+ if lowPrioCleanup and not self.status_loading and get_used_storage_gb(CACHE_HF_ROOT) > (storage_floor_gb * 2):
189
+ print("Cleaning up Hugging Face cache...")
190
+ clear_hf_cache()
191
+ self.inventory = [
192
+ m for m in self.inventory if os.path.exists(m)
193
+ ]
194
+
195
  def update_inventory(self, model_name):
196
  if model_name not in single_file_model_list:
197
  self.inventory = [
 
202
  def load_new_model(self, model_name, vae_model, task, controlnet_model, progress=gr.Progress(track_tqdm=True)):
203
 
204
  # download link model > model_name
205
+ if model_name.startswith("http"): #
206
+ yield f"Downloading model: {model_name}"
207
+ #model_name = download_things(DIRECTORY_MODELS, model_name, HF_TOKEN, CIVITAI_API_KEY)
208
  model_name, model_type = download_link_model(model_name, DIRECTORY_MODELS) #
209
+ if not model_name:
210
+ raise ValueError("Error retrieving model information from URL")
211
  is_link_model = True #
212
  else: is_link_model = False #
213
 
214
+ if IS_ZERO_GPU:
215
+ self.update_storage_models()
216
 
217
  vae_model = vae_model if vae_model != "None" else None
218
  model_type = get_model_type(model_name) if not is_link_model else model_type #
219
  dtype_model = torch.bfloat16 if model_type == "FLUX" else torch.float16
220
 
221
  if not os.path.exists(model_name):
222
+ logger.debug(f"model_name={model_name}, vae_model={vae_model}, task={task}, controlnet_model={controlnet_model}")
223
  _ = download_diffuser_repo(
224
  repo_name=model_name,
225
  model_type=model_type,
 
264
  type_model_precision=dtype_model,
265
  retain_task_model_in_cache=False,
266
  controlnet_model=controlnet_model,
267
+ device="cpu" if IS_ZERO_GPU else None,
268
  env_components=components,
269
  )
270
+ self.model.advanced_params(image_preprocessor_cuda_active=IS_GPU_MODE)
271
  else:
272
  if self.model.base_model_id != model_name:
273
  load_now_time = datetime.now()
 
277
  print("Waiting for the previous model's time ops...")
278
  time.sleep(9 - elapsed_time)
279
 
280
+ if IS_ZERO_GPU:
281
+ self.model.device = torch.device("cpu")
282
  self.model.load_pipe(
283
  model_name,
284
  task_name=TASK_STABLEPY[task],
 
438
  lora_model_list = get_lora_model_list()
439
  ## END MOD
440
 
441
+ logger.debug(f"Config model: {model_name}, {vae_model}, {loras_list}")
442
 
443
  task = TASK_STABLEPY[task]
444
 
 
536
  "distance_threshold": distance_threshold,
537
  "recolor_gamma_correction": float(recolor_gamma_correction),
538
  "tile_blur_sigma": int(tile_blur_sigma),
539
+ "lora_A": lora_chk(lora1),
540
  "lora_scale_A": lora_scale1,
541
+ "lora_B": lora_chk(lora2),
542
  "lora_scale_B": lora_scale2,
543
+ "lora_C": lora_chk(lora3),
544
  "lora_scale_C": lora_scale3,
545
+ "lora_D": lora_chk(lora4),
546
  "lora_scale_D": lora_scale4,
547
+ "lora_E": lora_chk(lora5),
548
  "lora_scale_E": lora_scale5,
549
+ "lora_F": lora_chk(lora6),
550
  "lora_scale_F": lora_scale6,
551
+ "lora_G": lora_chk(lora7),
552
  "lora_scale_G": lora_scale7,
553
  ## BEGIN MOD
554
  "textual_inversion": get_embed_list(self.model.class_name) if textual_inversion else [],
 
608
  # kwargs for diffusers pipeline
609
  if guidance_rescale:
610
  pipe_params["guidance_rescale"] = guidance_rescale
611
+ if IS_ZERO_GPU:
612
+ self.model.device = torch.device("cuda:0")
613
+ if hasattr(self.model.pipe, "transformer") and loras_list != ["None"] * self.model.num_loras:
614
+ self.model.pipe.transformer.to(self.model.device)
615
+ logger.debug("transformer to cuda")
616
 
617
  actual_progress = 0
618
  info_images = gr.update()
 
642
 
643
  download_links = "<br>".join(
644
  [
645
+ f'<a href="{path.replace("/images/", f"/gradio_api/file={allowed_path}/")}" download="{os.path.basename(path)}">Download Image {i + 1}</a>'
646
  for i, path in enumerate(image_path)
647
  ]
648
  )
 
754
 
755
  @spaces.GPU(duration=15)
756
  def process_upscale(image, upscaler_name, upscaler_size):
757
+ if image is None:
758
+ return None
759
 
760
  from stablepy.diffusers_vanilla.utils import save_pil_image_with_metadata
761
  from stablepy import load_upscaler_model
 
772
 
773
  name_upscaler = f"./{DIRECTORY_UPSCALERS}/{name_upscaler.split('/')[-1]}"
774
 
775
+ scaler_beta = load_upscaler_model(model=name_upscaler, tile=(0 if IS_ZERO_GPU else 192), tile_overlap=8, device=("cuda" if IS_GPU_MODE else "cpu"), half=IS_GPU_MODE)
776
  image_up = scaler_beta.upscale(image, upscaler_size, True)
777
 
778
  image_path = save_pil_image_with_metadata(image_up, f'{os.getcwd()}/up_images', exif_image)
 
781
 
782
 
783
  # https://huggingface.co/spaces/BestWishYsh/ConsisID-preview-Space/discussions/1#674969a022b99c122af5d407
784
+ # dynamic_gpu_duration.zerogpu = True
785
+ # sd_gen_generate_pipeline.zerogpu = True
786
  sd_gen = GuiSD()
787
 
788
 
env.py CHANGED
@@ -71,6 +71,7 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
71
  'Raelina/Raehoshi-illust-XL-4',
72
  'Raelina/Raehoshi-illust-XL-5',
73
  'Raelina/Raehoshi-illust-XL-5.1',
 
74
  'camenduru/FLUX.1-dev-diffusers',
75
  'black-forest-labs/FLUX.1-schnell',
76
  'sayakpaul/FLUX.1-merged',
 
71
  'Raelina/Raehoshi-illust-XL-4',
72
  'Raelina/Raehoshi-illust-XL-5',
73
  'Raelina/Raehoshi-illust-XL-5.1',
74
+ 'Raelina/Raehoshi-illust-XL-6',
75
  'camenduru/FLUX.1-dev-diffusers',
76
  'black-forest-labs/FLUX.1-schnell',
77
  'sayakpaul/FLUX.1-merged',
image_processor.py CHANGED
@@ -92,8 +92,8 @@ def preprocessor_tab():
92
  pre_processor_resolution = gr.Slider(minimum=64, maximum=2048, step=64, value=512, label="Preprocessor Resolution")
93
  pre_low_threshold = gr.Slider(minimum=1, maximum=255, step=1, value=100, label="'CANNY' low threshold")
94
  pre_high_threshold = gr.Slider(minimum=1, maximum=255, step=1, value=200, label="'CANNY' high threshold")
95
- pre_value_threshold = gr.Slider(minimum=1, maximum=2.0, step=0.01, value=0.1, label="'MLSD' Hough value threshold")
96
- pre_distance_threshold = gr.Slider(minimum=1, maximum=20.0, step=0.01, value=0.1, label="'MLSD' Hough distance threshold")
97
  pre_recolor_mode = gr.Dropdown(label="'RECOLOR' mode", choices=["luminance", "intensity"], value="luminance")
98
  pre_recolor_gamma_correction = gr.Number(minimum=0., maximum=25., value=1., step=0.001, label="'RECOLOR' gamma correction")
99
  pre_blur_k_size = gr.Number(minimum=0, maximum=100, value=9, step=1, label="'BLUR' sigma")
 
92
  pre_processor_resolution = gr.Slider(minimum=64, maximum=2048, step=64, value=512, label="Preprocessor Resolution")
93
  pre_low_threshold = gr.Slider(minimum=1, maximum=255, step=1, value=100, label="'CANNY' low threshold")
94
  pre_high_threshold = gr.Slider(minimum=1, maximum=255, step=1, value=200, label="'CANNY' high threshold")
95
+ pre_value_threshold = gr.Slider(minimum=0., maximum=2.0, step=0.01, value=0.1, label="'MLSD' Hough value threshold")
96
+ pre_distance_threshold = gr.Slider(minimum=0., maximum=20.0, step=0.01, value=0.1, label="'MLSD' Hough distance threshold")
97
  pre_recolor_mode = gr.Dropdown(label="'RECOLOR' mode", choices=["luminance", "intensity"], value="luminance")
98
  pre_recolor_gamma_correction = gr.Number(minimum=0., maximum=25., value=1., step=0.001, label="'RECOLOR' gamma correction")
99
  pre_blur_k_size = gr.Number(minimum=0, maximum=100, value=9, step=1, label="'BLUR' sigma")
llmdolphin.py CHANGED
@@ -7,6 +7,7 @@ import gc
7
  import os
8
  import urllib
9
  from typing import Any
 
10
  from huggingface_hub import hf_hub_download, HfApi
11
  from llama_cpp import Llama
12
  from llama_cpp_agent import LlamaCppAgent, MessagesFormatterType
@@ -376,10 +377,10 @@ def get_raw_prompt(msg: str):
376
 
377
  # https://llama-cpp-python.readthedocs.io/en/latest/api-reference/
378
  @torch.inference_mode()
379
- @spaces.GPU(duration=59)
380
  def dolphin_respond(
381
  message: str,
382
- history: list[tuple[str, str]],
383
  model: str = default_llm_model_filename,
384
  system_message: str = get_dolphin_sysprompt(),
385
  max_tokens: int = 1024,
@@ -434,16 +435,12 @@ def dolphin_respond(
434
  messages = BasicChatHistory()
435
 
436
  for msn in history:
437
- user = {
438
- 'role': Roles.user,
439
- 'content': msn[0]
440
- }
441
- assistant = {
442
- 'role': Roles.assistant,
443
- 'content': msn[1]
444
- }
445
- messages.add_message(user)
446
- messages.add_message(assistant)
447
 
448
  stream = agent.get_chat_response(
449
  message,
@@ -455,28 +452,28 @@ def dolphin_respond(
455
 
456
  progress(0.5, desc="Processing...")
457
 
458
- outputs = ""
 
459
  for output in stream:
460
- outputs += output
461
- yield [(outputs, None)]
462
  except Exception as e:
463
  print(e)
464
  raise gr.Error(f"Error: {e}")
465
- #yield [("", None)]
466
  finally:
467
  torch.cuda.empty_cache()
468
  gc.collect()
469
 
470
 
471
  def dolphin_parse(
472
- history: list[tuple[str, str]],
473
  state: dict,
474
  ):
475
  try:
476
  dolphin_sysprompt_mode = get_state(state, "dolphin_sysprompt_mode")
477
  if dolphin_sysprompt_mode == "Chat with LLM" or not history or len(history) < 1:
478
  return "", gr.update(), gr.update()
479
- msg = history[-1][0]
480
  raw_prompt = get_raw_prompt(msg)
481
  prompts = []
482
  if dolphin_sysprompt_mode == "Japanese to Danbooru Dictionary" and is_japanese(raw_prompt):
@@ -490,10 +487,10 @@ def dolphin_parse(
490
 
491
 
492
  @torch.inference_mode()
493
- @spaces.GPU(duration=59)
494
  def dolphin_respond_auto(
495
  message: str,
496
- history: list[tuple[str, str]],
497
  model: str = default_llm_model_filename,
498
  system_message: str = get_dolphin_sysprompt(),
499
  max_tokens: int = 1024,
@@ -549,16 +546,12 @@ def dolphin_respond_auto(
549
  messages = BasicChatHistory()
550
 
551
  for msn in history:
552
- user = {
553
- 'role': Roles.user,
554
- 'content': msn[0]
555
- }
556
- assistant = {
557
- 'role': Roles.assistant,
558
- 'content': msn[1]
559
- }
560
- messages.add_message(user)
561
- messages.add_message(assistant)
562
 
563
  progress(0, desc="Translating...")
564
  stream = agent.get_chat_response(
@@ -571,13 +564,16 @@ def dolphin_respond_auto(
571
 
572
  progress(0.5, desc="Processing...")
573
 
574
- outputs = ""
 
575
  for output in stream:
576
- outputs += output
577
- yield [(outputs, None)], gr.update(), gr.update()
578
  except Exception as e:
579
  print(e)
580
- yield [("", None)], gr.update(), gr.update()
 
 
581
  finally:
582
  torch.cuda.empty_cache()
583
  gc.collect()
@@ -585,14 +581,14 @@ def dolphin_respond_auto(
585
 
586
  def dolphin_parse_simple(
587
  message: str,
588
- history: list[tuple[str, str]],
589
  state: dict,
590
  ):
591
  try:
592
  #if not is_japanese(message): return message
593
  dolphin_sysprompt_mode = get_state(state, "dolphin_sysprompt_mode")
594
  if dolphin_sysprompt_mode == "Chat with LLM" or not history or len(history) < 1: return message
595
- msg = history[-1][0]
596
  raw_prompt = get_raw_prompt(msg)
597
  prompts = []
598
  if dolphin_sysprompt_mode == "Japanese to Danbooru Dictionary" and is_japanese(raw_prompt):
@@ -611,10 +607,10 @@ cv2.setNumThreads(1)
611
 
612
 
613
  @torch.inference_mode()
614
- @spaces.GPU(duration=59)
615
  def respond_playground(
616
  message: str,
617
- history: list[tuple[str, str]],
618
  model: str = default_llm_model_filename,
619
  system_message: str = get_dolphin_sysprompt(),
620
  max_tokens: int = 1024,
@@ -669,10 +665,12 @@ def respond_playground(
669
 
670
  # Add user and assistant messages to the history
671
  for msn in history:
672
- user = {'role': Roles.user, 'content': msn[0]}
673
- assistant = {'role': Roles.assistant, 'content': msn[1]}
674
- messages.add_message(user)
675
- messages.add_message(assistant)
 
 
676
 
677
  # Stream the response
678
  stream = agent.get_chat_response(
@@ -683,14 +681,14 @@ def respond_playground(
683
  print_output=False
684
  )
685
 
686
- outputs = ""
 
687
  for output in stream:
688
- outputs += output
689
- yield outputs
690
  except Exception as e:
691
  print(e)
692
  raise gr.Error(f"Error: {e}")
693
- #yield ""
694
  finally:
695
  torch.cuda.empty_cache()
696
  gc.collect()
 
7
  import os
8
  import urllib
9
  from typing import Any
10
+ from gradio import MessageDict
11
  from huggingface_hub import hf_hub_download, HfApi
12
  from llama_cpp import Llama
13
  from llama_cpp_agent import LlamaCppAgent, MessagesFormatterType
 
377
 
378
  # https://llama-cpp-python.readthedocs.io/en/latest/api-reference/
379
  @torch.inference_mode()
380
+ @spaces.GPU(duration=30)
381
  def dolphin_respond(
382
  message: str,
383
+ history: list[MessageDict],
384
  model: str = default_llm_model_filename,
385
  system_message: str = get_dolphin_sysprompt(),
386
  max_tokens: int = 1024,
 
435
  messages = BasicChatHistory()
436
 
437
  for msn in history:
438
+ if msn["role"] == "user":
439
+ user = {'role': Roles.user, 'content': msn["content"]}
440
+ messages.add_message(user)
441
+ elif msn["role"] == "assistant":
442
+ assistant = {'role': Roles.assistant, 'content': msn["content"]}
443
+ messages.add_message(assistant)
 
 
 
 
444
 
445
  stream = agent.get_chat_response(
446
  message,
 
452
 
453
  progress(0.5, desc="Processing...")
454
 
455
+ history.append({"role": "user", "content": message})
456
+ history.append({"role": "assistant", "content": ""})
457
  for output in stream:
458
+ history[-1]['content'] += output
459
+ yield history
460
  except Exception as e:
461
  print(e)
462
  raise gr.Error(f"Error: {e}")
 
463
  finally:
464
  torch.cuda.empty_cache()
465
  gc.collect()
466
 
467
 
468
  def dolphin_parse(
469
+ history: list[MessageDict],
470
  state: dict,
471
  ):
472
  try:
473
  dolphin_sysprompt_mode = get_state(state, "dolphin_sysprompt_mode")
474
  if dolphin_sysprompt_mode == "Chat with LLM" or not history or len(history) < 1:
475
  return "", gr.update(), gr.update()
476
+ msg = history[-1]["content"]
477
  raw_prompt = get_raw_prompt(msg)
478
  prompts = []
479
  if dolphin_sysprompt_mode == "Japanese to Danbooru Dictionary" and is_japanese(raw_prompt):
 
487
 
488
 
489
  @torch.inference_mode()
490
+ @spaces.GPU(duration=30)
491
  def dolphin_respond_auto(
492
  message: str,
493
+ history: list[MessageDict],
494
  model: str = default_llm_model_filename,
495
  system_message: str = get_dolphin_sysprompt(),
496
  max_tokens: int = 1024,
 
546
  messages = BasicChatHistory()
547
 
548
  for msn in history:
549
+ if msn["role"] == "user":
550
+ user = {'role': Roles.user, 'content': msn["content"]}
551
+ messages.add_message(user)
552
+ elif msn["role"] == "assistant":
553
+ assistant = {'role': Roles.assistant, 'content': msn["content"]}
554
+ messages.add_message(assistant)
 
 
 
 
555
 
556
  progress(0, desc="Translating...")
557
  stream = agent.get_chat_response(
 
564
 
565
  progress(0.5, desc="Processing...")
566
 
567
+ history.append({"role": "user", "content": message})
568
+ history.append({"role": "assistant", "content": ""})
569
  for output in stream:
570
+ history[-1]['content'] += output
571
+ yield history, gr.update(), gr.update()
572
  except Exception as e:
573
  print(e)
574
+ history.append({"role": "user", "content": message})
575
+ history.append({"role": "assistant", "content": message})
576
+ yield history, gr.update(), gr.update()
577
  finally:
578
  torch.cuda.empty_cache()
579
  gc.collect()
 
581
 
582
  def dolphin_parse_simple(
583
  message: str,
584
+ history: list[MessageDict],
585
  state: dict,
586
  ):
587
  try:
588
  #if not is_japanese(message): return message
589
  dolphin_sysprompt_mode = get_state(state, "dolphin_sysprompt_mode")
590
  if dolphin_sysprompt_mode == "Chat with LLM" or not history or len(history) < 1: return message
591
+ msg = history[-1]["content"]
592
  raw_prompt = get_raw_prompt(msg)
593
  prompts = []
594
  if dolphin_sysprompt_mode == "Japanese to Danbooru Dictionary" and is_japanese(raw_prompt):
 
607
 
608
 
609
  @torch.inference_mode()
610
+ @spaces.GPU(duration=30)
611
  def respond_playground(
612
  message: str,
613
+ history: list[MessageDict],
614
  model: str = default_llm_model_filename,
615
  system_message: str = get_dolphin_sysprompt(),
616
  max_tokens: int = 1024,
 
665
 
666
  # Add user and assistant messages to the history
667
  for msn in history:
668
+ if msn["role"] == "user":
669
+ user = {'role': Roles.user, 'content': msn["content"]}
670
+ messages.add_message(user)
671
+ elif msn["role"] == "assistant":
672
+ assistant = {'role': Roles.assistant, 'content': msn["content"]}
673
+ messages.add_message(assistant)
674
 
675
  # Stream the response
676
  stream = agent.get_chat_response(
 
681
  print_output=False
682
  )
683
 
684
+ history.append({"role": "user", "content": message})
685
+ history.append({"role": "assistant", "content": ""})
686
  for output in stream:
687
+ history[-1]['content'] += output
688
+ yield history
689
  except Exception as e:
690
  print(e)
691
  raise gr.Error(f"Error: {e}")
 
692
  finally:
693
  torch.cuda.empty_cache()
694
  gc.collect()
llmenv.py CHANGED
@@ -14,6 +14,7 @@ llm_models = {
14
  "mn-12b-lyra-v2a1-q5_k_m.gguf": ["HalleyStarbun/MN-12B-Lyra-v2a1-Q5_K_M-GGUF", MessagesFormatterType.CHATML],
15
  "L3-8B-Tamamo-v1.i1-Q5_K_M.gguf": ["mradermacher/L3-8B-Tamamo-v1-i1-GGUF", MessagesFormatterType.LLAMA_3],
16
  "MN-Chinofun-12B-2.i1-Q4_K_M.gguf": ["mradermacher/MN-Chinofun-12B-2-i1-GGUF", MessagesFormatterType.MISTRAL],
 
17
  "Dolphin-5.1-7b.Q5_K_M.gguf": ["mradermacher/Dolphin-5.1-7b-GGUF", MessagesFormatterType.MISTRAL],
18
  "Captain-Eris-Diogenes_Twighlight-V0.420-12B.i1-Q4_K_M.gguf": ["mradermacher/Captain-Eris-Diogenes_Twighlight-V0.420-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
19
  "Captain-Eris-Diogenes_Twilight-V0.420-12B.i1-Q4_K_M.gguf": ["mradermacher/Captain-Eris-Diogenes_Twilight-V0.420-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
@@ -129,9 +130,41 @@ llm_models = {
129
  "Aurora-SCE-12B.Q4_K_M.gguf": ["mradermacher/Aurora-SCE-12B-GGUF", MessagesFormatterType.MISTRAL],
130
  "Rune-14b.Q4_K_M.gguf": ["mradermacher/Rune-14b-GGUF", MessagesFormatterType.OPEN_CHAT],
131
  "L3.1-Athena-e-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Athena-e-8B-GGUF", MessagesFormatterType.LLAMA_3],
 
132
  "PR2-14B-Instruct.Q4_K_M.gguf": ["mradermacher/PR2-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
 
 
 
133
  "Nemo-12b-Humanize-KTO-Experimental-Latest.Q4_K_M.gguf": ["mradermacher/Nemo-12b-Humanize-KTO-Experimental-Latest-GGUF", MessagesFormatterType.CHATML],
134
  "Baptist-Christian-Bible-Expert-v1.1-12B.Q4_K_M.gguf": ["mradermacher/Baptist-Christian-Bible-Expert-v1.1-12B-GGUF", MessagesFormatterType.MISTRAL],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
135
  "openbuddy-qwen2.5llamaify-14b-v23.1-200k.i1-Q4_K_M.gguf": ["mradermacher/openbuddy-qwen2.5llamaify-14b-v23.1-200k-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
136
  "Trinas_Nectar-8B-model_stock.i1-Q4_K_M.gguf": ["mradermacher/Trinas_Nectar-8B-model_stock-i1-GGUF", MessagesFormatterType.MISTRAL],
137
  "ChatWaifu_Magnum_V0.2.Q4_K_M.gguf": ["mradermacher/ChatWaifu_Magnum_V0.2-GGUF", MessagesFormatterType.MISTRAL],
@@ -145,6 +178,715 @@ llm_models = {
145
  #"": ["", MessagesFormatterType.OPEN_CHAT],
146
  #"": ["", MessagesFormatterType.CHATML],
147
  #"": ["", MessagesFormatterType.PHI_3],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
148
  "Hush-Qwen2.5-7B-MST-v1.1.Q5_K_M.gguf": ["mradermacher/Hush-Qwen2.5-7B-MST-v1.1-GGUF", MessagesFormatterType.OPEN_CHAT],
149
  "MFANNv0.22.1.Q5_K_M.gguf": ["mradermacher/MFANNv0.22.1-GGUF", MessagesFormatterType.LLAMA_3],
150
  "gemma2-9b-cpt-sea-lionv3-instruct.Q4_K_M.gguf": ["mradermacher/gemma2-9b-cpt-sea-lionv3-instruct-GGUF", MessagesFormatterType.ALPACA],
 
14
  "mn-12b-lyra-v2a1-q5_k_m.gguf": ["HalleyStarbun/MN-12B-Lyra-v2a1-Q5_K_M-GGUF", MessagesFormatterType.CHATML],
15
  "L3-8B-Tamamo-v1.i1-Q5_K_M.gguf": ["mradermacher/L3-8B-Tamamo-v1-i1-GGUF", MessagesFormatterType.LLAMA_3],
16
  "MN-Chinofun-12B-2.i1-Q4_K_M.gguf": ["mradermacher/MN-Chinofun-12B-2-i1-GGUF", MessagesFormatterType.MISTRAL],
17
+ "Shirayukihime-12B.Q4_K_M.gguf": ["mradermacher/Shirayukihime-12B-GGUF", MessagesFormatterType.CHATML],
18
  "Dolphin-5.1-7b.Q5_K_M.gguf": ["mradermacher/Dolphin-5.1-7b-GGUF", MessagesFormatterType.MISTRAL],
19
  "Captain-Eris-Diogenes_Twighlight-V0.420-12B.i1-Q4_K_M.gguf": ["mradermacher/Captain-Eris-Diogenes_Twighlight-V0.420-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
20
  "Captain-Eris-Diogenes_Twilight-V0.420-12B.i1-Q4_K_M.gguf": ["mradermacher/Captain-Eris-Diogenes_Twilight-V0.420-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
 
130
  "Aurora-SCE-12B.Q4_K_M.gguf": ["mradermacher/Aurora-SCE-12B-GGUF", MessagesFormatterType.MISTRAL],
131
  "Rune-14b.Q4_K_M.gguf": ["mradermacher/Rune-14b-GGUF", MessagesFormatterType.OPEN_CHAT],
132
  "L3.1-Athena-e-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Athena-e-8B-GGUF", MessagesFormatterType.LLAMA_3],
133
+ "ElitePhi.i1-Q4_K_M.gguf": ["mradermacher/ElitePhi-i1-GGUF", MessagesFormatterType.LLAMA_3],
134
  "PR2-14B-Instruct.Q4_K_M.gguf": ["mradermacher/PR2-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
135
+ "Nemo_Pony_2-merged.Q4_K_M.gguf": ["mradermacher/Nemo_Pony_2-merged-GGUF", MessagesFormatterType.MISTRAL],
136
+ "MN-12B-Mimicore-Nocturne.Q4_K_M.gguf": ["mradermacher/MN-12B-Mimicore-Nocturne-GGUF", MessagesFormatterType.MISTRAL],
137
+ "Orca-2-13b-abliterated.Q4_K_M.gguf": ["mradermacher/Orca-2-13b-abliterated-GGUF", MessagesFormatterType.OPEN_CHAT],
138
  "Nemo-12b-Humanize-KTO-Experimental-Latest.Q4_K_M.gguf": ["mradermacher/Nemo-12b-Humanize-KTO-Experimental-Latest-GGUF", MessagesFormatterType.CHATML],
139
  "Baptist-Christian-Bible-Expert-v1.1-12B.Q4_K_M.gguf": ["mradermacher/Baptist-Christian-Bible-Expert-v1.1-12B-GGUF", MessagesFormatterType.MISTRAL],
140
+ "MN-Hekate-Enodia-12B.Q4_K_M.gguf": ["mradermacher/MN-Hekate-Enodia-12B-GGUF", MessagesFormatterType.MISTRAL],
141
+ "MN-Hekate-Ekklesia-12B.Q4_K_M.gguf": ["mradermacher/MN-Hekate-Ekklesia-12B-GGUF", MessagesFormatterType.MISTRAL],
142
+ "KANN-I-0.1-7b.Q5_K_M.gguf": ["mradermacher/KANN-I-0.1-7b-GGUF", MessagesFormatterType.MISTRAL],
143
+ "UnslopNemo-12B-v1.Q4_K_M.gguf": ["mradermacher/UnslopNemo-12B-v1-GGUF", MessagesFormatterType.MISTRAL],
144
+ "Irix-12B-Model_Stock.Q4_K_M.gguf": ["mradermacher/Irix-12B-Model_Stock-GGUF", MessagesFormatterType.CHATML],
145
+ "MagnaMellRei-v1-12B.Q4_K_M.gguf": ["mradermacher/MagnaMellRei-v1-12B-GGUF", MessagesFormatterType.CHATML],
146
+ "HMS-Slerp-12B.Q4_K_M.gguf": ["mradermacher/HMS-Slerp-12B-GGUF", MessagesFormatterType.CHATML],
147
+ "SnowElf-12B-v2.Q4_K_M.gguf": ["mradermacher/SnowElf-12B-v2-GGUF", MessagesFormatterType.CHATML],
148
+ "Queen-2.5-14B-aka.Q4_K_M.gguf": ["mradermacher/Queen-2.5-14B-aka-GGUF", MessagesFormatterType.OPEN_CHAT],
149
+ "KnowledgeCore-12B.Q4_K_M.gguf": ["mradermacher/KnowledgeCore-12B-GGUF", MessagesFormatterType.CHATML],
150
+ "Mistral-Nemo-2407-Role-Playing-Final-4data_ga16_lr6e7.Q4_K_M.gguf": ["mradermacher/Mistral-Nemo-2407-Role-Playing-Final-4data_ga16_lr6e7-GGUF", MessagesFormatterType.CHATML],
151
+ "PatriSlush-DarkRPReign-12B.Q4_K_M.gguf": ["mradermacher/PatriSlush-DarkRPReign-12B-GGUF", MessagesFormatterType.MISTRAL],
152
+ "BianCang-Qwen2.5-14B-Instruct.Q4_K_M.gguf": ["mradermacher/BianCang-Qwen2.5-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
153
+ "Simulation_LLM_wiki_14B_V2.Q4_K_M.gguf": ["mradermacher/Simulation_LLM_wiki_14B_V2-GGUF", MessagesFormatterType.OPEN_CHAT],
154
+ "Neona-12B.i1-Q4_K_M.gguf": ["mradermacher/Neona-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
155
+ "NeoSage-12B.Q4_K_M.gguf": ["mradermacher/NeoSage-12B-GGUF", MessagesFormatterType.MISTRAL],
156
+ "Patricide-12B-Forgottenslop-Mell.i1-Q4_K_M.gguf": ["mradermacher/Patricide-12B-Forgottenslop-Mell-i1-GGUF", MessagesFormatterType.MISTRAL],
157
+ "Pinecone-Rune-12b.Q4_K_M.gguf": ["mradermacher/Pinecone-Rune-12b-GGUF", MessagesFormatterType.MISTRAL],
158
+ "mn-12b-rp-without-dumb.Q4_K_M.gguf": ["mradermacher/mn-12b-rp-without-dumb-GGUF", MessagesFormatterType.MISTRAL],
159
+ "Denker-mistral-nemo-12B.Q4_K_M.gguf": ["mradermacher/Denker-mistral-nemo-12B-GGUF", MessagesFormatterType.MISTRAL],
160
+ "Goldcide-12B-Forgottenslop-Mell.Q4_K_M.gguf": ["mradermacher/Goldcide-12B-Forgottenslop-Mell-GGUF", MessagesFormatterType.MISTRAL],
161
+ "claude-3.7-sonnet-reasoning-gemma3-12B.Q4_K_M.gguf": ["mradermacher/claude-3.7-sonnet-reasoning-gemma3-12B-GGUF", MessagesFormatterType.ALPACA],
162
+ "allura-org_MN-Lyrebird-12B-Q4_K_M.gguf": ["bartowski/allura-org_MN-Lyrebird-12B-GGUF", MessagesFormatterType.MISTRAL],
163
+ "ape-fiction-2-mistral-nemo.Q4_K_M.gguf": ["mradermacher/ape-fiction-2-mistral-nemo-GGUF", MessagesFormatterType.MISTRAL],
164
+ "Irixxed_Homunculus-12B-Q3T-v.0.3-Reasoner.Q4_K_M.gguf": ["mradermacher/Irixxed_Homunculus-12B-Q3T-v.0.3-Reasoner-GGUF", MessagesFormatterType.MISTRAL],
165
+ "Gemma-2-Llama-Swallow-9b-pt-v0.1.Q4_K_M.gguf": ["mradermacher/Gemma-2-Llama-Swallow-9b-pt-v0.1-GGUF", MessagesFormatterType.ALPACA],
166
+ "Qwen2.5-7B-base-french-bespoke-stratos-full-sft.i1-Q5_K_S.gguf": ["mradermacher/Qwen2.5-7B-base-french-bespoke-stratos-full-sft-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
167
+ "Protestant-Christian-Bible-Expert-v2.0-12B.Q4_K_M.gguf": ["mradermacher/Protestant-Christian-Bible-Expert-v2.0-12B-GGUF", MessagesFormatterType.MISTRAL],
168
  "openbuddy-qwen2.5llamaify-14b-v23.1-200k.i1-Q4_K_M.gguf": ["mradermacher/openbuddy-qwen2.5llamaify-14b-v23.1-200k-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
169
  "Trinas_Nectar-8B-model_stock.i1-Q4_K_M.gguf": ["mradermacher/Trinas_Nectar-8B-model_stock-i1-GGUF", MessagesFormatterType.MISTRAL],
170
  "ChatWaifu_Magnum_V0.2.Q4_K_M.gguf": ["mradermacher/ChatWaifu_Magnum_V0.2-GGUF", MessagesFormatterType.MISTRAL],
 
178
  #"": ["", MessagesFormatterType.OPEN_CHAT],
179
  #"": ["", MessagesFormatterType.CHATML],
180
  #"": ["", MessagesFormatterType.PHI_3],
181
+ "SauerHuatuoSkyworkDeepWatt-o1-Llama-3.1-8B.Q5_K_M.gguf": ["mradermacher/SauerHuatuoSkyworkDeepWatt-o1-Llama-3.1-8B-GGUF", MessagesFormatterType.LLAMA_3],
182
+ "care-japanese-llama3.1-8b.Q5_K_M.gguf": ["mradermacher/care-japanese-llama3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
183
+ "UltraPatriMerge-12B.Q4_K_M.gguf": ["mradermacher/UltraPatriMerge-12B-GGUF", MessagesFormatterType.MISTRAL],
184
+ "Llama-3.1-Amelia-MTFT-8B-v1.Q5_K_M.gguf": ["mradermacher/Llama-3.1-Amelia-MTFT-8B-v1-GGUF", MessagesFormatterType.LLAMA_3],
185
+ "llama3-archimate-merged.Q5_K_M.gguf": ["mradermacher/llama3-archimate-merged-GGUF", MessagesFormatterType.LLAMA_3],
186
+ "Mistral-Nemo-Base-2407-RP-Merge.Q4_K_M.gguf": ["mradermacher/Mistral-Nemo-Base-2407-RP-Merge-GGUF", MessagesFormatterType.CHATML],
187
+ "PatriMaid-12B-Forgottenslop-NeonMell.Q4_K_M.gguf": ["mradermacher/PatriMaid-12B-Forgottenslop-NeonMell-GGUF", MessagesFormatterType.MISTRAL],
188
+ "Magnolia-v3-medis-dilute-12B.Q4_K_M.gguf": ["mradermacher/Magnolia-v3-medis-dilute-12B-GGUF", MessagesFormatterType.MISTRAL],
189
+ "Magnolia-v3b-12B.Q4_K_M.gguf": ["mradermacher/Magnolia-v3b-12B-GGUF", MessagesFormatterType.MISTRAL],
190
+ "MN-Mystic-Rune-12B.Q4_K_S.gguf": ["mradermacher/MN-Mystic-Rune-12B-GGUF", MessagesFormatterType.MISTRAL],
191
+ "MarinaraSpaghetti-NemoMix-Unleashed-12B-chat.Q4_K_M.gguf": ["mradermacher/MarinaraSpaghetti-NemoMix-Unleashed-12B-chat-GGUF", MessagesFormatterType.MISTRAL],
192
+ "GoldFox-12B-Forgottenslop-Mell.i1-Q4_K_M.gguf": ["mradermacher/GoldFox-12B-Forgottenslop-Mell-i1-GGUF", MessagesFormatterType.MISTRAL],
193
+ "Magnolia-Mell-v1-12B.Q4_K_M.gguf": ["mradermacher/Magnolia-Mell-v1-12B-GGUF", MessagesFormatterType.MISTRAL],
194
+ "Shisa-DellaTest-12B.Q4_K_M.gguf": ["mradermacher/Shisa-DellaTest-12B-GGUF", MessagesFormatterType.MISTRAL],
195
+ "Q2.5-Coldbrew14B-FusionMix.i1-Q4_K_M.gguf": ["mradermacher/Q2.5-Coldbrew14B-FusionMix-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
196
+ "nemo-instruct-books-model-stock.Q4_K_M.gguf": ["mradermacher/nemo-instruct-books-model-stock-GGUF", MessagesFormatterType.MISTRAL],
197
+ "FoxCide-12B-Forgottenslop-Mell.Q4_K_M.gguf": ["mradermacher/FoxCide-12B-Forgottenslop-Mell-GGUF", MessagesFormatterType.MISTRAL],
198
+ "BMO-CaptianMaid-12B.i1-Q4_K_M.gguf": ["mradermacher/BMO-CaptianMaid-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
199
+ "Aurore-Reveil_Koto-Small-7B-IT-Q5_K_M.gguf": ["bartowski/Aurore-Reveil_Koto-Small-7B-IT-GGUF", MessagesFormatterType.OPEN_CHAT],
200
+ "Pinecone-Rune-12b-Token-Surgery-Chatml-v0.1a.i1-Q4_K_M.gguf": ["mradermacher/Pinecone-Rune-12b-Token-Surgery-Chatml-v0.1a-i1-GGUF", MessagesFormatterType.MISTRAL],
201
+ "Kitsune-Symphony-V0.0-12B.Q4_K_M.gguf": ["mradermacher/Kitsune-Symphony-V0.0-12B-GGUF", MessagesFormatterType.MISTRAL],
202
+ "Anora-12b.i1-Q4_K_M.gguf": ["mradermacher/Anora-12b-i1-GGUF", MessagesFormatterType.MISTRAL],
203
+ "Minor-Repo-12B-omg.Q4_K_M.gguf": ["mradermacher/Minor-Repo-12B-omg-GGUF", MessagesFormatterType.MISTRAL],
204
+ "Luna.i1-Q5_K_M.gguf": ["mradermacher/Luna-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
205
+ "OmegaMaid-DarkWay-FinalNoctis-12B.Q4_K_M.gguf": ["mradermacher/OmegaMaid-DarkWay-FinalNoctis-12B-GGUF", MessagesFormatterType.MISTRAL],
206
+ "CaptainMaid-12B-VioletMell-V0.420.Q4_K_M.gguf": ["mradermacher/CaptainMaid-12B-VioletMell-V0.420-GGUF", MessagesFormatterType.MISTRAL],
207
+ "LatentSoup-modelstock-8b.Q5_K_M.gguf": ["mradermacher/LatentSoup-modelstock-8b-GGUF", MessagesFormatterType.LLAMA_3],
208
+ "Erotic-Model.v1.Q5_K_M.gguf": ["mradermacher/Erotic-Model.v1-GGUF", MessagesFormatterType.MISTRAL],
209
+ "Llama-3.1-Amelia-CD-8B-v1.Q5_K_M.gguf": ["mradermacher/Llama-3.1-Amelia-CD-8B-v1-GGUF", MessagesFormatterType.LLAMA_3],
210
+ "funny-nemo-embedding-merged.Q4_K_M.gguf": ["mradermacher/funny-nemo-embedding-merged-GGUF", MessagesFormatterType.MISTRAL],
211
+ "EviOmni-nq_train-7B.Q5_K_M.gguf": ["mradermacher/EviOmni-nq_train-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
212
+ "NuMarkdown-8B-Thinking-fork.Q5_K_M.gguf": ["mradermacher/NuMarkdown-8B-Thinking-fork-GGUF", MessagesFormatterType.OPEN_CHAT],
213
+ "Comet_12B_V.7.Q4_K_M.gguf": ["mradermacher/Comet_12B_V.7-GGUF", MessagesFormatterType.ALPACA],
214
+ "Llama-3.1-8B-Instruct-wildfeedback-seed-RPO-0.1.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-Instruct-wildfeedback-seed-RPO-0.1-GGUF", MessagesFormatterType.LLAMA_3],
215
+ "NeuralMerge-9B-Dare.Q5_K_M.gguf": ["mradermacher/NeuralMerge-9B-Dare-GGUF", MessagesFormatterType.MISTRAL],
216
+ "CaptainErisNebula-12B-Chimera-v0.420.Q4_K_M.gguf": ["mradermacher/CaptainErisNebula-12B-Chimera-v0.420-GGUF", MessagesFormatterType.MISTRAL],
217
+ "Nemo-12B-OldSpice.Q4_K_M.gguf": ["mradermacher/Nemo-12B-OldSpice-GGUF", MessagesFormatterType.MISTRAL],
218
+ "funny-nemo-embedding-testing.Q4_K_M.gguf": ["mradermacher/funny-nemo-embedding-testing-GGUF", MessagesFormatterType.MISTRAL],
219
+ "francois-v3.Q4_K_M.gguf": ["mradermacher/francois-v3-GGUF", MessagesFormatterType.CHATML],
220
+ "KansenSakura-Eclipse-RP-12b.Q4_K_M.gguf": ["mradermacher/KansenSakura-Eclipse-RP-12b-GGUF", MessagesFormatterType.CHATML],
221
+ "Sugoi-14B-Ultra-HF.Q4_K_M.gguf": ["mradermacher/Sugoi-14B-Ultra-HF-GGUF", MessagesFormatterType.OPEN_CHAT],
222
+ "CaptainErisNebula-12B-Chimera-v1.1.i1-Q4_K_M.gguf": ["mradermacher/CaptainErisNebula-12B-Chimera-v1.1-i1-GGUF", MessagesFormatterType.MISTRAL],
223
+ "NOVER1-Qwen2.5-7B.i1-Q5_K_M.gguf": ["mradermacher/NOVER1-Qwen2.5-7B-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
224
+ "MT-Gen6-U-gemma-2-Tv3Tv2-9B.Q4_K_M.gguf": ["mradermacher/MT-Gen6-U-gemma-2-Tv3Tv2-9B-GGUF", MessagesFormatterType.ALPACA],
225
+ "SimpleChat-4B-V1.i1-Q5_K_M.gguf": ["mradermacher/SimpleChat-4B-V1-i1-GGUF", MessagesFormatterType.ALPACA],
226
+ "MathTutor-7B-MDPO_v0.1.Q5_K_M.gguf": ["mradermacher/MathTutor-7B-MDPO_v0.1-GGUF", MessagesFormatterType.OPEN_CHAT],
227
+ "MathTutor-7B-H_v0.0.1.Q5_K_M.gguf": ["mradermacher/MathTutor-7B-H_v0.0.1-GGUF", MessagesFormatterType.OPEN_CHAT],
228
+ "Wingless-Alice-RP-NSFW-8B.Q4_K_S.gguf": ["mradermacher/Wingless-Alice-RP-NSFW-8B-GGUF", MessagesFormatterType.LLAMA_3],
229
+ "Shisa-K-12B.Q4_K_M.gguf": ["mradermacher/Shisa-K-12B-GGUF", MessagesFormatterType.MISTRAL],
230
+ "UI-Venus-Ground-7B.Q5_K_M.gguf": ["mradermacher/UI-Venus-Ground-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
231
+ "Smilodon-9B-v0.5.i1-Q4_K_M.gguf": ["mradermacher/Smilodon-9B-v0.5-i1-GGUF", MessagesFormatterType.ALPACA],
232
+ "Qwen2.5-VL-7B-NSFW-Caption-V3.Q5_K_M.gguf": ["mradermacher/Qwen2.5-VL-7B-NSFW-Caption-V3-GGUF", MessagesFormatterType.OPEN_CHAT],
233
+ "Arrow-R1-7B.Q5_K_M.gguf": ["mradermacher/Arrow-R1-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
234
+ "Moondark-12B.i1-Q4_K_M.gguf": ["mradermacher/Moondark-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
235
+ "Moonviolet-12B.Q4_K_M.gguf": ["mradermacher/Moonviolet-12B-GGUF", MessagesFormatterType.MISTRAL],
236
+ "MoonMega-12B.Q4_K_M.gguf": ["mradermacher/MoonMega-12B-GGUF", MessagesFormatterType.MISTRAL],
237
+ "Qwen2.5-7B-Instruct-wildfeedback-iterDPO-iter2-4k.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-wildfeedback-iterDPO-iter2-4k-GGUF", MessagesFormatterType.OPEN_CHAT],
238
+ "Josiefied-Qwen3-4B-Instruct-2507-gabliterated-v1.i1-Q5_K_M.gguf": ["mradermacher/Josiefied-Qwen3-4B-Instruct-2507-gabliterated-v1-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
239
+ "Qwen2.5-14B-Instruct-wildfeedback-RPO-DRIFT-iter1-4k.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Instruct-wildfeedback-RPO-DRIFT-iter1-4k-GGUF", MessagesFormatterType.OPEN_CHAT],
240
+ "gemma3-12b-tolkien_v3.Q4_K_M.gguf": ["mradermacher/gemma3-12b-tolkien_v3-GGUF", MessagesFormatterType.ALPACA],
241
+ "DellaMix-12B.i1-Q4_K_M.gguf": ["mradermacher/DellaMix-12B-i1-GGUF", MessagesFormatterType.CHATML],
242
+ "EtherealAurora-12B-Lorablated.Q4_K_M.gguf": ["mradermacher/EtherealAurora-12B-Lorablated-GGUF", MessagesFormatterType.MISTRAL],
243
+ "Not-WizardLM-2-The-Omega-Directive-7b-Unslop-v2.1.i1-Q5_K_M.gguf": ["mradermacher/Not-WizardLM-2-The-Omega-Directive-7b-Unslop-v2.1-i1-GGUF", MessagesFormatterType.MISTRAL],
244
+ "gemma3-12b-tolkien_v2.i1-Q4_K_M.gguf": ["mradermacher/gemma3-12b-tolkien_v2-i1-GGUF", MessagesFormatterType.ALPACA],
245
+ "Qwen-2.5-3b-Evol-CoT.Q5_K_M.gguf": ["mradermacher/Qwen-2.5-3b-Evol-CoT-GGUF", MessagesFormatterType.OPEN_CHAT],
246
+ "VersaVid-R1.Q5_K_M.gguf": ["mradermacher/VersaVid-R1-GGUF", MessagesFormatterType.OPEN_CHAT],
247
+ "sphinxnautics-7b-kl-distilled.Q5_K_M.gguf": ["mradermacher/sphinxnautics-7b-kl-distilled-GGUF", MessagesFormatterType.OPEN_CHAT],
248
+ "gemma3-12b-tolkien.Q5_K_M.gguf": ["mradermacher/gemma3-12b-tolkien-GGUF", MessagesFormatterType.ALPACA],
249
+ "ThinkLite-Critic-7b.Q5_K_M.gguf": ["mradermacher/ThinkLite-Critic-7b-GGUF", MessagesFormatterType.OPEN_CHAT],
250
+ "Qwen2.5-nsfw.Q5_K_M.gguf": ["mradermacher/Qwen2.5-nsfw-GGUF", MessagesFormatterType.OPEN_CHAT],
251
+ "Rewiz-Tom-7B.Q5_K_M.gguf": ["mradermacher/Rewiz-Tom-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
252
+ "NextCoder-Mirage-sol-7B.i1-Q5_K_M.gguf": ["mradermacher/NextCoder-Mirage-sol-7B-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
253
+ "qwen2.5-7b-v1.Q5_K_M.gguf": ["mradermacher/qwen2.5-7b-v1-GGUF", MessagesFormatterType.OPEN_CHAT],
254
+ "CriticLeanGPT-Qwen2.5-7B-Instruct-SFT.i1-Q5_K_M.gguf": ["mradermacher/CriticLeanGPT-Qwen2.5-7B-Instruct-SFT-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
255
+ "OpenR1-Qwen-7B-Instruct-SFT-lora-ctx16k.Q5_K_M.gguf": ["mradermacher/OpenR1-Qwen-7B-Instruct-SFT-lora-ctx16k-GGUF", MessagesFormatterType.OPEN_CHAT],
256
+ "CriticLeanGPT-Qwen2.5-14B-Instruct-SFT-RL.Q4_K_M.gguf": ["mradermacher/CriticLeanGPT-Qwen2.5-14B-Instruct-SFT-RL-GGUF", MessagesFormatterType.OPEN_CHAT],
257
+ "Qwen3-14B-SFT.Q5_K_M.gguf": ["mradermacher/Qwen3-14B-SFT-GGUF", MessagesFormatterType.OPEN_CHAT],
258
+ "Blossom-V6.1-8B.Q5_K_M.gguf": ["mradermacher/Blossom-V6.1-8B-GGUF", MessagesFormatterType.OPEN_CHAT],
259
+ "Homunculus-abliterated.Q4_K_M.gguf": ["mradermacher/Homunculus-abliterated-GGUF", MessagesFormatterType.CHATML],
260
+ "GThinker-7B.i1-Q5_K_M.gguf": ["mradermacher/GThinker-7B-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
261
+ "TAMA-QWen2.5.i1-Q5_K_M.gguf": ["mradermacher/TAMA-QWen2.5-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
262
+ "Qwen3-psychological-reasoning-1.7B.Q5_K_M.gguf": ["mradermacher/Qwen3-psychological-reasoning-1.7B-GGUF", MessagesFormatterType.OPEN_CHAT],
263
+ "QuasiStarSynth-12B.Q4_K_M.gguf": ["mradermacher/QuasiStarSynth-12B-GGUF", MessagesFormatterType.MISTRAL],
264
+ "Solution-back-7B.Q5_K_M.gguf": ["mradermacher/Solution-back-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
265
+ "Qwen2.5-3B-UFO-hotpotqa-1turn.Q5_K_M.gguf": ["mradermacher/Qwen2.5-3B-UFO-hotpotqa-1turn-GGUF", MessagesFormatterType.OPEN_CHAT],
266
+ "AbyssSynth-12B.Q4_K_M.gguf": ["mradermacher/AbyssSynth-12B-GGUF", MessagesFormatterType.MISTRAL],
267
+ "E-Star-12B-v0.1.Q4_K_M.gguf": ["mradermacher/E-Star-12B-v0.1-GGUF", MessagesFormatterType.ALPACA],
268
+ "SingularitySynth-12B.i1-Q4_K_M.gguf": ["mradermacher/SingularitySynth-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
269
+ "MT2-Gen2_gemma-3-12B.Q4_K_M.gguf": ["mradermacher/MT2-Gen2_gemma-3-12B-GGUF", MessagesFormatterType.ALPACA],
270
+ "LilithCore-v1-12B.Q4_K_M.gguf": ["mradermacher/LilithCore-v1-12B-GGUF", MessagesFormatterType.MISTRAL],
271
+ "WeThink-Qwen2.5VL-7B.i1-Q5_K_M.gguf": ["mradermacher/WeThink-Qwen2.5VL-7B-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
272
+ "RuleReasoner-8B.Q5_K_M.gguf": ["mradermacher/RuleReasoner-8B-GGUF", MessagesFormatterType.OPEN_CHAT],
273
+ "Vision-Matters-7B.Q5_K_M.gguf": ["mradermacher/Vision-Matters-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
274
+ "gemma-3-cn-novel-4b.Q5_K_M.gguf": ["mradermacher/gemma-3-cn-novel-4b-GGUF", MessagesFormatterType.ALPACA],
275
+ "Qwen2.5-7B-Instruct-ultrafeedback-10k.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-ultrafeedback-10k-GGUF", MessagesFormatterType.OPEN_CHAT],
276
+ "RACRO-7B-CRO.i1-Q5_K_M.gguf": ["mradermacher/RACRO-7B-CRO-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
277
+ "QVikhr-3-4B-Instruction.Q6_K.gguf": ["mradermacher/QVikhr-3-4B-Instruction-GGUF", MessagesFormatterType.OPEN_CHAT],
278
+ "care-japanese-gemma2-9b.Q4_K_M.gguf": ["mradermacher/care-japanese-gemma2-9b-GGUF", MessagesFormatterType.ALPACA],
279
+ "Geo-SFT.i1-Q5_K_M.gguf": ["mradermacher/Geo-SFT-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
280
+ "Geo-R1-CI-0620.Q5_K_M.gguf": ["mradermacher/Geo-R1-CI-0620-GGUF", MessagesFormatterType.OPEN_CHAT],
281
+ "SpaceQwen2.5-VL-3B-Instruct.Q5_K_M.gguf": ["mradermacher/SpaceQwen2.5-VL-3B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
282
+ "Metis-RISE-7B.i1-Q5_K_M.gguf": ["mradermacher/Metis-RISE-7B-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
283
+ "OpenBuddy-R10528DistillQwen-14B-v27.1.i1-Q4_K_M.gguf": ["mradermacher/OpenBuddy-R10528DistillQwen-14B-v27.1-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
284
+ "sft_subtitles_fixer_v1_merged.Q4_K_M.gguf": ["mradermacher/sft_subtitles_fixer_v1_merged-GGUF", MessagesFormatterType.ALPACA],
285
+ "Qwen3-8B-Josiefied-iSMART.Q6_K.gguf": ["mradermacher/Qwen3-8B-Josiefied-iSMART-GGUF", MessagesFormatterType.OPEN_CHAT],
286
+ "Polar-G3-DPO-12b.i1-Q4_K_M.gguf": ["mradermacher/Polar-G3-DPO-12b-i1-GGUF", MessagesFormatterType.ALPACA],
287
+ "DREX-062225-exp.Q5_K_M.gguf": ["mradermacher/DREX-062225-exp-GGUF", MessagesFormatterType.OPEN_CHAT],
288
+ "Japanese-Qwen2.5-14B-Instruct-V1.Q4_K_M.gguf": ["mradermacher/Japanese-Qwen2.5-14B-Instruct-V1-GGUF", MessagesFormatterType.OPEN_CHAT],
289
+ "Qwen2.5-7B-Instruct-userfeedback-on-policy-iter2.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-userfeedback-on-policy-iter2-GGUF", MessagesFormatterType.OPEN_CHAT],
290
+ "Configurable-Hermes-3-Llama-3.1-8B.Q5_K_M.gguf": ["mradermacher/Configurable-Hermes-3-Llama-3.1-8B-GGUF", MessagesFormatterType.LLAMA_3],
291
+ "KiloNovaSynth-12B.Q4_K_M.gguf": ["mradermacher/KiloNovaSynth-12B-GGUF", MessagesFormatterType.MISTRAL],
292
+ "Qwen2.5-7B-Instruct-userfeedback-on-policy-iter1.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-userfeedback-on-policy-iter1-GGUF", MessagesFormatterType.OPEN_CHAT],
293
+ "Llama-3.1-Swallow-8B-Instruct-v0.5-Q5_K_M.gguf": ["mmnga/Llama-3.1-Swallow-8B-Instruct-v0.5-gguf", MessagesFormatterType.LLAMA_3],
294
+ "Bohdi-Qwen2.5-7B-Instruct.i1-Q5_K_M.gguf": ["mradermacher/Bohdi-Qwen2.5-7B-Instruct-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
295
+ "Q2.5-ColdBrew-RxV.Q4_K_M.gguf": ["mradermacher/Q2.5-ColdBrew-RxV-GGUF", MessagesFormatterType.OPEN_CHAT],
296
+ "SuperNovaSynth-12B.Q4_K_M.gguf": ["mradermacher/SuperNovaSynth-12B-GGUF", MessagesFormatterType.MISTRAL],
297
+ "HyperNovaSynth-12B.Q4_K_M.gguf": ["mradermacher/HyperNovaSynth-12B-GGUF", MessagesFormatterType.MISTRAL],
298
+ "Simulation_LLM_wiki_7B_V2.Q5_K_M.gguf": ["mradermacher/Simulation_LLM_wiki_7B_V2-GGUF", MessagesFormatterType.OPEN_CHAT],
299
+ "Qwen2.5-7B-Instruct-userfeedback-SPIN-iter1.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-userfeedback-SPIN-iter1-GGUF", MessagesFormatterType.OPEN_CHAT],
300
+ "Qwen2.5-7B-Instruct-userfeedback-SFT-SPIN-iter1.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-userfeedback-SFT-SPIN-iter1-GGUF", MessagesFormatterType.OPEN_CHAT],
301
+ "PatriSlush-DarkReign-12B.Q4_K_M.gguf": ["mradermacher/PatriSlush-DarkReign-12B-GGUF", MessagesFormatterType.MISTRAL],
302
+ "Hermes-2-Theta-Llama-3-8B-32k.i1-Q5_K_M.gguf": ["mradermacher/Hermes-2-Theta-Llama-3-8B-32k-i1-GGUF", MessagesFormatterType.LLAMA_3],
303
+ "QFFT-S1-7B.Q5_K_M.gguf": ["mradermacher/QFFT-S1-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
304
+ "QFFT-LIMO-7B.Q5_K_M.gguf": ["mradermacher/QFFT-LIMO-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
305
+ "Omega-Darker_The-Final-Directive-Longform-Stage2-ERP-12B-v0.2.Q4_K_M.gguf": ["mradermacher/Omega-Darker_The-Final-Directive-Longform-Stage2-ERP-12B-v0.2-GGUF", MessagesFormatterType.MISTRAL],
306
+ "Qwen2.5-7B-Instruct-userfeedback-sentiment-seed.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-userfeedback-sentiment-seed-GGUF", MessagesFormatterType.OPEN_CHAT],
307
+ "Irix-12B_Slush_V2.i1-Q4_K_M.gguf": ["mradermacher/Irix-12B_Slush_V2-i1-GGUF", MessagesFormatterType.MISTRAL],
308
+ "NeonViolin.Q4_K_M.gguf": ["mradermacher/NeonViolin-GGUF", MessagesFormatterType.MISTRAL],
309
+ "Qwen2.5-3B-rat-polished.Q5_K_M.gguf": ["mradermacher/Qwen2.5-3B-rat-polished-GGUF", MessagesFormatterType.OPEN_CHAT],
310
+ "Mistral-qwq-12b-merge.Q4_K_M.gguf": ["mradermacher/Mistral-qwq-12b-merge-GGUF", MessagesFormatterType.MISTRAL],
311
+ "Situations-Model-1.1.Q5_K_M.gguf": ["mradermacher/Situations-Model-1.1-GGUF", MessagesFormatterType.OPEN_CHAT],
312
+ "Ego-R1-SFT-3B.Q5_K_M.gguf": ["mradermacher/Ego-R1-SFT-3B-GGUF", MessagesFormatterType.OPEN_CHAT],
313
+ "Ego-R1-Agent-3B.Q5_K_M.gguf": ["mradermacher/Ego-R1-Agent-3B-GGUF", MessagesFormatterType.OPEN_CHAT],
314
+ "JudgeLRM-14B.Q4_K_M.gguf": ["mradermacher/JudgeLRM-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
315
+ "Irix_MemecapSlerp-12B.Q4_K_M.gguf": ["mradermacher/Irix_MemecapSlerp-12B-GGUF", MessagesFormatterType.MISTRAL],
316
+ "Irixxed-Magcap-12B-Slerp.Q4_K_M.gguf": ["mradermacher/Irixxed-Magcap-12B-Slerp-GGUF", MessagesFormatterType.MISTRAL],
317
+ "The-Omega-Directive-M-12B-Unslop-v2.0.Q4_K_M.gguf": ["mradermacher/The-Omega-Directive-M-12B-Unslop-v2.0-GGUF", MessagesFormatterType.MISTRAL],
318
+ "ZYH-LLM-Qwen2.5-14B-V5.Q4_K_M.gguf": ["mradermacher/ZYH-LLM-Qwen2.5-14B-V5-GGUF", MessagesFormatterType.OPEN_CHAT],
319
+ "LorablatedStock-12B.Q4_K_M.gguf": ["mradermacher/LorablatedStock-12B-GGUF", MessagesFormatterType.CHATML],
320
+ "Qwen2.5-7B-Chess-MMXL.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Chess-MMXL-GGUF", MessagesFormatterType.OPEN_CHAT],
321
+ "Mistral-Nemo-Gutenberg-Encore-12B.Q4_K_M.gguf": ["mradermacher/Mistral-Nemo-Gutenberg-Encore-12B-GGUF", MessagesFormatterType.MISTRAL],
322
+ "Qwen2.5-7B-Instruct-wildfeedback.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-wildfeedback-GGUF", MessagesFormatterType.OPEN_CHAT],
323
+ "Shisa-v2-Mistral-Nemo-12B-Lorablated.Q4_K_M.gguf": ["mradermacher/Shisa-v2-Mistral-Nemo-12B-Lorablated-GGUF", MessagesFormatterType.MISTRAL],
324
+ "L3-Dark-Planet-8B-wordstorm2.Q5_K_M.gguf": ["mradermacher/L3-Dark-Planet-8B-wordstorm2-GGUF", MessagesFormatterType.LLAMA_3],
325
+ "Qwen2.5-7B-Instruct-userfeedback.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-userfeedback-GGUF", MessagesFormatterType.OPEN_CHAT],
326
+ "L3-Dark-Planet-8B-wordstorm-cr2.Q5_K_M.gguf": ["mradermacher/L3-Dark-Planet-8B-wordstorm-cr2-GGUF", MessagesFormatterType.LLAMA_3],
327
+ "L3-Dark-Planet-8B-wordstorm-b6.i1-Q5_K_M.gguf": ["mradermacher/L3-Dark-Planet-8B-wordstorm-b6-i1-GGUF", MessagesFormatterType.LLAMA_3],
328
+ "Gilded-Tempest-12B.Q4_K_S.gguf": ["mradermacher/Gilded-Tempest-12B-GGUF", MessagesFormatterType.MISTRAL],
329
+ "Qwen2.5-7B-Instruct-GRPO-Nei-Tokens.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-GRPO-Nei-Tokens-GGUF", MessagesFormatterType.OPEN_CHAT],
330
+ "FusionEngine-12B.Q4_K_M.gguf": ["mradermacher/FusionEngine-12B-GGUF", MessagesFormatterType.CHATML],
331
+ "ShiZhen-V0.1.Q5_K_M.gguf": ["mradermacher/ShiZhen-V0.1-GGUF", MessagesFormatterType.OPEN_CHAT],
332
+ "HMS-Fusion-12B.Q4_K_M.gguf": ["mradermacher/HMS-Fusion-12B-GGUF", MessagesFormatterType.MISTRAL],
333
+ "TutorRL-7B.Q5_K_M.gguf": ["mradermacher/TutorRL-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
334
+ "Omega-Darker_The-Final-Directive-14B.Q4_K_M.gguf": ["mradermacher/Omega-Darker_The-Final-Directive-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
335
+ "Student-rl_Qwen2.5-7B-Instruct_v0.0.1.Q5_K_M.gguf": ["mradermacher/Student-rl_Qwen2.5-7B-Instruct_v0.0.1-GGUF", MessagesFormatterType.OPEN_CHAT],
336
+ "Contrl-Stheno-v1-8B.Q5_K_M.gguf": ["mradermacher/Contrl-Stheno-v1-8B-GGUF", MessagesFormatterType.LLAMA_3],
337
+ "fluxiia_14b.Q4_K_M.gguf": ["mradermacher/fluxiia_14b-GGUF", MessagesFormatterType.OPEN_CHAT],
338
+ "JwenJwen1-1.5b.Q5_K_M.gguf": ["mradermacher/JwenJwen1-1.5b-GGUF", MessagesFormatterType.OPEN_CHAT],
339
+ "SecGPT-14B-Reasoning.Q4_K_M.gguf": ["mradermacher/SecGPT-14B-Reasoning-GGUF", MessagesFormatterType.OPEN_CHAT],
340
+ "Dirty-Muse-Writer-v01-Uncensored-Erotica-NSFW.Q4_K_M.gguf": ["mradermacher/Dirty-Muse-Writer-v01-Uncensored-Erotica-NSFW-GGUF", MessagesFormatterType.ALPACA],
341
+ "SearchR1-nq_hotpotqa_train-qwen2.5-7b-em-ppo.Q4_K_M.gguf": ["mradermacher/SearchR1-nq_hotpotqa_train-qwen2.5-7b-em-ppo-GGUF", MessagesFormatterType.OPEN_CHAT],
342
+ "ForgottenMaid-12B-v2.Q4_K_M.gguf": ["mradermacher/ForgottenMaid-12B-v2-GGUF", MessagesFormatterType.MISTRAL],
343
+ "Socratic-Qwen2.5-7B-v2.Q5_K_M.gguf": ["mradermacher/Socratic-Qwen2.5-7B-v2-GGUF", MessagesFormatterType.OPEN_CHAT],
344
+ "EsotericSage-12B.i1-Q4_K_M.gguf": ["mradermacher/EsotericSage-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
345
+ "praxis-bookwriter-llama3.1-8b-sft.Q5_K_M.gguf": ["mradermacher/praxis-bookwriter-llama3.1-8b-sft-GGUF", MessagesFormatterType.LLAMA_3],
346
+ "LinearWriter-12B.Q4_K_M.gguf": ["mradermacher/LinearWriter-12B-GGUF", MessagesFormatterType.MISTRAL],
347
+ "Qwen2.5-CoderX-14B-v0.5.Q4_K_M.gguf": ["mradermacher/Qwen2.5-CoderX-14B-v0.5-GGUF", MessagesFormatterType.OPEN_CHAT],
348
+ "Violet_MagCap-Rebase-12B.Q4_K_M.gguf": ["mradermacher/Violet_MagCap-Rebase-12B-GGUF", MessagesFormatterType.MISTRAL],
349
+ "VeriReason-Qwen2.5-7b-SFT-Reasoning.i1-Q5_K_M.gguf": ["mradermacher/VeriReason-Qwen2.5-7b-SFT-Reasoning-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
350
+ "Emilia-Multislerp-12B.Q4_K_M.gguf": ["mradermacher/Emilia-Multislerp-12B-GGUF", MessagesFormatterType.CHATML],
351
+ "TaskRouter-1.5B.Q5_K_M.gguf": ["mradermacher/TaskRouter-1.5B-GGUF", MessagesFormatterType.OPEN_CHAT],
352
+ "NeonMaid-12B-v2.Q4_K_M.gguf": ["mradermacher/NeonMaid-12B-v2-GGUF", MessagesFormatterType.CHATML],
353
+ "Orihime-Gutenberg-12B.Q4_K_M.gguf": ["mradermacher/Orihime-Gutenberg-12B-GGUF", MessagesFormatterType.CHATML],
354
+ "SearchSimulation_3B.Q5_K_M.gguf": ["mradermacher/SearchSimulation_3B-GGUF", MessagesFormatterType.OPEN_CHAT],
355
+ "SearchSimulation_14B.Q4_K_M.gguf": ["mradermacher/SearchSimulation_14B-GGUF", MessagesFormatterType.OPEN_CHAT],
356
+ "NeonMaid-12B.Q4_K_M.gguf": ["mradermacher/NeonMaid-12B-GGUF", MessagesFormatterType.CHATML],
357
+ "Chief_chain_generator-Qwen2.5-7B-Instruct.Q5_K_M.gguf": ["mradermacher/Chief_chain_generator-Qwen2.5-7B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
358
+ "Emotion_inferencer-Qwen2.5-7B-Instruct.Q5_K_M.gguf": ["mradermacher/Emotion_inferencer-Qwen2.5-7B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
359
+ "ForgottenMaid-12B.Q4_K_M.gguf": ["mradermacher/ForgottenMaid-12B-GGUF", MessagesFormatterType.MISTRAL],
360
+ "AceCoder-Qwen2.5-Coder-7B-Ins-V1.1.Q5_K_M.gguf": ["mradermacher/AceCoder-Qwen2.5-Coder-7B-Ins-V1.1-GGUF", MessagesFormatterType.OPEN_CHAT],
361
+ "TheDrummer_Rivermind-Lux-12B-v1-Q4_K_M.gguf": ["bartowski/TheDrummer_Rivermind-Lux-12B-v1-GGUF", MessagesFormatterType.MISTRAL],
362
+ "Qwen2.5-GenX-14B.Q4_K_M.gguf": ["mradermacher/Qwen2.5-GenX-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
363
+ "vanilla-cn-roleplay-0.1.Q5_K_M.gguf": ["mradermacher/vanilla-cn-roleplay-0.1-GGUF", MessagesFormatterType.OPEN_CHAT],
364
+ "EsotericLight-12B.Q4_K_M.gguf": ["mradermacher/EsotericLight-12B-GGUF", MessagesFormatterType.CHATML],
365
+ "llama-3-Daredevil-Mahou-8B.Q5_K_M.gguf": ["mradermacher/llama-3-Daredevil-Mahou-8B-GGUF", MessagesFormatterType.LLAMA_3],
366
+ "Orihime-12B.Q4_K_M.gguf": ["mradermacher/Orihime-12B-GGUF", MessagesFormatterType.CHATML],
367
+ "YM-12B-Model_Stock.Q4_K_M.gguf": ["mradermacher/YM-12B-Model_Stock-GGUF", MessagesFormatterType.MISTRAL],
368
+ "Qwen2.5-7B-Instruct-HardLambda0.1-220.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-HardLambda0.1-220-GGUF", MessagesFormatterType.OPEN_CHAT],
369
+ "gemma-2-abliterated-Ifable-9B-untied.Q4_K_M.gguf": ["mradermacher/gemma-2-abliterated-Ifable-9B-untied-GGUF", MessagesFormatterType.ALPACA],
370
+ "Smoothie-Qwen2.5-3B-Instruct.i1-Q5_K_M.gguf": ["mradermacher/Smoothie-Qwen2.5-3B-Instruct-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
371
+ "Qwen2.5-14B-Aletheia-cogito.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Aletheia-cogito-GGUF", MessagesFormatterType.OPEN_CHAT],
372
+ "Qwen2.5-14B-EMPO-Natural-Reasoning_simple_full.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-EMPO-Natural-Reasoning_simple_full-GGUF", MessagesFormatterType.OPEN_CHAT],
373
+ "exurbia-advance-q4_k_m.gguf": ["ClaudioItaly/Exurbia-Advance-Q4_K_M-GGUF", MessagesFormatterType.ALPACA],
374
+ "Levelled-12B-Model_Stock.Q4_K_M.gguf": ["mradermacher/Levelled-12B-Model_Stock-GGUF", MessagesFormatterType.MISTRAL],
375
+ "sapie-gemma2-9B-IT.i1-Q4_K_M.gguf": ["mradermacher/sapie-gemma2-9B-IT-i1-GGUF", MessagesFormatterType.ALPACA],
376
+ "Persona-Qwen2.5-3B-Instruct.Q5_K_M.gguf": ["mradermacher/Persona-Qwen2.5-3B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
377
+ "Socratic-Qwen2.5-7B.Q5_K_M.gguf": ["mradermacher/Socratic-Qwen2.5-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
378
+ "qwen2.5-1.5B-base-abliterated.Q5_K_M.gguf": ["mradermacher/qwen2.5-1.5B-base-abliterated-GGUF", MessagesFormatterType.OPEN_CHAT],
379
+ "Rei-V3-KTO-12B.Q4_K_M.gguf": ["mradermacher/Rei-V3-KTO-12B-GGUF", MessagesFormatterType.CHATML],
380
+ "Qwen2.5-14B-Instinct-Talk.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Instinct-Talk-GGUF", MessagesFormatterType.OPEN_CHAT],
381
+ "Llama3.1-Aloe-Beta-8B.Q5_K_M.gguf": ["mradermacher/Llama3.1-Aloe-Beta-8B-GGUF", MessagesFormatterType.LLAMA_3],
382
+ "gemma-2-9B-it-blend.Q4_K_S.gguf": ["mradermacher/gemma-2-9B-it-blend-GGUF", MessagesFormatterType.ALPACA],
383
+ "recoilme-gemma-2-Ataraxy-9B-v0.1-t0.75.Q4_K_M.gguf": ["mradermacher/recoilme-gemma-2-Ataraxy-9B-v0.1-t0.75-GGUF", MessagesFormatterType.ALPACA],
384
+ "Smoothie-Qwen2.5-1.5B-Instruct.Q5_K_M.gguf": ["mradermacher/Smoothie-Qwen2.5-1.5B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
385
+ "ELYZA-Shortcut-1.0-Qwen-7B-Q5_K_M.gguf": ["mmnga/ELYZA-Shortcut-1.0-Qwen-7B-gguf", MessagesFormatterType.OPEN_CHAT],
386
+ "Qwen2.5-7B-Instruct-abliterated.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-abliterated-GGUF", MessagesFormatterType.OPEN_CHAT],
387
+ "Crispers-14B-v1.Q4_K_M.gguf": ["mradermacher/Crispers-14B-v1-GGUF", MessagesFormatterType.OPEN_CHAT],
388
+ "amoral-cogito-Zara-14B-ties.i1-Q4_K_M.gguf": ["mradermacher/amoral-cogito-Zara-14B-ties-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
389
+ "Qwen2.5-Kunoulise-D.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Kunoulise-D-GGUF", MessagesFormatterType.OPEN_CHAT],
390
+ "Qwen2.5-Kunoulise-C.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Kunoulise-C-GGUF", MessagesFormatterType.OPEN_CHAT],
391
+ "Qwen2.5-Kunoulise-B.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Kunoulise-B-GGUF", MessagesFormatterType.OPEN_CHAT],
392
+ "Qwen2.5-Kunoulise-A.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Kunoulise-A-GGUF", MessagesFormatterType.OPEN_CHAT],
393
+ "Crispers-7B-v1.i1-Q5_K_M.gguf": ["mradermacher/Crispers-7B-v1-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
394
+ "SnowElf-12B.Q4_K_M.gguf": ["mradermacher/SnowElf-12B-GGUF", MessagesFormatterType.CHATML],
395
+ "Fusion4-14B-Instruct.Q4_K_M.gguf": ["mradermacher/Fusion4-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
396
+ "VeriCoder_Qwen14B.Q4_K_M.gguf": ["mradermacher/VeriCoder_Qwen14B-GGUF", MessagesFormatterType.OPEN_CHAT],
397
+ "hms-slerp-12b-v2-q4_k_m.gguf": ["yamatazen/HMS-Slerp-12B-v2-Q4_K_M-GGUF", MessagesFormatterType.MISTRAL],
398
+ "grpo-qwen7b-triton-10ep.Q5_K_M.gguf": ["mradermacher/grpo-qwen7b-triton-10ep-GGUF", MessagesFormatterType.OPEN_CHAT],
399
+ "grpo-qwen7b-triton-5ep.Q5_K_M.gguf": ["mradermacher/grpo-qwen7b-triton-5ep-GGUF", MessagesFormatterType.OPEN_CHAT],
400
+ "Kuugo-LawLLM-14B.Q4_K_M.gguf": ["mradermacher/Kuugo-LawLLM-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
401
+ "T3Q-qwen2.5-14b-v1.2-e2.Q4_K_M.gguf": ["mradermacher/T3Q-qwen2.5-14b-v1.2-e2-GGUF", MessagesFormatterType.OPEN_CHAT],
402
+ "ablation-200-a163.finaldpo2.1e7-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-200-a163.finaldpo2.1e7-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
403
+ "ablation-204-a163.finaldpo2.2e7-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-204-a163.finaldpo2.2e7-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
404
+ "Vera-v0.1.Q5_K_M.gguf": ["mradermacher/Vera-v0.1-GGUF", MessagesFormatterType.MISTRAL],
405
+ "ablation-192-a175.finaldpo2.8e7-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-192-a175.finaldpo2.8e7-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
406
+ "MagSoup-v1-12B.Q4_K_M.gguf": ["mradermacher/MagSoup-v1-12B-GGUF", MessagesFormatterType.MISTRAL],
407
+ "LUFFY-Qwen-Instruct-7B.Q5_K_M.gguf": ["mradermacher/LUFFY-Qwen-Instruct-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
408
+ "VeriPrefer-Qwen2.5-Coder-7B.Q5_K_M.gguf": ["mradermacher/VeriPrefer-Qwen2.5-Coder-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
409
+ "ToolRL-Qwen2.5-3B.Q5_K_M.gguf": ["mradermacher/ToolRL-Qwen2.5-3B-GGUF", MessagesFormatterType.OPEN_CHAT],
410
+ "TaxoQwen2.5-7B.Q5_K_M.gguf": ["mradermacher/TaxoQwen2.5-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
411
+ "Magnolia-v8-12B.Q4_K_M.gguf": ["mradermacher/Magnolia-v8-12B-GGUF", MessagesFormatterType.MISTRAL],
412
+ "Llama3.2-3B-Reasoning-SFT-fTuned-v1.Q5_K_M.gguf": ["mradermacher/Llama3.2-3B-Reasoning-SFT-fTuned-v1-GGUF", MessagesFormatterType.LLAMA_3],
413
+ "Qwen2.5-Freyalise-A.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Freyalise-A-GGUF", MessagesFormatterType.OPEN_CHAT],
414
+ "Magnolia-v9-12B.Q4_K_M.gguf": ["mradermacher/Magnolia-v9-12B-GGUF", MessagesFormatterType.MISTRAL],
415
+ "Magnolia-v10-12B.Q4_K_M.gguf": ["mradermacher/Magnolia-v10-12B-GGUF", MessagesFormatterType.MISTRAL],
416
+ "sft_3b_wopolicy.Q5_K_M.gguf": ["mradermacher/sft_3b_wopolicy-GGUF", MessagesFormatterType.OPEN_CHAT],
417
+ "Shisa-v2-Mistral-Nemo-12B-Abliterated.Q4_K_M.gguf": ["mradermacher/Shisa-v2-Mistral-Nemo-12B-Abliterated-GGUF", MessagesFormatterType.MISTRAL],
418
+ "Qwen2.5-14B-ZaraTimpist.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-ZaraTimpist-GGUF", MessagesFormatterType.OPEN_CHAT],
419
+ "Chronicle-Qwen2.5-7B.Q5_K_M.gguf": ["mradermacher/Chronicle-Qwen2.5-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
420
+ "Llama-3.1-1million-ctx-Dark-Planet-v1.01-8B.Q5_K_M.gguf": ["mradermacher/Llama-3.1-1million-ctx-Dark-Planet-v1.01-8B-GGUF", MessagesFormatterType.LLAMA_3],
421
+ "MN-Hekate-Pyrtania-12B.Q4_K_M.gguf": ["mradermacher/MN-Hekate-Pyrtania-12B-GGUF", MessagesFormatterType.MISTRAL],
422
+ "Qwen2.5-14B-Instruct-LIMO-new.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Instruct-LIMO-new-GGUF", MessagesFormatterType.OPEN_CHAT],
423
+ "Qwen2.5-14B-Instruct-CognitiveSFT-v0.1.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Instruct-CognitiveSFT-v0.1-GGUF", MessagesFormatterType.OPEN_CHAT],
424
+ "Qwen2.5-14B-Instruct-o4.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Instruct-o4-GGUF", MessagesFormatterType.OPEN_CHAT],
425
+ "Twilight-SCE-12B-v2.Q4_K_S.gguf": ["mradermacher/Twilight-SCE-12B-v2-GGUF", MessagesFormatterType.CHATML],
426
+ "Twilight-SCE-12B.Q4_K_M.gguf": ["mradermacher/Twilight-SCE-12B-GGUF", MessagesFormatterType.CHATML],
427
+ "MN-Hekate-Limenoskopos-12B.Q4_K_M.gguf": ["mradermacher/MN-Hekate-Limenoskopos-12B-GGUF", MessagesFormatterType.MISTRAL],
428
+ "Reformed-Christian-Bible-Expert-v2.1-12B.Q4_K_M.gguf": ["mradermacher/Reformed-Christian-Bible-Expert-v2.1-12B-GGUF", MessagesFormatterType.MISTRAL],
429
+ "AnotherOne-Unslop-Mell-12B.Q4_K_M.gguf": ["mradermacher/AnotherOne-Unslop-Mell-12B-GGUF", MessagesFormatterType.MISTRAL],
430
+ "Golden-Curry-12B.Q4_K_M.gguf": ["mradermacher/Golden-Curry-12B-GGUF", MessagesFormatterType.MISTRAL],
431
+ "Violet_Magcap-12B.i1-Q4_K_M.gguf": ["mradermacher/Violet_Magcap-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
432
+ "StarrySky-12B.Q4_K_M.gguf": ["mradermacher/StarrySky-12B-GGUF", MessagesFormatterType.CHATML],
433
+ "kunoichi-lemon-royale-v2experiment2-32K-7B.i1-Q5_K_M.gguf": ["mradermacher/kunoichi-lemon-royale-v2experiment2-32K-7B-i1-GGUF", MessagesFormatterType.MISTRAL],
434
+ "BlueLight-12B.Q4_K_M.gguf": ["mradermacher/BlueLight-12B-GGUF", MessagesFormatterType.CHATML],
435
+ "kunoichi-lemon-royale-hamansu-v1-32k-7B.Q5_K_M.gguf": ["mradermacher/kunoichi-lemon-royale-hamansu-v1-32k-7B-GGUF", MessagesFormatterType.MISTRAL],
436
+ "Llama-3.1-128k-Dark-Planet-Uncensored-8B.Q5_K_M.gguf": ["mradermacher/Llama-3.1-128k-Dark-Planet-Uncensored-8B-GGUF", MessagesFormatterType.LLAMA_3],
437
+ "Llama-3.1-1-million-ctx-DeepHermes-Deep-Reasoning-8B.Q5_K_M.gguf": ["mradermacher/Llama-3.1-1-million-ctx-DeepHermes-Deep-Reasoning-8B-GGUF", MessagesFormatterType.LLAMA_3],
438
+ "Llama-3.1-1million-ctx-Dark-Planet-8B.Q5_K_M.gguf": ["mradermacher/Llama-3.1-1million-ctx-Dark-Planet-8B-GGUF", MessagesFormatterType.LLAMA_3],
439
+ "kunoichi-lemon-royale-v2experiment1-32K-7B.i1-Q5_K_M.gguf": ["mradermacher/kunoichi-lemon-royale-v2experiment1-32K-7B-i1-GGUF", MessagesFormatterType.MISTRAL],
440
+ "ABEJA-Qwen2.5-7b-Japanese-v0.1-Q5_K_M.gguf": ["mmnga/ABEJA-Qwen2.5-7b-Japanese-v0.1-gguf", MessagesFormatterType.OPEN_CHAT],
441
+ "Txt-Polisher-Douyin-Style.Q4_K_M.gguf": ["mradermacher/Txt-Polisher-Douyin-Style-GGUF", MessagesFormatterType.OPEN_CHAT],
442
+ "Llama-3.1-Nemotron-8B-UltraLong-1M-Instruct.Q4_K_S.gguf": ["mradermacher/Llama-3.1-Nemotron-8B-UltraLong-1M-Instruct-GGUF", MessagesFormatterType.LLAMA_3],
443
+ "dreamgen_lucid-v1-nemo-Q4_K_M.gguf": ["bartowski/dreamgen_lucid-v1-nemo-GGUF", MessagesFormatterType.MISTRAL],
444
+ "TheDrummer_Rivermind-12B-v1-Q4_K_M.gguf": ["bartowski/TheDrummer_Rivermind-12B-v1-GGUF", MessagesFormatterType.MISTRAL],
445
+ "Josiefied-Qwen2.5-Coder-14B-Instruct-abliterated-v1.Q4_K_M.gguf": ["mradermacher/Josiefied-Qwen2.5-Coder-14B-Instruct-abliterated-v1-GGUF", MessagesFormatterType.OPEN_CHAT],
446
+ "KunoRZN-Llama-3-3B.Q5_K_M.gguf": ["mradermacher/KunoRZN-Llama-3-3B-GGUF", MessagesFormatterType.LLAMA_3],
447
+ "L3-8B-Niitamamo-v0.1-della.Q5_K_M.gguf": ["mradermacher/L3-8B-Niitamamo-v0.1-della-GGUF", MessagesFormatterType.LLAMA_3],
448
+ "Qwen2.5-14B-Imstinct.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Imstinct-GGUF", MessagesFormatterType.OPEN_CHAT],
449
+ "Pearl-14B.Q4_K_M.gguf": ["mradermacher/Pearl-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
450
+ "Qwen2.5-14B-Instinct-RP.i1-Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Instinct-RP-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
451
+ "Qwen2.5-Recursive-Coder-14B-Instruct.i1-Q4_K_M.gguf": ["mradermacher/Qwen2.5-Recursive-Coder-14B-Instruct-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
452
+ "nemo-12b-instruct-creative.Q4_K_M.gguf": ["mradermacher/nemo-12b-instruct-creative-GGUF", MessagesFormatterType.MISTRAL],
453
+ "Qwen2.5-Slick-Coder-14B-Instruct.Q4_K_S.gguf": ["mradermacher/Qwen2.5-Slick-Coder-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
454
+ "Gemma2-Evelyn-9B.Q4_K_M.gguf": ["mradermacher/Gemma2-Evelyn-9B-GGUF", MessagesFormatterType.ALPACA],
455
+ "The-Omega-Abomination-M-12B-v1.0.Q4_K_M.gguf": ["mradermacher/The-Omega-Abomination-M-12B-v1.0-GGUF", MessagesFormatterType.MISTRAL],
456
+ "Qwen2.5-Wicked-Coder-14B-Instruct.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Wicked-Coder-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
457
+ "ElvenMaid-12B-Stock.Q4_K_M.gguf": ["mradermacher/ElvenMaid-12B-Stock-GGUF", MessagesFormatterType.CHATML],
458
+ "shisa-v2-unphi4-14b.Q4_K_M.gguf": ["mradermacher/shisa-v2-unphi4-14b-GGUF", MessagesFormatterType.LLAMA_3],
459
+ "L3-Urvashi-v3.5.Q5_K_M.gguf": ["mradermacher/L3-Urvashi-v3.5-GGUF", MessagesFormatterType.LLAMA_3],
460
+ "shisa-v2-qwen2.5-7b.Q5_K_M.gguf": ["mradermacher/shisa-v2-qwen2.5-7b-GGUF", MessagesFormatterType.OPEN_CHAT],
461
+ "Qwen2.5-14B-YOYO-V1.5.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-YOYO-V1.5-GGUF", MessagesFormatterType.OPEN_CHAT],
462
+ "ElvenMaid-12B-v3.Q4_K_M.gguf": ["mradermacher/ElvenMaid-12B-v3-GGUF", MessagesFormatterType.CHATML],
463
+ "Qwen2.5-14B-Instruct-abliterated-SFT.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Instruct-abliterated-SFT-GGUF", MessagesFormatterType.OPEN_CHAT],
464
+ "Kaolinite-Kitara-12B.Q4_K_M.gguf": ["mradermacher/Kaolinite-Kitara-12B-GGUF", MessagesFormatterType.CHATML],
465
+ "deneb-v1-7b.Q5_K_M.gguf": ["mradermacher/deneb-v1-7b-GGUF", MessagesFormatterType.OPEN_CHAT],
466
+ "Ice0.105-13.04-RP.Q4_K_S.gguf": ["mradermacher/Ice0.105-13.04-RP-GGUF", MessagesFormatterType.MISTRAL],
467
+ "Gemma2-BlueMoon-9B.i1-Q4_K_M.gguf": ["mradermacher/Gemma2-BlueMoon-9B-i1-GGUF", MessagesFormatterType.ALPACA],
468
+ "Himeyuri-Magnum-12B-v2.i1-Q4_K_M.gguf": ["mradermacher/Himeyuri-Magnum-12B-v2-i1-GGUF", MessagesFormatterType.CHATML],
469
+ "ElvenMaid-12B-v2.Q4_K_M.gguf": ["mradermacher/ElvenMaid-12B-v2-GGUF", MessagesFormatterType.CHATML],
470
+ "ToolMaestro-Distil-T1-7B.Q5_K_M.gguf": ["mradermacher/ToolMaestro-Distil-T1-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
471
+ "OpenThoughts-114k-Rewrite-Thoughts-Llama-3.2-3B-Instruct.Q5_K_M.gguf": ["mradermacher/OpenThoughts-114k-Rewrite-Thoughts-Llama-3.2-3B-Instruct-GGUF", MessagesFormatterType.LLAMA_3],
472
+ "La-Superba-14B-Y.2.Q4_K_M.gguf": ["mradermacher/La-Superba-14B-Y.2-GGUF", MessagesFormatterType.OPEN_CHAT],
473
+ "Qwen2.5-3B-Instruct-Gate.i1-Q5_K_M.gguf": ["mradermacher/Qwen2.5-3B-Instruct-Gate-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
474
+ "cogito-v1-preview-qwen-14B.Q4_K_M.gguf": ["mradermacher/cogito-v1-preview-qwen-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
475
+ "Magnolia-v3a-12B.Q4_K_M.gguf": ["mradermacher/Magnolia-v3a-12B-GGUF", MessagesFormatterType.MISTRAL],
476
+ "Qwen2.5-14B-YOYO-stock.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-YOYO-stock-GGUF", MessagesFormatterType.OPEN_CHAT],
477
+ "ablation-164-shisav2.if100-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-164-shisav2.if100-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
478
+ "ablation-190-a175.finaldpo2.7.2e7-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-190-a175.finaldpo2.7.2e7-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
479
+ "ablation-183-a177.finaldpo2.4.5e7-shisa-v2-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/ablation-183-a177.finaldpo2.4.5e7-shisa-v2-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
480
+ "Qwen2.5-Casual-Coder-14B-Instruct.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Casual-Coder-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
481
+ "Ice0.102-10.04-RP.Q4_K_S.gguf": ["mradermacher/Ice0.102-10.04-RP-GGUF", MessagesFormatterType.MISTRAL],
482
+ "Qwen2.5-Flow-Coder-14B-Instruct.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Flow-Coder-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
483
+ "Qwen2.5-CompositeFlow-Coder-14B-Instruct.Q4_K_M.gguf": ["mradermacher/Qwen2.5-CompositeFlow-Coder-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
484
+ "Gemma-Muesli-2-9B-it.Q4_K_M.gguf": ["mradermacher/Gemma-Muesli-2-9B-it-GGUF", MessagesFormatterType.ALPACA],
485
+ "Chat-KTO-4B.Q5_K_M.gguf": ["mradermacher/Chat-KTO-4B-GGUF", MessagesFormatterType.LLAMA_3],
486
+ "Gliese-B1257.Q4_K_M.gguf": ["mradermacher/Gliese-B1257-GGUF", MessagesFormatterType.OPEN_CHAT],
487
+ "Qwen2.5-Sharp-Coder-14B-Instruct.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Sharp-Coder-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
488
+ "Llama-3.1-8B-UltraLong-1M-Instruct.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-UltraLong-1M-Instruct-GGUF", MessagesFormatterType.LLAMA_3],
489
+ "Final-Llama-CMD-1L.Q5_K_M.gguf": ["mradermacher/Final-Llama-CMD-1L-GGUF", MessagesFormatterType.LLAMA_3],
490
+ "Final-Llama-Cyber-1L.Q5_K_M.gguf": ["mradermacher/Final-Llama-Cyber-1L-GGUF", MessagesFormatterType.LLAMA_3],
491
+ "MagnaRei-v2-12B.Q4_K_M.gguf": ["mradermacher/MagnaRei-v2-12B-GGUF", MessagesFormatterType.MISTRAL],
492
+ "ThinkEdit-deepseek-llama3-8b.i1-Q5_K_M.gguf": ["mradermacher/ThinkEdit-deepseek-llama3-8b-i1-GGUF", MessagesFormatterType.LLAMA_3],
493
+ "M-Prometheus-14B.Q4_K_M.gguf": ["mradermacher/M-Prometheus-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
494
+ "Llama-Breeze2-8B-Instruct-text-only.Q5_K_M.gguf": ["mradermacher/Llama-Breeze2-8B-Instruct-text-only-GGUF", MessagesFormatterType.LLAMA_3],
495
+ "Magnolia-v7-12B.Q4_K_M.gguf": ["mradermacher/Magnolia-v7-12B-GGUF", MessagesFormatterType.MISTRAL],
496
+ "Magnolia-v6-12B.Q4_K_M.gguf": ["mradermacher/Magnolia-v6-12B-GGUF", MessagesFormatterType.MISTRAL],
497
+ "L3.1-Smthv1-12B.Q4_K_M.gguf": ["mradermacher/L3.1-Smthv1-12B-GGUF", MessagesFormatterType.LLAMA_3],
498
+ "The-Omega-Directive-M-12B-v1.0.Q4_K_M.gguf": ["mradermacher/The-Omega-Directive-M-12B-v1.0-GGUF", MessagesFormatterType.MISTRAL],
499
+ "The-Omega-Directive-M-8B-v1.0.Q5_K_M.gguf": ["mradermacher/The-Omega-Directive-M-8B-v1.0-GGUF", MessagesFormatterType.MISTRAL],
500
+ "MagnaRei-v1-12B.Q4_K_M.gguf": ["mradermacher/MagnaRei-v1-12B-GGUF", MessagesFormatterType.MISTRAL],
501
+ "Hamanasu-Magnum-4B.Q5_K_M.gguf": ["mradermacher/Hamanasu-Magnum-4B-GGUF", MessagesFormatterType.LLAMA_3],
502
+ "kd1-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/kd1-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
503
+ "kd2-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/kd2-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
504
+ "ablation-146-a128.dpo.armorm.rp.tl.1e6-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-146-a128.dpo.armorm.rp.tl.1e6-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
505
+ "ablation-151-finalsft-shisa-v2-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/ablation-151-finalsft-shisa-v2-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
506
+ "ablation-147-a128.dpo.armorm.rp.tl.1.5e6-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-147-a128.dpo.armorm.rp.tl.1.5e6-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
507
+ "ablation-158-a128.dpo.finaldpo.if50-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-158-a128.dpo.finaldpo.if50-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
508
+ "ablation-167-finaldpo.3e7-shisa-v2-unphi-4-14b.Q4_K_M.gguf": ["mradermacher/ablation-167-finaldpo.3e7-shisa-v2-unphi-4-14b-GGUF", MessagesFormatterType.LLAMA_3],
509
+ "ablation-155-a151.dpofinal-shisa-v2-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/ablation-155-a151.dpofinal-shisa-v2-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
510
+ "ablation-156-a152.dpofinal-shisa-v2-mistral-nemo-japanese-12b.Q4_K_M.gguf": ["mradermacher/ablation-156-a152.dpofinal-shisa-v2-mistral-nemo-japanese-12b-GGUF", MessagesFormatterType.MISTRAL],
511
+ "Qwen2.5-7B-YOYO-super.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-YOYO-super-GGUF", MessagesFormatterType.OPEN_CHAT],
512
+ "ChaiML-Nemo-DPO-V17.Q4_K_M.gguf": ["mradermacher/ChaiML-Nemo-DPO-V17-GGUF", MessagesFormatterType.MISTRAL],
513
+ "dippy-roleplay-llama-v2.8.Q5_K_M.gguf": ["mradermacher/dippy-roleplay-llama-v2.8-GGUF", MessagesFormatterType.LLAMA_3],
514
+ "Proto-Harpy-Burst-v0.1-7B.Q5_K_M.gguf": ["mradermacher/Proto-Harpy-Burst-v0.1-7B-GGUF", MessagesFormatterType.MISTRAL],
515
+ "OncoLlama-3.1-8B-Instruct-0.6.Q5_K_M.gguf": ["mradermacher/OncoLlama-3.1-8B-Instruct-0.6-GGUF", MessagesFormatterType.LLAMA_3],
516
+ "Mistral-nemo-ja-rp-v0.2.Q4_K_M.gguf": ["mradermacher/Mistral-nemo-ja-rp-v0.2-GGUF", MessagesFormatterType.CHATML],
517
+ "L3.1-Dark-Reasoning-LewdPlay-evo-Hermes-R1-Uncensored-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Dark-Reasoning-LewdPlay-evo-Hermes-R1-Uncensored-8B-GGUF", MessagesFormatterType.LLAMA_3],
518
+ "L3.1-Dark-Reasoning-Unholy-Hermes-R1-Uncensored-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Dark-Reasoning-Unholy-Hermes-R1-Uncensored-8B-GGUF", MessagesFormatterType.LLAMA_3],
519
+ "Llama-3.1-8B-Fusion-8020.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-Fusion-8020-GGUF", MessagesFormatterType.LLAMA_3],
520
+ "Llama-3.1-8B-Fusion-9010.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-Fusion-9010-GGUF", MessagesFormatterType.LLAMA_3],
521
+ "Proto-Harpy-Avian-v0.1-7B.Q5_K_M.gguf": ["mradermacher/Proto-Harpy-Avian-v0.1-7B-GGUF", MessagesFormatterType.MISTRAL],
522
+ "Qwen2.5-14B-ties.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-ties-GGUF", MessagesFormatterType.OPEN_CHAT],
523
+ "Qwen2.5-14B-ties-1M.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-ties-1M-GGUF", MessagesFormatterType.OPEN_CHAT],
524
+ "Qwen2.5-14B-della.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-della-GGUF", MessagesFormatterType.OPEN_CHAT],
525
+ "RoLlama3-8b-Instruct-2024-10-09.Q5_K_M.gguf": ["mradermacher/RoLlama3-8b-Instruct-2024-10-09-GGUF", MessagesFormatterType.LLAMA_3],
526
+ "VietCoMath-01-14B.Q4_K_M.gguf": ["mradermacher/VietCoMath-01-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
527
+ "t0-s1-14B.Q4_K_M.gguf": ["mradermacher/t0-s1-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
528
+ "Gemma2-9b-it-Girl-v1.Q4_K_M.gguf": ["mradermacher/Gemma2-9b-it-Girl-v1-GGUF", MessagesFormatterType.ALPACA],
529
+ "Nemo-DPO-V22.Q4_K_S.gguf": ["mradermacher/Nemo-DPO-V22-GGUF", MessagesFormatterType.MISTRAL],
530
+ "LoyalMaid-12B.Q4_K_S.gguf": ["mradermacher/LoyalMaid-12B-GGUF", MessagesFormatterType.MISTRAL],
531
+ "t0-s1.1-3B.Q5_K_M.gguf": ["mradermacher/t0-s1.1-3B-GGUF", MessagesFormatterType.OPEN_CHAT],
532
+ "NQLSG-Qwen2.5-14B-MegaFusion-v9.4.Q4_K_M.gguf": ["mradermacher/NQLSG-Qwen2.5-14B-MegaFusion-v9.4-GGUF", MessagesFormatterType.OPEN_CHAT],
533
+ "Ultron-Summarizer-8B.Q5_K_M.gguf": ["mradermacher/Ultron-Summarizer-8B-GGUF", MessagesFormatterType.LLAMA_3],
534
+ "NQLSG-Qwen2.5-14B-MegaFusion-v9.4-Coder.Q4_K_M.gguf": ["mradermacher/NQLSG-Qwen2.5-14B-MegaFusion-v9.4-Coder-GGUF", MessagesFormatterType.OPEN_CHAT],
535
+ "ElvenMaid-12B.Q4_K_M.gguf": ["mradermacher/ElvenMaid-12B-GGUF", MessagesFormatterType.MISTRAL],
536
+ "Llama-3.1-8B-Fusion-7030.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-Fusion-7030-GGUF", MessagesFormatterType.LLAMA_3],
537
+ "MedicalEDI-14b-EDI-Reasoning-5-bf16.Q4_K_M.gguf": ["mradermacher/MedicalEDI-14b-EDI-Reasoning-5-bf16-GGUF", MessagesFormatterType.OPEN_CHAT],
538
+ "Llama-3.1-8B-Athena-Apollo-exp.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-Athena-Apollo-exp-GGUF", MessagesFormatterType.LLAMA_3],
539
+ "Llama-3.1-8B-Fusion-5050.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-Fusion-5050-GGUF", MessagesFormatterType.LLAMA_3],
540
+ "Nemo-DPO-V21.Q4_K_M.gguf": ["mradermacher/Nemo-DPO-V21-GGUF", MessagesFormatterType.MISTRAL],
541
+ "ChatWaifu_32B_reasoning.Q4_K_M.gguf": ["mradermacher/ChatWaifu_32B_reasoning-GGUF", MessagesFormatterType.OPEN_CHAT],
542
+ "Llama3_Storyteller.Q5_K_M.gguf": ["mradermacher/Llama3_Storyteller-GGUF", MessagesFormatterType.LLAMA_3],
543
+ "ablation-110-openrlhf.dpo.a106.5e7ax-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-110-openrlhf.dpo.a106.5e7ax-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
544
+ "DeepResearcher-7b.Q5_K_M.gguf": ["mradermacher/DeepResearcher-7b-GGUF", MessagesFormatterType.OPEN_CHAT],
545
+ "ablation-141-a128.dpo.armorm.rp-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-141-a128.dpo.armorm.rp-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
546
+ "Bootes-Xuange-Qwen-14B.Q4_K_M.gguf": ["mradermacher/Bootes-Xuange-Qwen-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
547
+ "ablation-123-a114.dpo.armorm.rp-shisa-v2-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/ablation-123-a114.dpo.armorm.rp-shisa-v2-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
548
+ "ablation-138-shisav2.gbs128.1.6e5-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-138-shisav2.gbs128.1.6e5-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
549
+ "ablation-122-a114.dpo.armorm.rp-shisa-v2-unphi-4-14b.Q4_K_M.gguf": ["mradermacher/ablation-122-a114.dpo.armorm.rp-shisa-v2-unphi-4-14b-GGUF", MessagesFormatterType.MISTRAL],
550
+ "Qwen2.5-14B-stock-v3.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-stock-v3-GGUF", MessagesFormatterType.OPEN_CHAT],
551
+ "Humanize-Rei-Slerp.Q4_K_M.gguf": ["mradermacher/Humanize-Rei-Slerp-GGUF", MessagesFormatterType.MISTRAL],
552
+ "ablation-142-a128.dpo.armorm.rp.tl-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-142-a128.dpo.armorm.rp.tl-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
553
+ "ablation-139-shisav2.ds.gbs128.1e5-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-139-shisav2.ds.gbs128.1e5-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
554
+ "L3.1-Dark-Reasoning-Halu-Blackroot-Hermes-R1-Uncensored-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Dark-Reasoning-Halu-Blackroot-Hermes-R1-Uncensored-8B-GGUF", MessagesFormatterType.LLAMA_3],
555
+ "L3.1-Dark-Reasoning-Anjir-Hermes-R1-Uncensored-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Dark-Reasoning-Anjir-Hermes-R1-Uncensored-8B-GGUF", MessagesFormatterType.LLAMA_3],
556
+ "ablation-115-bestofn.atlo.rp.tlx.newmix2-shisa-v2-mistral-nemo-12b.Q4_K_S.gguf": ["mradermacher/ablation-115-bestofn.atlo.rp.tlx.newmix2-shisa-v2-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
557
+ "L3.1-Dark-Reasoning-Super-Nova-RP-Hermes-R1-Uncensored-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Dark-Reasoning-Super-Nova-RP-Hermes-R1-Uncensored-8B-GGUF", MessagesFormatterType.LLAMA_3],
558
+ "L3.1-Dark-Reasoning-Celeste-V1.2-Hermes-R1-Uncensored-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Dark-Reasoning-Celeste-V1.2-Hermes-R1-Uncensored-8B-GGUF", MessagesFormatterType.LLAMA_3],
559
+ "DeeperHermes3_R1_D_L3_8b.Q5_K_M.gguf": ["mradermacher/DeeperHermes3_R1_D_L3_8b-GGUF", MessagesFormatterType.LLAMA_3],
560
+ "Nemo-DPO-V20.Q4_K_S.gguf": ["mradermacher/Nemo-DPO-V20-GGUF", MessagesFormatterType.MISTRAL],
561
+ "Badudu-8B-instruct-v1.Q4_K_M.gguf": ["mradermacher/Badudu-8B-instruct-v1-GGUF", MessagesFormatterType.LLAMA_3],
562
+ "Himeyuri-Magnum-12B.Q4_K_M.gguf": ["mradermacher/Himeyuri-Magnum-12B-GGUF", MessagesFormatterType.OPEN_CHAT],
563
+ "Aura-12B-FT.i1-Q4_K_M.gguf": ["mradermacher/Aura-12B-FT-i1-GGUF", MessagesFormatterType.CHATML],
564
+ "NarrowMaid-8B.Q5_K_M.gguf": ["mradermacher/NarrowMaid-8B-GGUF", MessagesFormatterType.LLAMA_3],
565
+ "L3.1-Dark-Reasoning-Dark-Planet-Hermes-R1-Uncensored-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Dark-Reasoning-Dark-Planet-Hermes-R1-Uncensored-8B-GGUF", MessagesFormatterType.LLAMA_3],
566
+ "Qwen2.5-Vibe-Coder-14B-Instruct.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Vibe-Coder-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
567
+ "josie-7b-v6.0.Q5_K_M.gguf": ["mradermacher/josie-7b-v6.0-GGUF", MessagesFormatterType.OPEN_CHAT],
568
+ "NeuDistRo-a1-laser.Q5_K_M.gguf": ["mradermacher/NeuDistRo-a1-laser-GGUF", MessagesFormatterType.MISTRAL],
569
+ "Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v1.Q5_K_M.gguf": ["mradermacher/Josiefied-Qwen2.5-1.5B-Instruct-abliterated-v1-GGUF", MessagesFormatterType.OPEN_CHAT],
570
+ "DreadMix.Q5_K_M.gguf": ["mradermacher/DreadMix-GGUF", MessagesFormatterType.LLAMA_3],
571
+ "EvolCodeLlama-3.1-8B-Instruct.Q5_K_M.gguf": ["mradermacher/EvolCodeLlama-3.1-8B-Instruct-GGUF", MessagesFormatterType.LLAMA_3],
572
+ "Soaring-8b.Q5_K_M.gguf": ["mradermacher/Soaring-8b-GGUF", MessagesFormatterType.LLAMA_3],
573
+ "RoLlama3.1-8b-Instruct-DPO.Q5_K_M.gguf": ["mradermacher/RoLlama3.1-8b-Instruct-DPO-GGUF", MessagesFormatterType.LLAMA_3],
574
+ "qwen-14b-chisel.Q4_K_M.gguf": ["mradermacher/qwen-14b-chisel-GGUF", MessagesFormatterType.OPEN_CHAT],
575
+ "HarmReporter.Q5_K_M.gguf": ["mradermacher/HarmReporter-GGUF", MessagesFormatterType.LLAMA_3],
576
+ "Gemma2-Alicia-9B.Q4_K_M.gguf": ["mradermacher/Gemma2-Alicia-9B-GGUF", MessagesFormatterType.ALPACA],
577
+ "bad-apple-8b-v1.Q5_K_M.gguf": ["mradermacher/bad-apple-8b-v1-GGUF", MessagesFormatterType.LLAMA_3],
578
+ "Eris-Daturamix-Neural-7b-slerp.Q5_K_M.gguf": ["mradermacher/Eris-Daturamix-Neural-7b-slerp-GGUF", MessagesFormatterType.MISTRAL],
579
+ "miscii-1023-001-200.Q4_K_M.gguf": ["mradermacher/miscii-1023-001-200-GGUF", MessagesFormatterType.OPEN_CHAT],
580
+ "Faheng-14B.Q4_K_M.gguf": ["mradermacher/Faheng-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
581
+ "sft-math-llama3-8b-inst.Q5_K_M.gguf": ["mradermacher/sft-math-llama3-8b-inst-GGUF", MessagesFormatterType.LLAMA_3],
582
+ "ablation-104-openrlhf.dpo.a84.5e7ring-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-104-openrlhf.dpo.a84.5e7ring-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
583
+ "ablation-102-openrlhf.dpo.a84.5e7-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-102-openrlhf.dpo.a84.5e7-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
584
+ "ablation-105-bestofn.atlo.rp.tlx25.newmixall-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-105-bestofn.atlo.rp.tlx25.newmixall-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
585
+ "ablation-103-openrlhf.dpo.a84.5e7ax-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-103-openrlhf.dpo.a84.5e7ax-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
586
+ "Qwen2.5-1.5B-Instruct-Multiple-Choice-Maker.Q5_K_M.gguf": ["mradermacher/Qwen2.5-1.5B-Instruct-Multiple-Choice-Maker-GGUF", MessagesFormatterType.OPEN_CHAT],
587
+ "Nemo-DPO-V18.Q4_K_M.gguf": ["mradermacher/Nemo-DPO-V18-GGUF", MessagesFormatterType.MISTRAL],
588
+ "L3-8B-sunfall-v0.3.i1-Q5_K_M.gguf": ["mradermacher/L3-8B-sunfall-v0.3-i1-GGUF", MessagesFormatterType.LLAMA_3],
589
+ "bad-apple-8b-v2.Q5_K_M.gguf": ["mradermacher/bad-apple-8b-v2-GGUF", MessagesFormatterType.LLAMA_3],
590
+ "dippy-roleplay-llama-v3.1-0-10.Q5_K_M.gguf": ["mradermacher/dippy-roleplay-llama-v3.1-0-10-GGUF", MessagesFormatterType.LLAMA_3],
591
+ "CatunaMayo.Q5_K_M.gguf": ["mradermacher/CatunaMayo-GGUF", MessagesFormatterType.MISTRAL],
592
+ "DeepRetrieval-PubMed-3B.Q5_K_M.gguf": ["mradermacher/DeepRetrieval-PubMed-3B-GGUF", MessagesFormatterType.OPEN_CHAT],
593
+ "Gemma_Writer-9b.Q4_K_M.gguf": ["mradermacher/Gemma_Writer-9b-GGUF", MessagesFormatterType.ALPACA],
594
+ "Nemo-DPO-V19.Q4_K_M.gguf": ["mradermacher/Nemo-DPO-V19-GGUF", MessagesFormatterType.MISTRAL],
595
+ "supermario-slerp-v3.Q5_K_M.gguf": ["mradermacher/supermario-slerp-v3-GGUF", MessagesFormatterType.MISTRAL],
596
+ "RoLlama3-8b-Instruct-DPO-2024-10-09.Q5_K_M.gguf": ["mradermacher/RoLlama3-8b-Instruct-DPO-2024-10-09-GGUF", MessagesFormatterType.LLAMA_3],
597
+ "Atlantum-12B-v1.Q4_K_M.gguf": ["mradermacher/Atlantum-12B-v1-GGUF", MessagesFormatterType.MISTRAL],
598
+ "Neural-Eris-Daturamix-Chat-7b-slerp.Q5_K_M.gguf": ["mradermacher/Neural-Eris-Daturamix-Chat-7b-slerp-GGUF", MessagesFormatterType.MISTRAL],
599
+ "supermario-v2.Q5_K_M.gguf": ["mradermacher/supermario-v2-GGUF", MessagesFormatterType.MISTRAL],
600
+ "rainbowfish-7B-v9.Q5_K_M.gguf": ["mradermacher/rainbowfish-7B-v9-GGUF", MessagesFormatterType.MISTRAL],
601
+ "Con-J-Qwen2-7B.Q5_K_M.gguf": ["mradermacher/Con-J-Qwen2-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
602
+ "Gemma2-9b-it-Youki-v1.Q4_K_M.gguf": ["mradermacher/Gemma2-9b-it-Youki-v1-GGUF", MessagesFormatterType.ALPACA],
603
+ "Reasoning-Mistral-Nemo.Q4_K_M.gguf": ["mradermacher/Reasoning-Mistral-Nemo-GGUF", MessagesFormatterType.MISTRAL],
604
+ "DeepRetrieval-SQL-7B.Q5_K_M.gguf": ["mradermacher/DeepRetrieval-SQL-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
605
+ "Qwen2.5-14B-dpo-it-della.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-dpo-it-della-GGUF", MessagesFormatterType.OPEN_CHAT],
606
+ "Aura-12B.Q4_K_M.gguf": ["mradermacher/Aura-12B-GGUF", MessagesFormatterType.MISTRAL],
607
+ "JudgeLRM-7B.Q5_K_M.gguf": ["mradermacher/JudgeLRM-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
608
+ "LLaMAntino-3-SLIMER-IT-PARALLEL.Q5_K_M.gguf": ["mradermacher/LLaMAntino-3-SLIMER-IT-PARALLEL-GGUF", MessagesFormatterType.LLAMA_3],
609
+ "Gemma-Spoon-2-9B-it.i1-Q4_K_M.gguf": ["mradermacher/Gemma-Spoon-2-9B-it-i1-GGUF", MessagesFormatterType.ALPACA],
610
+ "L3-GothicMaid-8B.Q5_K_M.gguf": ["mradermacher/L3-GothicMaid-8B-GGUF", MessagesFormatterType.LLAMA_3],
611
+ "Neuronal-7b-FT.Q5_K_M.gguf": ["mradermacher/Neuronal-7b-FT-GGUF", MessagesFormatterType.MISTRAL],
612
+ "L3.1-SemiEvol-MMLUPro-8B.Q5_K_M.gguf": ["mradermacher/L3.1-SemiEvol-MMLUPro-8B-GGUF", MessagesFormatterType.LLAMA_3],
613
+ "meme-12b-e2.Q4_K_M.gguf": ["mradermacher/meme-12b-e2-GGUF", MessagesFormatterType.MISTRAL],
614
+ "ablation-100-bestofn.atlo.rp.tlx50-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-100-bestofn.atlo.rp.tlx50-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
615
+ "ablation-101-bestofn.atlo.rp.tlx25-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-101-bestofn.atlo.rp.tlx25-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
616
+ "TaCIE-Qwen2-7B.Q5_K_M.gguf": ["mradermacher/TaCIE-Qwen2-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
617
+ "Qwen2.5-Coder-7B-Instruct-Uncensored.i1-Q5_K_M.gguf": ["mradermacher/Qwen2.5-Coder-7B-Instruct-Uncensored-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
618
+ "Korean-llama-3.1-8B.Q5_K_M.gguf": ["mradermacher/Korean-llama-3.1-8B-GGUF", MessagesFormatterType.LLAMA_3],
619
+ "Gr3.1-8B-Desert-Rose.Q5_K_M.gguf": ["mradermacher/Gr3.1-8B-Desert-Rose-GGUF", MessagesFormatterType.OPEN_CHAT],
620
+ "L3-Tyche-8B-v1.0.Q5_K_M.gguf": ["mradermacher/L3-Tyche-8B-v1.0-GGUF", MessagesFormatterType.LLAMA_3],
621
+ "Eris-LelantaclesV2-7b.Q5_K_M.gguf": ["mradermacher/Eris-LelantaclesV2-7b-GGUF", MessagesFormatterType.MISTRAL],
622
+ "Eris-Lelantacles-7b.Q5_K_M.gguf": ["mradermacher/Eris-Lelantacles-7b-GGUF", MessagesFormatterType.MISTRAL],
623
+ "jessi-v0.1-qwen2.5-7b-instruct.Q5_K_M.gguf": ["mradermacher/jessi-v0.1-qwen2.5-7b-instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
624
+ "Qwen2.5-7B-Latent_Verification.i1-Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Latent_Verification-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
625
+ "Agentic-Qwen7b-GRPO.Q5_K_M.gguf": ["mradermacher/Agentic-Qwen7b-GRPO-GGUF", MessagesFormatterType.OPEN_CHAT],
626
+ "ablation-95-a85.dpo.armorm.rp-shisa-v2-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/ablation-95-a85.dpo.armorm.rp-shisa-v2-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
627
+ "ablation-89-rp.dpo.test-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-89-rp.dpo.test-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
628
+ "Gemma-Noon-2-9B-it.Q4_K_M.gguf": ["mradermacher/Gemma-Noon-2-9B-it-GGUF", MessagesFormatterType.ALPACA],
629
+ "ablation-99-bestofn.atlo.rp.tlx-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-99-bestofn.atlo.rp.tlx-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
630
+ "ablation-86-bestofn.atlo.rp.tl-shisa-v2-mistral-nemo-12b-5e6.Q4_K_M.gguf": ["mradermacher/ablation-86-bestofn.atlo.rp.tl-shisa-v2-mistral-nemo-12b-5e6-GGUF", MessagesFormatterType.MISTRAL],
631
+ "ablation-85-bestofn.atlo.rp.tl-shisa-v2-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/ablation-85-bestofn.atlo.rp.tl-shisa-v2-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
632
+ "MistThena7B.Q5_K_M.gguf": ["mradermacher/MistThena7B-GGUF", MessagesFormatterType.MISTRAL],
633
+ "Pervepsion-12B-x1.Q4_K_M.gguf": ["mradermacher/Pervepsion-12B-x1-GGUF", MessagesFormatterType.MISTRAL],
634
+ "ablation-98-bestofn.atlo.rp.tl.geniac-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-98-bestofn.atlo.rp.tl.geniac-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
635
+ "Dark-Triad-Adam-Llama3.2-3B-v2.Q5_K_M.gguf": ["mradermacher/Dark-Triad-Adam-Llama3.2-3B-v2-GGUF", MessagesFormatterType.LLAMA_3],
636
+ "Qandora-2.5-7B.Q5_K_M.gguf": ["mradermacher/Qandora-2.5-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
637
+ "Tesslate_Tessa-T1-14B-Q4_K_M.gguf": ["bartowski/Tesslate_Tessa-T1-14B-GGUF", MessagesFormatterType.CHATML],
638
+ "Tesslate_Tessa-T1-7B-Q5_K_M.gguf": ["bartowski/Tesslate_Tessa-T1-7B-GGUF", MessagesFormatterType.CHATML],
639
+ "tabletop_game_design.Q5_K_M.gguf": ["mradermacher/tabletop_game_design-GGUF", MessagesFormatterType.OPEN_CHAT],
640
+ "mergekit-model_stock-tiwlqms.Q4_K_M.gguf": ["mradermacher/mergekit-model_stock-tiwlqms-GGUF", MessagesFormatterType.MISTRAL],
641
+ "Buttocks-7B-v2.1.Q5_K_M.gguf": ["mradermacher/Buttocks-7B-v2.1-GGUF", MessagesFormatterType.MISTRAL],
642
+ "MN-Hekate-Geneteira-12B.Q4_K_M.gguf": ["mradermacher/MN-Hekate-Geneteira-12B-GGUF", MessagesFormatterType.CHATML],
643
+ "Agentic-Qwen7b.Q5_K_M.gguf": ["mradermacher/Agentic-Qwen7b-GGUF", MessagesFormatterType.OPEN_CHAT],
644
+ "Mistral-Nemo-12B-Instruct.Q4_K_M.gguf": ["mradermacher/Mistral-Nemo-12B-Instruct-GGUF", MessagesFormatterType.MISTRAL],
645
+ "EvaMedius-14B-Instruct.Q4_K_M.gguf": ["mradermacher/EvaMedius-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
646
+ "Captain_Eris_Twighlight-12B-v2.Q5_K_M.gguf": ["mradermacher/Captain_Eris_Twighlight-12B-v2-GGUF", MessagesFormatterType.MISTRAL],
647
+ "Llama-3-8B-WizardLM-196K.Q5_K_M.gguf": ["mradermacher/Llama-3-8B-WizardLM-196K-GGUF", MessagesFormatterType.LLAMA_3],
648
+ "zephyr-qwen2-7b-dpo.Q5_K_M.gguf": ["mradermacher/zephyr-qwen2-7b-dpo-GGUF", MessagesFormatterType.OPEN_CHAT],
649
+ "L3-Umbral-Mind-RP-v0.6.2-8B.i1-Q5_K_M.gguf": ["mradermacher/L3-Umbral-Mind-RP-v0.6.2-8B-i1-GGUF", MessagesFormatterType.LLAMA_3],
650
+ "Final-Llama-IoI-1E.Q5_K_M.gguf": ["mradermacher/Final-Llama-IoI-1E-GGUF", MessagesFormatterType.LLAMA_3],
651
+ "NQLSG-Qwen2.5-14B-MegaFusion-v9.3-Coder.Q4_K_M.gguf": ["mradermacher/NQLSG-Qwen2.5-14B-MegaFusion-v9.3-Coder-GGUF", MessagesFormatterType.OPEN_CHAT],
652
+ "ReSearch-Qwen-7B-Instruct.Q5_K_M.gguf": ["mradermacher/ReSearch-Qwen-7B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
653
+ "MN-Hekate-Panopaia-12B.Q4_K_S.gguf": ["mradermacher/MN-Hekate-Panopaia-12B-GGUF", MessagesFormatterType.MISTRAL],
654
+ "Hathor_Ultra-Mix-Final-0.420-L3-8B.Q5_K_M.gguf": ["mradermacher/Hathor_Ultra-Mix-Final-0.420-L3-8B-GGUF", MessagesFormatterType.LLAMA_3],
655
+ "Nova-1.5-Conversational-7B.Q5_K_M.gguf": ["mradermacher/Nova-1.5-Conversational-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
656
+ "Raptor-X1.Q4_K_M.gguf": ["mradermacher/Raptor-X1-GGUF", MessagesFormatterType.OPEN_CHAT],
657
+ "Noxis-8B-LINEAR.Q5_K_M.gguf": ["mradermacher/Noxis-8B-LINEAR-GGUF", MessagesFormatterType.LLAMA_3],
658
+ "Q2.5-Instruct-1M_Harmony.Q5_K_M.gguf": ["mradermacher/Q2.5-Instruct-1M_Harmony-GGUF", MessagesFormatterType.OPEN_CHAT],
659
+ "Qwen2.5-14B-Instruct-1M-Unalign.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Instruct-1M-Unalign-GGUF", MessagesFormatterType.OPEN_CHAT],
660
+ "Llama-3.2-3B-Instruct-Explainable-Propaganda-Detection.Q5_K_M.gguf": ["mradermacher/Llama-3.2-3B-Instruct-Explainable-Propaganda-Detection-GGUF", MessagesFormatterType.LLAMA_3],
661
+ "Qwen-2.5-7B-GRA-WizardLM.Q5_K_M.gguf": ["mradermacher/Qwen-2.5-7B-GRA-WizardLM-GGUF", MessagesFormatterType.OPEN_CHAT],
662
+ "dpo_plroleplay.Q4_K_M.gguf": ["mradermacher/dpo_plroleplay-GGUF", MessagesFormatterType.MISTRAL],
663
+ "YinkaMath-14B.Q4_K_M.gguf": ["mradermacher/YinkaMath-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
664
+ "CosmoCoder.Q4_K_M.gguf": ["mradermacher/CosmoCoder-GGUF", MessagesFormatterType.OPEN_CHAT],
665
+ "Sinner_12B_V.1.Q4_K_M.gguf": ["mradermacher/Sinner_12B_V.1-GGUF", MessagesFormatterType.MISTRAL],
666
+ "Nova-0.5-e3-7B.i1-Q5_K_M.gguf": ["mradermacher/Nova-0.5-e3-7B-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
667
+ "ChatingDeveloper-7B-slerp.Q5_K_M.gguf": ["mradermacher/ChatingDeveloper-7B-slerp-GGUF", MessagesFormatterType.MISTRAL],
668
+ "Neo-1-16B.Q5_K_M.gguf": ["mradermacher/Neo-1-16B-GGUF", MessagesFormatterType.OPEN_CHAT],
669
+ "S5-0014-Plus-81-14B-Exp.Q4_K_M.gguf": ["mradermacher/S5-0014-Plus-81-14B-Exp-GGUF", MessagesFormatterType.OPEN_CHAT],
670
+ "Ganymede-Llama-3.3-3B-Preview.Q5_K_M.gguf": ["mradermacher/Ganymede-Llama-3.3-3B-Preview-GGUF", MessagesFormatterType.LLAMA_3],
671
+ "Alpha-UMa-Qwen-14B.Q4_K_M.gguf": ["mradermacher/Alpha-UMa-Qwen-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
672
+ "Ice0.101-20.03-RP-GRPO-2.Q5_K_M.gguf": ["mradermacher/Ice0.101-20.03-RP-GRPO-2-GGUF", MessagesFormatterType.MISTRAL],
673
+ "Dirty-Shirley-Writer-v01-Uncensored.i1-Q4_K_M.gguf": ["mradermacher/Dirty-Shirley-Writer-v01-Uncensored-i1-GGUF", MessagesFormatterType.ALPACA],
674
+ "Dirty-Shirley-Writer-v0-abliterated.Q4_K_M.gguf": ["mradermacher/Dirty-Shirley-Writer-v0-abliterated-GGUF", MessagesFormatterType.ALPACA],
675
+ "Wicked-Love-12B-v2.Q4_K_M.gguf": ["mradermacher/Wicked-Love-12B-v2-GGUF", MessagesFormatterType.MISTRAL],
676
+ "Qwen2.5-14B-YOYO-V5.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-YOYO-V5-GGUF", MessagesFormatterType.OPEN_CHAT],
677
+ "gemma-2-9b-slimorca20k.Q4_K_M.gguf": ["mradermacher/gemma-2-9b-slimorca20k-GGUF", MessagesFormatterType.ALPACA],
678
+ "oh-dcft-v3.1-gemini-1.5-pro.i1-Q5_K_M.gguf": ["mradermacher/oh-dcft-v3.1-gemini-1.5-pro-i1-GGUF", MessagesFormatterType.LLAMA_3],
679
+ "Community_Request-03-12B.Q4_K_M.gguf": ["mradermacher/Community_Request-03-12B-GGUF", MessagesFormatterType.MISTRAL],
680
+ "Community_Request-01-12B.i1-Q4_K_M.gguf": ["mradermacher/Community_Request-01-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
681
+ "Community_Request-02-12B.i1-Q4_K_M.gguf": ["mradermacher/Community_Request-02-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
682
+ "mt-merge9-gemma-2-9b-q4_k_m.gguf": ["Otakadelic/MT-Merge9-gemma-2-9B-Q4_K_M-GGUF", MessagesFormatterType.ALPACA],
683
+ "Qwen2.5-14B-YOYO-V2.i1-Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-YOYO-V2-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
684
+ "AV-FI-Qwen2.5-1.5B-PT-BR-Instruct.Q5_K_M.gguf": ["mradermacher/AV-FI-Qwen2.5-1.5B-PT-BR-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
685
+ "Qwen2.5-7B-YOYO.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-YOYO-GGUF", MessagesFormatterType.OPEN_CHAT],
686
+ "Ice0.101-20.03-RP.Q5_K_M.gguf": ["mradermacher/Ice0.101-20.03-RP-GGUF", MessagesFormatterType.MISTRAL],
687
+ "Qwen2.5-Osiris-7B-Instruct.Q5_K_M.gguf": ["mradermacher/Qwen2.5-Osiris-7B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
688
+ "Forgotten-Abomination-8B-v4.0.Q5_K_M.gguf": ["mradermacher/Forgotten-Abomination-8B-v4.0-GGUF", MessagesFormatterType.MISTRAL],
689
+ "Qwen2.5-7B-sft-ultrachat-all-pool.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-sft-ultrachat-all-pool-GGUF", MessagesFormatterType.OPEN_CHAT],
690
+ "Qwen2.5-7B-sft-ultrachat-SPIN-Qwen2.5-72B-Instruct.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-sft-ultrachat-SPIN-Qwen2.5-72B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
691
+ "Qwen2.5-1.5B-YOYO.Q5_K_M.gguf": ["mradermacher/Qwen2.5-1.5B-YOYO-GGUF", MessagesFormatterType.OPEN_CHAT],
692
+ "Qwen2.5-3B-YOYO.Q5_K_M.gguf": ["mradermacher/Qwen2.5-3B-YOYO-GGUF", MessagesFormatterType.OPEN_CHAT],
693
+ "NeMo-NeMo-12B.Q4_K_M.gguf": ["mradermacher/NeMo-NeMo-12B-GGUF", MessagesFormatterType.MISTRAL],
694
+ "Qwen2.5-Coder-14B-YOYO.Q4_K_M.gguf": ["mradermacher/Qwen2.5-Coder-14B-YOYO-GGUF", MessagesFormatterType.OPEN_CHAT],
695
+ "Qwen2.5-14B-YOYO.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-YOYO-GGUF", MessagesFormatterType.OPEN_CHAT],
696
+ "Qwen2.5-14B-YOYO-V3.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-YOYO-V3-GGUF", MessagesFormatterType.OPEN_CHAT],
697
+ "Saanvi-C0-12B.Q4_K_M.gguf": ["mradermacher/Saanvi-C0-12B-GGUF", MessagesFormatterType.MISTRAL],
698
+ "Celestial-Harmony-14b-v1.0-Experimental-1015.Q4_K_M.gguf": ["mradermacher/Celestial-Harmony-14b-v1.0-Experimental-1015-GGUF", MessagesFormatterType.OPEN_CHAT],
699
+ "Qwen2.5-3B-EMPO-TQA.Q5_K_M.gguf": ["mradermacher/Qwen2.5-3B-EMPO-TQA-GGUF", MessagesFormatterType.OPEN_CHAT],
700
+ "Wicked-Love-12B.Q4_K_M.gguf": ["mradermacher/Wicked-Love-12B-GGUF", MessagesFormatterType.MISTRAL],
701
+ "UnslopNemo-Mag-Mell_T-2.Q4_K_M.gguf": ["mradermacher/UnslopNemo-Mag-Mell_T-2-GGUF", MessagesFormatterType.MISTRAL],
702
+ "Neporalito-12B-x2.Q4_K_M.gguf": ["mradermacher/Neporalito-12B-x2-GGUF", MessagesFormatterType.MISTRAL],
703
+ "Qwen2.5-7B-sft-ultrachat-peers-pool.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-sft-ultrachat-peers-pool-GGUF", MessagesFormatterType.OPEN_CHAT],
704
+ "Qwen2-7B-sft-ultrachat-peers-pool.Q5_K_M.gguf": ["mradermacher/Qwen2-7B-sft-ultrachat-peers-pool-GGUF", MessagesFormatterType.OPEN_CHAT],
705
+ "Fiendish_LLAMA_3B.Q5_K_M.gguf": ["mradermacher/Fiendish_LLAMA_3B-GGUF", MessagesFormatterType.LLAMA_3],
706
+ "Qwen3b-bs128-GRPO.Q5_K_M.gguf": ["mradermacher/Qwen3b-bs128-GRPO-GGUF", MessagesFormatterType.OPEN_CHAT],
707
+ "UnslopNemo-Mag-Mell_T-1.Q4_K_M.gguf": ["mradermacher/UnslopNemo-Mag-Mell_T-1-GGUF", MessagesFormatterType.MISTRAL],
708
+ "llama3.1-8b-instruct-hasanpiker-abliterated.Q5_K_M.gguf": ["mradermacher/llama3.1-8b-instruct-hasanpiker-abliterated-GGUF", MessagesFormatterType.LLAMA_3],
709
+ "qwen-writerdemo-7b-s500.Q5_K_M.gguf": ["mradermacher/qwen-writerdemo-7b-s500-GGUF", MessagesFormatterType.OPEN_CHAT],
710
+ "AlignXpert-PBA-Subset.Q5_K_M.gguf": ["mradermacher/AlignXpert-PBA-Subset-GGUF", MessagesFormatterType.LLAMA_3],
711
+ "Glowing-Forest-2-12B.Q4_K_M.gguf": ["mradermacher/Glowing-Forest-2-12B-GGUF", MessagesFormatterType.MISTRAL],
712
+ "Kosmos-EVAA-gamma-v16-8B.Q5_K_M.gguf": ["mradermacher/Kosmos-EVAA-gamma-v16-8B-GGUF", MessagesFormatterType.LLAMA_3],
713
+ "gemma-2-aeria-infinity-9b.Q4_K_M.gguf": ["mradermacher/gemma-2-aeria-infinity-9b-GGUF", MessagesFormatterType.ALPACA],
714
+ "Kosmos-EVAA-gamma-v13-8B.Q5_K_M.gguf": ["mradermacher/Kosmos-EVAA-gamma-v13-8B-GGUF", MessagesFormatterType.LLAMA_3],
715
+ "DeepHermes-3-Llama-3-3B-Preview-abliterated.i1-Q5_K_M.gguf": ["mradermacher/DeepHermes-3-Llama-3-3B-Preview-abliterated-i1-GGUF", MessagesFormatterType.LLAMA_3],
716
+ "GeneralThought-Qwen-7B-instruct-SFT.Q5_K_M.gguf": ["mradermacher/GeneralThought-Qwen-7B-instruct-SFT-GGUF", MessagesFormatterType.OPEN_CHAT],
717
+ "MN-Hekate-Kleidoukhos-12B.Q4_K_M.gguf": ["mradermacher/MN-Hekate-Kleidoukhos-12B-GGUF", MessagesFormatterType.MISTRAL],
718
+ "Forgotten-Abomination-12B-v4.0.Q4_K_M.gguf": ["mradermacher/Forgotten-Abomination-12B-v4.0-GGUF", MessagesFormatterType.MISTRAL],
719
+ "Kosmos-EVAA-gamma-v18-8B.Q5_K_M.gguf": ["Kosmos-EVAA-gamma-v18-8B.Q5_K_M.gguf", MessagesFormatterType.LLAMA_3],
720
+ "Llama-3.2-3B-Instruct-abliterated-finetuned.Q5_K_M.gguf": ["mradermacher/Llama-3.2-3B-Instruct-abliterated-finetuned-GGUF", MessagesFormatterType.LLAMA_3],
721
+ "Neo-1-7B-Flash.Q5_K_M.gguf": ["mradermacher/Neo-1-7B-Flash-GGUF", MessagesFormatterType.OPEN_CHAT],
722
+ "NightWind-12B.Q4_K_M.gguf": ["mradermacher/NightWind-12B-GGUF", MessagesFormatterType.CHATML],
723
+ "ablation-81-bestofn.athenev2-shisa-v2-llama-3.1-8b-lr8e6.Q5_K_M.gguf": ["mradermacher/ablation-81-bestofn.athenev2-shisa-v2-llama-3.1-8b-lr8e6-GGUF", MessagesFormatterType.LLAMA_3],
724
+ "formula-bs-mix-7B-level-1-0318.Q5_K_M.gguf": ["mradermacher/formula-bs-mix-7B-level-1-0318-GGUF", MessagesFormatterType.OPEN_CHAT],
725
+ "nano-r1-model.Q5_K_M.gguf": ["mradermacher/nano-r1-model-GGUF", MessagesFormatterType.OPEN_CHAT],
726
+ "NM-12B-Lyris-dev-2.Q4_K_M.gguf": ["mradermacher/NM-12B-Lyris-dev-2-GGUF", MessagesFormatterType.MISTRAL],
727
+ "RAt0.25-gemma-2-RI-9B.Q4_K_M.gguf": ["mradermacher/RAt0.25-gemma-2-RI-9B-GGUF", MessagesFormatterType.ALPACA],
728
+ "MN-Hecate-Chthonia-12B.Q4_K_M.gguf": ["mradermacher/MN-Hecate-Chthonia-12B-GGUF", MessagesFormatterType.MISTRAL],
729
+ "Dark-Science-Forgotten-Safeword-12B.Q4_K_M.gguf": ["mradermacher/Dark-Science-Forgotten-Safeword-12B-GGUF", MessagesFormatterType.MISTRAL],
730
+ "BBmath-Qwen2.5-3B-GRPO.Q5_K_M.gguf": ["mradermacher/BBmath-Qwen2.5-3B-GRPO-GGUF", MessagesFormatterType.OPEN_CHAT],
731
+ "Delta-Pavonis-Qwen-14B.Q4_K_S.gguf": ["mradermacher/Delta-Pavonis-Qwen-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
732
+ "BeaverAI_MN-2407-DSK-QwQify-v0.1-12B-Q4_K_M.gguf": ["bartowski/BeaverAI_MN-2407-DSK-QwQify-v0.1-12B-GGUF", MessagesFormatterType.MISTRAL],
733
+ "EtherealNight-12B.Q4_K_S.gguf": ["mradermacher/EtherealNight-12B-GGUF", MessagesFormatterType.CHATML],
734
+ "MN-Nyx-Chthonia-12B.i1-IQ3_M.gguf": ["mradermacher/MN-Nyx-Chthonia-12B-i1-GGUF", MessagesFormatterType.CHATML],
735
+ "Nu2-Lupi-Qwen-14B.i1-Q4_K_S.gguf": ["mradermacher/Nu2-Lupi-Qwen-14B-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
736
+ "KPT-7b-v0.2_stock.Q5_K_M.gguf": ["mradermacher/KPT-7b-v0.2_stock-GGUF", MessagesFormatterType.MISTRAL],
737
+ "Violet-Lyra-Gutenberg-v4.Q4_K_M.gguf": ["mradermacher/Violet-Lyra-Gutenberg-v4-GGUF", MessagesFormatterType.MISTRAL],
738
+ "Kosmos-EVAA-v12-8B.Q5_K_M.gguf": ["mradermacher/Kosmos-EVAA-v12-8B-GGUF", MessagesFormatterType.LLAMA_3],
739
+ "mistral-nemo-instruct-2407-Reasoning.Q4_K_M.gguf": ["mradermacher/mistral-nemo-instruct-2407-Reasoning-GGUF", MessagesFormatterType.MISTRAL],
740
+ "MT-Merge-gemma-2-9B.Q4_K_M.gguf": ["mradermacher/MT-Merge-gemma-2-9B-GGUF", MessagesFormatterType.ALPACA],
741
+ "Blagoveshchensk_14B_V4.Q4_K_M.gguf": ["mradermacher/Blagoveshchensk_14B_V4-GGUF", MessagesFormatterType.OPEN_CHAT],
742
+ "ablation-69-a67.dpo.armorm-ablation-67-nemo-12b.Q4_K_S.gguf": ["mradermacher/ablation-69-a67.dpo.armorm-ablation-67-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
743
+ "ablation-07-raftulu70b-shisa-v2-llama3.1-8b-lr8e6.Q5_K_M.gguf": ["mradermacher/ablation-07-raftulu70b-shisa-v2-llama3.1-8b-lr8e6-GGUF", MessagesFormatterType.LLAMA_3],
744
+ "ablation-60-a55.dpo.enja-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-60-a55.dpo.enja-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
745
+ "ablation-53-rafathenev2.rp-shisa-v2-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/ablation-53-rafathenev2.rp-shisa-v2-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
746
+ "ablation-61-a55.dpo.enjanot-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-61-a55.dpo.enjanot-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
747
+ "ablation-32-rafbestseq.reasoning-shisa-v2-llama-3.1-8b.Q5_K_M.gguf": ["mradermacher/ablation-32-rafbestseq.reasoning-shisa-v2-llama-3.1-8b-GGUF", MessagesFormatterType.LLAMA_3],
748
+ "ablation-42-rafathenev2.cmrmixen.masked-shisa-v2-llama-3.1-8b-lr8e6.Q5_K_M.gguf": ["mradermacher/ablation-42-rafathenev2.cmrmixen.masked-shisa-v2-llama-3.1-8b-lr8e6-GGUF", MessagesFormatterType.LLAMA_3],
749
+ "ablation-23-rafbestseqshuffle.2epoch-shisa-v2-llama-3.1-8b-lr8e6.Q5_K_M.gguf": ["mradermacher/ablation-23-rafbestseqshuffle.2epoch-shisa-v2-llama-3.1-8b-lr8e6-GGUF", MessagesFormatterType.LLAMA_3],
750
+ "ablation-67-rafathenev2.rp.tl.yahoo-mistral-nemo-12b.Q4_K_M.gguf": ["mradermacher/ablation-67-rafathenev2.rp.tl.yahoo-mistral-nemo-12b-GGUF", MessagesFormatterType.MISTRAL],
751
+ "EtherealMoon-12B.Q4_K_M.gguf": ["mradermacher/EtherealMoon-12B-GGUF", MessagesFormatterType.CHATML],
752
+ "ablation-56-rafathenev2.rp.tl.gamma-shisa-v2-gamma-7b.Q5_K_M.gguf": ["mradermacher/ablation-56-rafathenev2.rp.tl.gamma-shisa-v2-gamma-7b-GGUF", MessagesFormatterType.MISTRAL],
753
+ "ablation-11-rafathenev2-shisa-v2-llama-3.1-8b-lr8e6.Q5_K_M.gguf": ["mradermacher/ablation-11-rafathenev2-shisa-v2-llama-3.1-8b-lr8e6-GGUF", MessagesFormatterType.LLAMA_3],
754
+ "ablation-36-cmrmix-shisa-v2-llama-3.1-8b-lr8e6.Q5_K_M.gguf": ["mradermacher/ablation-36-cmrmix-shisa-v2-llama-3.1-8b-lr8e6-GGUF", MessagesFormatterType.LLAMA_3],
755
+ "L3.1-Athena-n-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Athena-n-8B-GGUF", MessagesFormatterType.LLAMA_3],
756
+ "NQLSG-Qwen2.5-14B-MegaFusion-v9.3.Q4_K_M.gguf": ["mradermacher/NQLSG-Qwen2.5-14B-MegaFusion-v9.3-GGUF", MessagesFormatterType.OPEN_CHAT],
757
+ "NQLSG-Qwen2.5-14B-MegaFusion-v9.2-fusechat-sft.Q4_K_M.gguf": ["mradermacher/NQLSG-Qwen2.5-14B-MegaFusion-v9.2-fusechat-sft-GGUF", MessagesFormatterType.OPEN_CHAT],
758
+ "TopicalChat_gemma.Q4_K_M.gguf": ["mradermacher/TopicalChat_gemma-GGUF", MessagesFormatterType.ALPACA],
759
+ "general7Bv2-ECE-PRYMMAL-Martial.Q5_K_M.gguf": ["mradermacher/general7Bv2-ECE-PRYMMAL-Martial-GGUF", MessagesFormatterType.OPEN_CHAT],
760
+ "NM-12B-Lyris-dev-3.Q4_K_M.gguf": ["mradermacher/NM-12B-Lyris-dev-3-GGUF", MessagesFormatterType.CHATML],
761
+ "DeepHermes-3-Llama-3-3B-Preview.i1-Q5_K_M.gguf": ["mradermacher/DeepHermes-3-Llama-3-3B-Preview-i1-GGUF", MessagesFormatterType.LLAMA_3],
762
+ "Kosmos-EVAA-v10-8B.Q5_K_M.gguf": ["mradermacher/Kosmos-EVAA-v10-8B-GGUF", MessagesFormatterType.LLAMA_3],
763
+ "Scrollplay.Q5_K_M.gguf": ["mradermacher/Scrollplay-GGUF", MessagesFormatterType.MISTRAL],
764
+ "Kosmos-EVAA-gamma-v17-8B.Q5_K_M.gguf": ["mradermacher/Kosmos-EVAA-gamma-v17-8B-GGUF", MessagesFormatterType.LLAMA_3],
765
+ "Cheng-2-Base.i1-Q4_K_M.gguf": ["mradermacher/Cheng-2-Base-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
766
+ "Cheng-2.i1-Q4_K_M.gguf": ["mradermacher/Cheng-2-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
767
+ "Perpicon-Experimental-3.1.Q5_K_M.gguf": ["mradermacher/Perpicon-Experimental-3.1-GGUF", MessagesFormatterType.LLAMA_3],
768
+ "Gemma-2-Ataraxy-v3b-9B.Q4_K_M.gguf": ["mradermacher/Gemma-2-Ataraxy-v3b-9B-GGUF", MessagesFormatterType.ALPACA],
769
+ "llama-3-1-8b-answer-generator-ft-2.Q5_K_M.gguf": ["mradermacher/llama-3-1-8b-answer-generator-ft-2-GGUF", MessagesFormatterType.LLAMA_3],
770
+ "nsfw-back-to-model-stock.Q5_K_M.gguf": ["mradermacher/nsfw-back-to-model-stock-GGUF", MessagesFormatterType.LLAMA_3],
771
+ "MN-Chthonia-12B.i1-Q4_K_M.gguf": ["mradermacher/MN-Chthonia-12B-i1-GGUF", MessagesFormatterType.MISTRAL],
772
+ "exurbia-enhanced-q4_k_m.gguf": ["ClaudioItaly/Exurbia-Enhanced-Q4_K_M-GGUF", MessagesFormatterType.ALPACA],
773
+ "Gemma-2-Ataraxy-v3j-9B.Q4_K_M.gguf": ["mradermacher/Gemma-2-Ataraxy-v3j-9B-GGUF", MessagesFormatterType.ALPACA],
774
+ "Abus-14B-Instruct.Q4_K_M.gguf": ["mradermacher/Abus-14B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
775
+ "Qwen2.5-7B-task8.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-task8-GGUF", MessagesFormatterType.OPEN_CHAT],
776
+ "Cheng-2-v1.1.Q4_K_M.gguf": ["mradermacher/Cheng-2-v1.1-GGUF", MessagesFormatterType.OPEN_CHAT],
777
+ "Qwen2.5-7B-Fuse-Exp.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Fuse-Exp-GGUF", MessagesFormatterType.OPEN_CHAT],
778
+ "Llama-3.1-8B-sft-ultrachat-gen-dpo-10k.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-sft-ultrachat-gen-dpo-10k-GGUF", MessagesFormatterType.LLAMA_3],
779
+ "NousResearch_DeepHermes-3-Llama-3-3B-Preview-Q5_K_M.gguf": ["bartowski/NousResearch_DeepHermes-3-Llama-3-3B-Preview-GGUF", MessagesFormatterType.LLAMA_3],
780
+ "Romulus-cpt-Llama-3.1-8B-v0.1-Instruct.Q5_K_M.gguf": ["mradermacher/Romulus-cpt-Llama-3.1-8B-v0.1-Instruct-GGUF", MessagesFormatterType.LLAMA_3],
781
+ "caliburn-v2-12b.Q4_K_M.gguf": ["mradermacher/caliburn-v2-12b-GGUF", MessagesFormatterType.MISTRAL],
782
+ "hermes-llama3-roleplay-4000-v2.Q5_K_M.gguf": ["mradermacher/hermes-llama3-roleplay-4000-v2-GGUF", MessagesFormatterType.LLAMA_3],
783
+ "hercule-bn.Q5_K_M.gguf": ["mradermacher/hercule-bn-GGUF", MessagesFormatterType.LLAMA_3],
784
+ "G2-GSHT-32K.Q4_K_M.gguf": ["mradermacher/G2-GSHT-32K-GGUF", MessagesFormatterType.ALPACA],
785
+ "krx-qwen2.5-7B-pt.Q5_K_M.gguf": ["mradermacher/krx-qwen2.5-7B-pt-GGUF", MessagesFormatterType.OPEN_CHAT],
786
+ "etherealaurora-12b-v3-q4_k_m.gguf": ["yamatazen/EtherealAurora-12B-v3-Q4_K_M-GGUF", MessagesFormatterType.MISTRAL],
787
+ "4PRYMMAL-GEMMA2-9B-SLERP.Q4_K_M.gguf": ["mradermacher/4PRYMMAL-GEMMA2-9B-SLERP-GGUF", MessagesFormatterType.ALPACA],
788
+ "gemma-2-aeria-9b.Q4_K_M.gguf": ["mradermacher/gemma-2-aeria-9b-GGUF", MessagesFormatterType.ALPACA],
789
+ "gemma-2-inko-9b.Q4_K_M.gguf": ["mradermacher/gemma-2-inko-9b-GGUF", MessagesFormatterType.ALPACA],
790
+ "krx-qwen2.5-7b-it-prompt.Q5_K_M.gguf": ["mradermacher/krx-qwen2.5-7b-it-prompt-GGUF", MessagesFormatterType.OPEN_CHAT],
791
+ "EmojiLlama-3.1-8B.Q5_K_M.gguf": ["mradermacher/EmojiLlama-3.1-8B-GGUF", MessagesFormatterType.LLAMA_3],
792
+ "Kosmos-EVAA-gamma-v15-8B.Q5_K_M.gguf": ["mradermacher/Kosmos-EVAA-gamma-v15-8B-GGUF", MessagesFormatterType.LLAMA_3],
793
+ "Qwen2.5-7B-Instruct-preference.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-Instruct-preference-GGUF", MessagesFormatterType.OPEN_CHAT],
794
+ "ZYH-LLM-Qwen2.5-14B-V4.Q4_K_M.gguf": ["mradermacher/ZYH-LLM-Qwen2.5-14B-V4-GGUF", MessagesFormatterType.OPEN_CHAT],
795
+ "Agente-GPT-Qwen2.5-3B-Instruct-Asistant-16bit-v2.Q5_K_M.gguf": ["mradermacher/Agente-GPT-Qwen2.5-3B-Instruct-Asistant-16bit-v2-GGUF", MessagesFormatterType.OPEN_CHAT],
796
+ "TopicalChat_mistral_nemo.Q4_K_M.gguf": ["mradermacher/TopicalChat_mistral_nemo-GGUF", MessagesFormatterType.MISTRAL],
797
+ "Llama-3.2-3B-Fluxed-uncensored.Q5_K_M.gguf": ["mradermacher/Llama-3.2-3B-Fluxed-uncensored-GGUF", MessagesFormatterType.LLAMA_3],
798
+ "Qwen2.5-Coder-3B-few-shot-model.Q5_K_M.gguf": ["mradermacher/Qwen2.5-Coder-3B-few-shot-model-GGUF", MessagesFormatterType.OPEN_CHAT],
799
+ "Gigantes-v2-gemma2-9b-it.Q4_K_M.gguf": ["mradermacher/Gigantes-v2-gemma2-9b-it-GGUF", MessagesFormatterType.ALPACA],
800
+ "Unichat-llama3.1-Chinese-8B.Q5_K_M.gguf": ["mradermacher/Unichat-llama3.1-Chinese-8B-GGUF", MessagesFormatterType.LLAMA_3],
801
+ "HollandV2R1.Q5_K_M.gguf": ["mradermacher/HollandV2R1-GGUF", MessagesFormatterType.LLAMA_3],
802
+ "Qwen-2.5-3b-Text_to_SQL.Q5_K_M.gguf": ["mradermacher/Qwen-2.5-3b-Text_to_SQL-GGUF", MessagesFormatterType.OPEN_CHAT],
803
+ "TableGPT2-7B.Q5_K_M.gguf": ["mradermacher/TableGPT2-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
804
+ "Kosmos-EVAA-gamma-v14-8B.Q5_K_M.gguf": ["mradermacher/Kosmos-EVAA-gamma-v14-8B-GGUF", MessagesFormatterType.LLAMA_3],
805
+ "KPT-7b-v0.3.Q5_K_M.gguf": ["mradermacher/KPT-7b-v0.3-GGUF", MessagesFormatterType.MISTRAL],
806
+ "qwen2.5-coder-7b-instruct-fma.Q5_K_M.gguf": ["mradermacher/qwen2.5-coder-7b-instruct-fma-GGUF", MessagesFormatterType.OPEN_CHAT],
807
+ "oh_v1.3_unnatural_instructions_x8.Q5_K_M.gguf": ["mradermacher/oh_v1.3_unnatural_instructions_x8-GGUF", MessagesFormatterType.LLAMA_3],
808
+ "Qwen2.5-Taiwan-7B-Instruct.Q5_K_M.gguf": ["mradermacher/Qwen2.5-Taiwan-7B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
809
+ "NQLSG-Qwen2.5-14B-MegaFusion-v9.2.Q4_K_M.gguf": ["mradermacher/NQLSG-Qwen2.5-14B-MegaFusion-v9.2-GGUF", MessagesFormatterType.OPEN_CHAT],
810
+ "Ext2Gen-8B-R2.Q5_K_M.gguf": ["mradermacher/Ext2Gen-8B-R2-GGUF", MessagesFormatterType.LLAMA_3],
811
+ "Meta-Llama-3-8B-Instruct-abliterated.i1-Q5_K_M.gguf": ["mradermacher/Meta-Llama-3-8B-Instruct-abliterated-i1-GGUF", MessagesFormatterType.LLAMA_3],
812
+ "Dirty-Shirley-Writer-v3-9B.i1-Q4_K_M.gguf": ["mradermacher/Dirty-Shirley-Writer-v3-9B-i1-GGUF", MessagesFormatterType.ALPACA],
813
+ "EtherealAurora-12B-v2.Q4_K_M.gguf": ["mradermacher/EtherealAurora-12B-v2-GGUF", MessagesFormatterType.MISTRAL],
814
+ "Galactic-Qwen-14B-Exp2.Q4_K_M.gguf": ["mradermacher/Galactic-Qwen-14B-Exp2-GGUF", MessagesFormatterType.OPEN_CHAT],
815
+ "Qwen2.5-14B-Verged.Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Verged-GGUF", MessagesFormatterType.OPEN_CHAT],
816
+ "Galactic-Qwen-14B-Exp1.Q4_K_M.gguf": ["mradermacher/Galactic-Qwen-14B-Exp1-GGUF", MessagesFormatterType.OPEN_CHAT],
817
+ "Volans-Opus-14B-Exp.Q4_K_M.gguf": ["mradermacher/Volans-Opus-14B-Exp-GGUF", MessagesFormatterType.OPEN_CHAT],
818
+ "MN-Anathema-12B.Q4_K_M.gguf": ["mradermacher/MN-Anathema-12B-GGUF", MessagesFormatterType.CHATML],
819
+ "Gaea-Opus-14B-Exp.Q4_K_M.gguf": ["mradermacher/Gaea-Opus-14B-Exp-GGUF", MessagesFormatterType.OPEN_CHAT],
820
+ "OnlyQ.Q5_K_M.gguf": ["mradermacher/OnlyQ-GGUF", MessagesFormatterType.OPEN_CHAT],
821
+ "asteroid-14b-v0.1.Q4_K_M.gguf": ["mradermacher/asteroid-14b-v0.1-GGUF", MessagesFormatterType.OPEN_CHAT],
822
+ "Forgotten-Safeword-8B-V3.0.Q5_K_M.gguf": ["mradermacher/Forgotten-Safeword-8B-V3.0-GGUF", MessagesFormatterType.MISTRAL],
823
+ "Qwen2.5-7B-PubMeQA-Distill.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-PubMeQA-Distill-GGUF", MessagesFormatterType.OPEN_CHAT],
824
+ "Llama-3.2-3b-RP-Toxic-Fuse.Q5_K_M.gguf": ["mradermacher/Llama-3.2-3b-RP-Toxic-Fuse-GGUF", MessagesFormatterType.LLAMA_3],
825
+ "v1-deepseek-instruct.Q5_K_M.gguf": ["mradermacher/v1-deepseek-instruct-GGUF", MessagesFormatterType.LLAMA_3],
826
+ "L3.1-Athena-l2-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Athena-l2-8B-GGUF", MessagesFormatterType.LLAMA_3],
827
+ "L3.1-Athena-l4-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Athena-l4-8B-GGUF", MessagesFormatterType.LLAMA_3],
828
+ "L3.1-Athena-l3-8B.Q5_K_M.gguf": ["mradermacher/L3.1-Athena-l3-8B-GGUF", MessagesFormatterType.LLAMA_3],
829
+ "Llama-3.1-8B-AthenaSky-MegaMix.i1-Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-AthenaSky-MegaMix-i1-GGUF", MessagesFormatterType.LLAMA_3],
830
+ "Writing-Model-Qwen-7B.Q5_K_M.gguf": ["mradermacher/Writing-Model-Qwen-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
831
+ "open-r1_OlympicCoder-7B-Q5_K_M.gguf": ["bartowski/open-r1_OlympicCoder-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
832
+ "Qwen-2.5-7B-deepscale-RL.Q5_K_M.gguf": ["mradermacher/Qwen-2.5-7B-deepscale-RL-GGUF", MessagesFormatterType.OPEN_CHAT],
833
+ "L3.1-Athena-l-8B.i1-Q5_K_M.gguf": ["mradermacher/L3.1-Athena-l-8B-i1-GGUF", MessagesFormatterType.LLAMA_3],
834
+ "Llama-3.2-3B-Instruct-lyrics.Q5_K_M.gguf": ["mradermacher/Llama-3.2-3B-Instruct-lyrics-GGUF", MessagesFormatterType.LLAMA_3],
835
+ "Llama-3-Base-8B-DICE-Iter2.Q5_K_M.gguf": ["mradermacher/Llama-3-Base-8B-DICE-Iter2-GGUF", MessagesFormatterType.LLAMA_3],
836
+ "Llama-3-Base-8B-DICE-Iter1.Q5_K_M.gguf": ["mradermacher/Llama-3-Base-8B-DICE-Iter1-GGUF", MessagesFormatterType.LLAMA_3],
837
+ "Llama-3-8B-Instruct-Coding-Expert.Q5_K_M.gguf": ["mradermacher/Llama-3-8B-Instruct-Coding-Expert-GGUF", MessagesFormatterType.LLAMA_3],
838
+ "Llama-O-3.1-8B-levent.Q5_K_M.gguf": ["mradermacher/Llama-O-3.1-8B-levent-GGUF", MessagesFormatterType.LLAMA_3],
839
+ "Qwen2.5-7B-task6.Q5_K_M.gguf": ["mradermacher/Qwen2.5-7B-task6-GGUF", MessagesFormatterType.OPEN_CHAT],
840
+ "finance-Llama3-8B.Q5_K_M.gguf": ["mradermacher/finance-Llama3-8B-GGUF", MessagesFormatterType.LLAMA_3],
841
+ "OpenPipe_Deductive-Reasoning-Qwen-14B-Q4_K_M.gguf": ["bartowski/OpenPipe_Deductive-Reasoning-Qwen-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
842
+ "Dirty-Shirley-Writer-v2-9B-Uncensored.i1-Q4_K_M.gguf": ["mradermacher/Dirty-Shirley-Writer-v2-9B-Uncensored-i1-GGUF", MessagesFormatterType.ALPACA],
843
+ "ichor_1.1-8B-Model_Stock.Q5_K_M.gguf": ["mradermacher/ichor_1.1-8B-Model_Stock-GGUF", MessagesFormatterType.LLAMA_3],
844
+ "gemma-2-2b-fa-v2.Q5_K_M.gguf": ["mradermacher/gemma-2-2b-fa-v2-GGUF", MessagesFormatterType.ALPACA],
845
+ "H_the_eighth-8B-LINEAR.Q5_K_M.gguf": ["mradermacher/H_the_eighth-8B-LINEAR-GGUF", MessagesFormatterType.LLAMA_3],
846
+ "granite-3.2-8b-instruct.Q5_K_M.gguf": ["mradermacher/granite-3.2-8b-instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
847
+ "tempestissimo-14b-0309.Q4_K_M.gguf": ["mradermacher/tempestissimo-14b-0309-GGUF", MessagesFormatterType.OPEN_CHAT],
848
+ "Deepseek-R1-Distill-NSFW-RP-vRedux-Proper.Q5_K_M.gguf": ["mradermacher/Deepseek-R1-Distill-NSFW-RP-vRedux-Proper-GGUF", MessagesFormatterType.OPEN_CHAT],
849
+ "inexpertus_1.1-8B-LINEAR.Q5_K_M.gguf": ["mradermacher/inexpertus_1.1-8B-LINEAR-GGUF", MessagesFormatterType.LLAMA_3],
850
+ "Llama-3.1-8B-RainbowLight-EtherealMix.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-RainbowLight-EtherealMix-GGUF", MessagesFormatterType.LLAMA_3],
851
+ "Llama-3.1-8B-SuperNova-EtherealHermes.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-SuperNova-EtherealHermes-GGUF", MessagesFormatterType.LLAMA_3],
852
+ "Dirty-Shirley-Quill-v1-gemma-2-Ifable-9B-Uncensored-slerp.Q4_K_M.gguf": ["mradermacher/Dirty-Shirley-Quill-v1-gemma-2-Ifable-9B-Uncensored-slerp-GGUF", MessagesFormatterType.ALPACA],
853
+ "Llama3.1-SuperHawk-8B.Q5_K_M.gguf": ["mradermacher/Llama3.1-SuperHawk-8B-GGUF", MessagesFormatterType.LLAMA_3],
854
+ "Qwen-2.5-7B-Reasoning.Q5_K_M.gguf": ["mradermacher/Qwen-2.5-7B-Reasoning-GGUF", MessagesFormatterType.OPEN_CHAT],
855
+ "nsfw-sce-test-2-redux.Q5_K_M.gguf": ["mradermacher/nsfw-sce-test-2-redux-GGUF", MessagesFormatterType.LLAMA_3],
856
+ "DCFT-camel_physics_seed_science-etash.Q5_K_M.gguf": ["mradermacher/DCFT-camel_physics_seed_science-etash-GGUF", MessagesFormatterType.OPEN_CHAT],
857
+ "Qwen2.5-Coder-7B-Instruct-Ties.Q5_K_M.gguf": ["mradermacher/Qwen2.5-Coder-7B-Instruct-Ties-GGUF", MessagesFormatterType.OPEN_CHAT],
858
+ "Meta-Llama-3-8B-Instruct-abliterated.Q5_K_M.gguf": ["mradermacher/Meta-Llama-3-8B-Instruct-abliterated-GGUF", MessagesFormatterType.LLAMA_3],
859
+ "MFANN-llama3.1-abliterated-v2.Q5_K_M.gguf": ["mradermacher/MFANN-llama3.1-abliterated-v2-GGUF", MessagesFormatterType.LLAMA_3],
860
+ "lamarckvergence-14b-tensopolis-v1.Q4_K_M.gguf": ["mradermacher/lamarckvergence-14b-tensopolis-v1-GGUF", MessagesFormatterType.OPEN_CHAT],
861
+ "NQLSG-Qwen2.5-14B-MegaFusion-v9.1.i1-Q4_K_M.gguf": ["mradermacher/NQLSG-Qwen2.5-14B-MegaFusion-v9.1-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
862
+ "s1-Qwen2.5-7B-Instruct.Q5_K_M.gguf": ["mradermacher/s1-Qwen2.5-7B-Instruct-GGUF", MessagesFormatterType.OPEN_CHAT],
863
+ "s1-Qwen2.5-7B.Q5_K_M.gguf": ["mradermacher/s1-Qwen2.5-7B-GGUF", MessagesFormatterType.OPEN_CHAT],
864
+ "Llama3.1-IgneousIguana-8B.Q5_K_M.gguf": ["mradermacher/Llama3.1-IgneousIguana-8B-GGUF", MessagesFormatterType.LLAMA_3],
865
+ "Llama-3-Instruct-8B-SPPO-Iter3-SimPO-merge.Q5_K_M.gguf": ["mradermacher/Llama-3-Instruct-8B-SPPO-Iter3-SimPO-merge-GGUF", MessagesFormatterType.LLAMA_3],
866
+ "recoilme-gemma-2-Ataraxy-9B-v0.1.i1-Q4_K_M.gguf": ["mradermacher/recoilme-gemma-2-Ataraxy-9B-v0.1-i1-GGUF", MessagesFormatterType.ALPACA],
867
+ "Gemma-Moon-2-9B-it.Q4_K_M.gguf": ["mradermacher/Gemma-Moon-2-9B-it-GGUF", MessagesFormatterType.ALPACA],
868
+ "CardProjector-7B-v2.Q5_K_M.gguf": ["mradermacher/CardProjector-7B-v2-GGUF", MessagesFormatterType.OPEN_CHAT],
869
+ "PerpetualNight-12B.Q4_K_M.gguf": ["mradermacher/PerpetualNight-12B-GGUF", MessagesFormatterType.CHATML],
870
+ "Llama3.2-3B-Instruct-KAI.Q5_K_M.gguf": ["mradermacher/Llama3.2-3B-Instruct-KAI-GGUF", MessagesFormatterType.LLAMA_3],
871
+ "Llama-3.1-8B-SuperTulu-LexiNova.Q5_K_M.gguf": ["mradermacher/Llama-3.1-8B-SuperTulu-LexiNova-GGUF", MessagesFormatterType.LLAMA_3],
872
+ "Qwen2-1.5B-Ita_v5.Q5_K_M.gguf": ["mradermacher/Qwen2-1.5B-Ita_v5-GGUF", MessagesFormatterType.OPEN_CHAT],
873
+ "Llama-VARCO-8B-Instruct-LLaMAfied.i1-Q5_K_M.gguf": ["mradermacher/Llama-VARCO-8B-Instruct-LLaMAfied-i1-GGUF", MessagesFormatterType.LLAMA_3],
874
+ "llama3-1_8b_mlfoundations-dev-stackexchange_scicomp.Q5_K_M.gguf": ["mradermacher/llama3-1_8b_mlfoundations-dev-stackexchange_scicomp-GGUF", MessagesFormatterType.LLAMA_3],
875
+ "Gemma-Ko-Merge.i1-Q4_K_M.gguf": ["mradermacher/Gemma-Ko-Merge-i1-GGUF", MessagesFormatterType.ALPACA],
876
+ "gemma-2-9b-it-russian-strict-function-calling-DPO.Q4_K_M.gguf": ["mradermacher/gemma-2-9b-it-russian-strict-function-calling-DPO-GGUF", MessagesFormatterType.ALPACA],
877
+ "Llama-3.1-Supernova-Lite-Instruct-merge-abliterated-8B.Q5_K_M.gguf": ["mradermacher/Llama-3.1-Supernova-Lite-Instruct-merge-abliterated-8B-GGUF", MessagesFormatterType.LLAMA_3],
878
+ "Qwenslerp2-14B-F32.Q4_K_M.gguf": ["mradermacher/Qwenslerp2-14B-F32-GGUF", MessagesFormatterType.OPEN_CHAT],
879
+ "Llama-3-Instruct-8B-SimPO-SPPO-Iter3-merge.Q5_K_M.gguf": ["mradermacher/Llama-3-Instruct-8B-SimPO-SPPO-Iter3-merge-GGUF", MessagesFormatterType.LLAMA_3],
880
+ "gemma-2-baku-2b-it.i1-Q5_K_M.gguf": ["mradermacher/gemma-2-baku-2b-it-i1-GGUF", MessagesFormatterType.ALPACA],
881
+ "Nemo-Picaro-12B.i1-Q4_K_M.gguf": ["mradermacher/Nemo-Picaro-12B-i1-GGUF", MessagesFormatterType.CHATML],
882
+ "OpenCrystal-L3-15B-v2.1.Q4_K_M.gguf": ["mradermacher/OpenCrystal-L3-15B-v2.1-GGUF", MessagesFormatterType.LLAMA_3],
883
+ "LLMEvoLLaMA-3.1-8B-v0.1.Q5_K_M.gguf": ["mradermacher/LLMEvoLLaMA-3.1-8B-v0.1-GGUF", MessagesFormatterType.LLAMA_3],
884
+ "Peach-2.0-9B-8k-Roleplay.Q4_K_M.gguf": ["mradermacher/Peach-2.0-9B-8k-Roleplay-GGUF", MessagesFormatterType.CHATML],
885
+ "llama-3.1-8b-sarcasm.Q5_K_M.gguf": ["mradermacher/llama-3.1-8b-sarcasm-GGUF", MessagesFormatterType.LLAMA_3],
886
+ "walrus-gemma-2-9b-it-v01.Q4_K_M.gguf": ["mradermacher/walrus-gemma-2-9b-it-v01-GGUF", MessagesFormatterType.ALPACA],
887
+ "L3.1-Diffbot-Small-2412.Q5_K_M.gguf": ["mradermacher/L3.1-Diffbot-Small-2412-GGUF", MessagesFormatterType.LLAMA_3],
888
+ "s1.1-14B.Q4_K_M.gguf": ["mradermacher/s1.1-14B-GGUF", MessagesFormatterType.OPEN_CHAT],
889
+ "Lix-14B-v0.1-Uncensored-DeLMAT.Q4_K_M.gguf": ["mradermacher/Lix-14B-v0.1-Uncensored-DeLMAT-GGUF", MessagesFormatterType.OPEN_CHAT],
890
  "Hush-Qwen2.5-7B-MST-v1.1.Q5_K_M.gguf": ["mradermacher/Hush-Qwen2.5-7B-MST-v1.1-GGUF", MessagesFormatterType.OPEN_CHAT],
891
  "MFANNv0.22.1.Q5_K_M.gguf": ["mradermacher/MFANNv0.22.1-GGUF", MessagesFormatterType.LLAMA_3],
892
  "gemma2-9b-cpt-sea-lionv3-instruct.Q4_K_M.gguf": ["mradermacher/gemma2-9b-cpt-sea-lionv3-instruct-GGUF", MessagesFormatterType.ALPACA],
lora_dict.json CHANGED
@@ -8882,6 +8882,13 @@
8882
  "https://civitai.com/models/491760",
8883
  "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/36267afb-7093-4c71-ae47-516c7219fabd/width=450/14403049.jpeg"
8884
  ],
 
 
 
 
 
 
 
8885
  "triple_oral_-_triple_barrel_fellatio_-_triple_fellatio-000014": [
8886
  "tripplebj / read description",
8887
  "Pony",
@@ -9470,6 +9477,13 @@
9470
  "https://civitai.com/models/850497",
9471
  "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/0a48013a-809a-41a5-823d-e9183eae0dd0/width=450/34362448.jpeg"
9472
  ],
 
 
 
 
 
 
 
9473
  "\u7523\u5375\u65ad\u9762\u56f3": [
9474
  "girl, egg laying, eggs in uterus",
9475
  "Pony",
 
8882
  "https://civitai.com/models/491760",
8883
  "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/36267afb-7093-4c71-ae47-516c7219fabd/width=450/14403049.jpeg"
8884
  ],
8885
+ "tribbie-hsrIL-v1": [
8886
+ "tribbie, wings, detached collar, puffy long sleeves, black gloves, black and white boots",
8887
+ "Illustrious",
8888
+ "[IL & PONY & SD1.5] Honkai Star Rail - Tribbie | \u7f07\u5b9d 3+1",
8889
+ "https://civitai.com/models/1340805",
8890
+ "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/f21f1b54-0b5a-4e92-ae08-6bf81eb2f0c2/width=1688/62634838.jpeg"
8891
+ ],
8892
  "triple_oral_-_triple_barrel_fellatio_-_triple_fellatio-000014": [
8893
  "tripplebj / read description",
8894
  "Pony",
 
9477
  "https://civitai.com/models/850497",
9478
  "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/0a48013a-809a-41a5-823d-e9183eae0dd0/width=450/34362448.jpeg"
9479
  ],
9480
+ "\u7231\u8bfa-Illustrious": [
9481
+ "aino (genshin impact), pink hair, grey eyes, bangs pinned back, hair ornament, green dress, white coat, oversized clothes, mechanical tail, sleeves past fingers, mismatched legwear, asymmetrical legwear, open coat, sleeves past wrists",
9482
+ "Illustrious",
9483
+ "\u3010Genshin Impact\u3011\u7231\u8bfa\u4e28Aino\u4e28\u30a2\u30a4\u30ce",
9484
+ "https://civitai.com/models/1792295",
9485
+ "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/ced12386-d16d-42b2-ad02-2ef70e736f5e/width=1536/94782692.jpeg"
9486
+ ],
9487
  "\u7523\u5375\u65ad\u9762\u56f3": [
9488
  "girl, egg laying, eggs in uterus",
9489
  "Pony",
modutils.py CHANGED
@@ -1730,7 +1730,7 @@ EXAMPLES_GUI = [
1730
  "votepurchase/ponyDiffusionV6XL",
1731
  ],
1732
  [
1733
- "1girl, oomuro sakurako, yuru yuri, official art, school uniform, anime artwork, anime style, vibrant, studio anime, highly detailed, masterpiece, best quality, very aesthetic, absurdres",
1734
  "photo, deformed, black and white, realism, disfigured, low contrast, lowres, (bad), text, error, fewer, extra, missing, worst quality, jpeg artifacts, low quality, watermark, unfinished, displeasing, oldest, early, chromatic aberration, signature, extra digits, artistic error, username, scan, [abstract]",
1735
  1,
1736
  40,
@@ -1743,17 +1743,17 @@ EXAMPLES_GUI = [
1743
  "Raelina/Rae-Diffusion-XL-V2",
1744
  ],
1745
  [
1746
- "1girl, akaza akari, yuru yuri, official art, anime artwork, anime style, vibrant, studio anime, highly detailed, masterpiece, best quality, very aesthetic, absurdres",
1747
- "photo, deformed, black and white, realism, disfigured, low contrast, lowres, (bad), text, error, fewer, extra, missing, worst quality, jpeg artifacts, low quality, watermark, unfinished, displeasing, oldest, early, chromatic aberration, signature, extra digits, artistic error, username, scan, [abstract]",
1748
  1,
1749
- 35,
1750
- 7.0,
1751
  True,
1752
  -1,
1753
  "Euler",
1754
  1024,
1755
  1024,
1756
- "Raelina/Raemu-XL-V5",
1757
  ],
1758
  [
1759
  "yoshida yuuko, machikado mazoku, 1girl, solo, demon horns,horns, school uniform, long hair, open mouth, skirt, demon girl, ahoge, shiny, shiny hair, anime artwork",
 
1730
  "votepurchase/ponyDiffusionV6XL",
1731
  ],
1732
  [
1733
+ "1girl, oomuro sakurako, yuru yuri, official art, school uniform, anime artwork, anime style, studio anime, highly detailed, masterpiece, best quality, very aesthetic, absurdres",
1734
  "photo, deformed, black and white, realism, disfigured, low contrast, lowres, (bad), text, error, fewer, extra, missing, worst quality, jpeg artifacts, low quality, watermark, unfinished, displeasing, oldest, early, chromatic aberration, signature, extra digits, artistic error, username, scan, [abstract]",
1735
  1,
1736
  40,
 
1743
  "Raelina/Rae-Diffusion-XL-V2",
1744
  ],
1745
  [
1746
+ "1girl, akaza akari, yuru yuri, official art, anime screencap, anime coloring, masterpiece, best quality, absurdres",
1747
+ "bad quality, worst quality, poorly drawn, sketch, multiple views, bad anatomy, bad hands, missing fingers, extra fingers, extra digits, fewer digits, signature, watermark, username",
1748
  1,
1749
+ 28,
1750
+ 5.5,
1751
  True,
1752
  -1,
1753
  "Euler",
1754
  1024,
1755
  1024,
1756
+ "Raelina/Raehoshi-illust-XL-6",
1757
  ],
1758
  [
1759
  "yoshida yuuko, machikado mazoku, 1girl, solo, demon horns,horns, school uniform, long hair, open mouth, skirt, demon girl, ahoge, shiny, shiny hair, anime artwork",
packages.txt CHANGED
@@ -1 +1 @@
1
- git-lfs aria2 -y ffmpeg
 
1
+ git-lfs aria2 ffmpeg
requirements.txt CHANGED
@@ -1,10 +1,8 @@
1
  stablepy==0.6.2
2
- #diffusers==0.31.0
3
- #transformers==4.47.1
4
  accelerate
5
- invisible_watermark
6
- datasets
7
- torch==2.4.0
8
  numpy<2
9
  gdown
10
  opencv-python
@@ -12,17 +10,19 @@ huggingface_hub
12
  hf_xet
13
  hf_transfer
14
  scikit-build-core
15
- https://github.com/abetlen/llama-cpp-python/releases/download/v0.3.4-cu124/llama_cpp_python-0.3.4-cp310-cp310-linux_x86_64.whl
16
- git+https://github.com/Maximilian-Winter/llama-cpp-agent
17
  pybind11>=2.12
18
  rapidfuzz
19
  torchvision
20
  optimum[onnxruntime]
21
- dartrs
 
22
  translatepy
23
  timm
24
  wrapt-timeout-decorator
25
  sentencepiece
26
  unidecode
27
- ultralytics>=8.3.47
 
28
  pydantic==2.10.6
 
1
  stablepy==0.6.2
2
+ diffusers
3
+ transformers<=4.48.3
4
  accelerate
5
+ torch==2.5.1
 
 
6
  numpy<2
7
  gdown
8
  opencv-python
 
10
  hf_xet
11
  hf_transfer
12
  scikit-build-core
13
+ https://github.com/abetlen/llama-cpp-python/releases/download/v0.3.16-cu124/llama_cpp_python-0.3.16-cp310-cp310-linux_x86_64.whl
14
+ git+https://github.com/John6666cat/llama-cpp-agent
15
  pybind11>=2.12
16
  rapidfuzz
17
  torchvision
18
  optimum[onnxruntime]
19
+ #dartrs
20
+ git+https://github.com/John6666cat/dartrs
21
  translatepy
22
  timm
23
  wrapt-timeout-decorator
24
  sentencepiece
25
  unidecode
26
+ matplotlib-inline
27
+ https://github.com/mjun0812/flash-attention-prebuild-wheels/releases/download/v0.4.11/flash_attn-2.8.3+cu124torch2.5-cp310-cp310-linux_x86_64.whl
28
  pydantic==2.10.6
tagger/fl2sd3longcap.py CHANGED
@@ -1,16 +1,22 @@
1
- from transformers import AutoProcessor, AutoModelForCausalLM
2
  import spaces
 
3
  import re
4
  from PIL import Image
5
  import torch
6
 
7
- import subprocess
8
- subprocess.run('pip install flash-attn --no-build-isolation', env={'FLASH_ATTENTION_SKIP_CUDA_BUILD': "TRUE"}, shell=True)
 
 
9
 
10
  device = "cuda" if torch.cuda.is_available() else "cpu"
11
- fl_model = AutoModelForCausalLM.from_pretrained('gokaygokay/Florence-2-SD3-Captioner', trust_remote_code=True).to("cpu").eval()
12
- fl_processor = AutoProcessor.from_pretrained('gokaygokay/Florence-2-SD3-Captioner', trust_remote_code=True)
13
 
 
 
 
 
 
 
14
 
15
  def fl_modify_caption(caption: str) -> str:
16
  """
@@ -41,7 +47,7 @@ def fl_modify_caption(caption: str) -> str:
41
  return modified_caption if modified_caption != caption else caption
42
 
43
 
44
- @spaces.GPU(duration=30)
45
  def fl_run_example(image):
46
  task_prompt = "<DESCRIPTION>"
47
  prompt = task_prompt + "Describe this image in great detail."
 
 
1
  import spaces
2
+ from transformers import AutoProcessor, AutoModelForCausalLM
3
  import re
4
  from PIL import Image
5
  import torch
6
 
7
+ from transformers.utils import is_flash_attn_2_available
8
+ if not is_flash_attn_2_available():
9
+ import subprocess
10
+ subprocess.run('pip install flash-attn --no-build-isolation', env={'FLASH_ATTENTION_SKIP_CUDA_BUILD': "TRUE"}, shell=True)
11
 
12
  device = "cuda" if torch.cuda.is_available() else "cpu"
 
 
13
 
14
+ try:
15
+ fl_model = AutoModelForCausalLM.from_pretrained('gokaygokay/Florence-2-SD3-Captioner', trust_remote_code=True).to("cpu").eval()
16
+ fl_processor = AutoProcessor.from_pretrained('gokaygokay/Florence-2-SD3-Captioner', trust_remote_code=True)
17
+ except Exception as e:
18
+ print(e)
19
+ fl_model = fl_processor = None
20
 
21
  def fl_modify_caption(caption: str) -> str:
22
  """
 
47
  return modified_caption if modified_caption != caption else caption
48
 
49
 
50
+ @spaces.GPU(duration=10)
51
  def fl_run_example(image):
52
  task_prompt = "<DESCRIPTION>"
53
  prompt = task_prompt + "Describe this image in great detail."
tagger/tagger.py CHANGED
@@ -504,7 +504,7 @@ def gen_prompt(rating: list[str], character: list[str], general: list[str]):
504
  return ", ".join(all_tags)
505
 
506
 
507
- @spaces.GPU(duration=30)
508
  def predict_tags(image: Image.Image, general_threshold: float = 0.3, character_threshold: float = 0.8):
509
  inputs = wd_processor.preprocess(image, return_tensors="pt")
510
 
 
504
  return ", ".join(all_tags)
505
 
506
 
507
+ @spaces.GPU(duration=10)
508
  def predict_tags(image: Image.Image, general_threshold: float = 0.3, character_threshold: float = 0.8):
509
  inputs = wd_processor.preprocess(image, return_tensors="pt")
510
 
tagger/v2.py CHANGED
@@ -1,4 +1,4 @@
1
- import time
2
  import torch
3
  from typing import Callable
4
  from pathlib import Path
@@ -16,16 +16,15 @@ from dartrs.v2 import (
16
  from dartrs.dartrs import DartTokenizer
17
  from dartrs.utils import get_generation_config
18
 
19
-
20
  import gradio as gr
21
  from gradio.components import Component
22
 
23
-
24
  try:
25
  from output import UpsamplingOutput
26
  except:
27
  from .output import UpsamplingOutput
28
 
 
29
 
30
  V2_ALL_MODELS = {
31
  "dart-v2-moe-sft": {
@@ -43,8 +42,8 @@ V2_ALL_MODELS = {
43
 
44
  def prepare_models(model_config: dict):
45
  model_name = model_config["repo"]
46
- tokenizer = DartTokenizer.from_pretrained(model_name)
47
- model = model_config["class"].from_pretrained(model_name)
48
 
49
  return {
50
  "tokenizer": tokenizer,
 
1
+ import time, os
2
  import torch
3
  from typing import Callable
4
  from pathlib import Path
 
16
  from dartrs.dartrs import DartTokenizer
17
  from dartrs.utils import get_generation_config
18
 
 
19
  import gradio as gr
20
  from gradio.components import Component
21
 
 
22
  try:
23
  from output import UpsamplingOutput
24
  except:
25
  from .output import UpsamplingOutput
26
 
27
+ HF_TOKEN = os.getenv("HF_TOKEN", None)
28
 
29
  V2_ALL_MODELS = {
30
  "dart-v2-moe-sft": {
 
42
 
43
  def prepare_models(model_config: dict):
44
  model_name = model_config["repo"]
45
+ tokenizer = DartTokenizer.from_pretrained(model_name, auth_token=HF_TOKEN)
46
+ model = model_config["class"].from_pretrained(model_name, auth_token=HF_TOKEN)
47
 
48
  return {
49
  "tokenizer": tokenizer,
utils.py CHANGED
@@ -9,6 +9,7 @@ from constants import (
9
  DIRECTORY_LORAS,
10
  DIRECTORY_MODELS,
11
  DIFFUSECRAFT_CHECKPOINT_NAME,
 
12
  CACHE_HF,
13
  STORAGE_ROOT,
14
  )
@@ -28,6 +29,7 @@ from urllib3.util import Retry
28
  import shutil
29
  import subprocess
30
 
 
31
  USER_AGENT = 'Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:127.0) Gecko/20100101 Firefox/127.0'
32
 
33
 
@@ -66,7 +68,8 @@ class ModelInformation:
66
  )
67
  self.filename_url = self.filename_url if self.filename_url else ""
68
  self.description = json_data.get("description", "")
69
- if self.description is None: self.description = ""
 
70
  self.model_name = json_data.get("model", {}).get("name", "")
71
  self.model_type = json_data.get("model", {}).get("type", "")
72
  self.nsfw = json_data.get("model", {}).get("nsfw", False)
@@ -76,118 +79,175 @@ class ModelInformation:
76
  self.original_json = copy.deepcopy(json_data)
77
 
78
 
79
- def retrieve_model_info(url):
80
- json_data = request_json_data(url)
81
- if not json_data:
82
- return None
83
- model_descriptor = ModelInformation(json_data)
84
- return model_descriptor
 
 
 
 
 
85
 
86
 
87
- def download_things(directory, url, hf_token="", civitai_api_key="", romanize=False):
88
- url = url.strip()
89
- downloaded_file_path = None
90
 
91
- if "drive.google.com" in url:
92
- original_dir = os.getcwd()
93
- os.chdir(directory)
94
- os.system(f"gdown --fuzzy {url}")
95
- os.chdir(original_dir)
96
- elif "huggingface.co" in url:
97
- url = url.replace("?download=true", "")
98
- # url = urllib.parse.quote(url, safe=':/') # fix encoding
99
- if "/blob/" in url:
100
- url = url.replace("/blob/", "/resolve/")
101
- user_header = f'"Authorization: Bearer {hf_token}"'
102
 
103
- filename = unidecode(url.split('/')[-1]) if romanize else url.split('/')[-1]
 
 
 
 
 
104
 
105
- if hf_token:
106
- os.system(f"aria2c --console-log-level=error --summary-interval=10 --header={user_header} -c -x 16 -k 1M -s 16 {url} -d {directory} -o {filename}")
107
- else:
108
- os.system(f"aria2c --optimize-concurrent-downloads --console-log-level=error --summary-interval=10 -c -x 16 -k 1M -s 16 {url} -d {directory} -o {filename}")
109
 
110
- downloaded_file_path = os.path.join(directory, filename)
 
111
 
112
- elif "civitai.com" in url:
 
113
 
114
- if not civitai_api_key:
115
- print("\033[91mYou need an API key to download Civitai models.\033[0m")
116
-
117
- model_profile = retrieve_model_info(url)
118
- if (
119
- model_profile is not None
120
- and model_profile.download_url
121
- and model_profile.filename_url
122
- ):
123
- url = model_profile.download_url
124
- filename = unidecode(model_profile.filename_url) if romanize else model_profile.filename_url
125
- else:
126
- if "?" in url:
127
- url = url.split("?")[0]
128
- filename = ""
129
 
130
- url_dl = url + f"?token={civitai_api_key}"
131
- print(f"Filename: {filename}")
132
 
133
- param_filename = ""
134
- if filename:
135
- param_filename = f"-o '{filename}'"
 
 
136
 
137
- aria2_command = (
138
- f'aria2c --console-log-level=error --summary-interval=10 -c -x 16 '
139
- f'-k 1M -s 16 -d "{directory}" {param_filename} "{url_dl}"'
140
- )
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
141
  os.system(aria2_command)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
142
 
143
- if param_filename and os.path.exists(os.path.join(directory, filename)):
144
- downloaded_file_path = os.path.join(directory, filename)
145
-
146
- # # PLAN B
147
- # # Follow the redirect to get the actual download URL
148
- # curl_command = (
149
- # f'curl -L -sI --connect-timeout 5 --max-time 5 '
150
- # f'-H "Content-Type: application/json" '
151
- # f'-H "Authorization: Bearer {civitai_api_key}" "{url}"'
152
- # )
153
-
154
- # headers = os.popen(curl_command).read()
155
-
156
- # # Look for the redirected "Location" URL
157
- # location_match = re.search(r'location: (.+)', headers, re.IGNORECASE)
158
-
159
- # if location_match:
160
- # redirect_url = location_match.group(1).strip()
161
-
162
- # # Extract the filename from the redirect URL's "Content-Disposition"
163
- # filename_match = re.search(r'filename%3D%22(.+?)%22', redirect_url)
164
- # if filename_match:
165
- # encoded_filename = filename_match.group(1)
166
- # # Decode the URL-encoded filename
167
- # decoded_filename = urllib.parse.unquote(encoded_filename)
168
-
169
- # filename = unidecode(decoded_filename) if romanize else decoded_filename
170
- # print(f"Filename: {filename}")
171
-
172
- # aria2_command = (
173
- # f'aria2c --console-log-level=error --summary-interval=10 -c -x 16 '
174
- # f'-k 1M -s 16 -d "{directory}" -o "{filename}" "{redirect_url}"'
175
- # )
176
- # return_code = os.system(aria2_command)
177
-
178
- # # if return_code != 0:
179
- # # raise RuntimeError(f"Failed to download file: {filename}. Error code: {return_code}")
180
- # downloaded_file_path = os.path.join(directory, filename)
181
- # if not os.path.exists(downloaded_file_path):
182
- # downloaded_file_path = None
183
-
184
- # if not downloaded_file_path:
185
- # # Old method
186
- # if "?" in url:
187
- # url = url.split("?")[0]
188
- # url = url + f"?token={civitai_api_key}"
189
- # os.system(f"aria2c --console-log-level=error --summary-interval=10 -c -x 16 -k 1M -s 16 -d {directory} {url}")
190
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
191
  else:
192
  os.system(f"aria2c --console-log-level=error --summary-interval=10 -c -x 16 -k 1M -s 16 -d {directory} {url}")
193
 
@@ -216,14 +276,15 @@ def extract_parameters(input_string):
216
  if "Steps:" in input_string:
217
  input_string = input_string.replace("Steps:", "Negative prompt: Steps:")
218
  else:
219
- print("Invalid metadata")
 
 
220
  parameters["prompt"] = input_string
221
  return parameters
222
 
223
  parm = input_string.split("Negative prompt:")
224
  parameters["prompt"] = parm[0].strip()
225
  if "Steps:" not in parm[1]:
226
- print("Steps not detected")
227
  parameters["neg_prompt"] = parm[1].strip()
228
  return parameters
229
  parm = parm[1].split("Steps:")
@@ -306,7 +367,8 @@ def get_model_type(repo_id: str):
306
  model = api.model_info(repo_id=repo_id, timeout=5.0)
307
  tags = model.tags
308
  for tag in tags:
309
- if tag in MODEL_TYPE_CLASS.keys(): return MODEL_TYPE_CLASS.get(tag, default)
 
310
 
311
  except Exception:
312
  return default
@@ -433,9 +495,9 @@ def get_folder_size_gb(folder_path):
433
  return total_size_gb
434
 
435
 
436
- def get_used_storage_gb():
437
  try:
438
- used_gb = get_folder_size_gb(STORAGE_ROOT)
439
  print(f"Used Storage: {used_gb:.2f} GB")
440
  except Exception as e:
441
  used_gb = 999
@@ -455,6 +517,21 @@ def delete_model(removal_candidate):
455
  shutil.rmtree(diffusers_model)
456
 
457
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
458
  def progress_step_bar(step, total):
459
  # Calculate the percentage for the progress bar width
460
  percentage = min(100, ((step / total) * 100))
 
9
  DIRECTORY_LORAS,
10
  DIRECTORY_MODELS,
11
  DIFFUSECRAFT_CHECKPOINT_NAME,
12
+ CACHE_HF_ROOT,
13
  CACHE_HF,
14
  STORAGE_ROOT,
15
  )
 
29
  import shutil
30
  import subprocess
31
 
32
+ IS_ZERO_GPU = bool(os.getenv("SPACES_ZERO_GPU"))
33
  USER_AGENT = 'Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:127.0) Gecko/20100101 Firefox/127.0'
34
 
35
 
 
68
  )
69
  self.filename_url = self.filename_url if self.filename_url else ""
70
  self.description = json_data.get("description", "")
71
+ if self.description is None:
72
+ self.description = ""
73
  self.model_name = json_data.get("model", {}).get("name", "")
74
  self.model_type = json_data.get("model", {}).get("type", "")
75
  self.nsfw = json_data.get("model", {}).get("nsfw", False)
 
79
  self.original_json = copy.deepcopy(json_data)
80
 
81
 
82
+ def get_civit_params(url):
83
+ try:
84
+ json_data = request_json_data(url)
85
+ mdc = ModelInformation(json_data)
86
+ if mdc.download_url and mdc.filename_url:
87
+ return mdc.download_url, mdc.filename_url, mdc.model_url
88
+ else:
89
+ ValueError("Invalid Civitai model URL")
90
+ except Exception as e:
91
+ print(f"Error retrieving Civitai metadata: {e} — fallback to direct download")
92
+ return url, None, None
93
 
94
 
95
+ def civ_redirect_down(url, dir_, civitai_api_key, romanize, alternative_name):
96
+ filename_base = filename = None
 
97
 
98
+ if alternative_name:
99
+ output_path = os.path.join(dir_, alternative_name)
100
+ if os.path.exists(output_path):
101
+ return output_path, alternative_name
 
 
 
 
 
 
 
102
 
103
+ # Follow the redirect to get the actual download URL
104
+ curl_command = (
105
+ f'curl -L -sI --connect-timeout 5 --max-time 5 '
106
+ f'-H "Content-Type: application/json" '
107
+ f'-H "Authorization: Bearer {civitai_api_key}" "{url}"'
108
+ )
109
 
110
+ headers = os.popen(curl_command).read()
 
 
 
111
 
112
+ # Look for the redirected "Location" URL
113
+ location_match = re.search(r'location: (.+)', headers, re.IGNORECASE)
114
 
115
+ if location_match:
116
+ redirect_url = location_match.group(1).strip()
117
 
118
+ # Extract the filename from the redirect URL's "Content-Disposition"
119
+ filename_match = re.search(r'filename%3D%22(.+?)%22', redirect_url)
120
+ if filename_match:
121
+ encoded_filename = filename_match.group(1)
122
+ # Decode the URL-encoded filename
123
+ decoded_filename = urllib.parse.unquote(encoded_filename)
 
 
 
 
 
 
 
 
 
124
 
125
+ filename = unidecode(decoded_filename) if romanize else decoded_filename
126
+ # print(f"Filename redirect: {filename}")
127
 
128
+ filename_base = alternative_name if alternative_name else filename
129
+ if not filename_base:
130
+ return None, None
131
+ elif os.path.exists(os.path.join(dir_, filename_base)):
132
+ return os.path.join(dir_, filename_base), filename_base
133
 
134
+ aria2_command = (
135
+ f'aria2c --console-log-level=error --summary-interval=10 -c -x 16 '
136
+ f'-k 1M -s 16 -d "{dir_}" -o "{filename_base}" "{redirect_url}"'
137
+ )
138
+ r_code = os.system(aria2_command) # noqa
139
+
140
+ # if r_code != 0:
141
+ # raise RuntimeError(f"Failed to download file: {filename_base}. Error code: {r_code}")
142
+
143
+ output_path = os.path.join(dir_, filename_base)
144
+ if not os.path.exists(output_path):
145
+ return None, filename_base
146
+
147
+ return output_path, filename_base
148
+
149
+
150
+ def civ_api_down(url, dir_, civitai_api_key, civ_filename):
151
+ """
152
+ This method is susceptible to being blocked because it generates a lot of temp redirect links with aria2c.
153
+ If an API key limit is reached, generating a new API key and using it can fix the issue.
154
+ """
155
+ output_path = None
156
+
157
+ url_dl = url + f"?token={civitai_api_key}"
158
+ if not civ_filename:
159
+ aria2_command = f'aria2c -c -x 1 -s 1 -d "{dir_}" "{url_dl}"'
160
  os.system(aria2_command)
161
+ else:
162
+ output_path = os.path.join(dir_, civ_filename)
163
+ if not os.path.exists(output_path):
164
+ aria2_command = (
165
+ f'aria2c --console-log-level=error --summary-interval=10 -c -x 16 '
166
+ f'-k 1M -s 16 -d "{dir_}" -o "{civ_filename}" "{url_dl}"'
167
+ )
168
+ os.system(aria2_command)
169
+
170
+ return output_path
171
+
172
+
173
+ def drive_down(url, dir_):
174
+ import gdown
175
+
176
+ output_path = None
177
 
178
+ drive_id, _ = gdown.parse_url.parse_url(url, warning=False)
179
+ dir_files = os.listdir(dir_)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
180
 
181
+ for dfile in dir_files:
182
+ if drive_id in dfile:
183
+ output_path = os.path.join(dir_, dfile)
184
+ break
185
+
186
+ if not output_path:
187
+ original_path = gdown.download(url, f"{dir_}/", fuzzy=True)
188
+
189
+ dir_name, base_name = os.path.split(original_path)
190
+ name, ext = base_name.rsplit(".", 1)
191
+ new_name = f"{name}_{drive_id}.{ext}"
192
+ output_path = os.path.join(dir_name, new_name)
193
+
194
+ os.rename(original_path, output_path)
195
+
196
+ return output_path
197
+
198
+
199
+ def hf_down(url, dir_, hf_token, romanize):
200
+ url = url.replace("?download=true", "")
201
+ # url = urllib.parse.quote(url, safe=':/') # fix encoding
202
+
203
+ filename = unidecode(url.split('/')[-1]) if romanize else url.split('/')[-1]
204
+ output_path = os.path.join(dir_, filename)
205
+
206
+ if os.path.exists(output_path):
207
+ return output_path
208
+
209
+ if "/blob/" in url:
210
+ url = url.replace("/blob/", "/resolve/")
211
+
212
+ if hf_token:
213
+ user_header = f'"Authorization: Bearer {hf_token}"'
214
+ os.system(f"aria2c --console-log-level=error --summary-interval=10 --header={user_header} -c -x 16 -k 1M -s 16 {url} -d {dir_} -o {filename}")
215
+ else:
216
+ os.system(f"aria2c --optimize-concurrent-downloads --console-log-level=error --summary-interval=10 -c -x 16 -k 1M -s 16 {url} -d {dir_} -o {filename}")
217
+
218
+ return output_path
219
+
220
+
221
+ def download_things(directory, url, hf_token="", civitai_api_key="", romanize=False):
222
+ url = url.strip()
223
+ downloaded_file_path = None
224
+
225
+ if "drive.google.com" in url:
226
+ downloaded_file_path = drive_down(url, directory)
227
+ elif "huggingface.co" in url:
228
+ downloaded_file_path = hf_down(url, directory, hf_token, romanize)
229
+ elif "civitai.com" in url:
230
+ if not civitai_api_key:
231
+ msg = "You need an API key to download Civitai models."
232
+ print(f"\033[91m{msg}\033[0m")
233
+ gr.Warning(msg)
234
+ return None
235
+
236
+ url, civ_filename, civ_page = get_civit_params(url)
237
+ if civ_page and not IS_ZERO_GPU:
238
+ print(f"\033[92mCivitai model: {civ_filename} [page: {civ_page}]\033[0m")
239
+
240
+ downloaded_file_path, civ_filename = civ_redirect_down(url, directory, civitai_api_key, romanize, civ_filename)
241
+
242
+ if not downloaded_file_path:
243
+ msg = (
244
+ "Download failed.\n"
245
+ "If this is due to an API limit, generating a new API key may resolve the issue.\n"
246
+ "Attempting to download using the old method..."
247
+ )
248
+ print(msg)
249
+ gr.Warning(msg)
250
+ downloaded_file_path = civ_api_down(url, directory, civitai_api_key, civ_filename)
251
  else:
252
  os.system(f"aria2c --console-log-level=error --summary-interval=10 -c -x 16 -k 1M -s 16 -d {directory} {url}")
253
 
 
276
  if "Steps:" in input_string:
277
  input_string = input_string.replace("Steps:", "Negative prompt: Steps:")
278
  else:
279
+ msg = "Generation data is invalid."
280
+ gr.Warning(msg)
281
+ print(msg)
282
  parameters["prompt"] = input_string
283
  return parameters
284
 
285
  parm = input_string.split("Negative prompt:")
286
  parameters["prompt"] = parm[0].strip()
287
  if "Steps:" not in parm[1]:
 
288
  parameters["neg_prompt"] = parm[1].strip()
289
  return parameters
290
  parm = parm[1].split("Steps:")
 
367
  model = api.model_info(repo_id=repo_id, timeout=5.0)
368
  tags = model.tags
369
  for tag in tags:
370
+ if tag in MODEL_TYPE_CLASS.keys():
371
+ return MODEL_TYPE_CLASS.get(tag, default)
372
 
373
  except Exception:
374
  return default
 
495
  return total_size_gb
496
 
497
 
498
+ def get_used_storage_gb(path_storage=STORAGE_ROOT):
499
  try:
500
+ used_gb = get_folder_size_gb(path_storage)
501
  print(f"Used Storage: {used_gb:.2f} GB")
502
  except Exception as e:
503
  used_gb = 999
 
517
  shutil.rmtree(diffusers_model)
518
 
519
 
520
+ def clear_hf_cache():
521
+ """
522
+ Clears the entire Hugging Face cache at ~/.cache/huggingface.
523
+ Hugging Face will re-download models as needed later.
524
+ """
525
+ try:
526
+ if os.path.exists(CACHE_HF_ROOT):
527
+ shutil.rmtree(CACHE_HF_ROOT, ignore_errors=True)
528
+ print(f"Hugging Face cache cleared: {CACHE_HF_ROOT}")
529
+ else:
530
+ print(f"No Hugging Face cache found at: {CACHE_HF_ROOT}")
531
+ except Exception as e:
532
+ print(f"Error clearing Hugging Face cache: {e}")
533
+
534
+
535
  def progress_step_bar(step, total):
536
  # Calculate the percentage for the progress bar width
537
  percentage = min(100, ((step / total) * 100))