John6666 commited on
Commit
cac72ef
β€’
1 Parent(s): badf2fd

Upload 2 files

Browse files
Files changed (2) hide show
  1. llmdolphin.py +25 -18
  2. llmenv.py +74 -0
llmdolphin.py CHANGED
@@ -219,7 +219,7 @@ def get_raw_prompt(msg: str):
219
 
220
 
221
  @torch.inference_mode()
222
- @spaces.GPU(duration=60)
223
  def dolphin_respond(
224
  message: str,
225
  history: list[tuple[str, str]],
@@ -234,13 +234,15 @@ def dolphin_respond(
234
  progress=gr.Progress(track_tqdm=True),
235
  ):
236
  try:
 
 
237
  progress(0, desc="Processing...")
238
  override_llm_format = get_state(state, "override_llm_format")
239
  if override_llm_format: chat_template = override_llm_format
240
  else: chat_template = llm_models[model][1]
241
 
242
  llm = Llama(
243
- model_path=str(Path(f"{llm_models_dir}/{model}")),
244
  flash_attn=True,
245
  n_gpu_layers=81, # 81
246
  n_batch=1024,
@@ -294,8 +296,8 @@ def dolphin_respond(
294
  yield [(outputs, None)]
295
  except Exception as e:
296
  print(e)
297
- gr.Warning(f"Error: {e}")
298
- yield [("", None)]
299
 
300
 
301
  def dolphin_parse(
@@ -320,7 +322,7 @@ def dolphin_parse(
320
 
321
 
322
  @torch.inference_mode()
323
- @spaces.GPU(duration=60)
324
  def dolphin_respond_auto(
325
  message: str,
326
  history: list[tuple[str, str]],
@@ -335,6 +337,7 @@ def dolphin_respond_auto(
335
  progress=gr.Progress(track_tqdm=True),
336
  ):
337
  try:
 
338
  #if not is_japanese(message): return [(None, None)]
339
  progress(0, desc="Processing...")
340
 
@@ -343,7 +346,7 @@ def dolphin_respond_auto(
343
  else: chat_template = llm_models[model][1]
344
 
345
  llm = Llama(
346
- model_path=str(Path(f"{llm_models_dir}/{model}")),
347
  flash_attn=True,
348
  n_gpu_layers=81, # 81
349
  n_batch=1024,
@@ -429,26 +432,29 @@ cv2.setNumThreads(1)
429
 
430
 
431
  @torch.inference_mode()
432
- @spaces.GPU()
433
  def respond_playground(
434
- message,
435
  history: list[tuple[str, str]],
436
- model,
437
- system_message,
438
- max_tokens,
439
- temperature,
440
- top_p,
441
- top_k,
442
- repeat_penalty,
443
- state,
 
444
  ):
445
  try:
 
 
446
  override_llm_format = get_state(state, "override_llm_format")
447
  if override_llm_format: chat_template = override_llm_format
448
  else: chat_template = llm_models[model][1]
449
 
450
  llm = Llama(
451
- model_path=str(Path(f"{llm_models_dir}/{model}")),
452
  flash_attn=True,
453
  n_gpu_layers=81, # 81
454
  n_batch=1024,
@@ -496,4 +502,5 @@ def respond_playground(
496
  yield outputs
497
  except Exception as e:
498
  print(e)
499
- yield ""
 
 
219
 
220
 
221
  @torch.inference_mode()
222
+ @spaces.GPU(duration=59)
223
  def dolphin_respond(
224
  message: str,
225
  history: list[tuple[str, str]],
 
234
  progress=gr.Progress(track_tqdm=True),
235
  ):
236
  try:
237
+ model_path = Path(f"{llm_models_dir}/{model}")
238
+ if not model_path.exists(): raise gr.Error(f"Model file not found: {str(model_path)}")
239
  progress(0, desc="Processing...")
240
  override_llm_format = get_state(state, "override_llm_format")
241
  if override_llm_format: chat_template = override_llm_format
242
  else: chat_template = llm_models[model][1]
243
 
244
  llm = Llama(
245
+ model_path=str(model_path),
246
  flash_attn=True,
247
  n_gpu_layers=81, # 81
248
  n_batch=1024,
 
296
  yield [(outputs, None)]
297
  except Exception as e:
298
  print(e)
299
+ raise gr.Error(f"Error: {e}")
300
+ #yield [("", None)]
301
 
302
 
303
  def dolphin_parse(
 
322
 
323
 
324
  @torch.inference_mode()
325
+ @spaces.GPU(duration=59)
326
  def dolphin_respond_auto(
327
  message: str,
328
  history: list[tuple[str, str]],
 
337
  progress=gr.Progress(track_tqdm=True),
338
  ):
339
  try:
340
+ model_path = Path(f"{llm_models_dir}/{model}")
341
  #if not is_japanese(message): return [(None, None)]
342
  progress(0, desc="Processing...")
343
 
 
346
  else: chat_template = llm_models[model][1]
347
 
348
  llm = Llama(
349
+ model_path=str(model_path),
350
  flash_attn=True,
351
  n_gpu_layers=81, # 81
352
  n_batch=1024,
 
432
 
433
 
434
  @torch.inference_mode()
435
+ @spaces.GPU(duration=59)
436
  def respond_playground(
437
+ message: str,
438
  history: list[tuple[str, str]],
439
+ model: str = default_llm_model_filename,
440
+ system_message: str = get_dolphin_sysprompt(),
441
+ max_tokens: int = 1024,
442
+ temperature: float = 0.7,
443
+ top_p: float = 0.95,
444
+ top_k: int = 40,
445
+ repeat_penalty: float = 1.1,
446
+ state: dict = {},
447
+ progress=gr.Progress(track_tqdm=True),
448
  ):
449
  try:
450
+ model_path = Path(f"{llm_models_dir}/{model}")
451
+ if not model_path.exists(): raise gr.Error(f"Model file not found: {str(model_path)}")
452
  override_llm_format = get_state(state, "override_llm_format")
453
  if override_llm_format: chat_template = override_llm_format
454
  else: chat_template = llm_models[model][1]
455
 
456
  llm = Llama(
457
+ model_path=str(model_path),
458
  flash_attn=True,
459
  n_gpu_layers=81, # 81
460
  n_batch=1024,
 
502
  yield outputs
503
  except Exception as e:
504
  print(e)
505
+ raise gr.Error(f"Error: {e}")
506
+ #yield ""
llmenv.py CHANGED
@@ -67,6 +67,7 @@ llm_models = {
67
  "MISCHIEVOUS-12B-Mix_0.1v.Q4_K_M.gguf": ["mradermacher/MISCHIEVOUS-12B-Mix_0.1v-GGUF", MessagesFormatterType.MISTRAL],
68
  "Canidori-12B-v1.i1-Q4_K_M.gguf": ["mradermacher/Canidori-12B-v1-i1-GGUF", MessagesFormatterType.MISTRAL],
69
  "MT2-Gen4-MM-gemma-2-Rv0.4MTM-9B.Q4_K_M.gguf": ["mradermacher/MT2-Gen4-MM-gemma-2-Rv0.4MTM-9B-GGUF", MessagesFormatterType.ALPACA],
 
70
  "Trinas_Nectar-8B-model_stock.i1-Q4_K_M.gguf": ["mradermacher/Trinas_Nectar-8B-model_stock-i1-GGUF", MessagesFormatterType.MISTRAL],
71
  "ChatWaifu_Magnum_V0.2.Q4_K_M.gguf": ["mradermacher/ChatWaifu_Magnum_V0.2-GGUF", MessagesFormatterType.MISTRAL],
72
  "ChatWaifu_12B_v2.0.Q5_K_M.gguf": ["mradermacher/ChatWaifu_12B_v2.0-GGUF", MessagesFormatterType.MISTRAL],
@@ -79,6 +80,79 @@ llm_models = {
79
  #"": ["", MessagesFormatterType.OPEN_CHAT],
80
  #"": ["", MessagesFormatterType.CHATML],
81
  #"": ["", MessagesFormatterType.PHI_3],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
82
  "Mistral-7B-Instruct-v0.2-Neural-Story.Q5_K_M.gguf": ["mradermacher/Mistral-7B-Instruct-v0.2-Neural-Story-GGUF", MessagesFormatterType.MISTRAL],
83
  "Layris_9B.Q4_K_M.gguf": ["mradermacher/Layris_9B-GGUF", MessagesFormatterType.MISTRAL],
84
  "Eris_Remix_7B.Q5_K_M.gguf": ["mradermacher/Eris_Remix_7B-GGUF", MessagesFormatterType.MISTRAL],
 
67
  "MISCHIEVOUS-12B-Mix_0.1v.Q4_K_M.gguf": ["mradermacher/MISCHIEVOUS-12B-Mix_0.1v-GGUF", MessagesFormatterType.MISTRAL],
68
  "Canidori-12B-v1.i1-Q4_K_M.gguf": ["mradermacher/Canidori-12B-v1-i1-GGUF", MessagesFormatterType.MISTRAL],
69
  "MT2-Gen4-MM-gemma-2-Rv0.4MTM-9B.Q4_K_M.gguf": ["mradermacher/MT2-Gen4-MM-gemma-2-Rv0.4MTM-9B-GGUF", MessagesFormatterType.ALPACA],
70
+ "Flammen-Trismegistus-7B.i1-Q5_K_M.gguf": ["mradermacher/Flammen-Trismegistus-7B-i1-GGUF", MessagesFormatterType.MISTRAL],
71
  "Trinas_Nectar-8B-model_stock.i1-Q4_K_M.gguf": ["mradermacher/Trinas_Nectar-8B-model_stock-i1-GGUF", MessagesFormatterType.MISTRAL],
72
  "ChatWaifu_Magnum_V0.2.Q4_K_M.gguf": ["mradermacher/ChatWaifu_Magnum_V0.2-GGUF", MessagesFormatterType.MISTRAL],
73
  "ChatWaifu_12B_v2.0.Q5_K_M.gguf": ["mradermacher/ChatWaifu_12B_v2.0-GGUF", MessagesFormatterType.MISTRAL],
 
80
  #"": ["", MessagesFormatterType.OPEN_CHAT],
81
  #"": ["", MessagesFormatterType.CHATML],
82
  #"": ["", MessagesFormatterType.PHI_3],
83
+ "WONMSeverusDevilv2-TIES.i1-Q5_K_M.gguf": ["mradermacher/WONMSeverusDevilv2-TIES-i1-GGUF", MessagesFormatterType.MISTRAL],
84
+ "VirtuosoSmall-InstructModelStock.i1-Q4_K_M.gguf": ["mradermacher/VirtuosoSmall-InstructModelStock-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
85
+ "VICIOUS_MESH-12B-GAMMA.Q4_K_M.gguf": ["mradermacher/VICIOUS_MESH-12B-GAMMA-GGUF", MessagesFormatterType.MISTRAL],
86
+ "VICIOUS_MESH-12B-BETA.Q4_K_M.gguf": ["mradermacher/VICIOUS_MESH-12B-BETA-GGUF", MessagesFormatterType.MISTRAL],
87
+ "TQ2.5-14B-Aletheia-v1.i1-Q4_K_M.gguf": ["mradermacher/TQ2.5-14B-Aletheia-v1-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
88
+ "Tiger-7B-v0.1.i1-Q5_K_M.gguf": ["mradermacher/Tiger-7B-v0.1-i1-GGUF", MessagesFormatterType.MISTRAL],
89
+ "TaxoLlama3.1-8b-instruct.i1-Q5_K_M.gguf": ["mradermacher/TaxoLlama3.1-8b-instruct-i1-GGUF", MessagesFormatterType.LLAMA_3],
90
+ "RandomMergeNoNormWEIGHTED-7B-MODELSTOCK.i1-Q5_K_M.gguf": ["mradermacher/RandomMergeNoNormWEIGHTED-7B-MODELSTOCK-i1-GGUF", MessagesFormatterType.MISTRAL],
91
+ "RandomMergeNoNormWEIGHTED-7B-DARETIES.i1-Q5_K_M.gguf": ["mradermacher/RandomMergeNoNormWEIGHTED-7B-DARETIES-i1-GGUF", MessagesFormatterType.MISTRAL],
92
+ "RandomMergeNoNorm-7B-DARETIES.i1-Q5_K_M.gguf": ["mradermacher/RandomMergeNoNorm-7B-DARETIES-i1-GGUF", MessagesFormatterType.MISTRAL],
93
+ "R_Marco_polo.Q5_K_M.gguf": ["mradermacher/R_Marco_polo-GGUF", MessagesFormatterType.OPEN_CHAT],
94
+ "Qwenvergence-14B-v3-Reason.i1-Q4_K_M.gguf": ["mradermacher/Qwenvergence-14B-v3-Reason-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
95
+ "Qwenvergence-14B-v3-Prose.i1-Q4_K_M.gguf": ["mradermacher/Qwenvergence-14B-v3-Prose-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
96
+ "Qwenvergence-14B-v3.i1-Q4_K_M.gguf": ["mradermacher/Qwenvergence-14B-v3-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
97
+ "Qwen2.5-14B-SLERPv7.i1-Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-SLERPv7-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
98
+ "Qwen2.5-14B-Instruct-abliterated-v2.i1-Q4_K_M.gguf": ["mradermacher/Qwen2.5-14B-Instruct-abliterated-v2-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
99
+ "Prima-LelantaclesV5-7b.i1-Q5_K_M.gguf": ["mradermacher/Prima-LelantaclesV5-7b-i1-GGUF", MessagesFormatterType.MISTRAL],
100
+ "q2.5-veltha-14b-q5_k_m.gguf": ["djuna/Q2.5-Veltha-14B-Q5_K_M-GGUF", MessagesFormatterType.OPEN_CHAT],
101
+ "OgnoExperiment27-7B.i1-Q5_K_M.gguf": ["mradermacher/OgnoExperiment27-7B-i1-GGUF", MessagesFormatterType.MISTRAL],
102
+ "Neurotic-Jomainotrik-7b-slerp.i1-Q5_K_M.gguf": ["mradermacher/Neurotic-Jomainotrik-7b-slerp-i1-GGUF", MessagesFormatterType.MISTRAL],
103
+ "NeuralPearlBeagle.Q5_K_M.gguf": ["mradermacher/NeuralPearlBeagle-GGUF", MessagesFormatterType.MISTRAL],
104
+ "Mistral-7B-Instruct-Ukrainian.i1-Q5_K_M.gguf": ["mradermacher/Mistral-7B-Instruct-Ukrainian-i1-GGUF", MessagesFormatterType.MISTRAL],
105
+ "Maverick-7B.i1-Q5_K_M.gguf": ["mradermacher/Maverick-7B-i1-GGUF", MessagesFormatterType.MISTRAL],
106
+ "Hermes3-L3.1-DirtyHarry-8B.Q5_K_M.gguf": ["mradermacher/Hermes3-L3.1-DirtyHarry-8B-GGUF", MessagesFormatterType.LLAMA_3],
107
+ "Hermes3-L3.1-DarkPlanetSF-8B.i1-Q5_K_M.gguf": ["mradermacher/Hermes3-L3.1-DarkPlanetSF-8B-i1-GGUF", MessagesFormatterType.LLAMA_3],
108
+ "Hermes3-L3.1-BigTalker-8B.i1-Q5_K_M.gguf": ["mradermacher/Hermes3-L3.1-BigTalker-8B-i1-GGUF", MessagesFormatterType.LLAMA_3],
109
+ "Casuar-9B-Model_Stock.i1-Q4_K_M.gguf": ["mradermacher/Casuar-9B-Model_Stock-i1-GGUF", MessagesFormatterType.ALPACA],
110
+ "ZEUS-8B-V9.i1-Q5_K_M.gguf": ["mradermacher/ZEUS-8B-V9-i1-GGUF", MessagesFormatterType.LLAMA_3],
111
+ "vicious_mesh-12b-delta-q4_k_m.gguf": ["bamec66557/VICIOUS_MESH-12B-DELTA-Q4_K_M-GGUF", MessagesFormatterType.MISTRAL],
112
+ "Neural-Logical-Abstract-7B-slerp.Q5_K_M.gguf": ["mradermacher/Neural-Logical-Abstract-7B-slerp-GGUF", MessagesFormatterType.MISTRAL],
113
+ "NeuralDareBeagle-7B-slerp.Q5_K_M.gguf": ["mradermacher/NeuralDareBeagle-7B-slerp-GGUF", MessagesFormatterType.MISTRAL],
114
+ "Neural_Waifu_7b_V0.1.Q5_K_M.gguf": ["mradermacher/Neural_Waifu_7b_V0.1-GGUF", MessagesFormatterType.MISTRAL],
115
+ "Mistral-7B-Instruct_v0.2_UNA-TheBeagle-7b-v1.i1-Q5_K_M.gguf": ["mradermacher/Mistral-7B-Instruct_v0.2_UNA-TheBeagle-7b-v1-i1-GGUF", MessagesFormatterType.MISTRAL],
116
+ "MISCHIEVOUS-12B-Mix_Neo.i1-Q4_K_M.gguf": ["mradermacher/MISCHIEVOUS-12B-Mix_Neo-i1-GGUF", MessagesFormatterType.MISTRAL],
117
+ "MiaLatte-Indo-Mistral-7b.Q5_K_M.gguf": ["mradermacher/MiaLatte-Indo-Mistral-7b-GGUF", MessagesFormatterType.MISTRAL],
118
+ "MFANN-SFT-2.i1-Q5_K_M.gguf": ["mradermacher/MFANN-SFT-2-i1-GGUF", MessagesFormatterType.LLAMA_3],
119
+ "maestrale-chat-v0.3-beta-sft.i1-Q5_K_M.gguf": ["mradermacher/maestrale-chat-v0.3-beta-sft-i1-GGUF", MessagesFormatterType.MISTRAL],
120
+ "Llama-3-EZO-8b-Common-it.i1-Q5_K_M.gguf": ["mradermacher/Llama-3-EZO-8b-Common-it-i1-GGUF", MessagesFormatterType.LLAMA_3],
121
+ "Legal-Saul-Multiverse-7b.Q5_K_M.gguf": ["mradermacher/Legal-Saul-Multiverse-7b-GGUF", MessagesFormatterType.MISTRAL],
122
+ "kuno-royale-v3-7b.Q5_K_M.gguf": ["mradermacher/kuno-royale-v3-7b-GGUF", MessagesFormatterType.MISTRAL],
123
+ "Kuno-lake-slerp-7b.Q5_K_M.gguf": ["mradermacher/Kuno-lake-slerp-7b-GGUF", MessagesFormatterType.MISTRAL],
124
+ "KRONOS-8B-V6.i1-Q5_K_M.gguf": ["mradermacher/KRONOS-8B-V6-i1-GGUF", MessagesFormatterType.LLAMA_3],
125
+ "Homer-v1.0-Qwen2.5-7B.i1-Q5_K_M.gguf": ["mradermacher/Homer-v1.0-Qwen2.5-7B-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
126
+ "FuseCyberMix-Qwen-2.5-7B-Instruct.i1-Q5_K_M.gguf": ["mradermacher/FuseCyberMix-Qwen-2.5-7B-Instruct-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
127
+ "flammen7-mistral-7B.Q5_K_M.gguf": ["mradermacher/flammen7-mistral-7B-GGUF", MessagesFormatterType.MISTRAL],
128
+ "flammen6-mistral-7B.Q5_K_M.gguf": ["mradermacher/flammen6-mistral-7B-GGUF", MessagesFormatterType.MISTRAL],
129
+ "flammen3X-mistral-7B.i1-Q5_K_M.gguf": ["mradermacher/flammen3X-mistral-7B-i1-GGUF", MessagesFormatterType.MISTRAL],
130
+ "Eris-Floramix-7b.i1-Q5_K_M.gguf": ["mradermacher/Eris-Floramix-7b-i1-GGUF", MessagesFormatterType.MISTRAL],
131
+ "Eris-Daturamix-7b.i1-Q5_K_M.gguf": ["mradermacher/Eris-Daturamix-7b-i1-GGUF", MessagesFormatterType.MISTRAL],
132
+ "Eris_Remix_DPO_7B.i1-Q5_K_M.gguf": ["mradermacher/Eris_Remix_DPO_7B-i1-GGUF", MessagesFormatterType.MISTRAL],
133
+ "dolphin-2.6-mistral-7b-dpo-chat.Q5_K_M.gguf": ["mradermacher/dolphin-2.6-mistral-7b-dpo-chat-GGUF", MessagesFormatterType.MISTRAL],
134
+ "BuRP_7B.i1-Q5_K_M.gguf": ["mradermacher/BuRP_7B-i1-GGUF", MessagesFormatterType.MISTRAL],
135
+ "Blurred-Beagle-7b-slerp.Q5_K_M.gguf": ["mradermacher/Blurred-Beagle-7b-slerp-GGUF", MessagesFormatterType.MISTRAL],
136
+ "Blur-7B-slerp-v0.1.i1-Q5_K_M.gguf": ["mradermacher/Blur-7B-slerp-v0.1-i1-GGUF", MessagesFormatterType.MISTRAL],
137
+ "Biomistral-Clown-Slerp.Q5_K_M.gguf": ["mradermacher/Biomistral-Clown-Slerp-GGUF", MessagesFormatterType.MISTRAL],
138
+ "Deris-SSS-7B.Q5_K_M.gguf": ["mradermacher/Deris-SSS-7B-GGUF", MessagesFormatterType.MISTRAL],
139
+ "Dark-Waifu-7b.Q5_K_M.gguf": ["mradermacher/Dark-Waifu-7b-GGUF", MessagesFormatterType.MISTRAL],
140
+ "Brezn3.Q5_K_M.gguf": ["mradermacher/Brezn3-GGUF", MessagesFormatterType.MISTRAL],
141
+ "BreezeDolphin-SLERP-0.1.Q5_K_M.gguf": ["mradermacher/BreezeDolphin-SLERP-0.1-GGUF", MessagesFormatterType.MISTRAL],
142
+ "Blitz-v0.1.i1-Q5_K_M.gguf": ["mradermacher/Blitz-v0.1-i1-GGUF", MessagesFormatterType.MISTRAL],
143
+ "Asymmetric_Linearity-8B-Model_Stock.Q5_K_M.gguf": ["mradermacher/Asymmetric_Linearity-8B-Model_Stock-GGUF", MessagesFormatterType.LLAMA_3],
144
+ "KRONOS-8B-V5.i1-Q5_K_M.gguf": ["mradermacher/KRONOS-8B-V5-i1-GGUF", MessagesFormatterType.LLAMA_3],
145
+ "14b-Qwen2.5-Infermatic-Crea-v1.i1-Q4_K_M.gguf": ["mradermacher/14b-Qwen2.5-Infermatic-Crea-v1-i1-GGUF", MessagesFormatterType.OPEN_CHAT],
146
+ "TQ2.5-14B-Neon-v1-Q4_K_M.gguf": ["bartowski/TQ2.5-14B-Neon-v1-GGUF", MessagesFormatterType.OPEN_CHAT],
147
+ "TQ2.5-14B-Aletheia-v1-Q4_K_M.gguf": ["bartowski/TQ2.5-14B-Aletheia-v1-GGUF", MessagesFormatterType.OPEN_CHAT],
148
+ "SynthIQ-7b.i1-Q5_K_M.gguf": ["mradermacher/SynthIQ-7b-i1-GGUF", MessagesFormatterType.MISTRAL],
149
+ "MISCHIEVOUS-12B-Mix_III_IV_V.i1-Q4_K_M.gguf": ["mradermacher/MISCHIEVOUS-12B-Mix_III_IV_V-i1-GGUF", MessagesFormatterType.MISTRAL],
150
+ "KRONOS-8B-V4.i1-Q5_K_M.gguf": ["mradermacher/KRONOS-8B-V4-i1-GGUF", MessagesFormatterType.LLAMA_3],
151
+ "fratricide-12B-Unslop-Mell-DARKNESS.i1-Q4_K_M.gguf": ["mradermacher/fratricide-12B-Unslop-Mell-DARKNESS-i1-GGUF", MessagesFormatterType.MISTRAL],
152
+ "AngelSlayer-12B-Unslop-Mell-RPMax-DARKNESS-v2.i1-Q4_K_M.gguf": ["mradermacher/AngelSlayer-12B-Unslop-Mell-RPMax-DARKNESS-v2-i1-GGUF", MessagesFormatterType.MISTRAL],
153
+ "Llama-3-8B-ArliAI-Formax-v1.0.i1-Q5_K_M.gguf": ["mradermacher/Llama-3-8B-ArliAI-Formax-v1.0-i1-GGUF", MessagesFormatterType.LLAMA_3],
154
+ "MonarchLake-7B.i1-Q5_K_M.gguf": ["mradermacher/MonarchLake-7B-i1-GGUF", MessagesFormatterType.MISTRAL],
155
+ "Evangelion-7B.i1-Q5_K_M.gguf": ["mradermacher/Evangelion-7B-i1-GGUF", MessagesFormatterType.MISTRAL],
156
  "Mistral-7B-Instruct-v0.2-Neural-Story.Q5_K_M.gguf": ["mradermacher/Mistral-7B-Instruct-v0.2-Neural-Story-GGUF", MessagesFormatterType.MISTRAL],
157
  "Layris_9B.Q4_K_M.gguf": ["mradermacher/Layris_9B-GGUF", MessagesFormatterType.MISTRAL],
158
  "Eris_Remix_7B.Q5_K_M.gguf": ["mradermacher/Eris_Remix_7B-GGUF", MessagesFormatterType.MISTRAL],