diff --git "a/mergekit_config.yml" "b/mergekit_config.yml" new file mode 100644--- /dev/null +++ "b/mergekit_config.yml" @@ -0,0 +1,5903 @@ +slices: + - sources: + - layer_range: [0, 3] + model: lilmeaty/5 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/5 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/5 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/5 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/5 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/5 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/5 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/5 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/5 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/5 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/5 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/5 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/5 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/5 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ICEPVP8977/Uncensored_llama_3.2_3b_safetensors + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/2 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: lilmeaty/4 + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: chuanli11/Llama-3.2-3B-Instruct-uncensored + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Bllossom/llama-3.2-Korean-Bllossom-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B-Instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: Qwen/Qwen2.5-3B + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: stabilityai/stable-code-3b + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ibm-granite/granite-3b-code-base-2k + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + - sources: + - layer_range: [0, 3] + model: ministral/Ministral-3b-instruct + parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true +parameters: + normalize: true + int8_mask: true + density: 0.5 + weight: 0.1 + random_seed: 0 + temperature: 0.5 + top_p: 0.65 + inference: true + max_tokens: 999999999 + stream: true +normalize: true +int8_mask: true +density: 0.5 +weight: 0.1 +random_seed: 0 +temperature: 0.5 +top_p: 0.65 +inference: true +max_tokens: 999999999 +stream: true +merge_method: passthrough