An MLX bfloat16 FP16 model, 1M context length, uncensored.

Model Merge: DeepSeek-R1-Distill-Qwen-14B-abliterated-v2

Description:
This model is a tie merge of "huihui-ai/DeepSeek-R1-Distill-Qwen-14B-abliterated-v2" and "mkurman/Qwen2.5-14B-DeepSeek-R1-1M".

Recipes

Model Recipe:

  - model: "huihui-ai/DeepSeek-R1-Distill-Qwen-14B-abliterated-v2"
    parameters:
      weight: 1
      density: 1

merge_method: ties
base_model: "mkurman/Qwen2.5-14B-DeepSeek-R1-1M"
parameters:
  density: 1
  normalize: true
  int8_mask: true
dtype: bfloat16

Merged Model:
The model was merged using the "ties" method.

Conversion to MLX Format

The model was converted to the MLX format with brainfloat16 precision using the following command:

mlx_lm.convert --hf-path FiditeNemini/Qwen2.5-14B-DeepSeek-R1-1M --mlx-path ./Unhinged-Qwen2.5-R1.bf16 --dtype bfloat16 -q --q-bits 16

Usage Example

You can use this model with the following code snippet:

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("FiditeNemini/Unhinged-Qwen2.5-R1.bf16")
tokenizer = AutoTokenizer.from_pretrained("FiditeNemini/Unhinged-Qwen2.5-R1.bf16")

Details

  • Model type: CausalLM
  • Context length: 4096 tokens
  • License: Apache 2.0

Keywords

  • DeepSeek-R1-Distill
  • Qwen2.5
  • Abliterated
  • LLM
  • 1M context
Downloads last month
49
Safetensors
Model size
14.8B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model’s pipeline type.

Model tree for FiditeNemini/Unhinged-Qwen2.5-R1-1M-Uncensored-BF16