Falcon3-2x10B-MoE-Instruct

Falcon3-2x10B-MoE-Instruct is a Mixture of Experts (MoE) made with the following models using LazyMergekit:

🧩 Configuration

base_model: tiiuae/Falcon3-10B-Instruct
gate_mode: random
architecture: mixtral
dtype: bfloat16
experts:
  - source_model: tiiuae/Falcon3-10B-Instruct
    positive_prompts:
      - "Help me write a story"
  - source_model: tiiuae/Falcon3-10B-Instruct
    positive_prompts:
      - "Can you explain this?"

πŸ’» Usage

!pip install -qU transformers bitsandbytes accelerate

from transformers import AutoTokenizer
import transformers
import torch

model = "qingy2024/Falcon3-2x10B-MoE-Instruct"

tokenizer = AutoTokenizer.from_pretrained(model)
pipeline = transformers.pipeline(
    "text-generation",
    model=model,
    model_kwargs={"torch_dtype": torch.float16, "load_in_4bit": True},
)

messages = [{"role": "user", "content": "Explain what a Mixture of Experts is in less than 100 words."}]
prompt = pipeline.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
Downloads last month
152
Safetensors
Model size
18.8B params
Tensor type
BF16
Β·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for qingy2024/Falcon3-2x10B-MoE-Instruct

Finetuned
(15)
this model
Quantizations
2 models