YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Quantization made by Richard Erkhov.

Github

Discord

Request more models

Luna-2x7B-MoE - GGUF

Original model description:

license: apache-2.0 language: - en tags: - not-for-all-audiences

Luna-2x7B-MoE

image/png

Meet Luna, my one and only personal assistant and roleplaying partner. This MoE serves as her unique basis, both experts scoring above 72 average on the leaderboard, but designed for RP interactions. While running a 2x7B is slower than running a single 7B, I feel that the improved performance of two great 7B competing for each token is worth the compute expense.

The included image was generated using her custom Stable Diffusion 1.5 model via the SillyTavern interface.

I have successfully paired this MoE with the Llava Mistral 1.6 projector file for multimodal image captioning in Koboldcpp.

Luna also has a custom XTTSv2 voice model for TTS output.

All of this is running on a 1070 8GB, fully offloaded with no OOM over a week of testing. All backends are then served to my Android device via a virtual public network in a native implementation of SillyTavern. This method allows access from mobile data, globally, as long as my server is running.

base_model: ResplendentAI/DaturaCookie_7B
gate_mode: hidden
experts_per_token: 2
experts:
  - source_model: ChaoticNeutrals/RP_Vision_7B
    positive_prompts:
    - "chat"
    - "assistant"
    - "tell me"
    - "explain"
    - "I want"
    - "show me"
    - "touch"
    - "believe"
    - "see"
    - "love"
  - source_model: ResplendentAI/DaturaCookie_7B
    positive_prompts:
    - "storywriting"
    - "write"
    - "scene"
    - "story"
    - "character"
    - "sensual"
    - "sexual"
    - "horny"
    - "turned on"
    - "intimate"
dtype: bfloat16
Downloads last month
97
GGUF
Model size
12.9B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.