Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Quantization made by Richard Erkhov.

Github

Discord

Request more models

Frankenstein-MoE-en-10.7Bx4 - GGUF

Original model description:

language: - en license: cc-by-nc-sa-4.0

Frankenstein-MoE

Method

To initialize the gate projection weight of the MoE layer, the H6 trainset was sampled and used. We sampled 400 and selected the final 30 with low PPL.

trufulqa used gpt4 to generate data.

Evals

in progress

Downloads last month
18
GGUF
Model size
36.1B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference API
Unable to determine this model's library. Check the docs .