YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)
Quantization made by Richard Erkhov.
Frankenstein-MoE-en-10.7Bx4 - GGUF
- Model creator: https://huggingface.co/MoEMoEKKung/
- Original model: https://huggingface.co/MoEMoEKKung/Frankenstein-MoE-en-10.7Bx4/
Original model description:
language: - en license: cc-by-nc-sa-4.0
Frankenstein-MoE
Method
To initialize the gate projection weight of the MoE layer, the H6 trainset was sampled and used. We sampled 400 and selected the final 30 with low PPL.
trufulqa used gpt4 to generate data.
Evals
in progress
- Downloads last month
- 18