stories15M_MOE / README.md
ngxson's picture
guide for lora
6c8c407
|
raw
history blame contribute delete
No virus
763 Bytes
---
license: mit
---
# stories15M_MOE
This model is [ModelCloud/tinyllama-15M-stories](https://huggingface.co/ModelCloud/tinyllama-15M-stories) repeated 4 times to make 4 experts.
The model is used for testing, not intended to be used in production (unless your product is some kind of bedtime story teller)
Weight of router is initialized randomly
## shakespeare LoRA adapter
A LoRA adapter trained on first 100 paragraphs of shakespeare can be found inside `moe_shakespeare15M`
With input: `Look in thy glass`
- Original model generates: `Look in thy glass was a little girl. She was only three years old and she was three years old. She was`
- LoRA adapter generates: `Look in thy glass in love of the eye: That's when when the eye see thy on the sun'`