stories15M_MOE / README.md
ngxson's picture
guide for lora
6c8c407
|
raw
history blame contribute delete
No virus
763 Bytes
metadata
license: mit

stories15M_MOE

This model is ModelCloud/tinyllama-15M-stories repeated 4 times to make 4 experts.

The model is used for testing, not intended to be used in production (unless your product is some kind of bedtime story teller)

Weight of router is initialized randomly

shakespeare LoRA adapter

A LoRA adapter trained on first 100 paragraphs of shakespeare can be found inside moe_shakespeare15M

With input: Look in thy glass

  • Original model generates: Look in thy glass was a little girl. She was only three years old and she was three years old. She was
  • LoRA adapter generates: Look in thy glass in love of the eye: That's when when the eye see thy on the sun'