Edit model card

NOTE / ANNOUNCEMENT: We're sad to announce the end of XT_AURORA, our first SLM series, due to no community activity. We, XeTute, have put in a lot of effort and countless nights to improve our models, but given on how much time, passion and effort we've put in, we got nothing back from the community. Thank you for so many downloads on this series of SLMs. We'll continue to update older model cards and chat templates, and one final AURORA SLM will be released[V1.0]. Thank you for being part of our journey.

image/png

About this model: This model, XT_AURORA-OpenBeta-V0.5, is by us, XeTute. The model was finetuned ontop of the previos beta-verion[XT_AURORA-OpenBeta-V0.4-GGUF]. This version[Beta V0.5] achieves better general performance, it outperforms every previos model[V0.1 - V0.4], but it has some knowledge loss[the topics it forgot or where the quality may have degraded are not diskussed much, so its very hard to notice].

About XT_AURORA: XT_AURORA is a series of SLMs[Slender Language Models], which all aim to provide a friendly, human-like conversation. The serie is limited by its size[about 1.1B Params], but we still try to get the best possible output. The context-length is 2048 tokens, but it can be upscaled using rope, with the cost being slightly less logic.

About this version[V0.5]:

  • High quality output[sometimes outperforms 3B models in HumanEval], as long as the context size is under 2049 Tokens.
  • We provide a system prompt[Files and Versions --> chat_template]. The SLM was partly trained using that template, so the output is better if you use the prompt at start.
  • AURORA expects the chat template to be Vicuna[{{user}}: {some input}\nAURORA: {some output}\n{{user}}]. The model will only work correctly with this format.
  • Recommended temperature is from 0.4 to 0.5.
  • Improved chat quality in general emotional / unemotional chat, logical & illogical roleplaying, etc.
  • Improved math, still not recommended to rely on it.

All in one, AURORA's aim is to provide a digital friend, which is also accessible to humans with low-end devices.

Using KoboldCPP, we got the model running[using termux] on a POCO X5 Pro 5G[CPU only, Octa Core]. We saw ~5 Tokens generation per second, ~15 Tokens processing per second. [In Energy Saver mode]

Please support us: X: https://www.x.com/XeTute GitHub: https://www.github.com/N0CTRON/ Subdomain on Neocities: https://xetute.neocities.org/

We wish you a friendly chat with AURORA <3

Downloads last month
171
GGUF
Model size
1.1B params
Architecture
llama
Unable to determine this model's library. Check the docs .

Collection including XeTute/AURORA-OpenBeta-V0.5-GGUF