30B-Epsilon
Epsilon is an instruct based general purpose model assembled from hand picked models and LoRAs. There is no censorship and it follows instructions in the Alpaca format. This means you can create your own rules in the context memory of your inference system of choice [mainly KoboldAI or Text Generation Webui and chat UIs like SillyTavern and so on].
Composition:
This model is the result of an experimental use of LoRAs on language models and model merges. [] = applied as LoRA to a composite model | () = combined as composite models 30B-Epsilon = [SuperCOT[SuperHOT-prototype13b-8192[(wizardlmuncensored+((hippogriff+manticore)+(StoryV2))]
Alpaca's instruct format can be used to do many things, including control of the terms of behavior between a user and a response from an agent in chat. Below is an example of a command injected into memory.
### Instruction:
Make Narrator function as a text based adventure game that responds with verbose, detailed, and creative descriptions of what happens next after Player's response.
Make Player function as the player input for Narrator's text based adventure game, controlling a character named (insert character name here, their short bio, and
whatever quest or other information to keep consistent in the interaction).
### Response:
{an empty new line here}
All datasets from all models and LoRAs used were documented and reviewed as model candidates for merging. Model candidates were based on five core principles: creativity, logic, inference, instruction following, and longevity of trained responses. SuperHOT-prototype30b-8192 was used in this mix, not the 8K version; the prototype LoRA seems to have been removed [from HF] as of this writing. The GPT4Alpaca LoRA from Chansung was removed from this amalgam following a thorough review of where censorship and railroading the user came from in 33B-Lazarus. This is not a reflection of ChanSung's excellent work - it merely did not fit the purpose of this model.
Language Models and LoRAs Used Credits:
manticore-30b-chat-pyg-alpha [Epoch0.4] by openaccess-ai-collective
https://huggingface.co/openaccess-ai-collective/manticore-30b-chat-pyg-alpha
hippogriff-30b-chat by openaccess-ai-collective
https://huggingface.co/openaccess-ai-collective/hippogriff-30b-chat
WizardLM-33B-V1.0-Uncensored by ehartford
https://huggingface.co/ehartford/WizardLM-33B-V1.0-Uncensored
Storytelling-LLaMa-LoRA [30B, Version 2] by GamerUnTouch
https://huggingface.co/GamerUntouch/Storytelling-LLaMa-LoRAs
SuperCOT-LoRA [30B] by kaiokendev
https://huggingface.co/kaiokendev/SuperCOT-LoRA
SuperHOT-LoRA-prototype30b-8192 [30b, not 8K version, but a removed prototype] by kaiokendev
https://huggingface.co/kaiokendev/superhot-30b-8k-no-rlhf-test [Similar LoRA to one since removed that was used in making this model.]
Also thanks to Meta for LLaMA and to each and every one of you who developed these fine-tunes and LoRAs.
- Downloads last month
- 1,240