--- license: apache-2.0 language: - en datasets: - teknium/openhermes - jondurbin/airoboros-2.2 --- An attempt at a Generalist Model, using chunks spliced from the Airoboros and OpenHermes dataset, and some of my own stuff which I have since discarded (chat stuff, basic shit)forgot i added CollectiveCognition to the mix. Base Model used is a gradient merge between OpenHermes & Airoboros 2.2, lora trained on the two spliced datasets and my own one, which was reapplied to them (dumb move I know) Tests: Uncensored, not bad at RP I Guess, it felt decent. Good? who knows, tbh. try it urself or smth