This is a version of
hivemind/gpt-j-6B-8bit fine-tuned on the Wizard of Wikipedia dataset for 10k steps (just under an epoch) on an A100. it can be used as a chatbot. It is designed to be used with ai-msgbot to take advantage of the prompt engineering.
NOTE: this needs to be loaded via the special patching technique outlined in the hivemind model card (as with all 8bit models)
Examples of how to load the model correctly are already in place in the notebook linked above. A
.py of said notebook was uploaded to the repo for reference - link here
For details, please see this wandb report for both the daily-dialogues version and the WoW version.
- Downloads last month
Inference API has been turned off for this model.