--- license: apache-2.0 language: - en pipeline_tag: text-generation tags: - pytorch - mixtral - fine-tuned - moe --- # Mixtral 8x7B - Holodeck ## Model Description Mistral 7B-Holodeck is a finetune created using Mixtral's 8x7B model. ## Training data The training data contains around 3000 ebooks in various genres. Most parts of the dataset have been prepended using the following text: `[Genre: , ]` *** ### Limitations and Biases Based on known problems with NLP technology, potential relevant factors include bias (gender, profession, race and religion).