license: apache-2.0 | |
language: | |
- en | |
pipeline_tag: text-generation | |
tags: | |
- pytorch | |
- mixtral | |
- fine-tuned | |
- moe | |
# Mixtral 8x7B - Holodeck | |
## Model Description | |
Mistral 7B-Holodeck is a finetune created using Mixtral's 8x7B model. | |
## Training data | |
The training data contains around 3000 ebooks in various genres. | |
Most parts of the dataset have been prepended using the following text: `[Genre: <genre1>, <genre2>]` | |
*** | |
### Limitations and Biases | |
Based on known problems with NLP technology, potential relevant factors include bias (gender, profession, race and religion). |