--- license: other language: - en --- ## Model details MythoBoros-13b can be considered a sister model to [MythoLogic-13b](https://huggingface.co/Gryphe/MythoLogic-13b), sharing the same goals but having a different approach. Whereas the previous model was a series of experimental gradient merges, this one is a simple straight-up 66/34 merge of [Chronos](https://huggingface.co/elinas/chronos-13b) and the freshly released [Ouroboros](https://huggingface.co/CalderaAI/13B-Ouroboros), providing a very solid foundation for a well-performing roleplaying model. MythoBoros tends to be somewhat more formal with its responses in comparison to MythoLogic. My advice? Try both, see which one you prefer. Quantized models are available from TheBloke: [GGML](https://huggingface.co/TheBloke/MythoBoros-13B-GGML) - [GPTQ](https://huggingface.co/TheBloke/MythoBoros-13B-GPTQ) (You're the best!) ## Prompt Format This model works best with Alpaca formatting, so for optimal model performance, use: ``` ### Instruction: Your instruction or question here. For roleplay purposes, I suggest the following - Write 's next reply in a chat between and . Write a single reply only. ### Response: ``` # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Gryphe__MythoBoros-13b) | Metric | Value | |-----------------------|---------------------------| | Avg. | 47.27 | | ARC (25-shot) | 58.19 | | HellaSwag (10-shot) | 81.75 | | MMLU (5-shot) | 50.13 | | TruthfulQA (0-shot) | 48.93 | | Winogrande (5-shot) | 75.77 | | GSM8K (5-shot) | 8.64 | | DROP (3-shot) | 7.45 |