--- license: other language: - en pipeline_tag: text-generation inference: false tags: - transformers - gguf - imatrix - Chihiro-7B-v0.1 --- Quantizations of https://huggingface.co/yuuko-eth/Chihiro-7B-v0.1 ### Inference Clients/UIs * [llama.cpp](https://github.com/ggerganov/llama.cpp) * [KoboldCPP](https://github.com/LostRuins/koboldcpp) * [ollama](https://github.com/ollama/ollama) * [text-generation-webui](https://github.com/oobabooga/text-generation-webui) * [jan](https://github.com/janhq/jan) * [GPT4All](https://github.com/nomic-ai/gpt4all) --- # From original readme ![](https://i.imgur.com/UwNO4fS.png) This is an experimental Mistral-architecture SLERP merge with two brilliant base models. Zebrafish and Breeze were used together in this work. Model configuration is as follows: * [Breeze-7B-Instruct](https://huggingface.co/MediaTek-Research/Breeze-7B-Instruct-v0_1) as base. * [Zebrafish-7B](https://huggingface.co/mlabonne/Zebrafish-7B) as model 1. To use the model, please use either prompt templates suggested by the base models, or just slap the Mistral one on.

### Benchmarks Evaluation suite: OpenLLM | Model | ARC |HellaSwag| MMLU |TruthfulQA|Winogrande|GSM8K| |-------------------------------------------------------------------|----:|--------:|--------------------------|---------:|---------:|----:| |[Chihiro-7B-v0.1](https://huggingface.co/yuuko-eth/Chihiro-7B-v0.1)|68.52| 85.95| (not yet evaluated) | 63.81| 81.77|64.22| Evaluation suite: Nous | Model |AGIEval|GPT4All|TruthfulQA|Bigbench|Average| |-------------------------------------------------------------------|------:|------:|---------:|-------:|------:| |[Chihiro-7B-v0.1](https://huggingface.co/yuuko-eth/Chihiro-7B-v0.1)| 45.16| 75.26| 63.82| 47.38| 57.91| Average: 47.38% Average score: 57.91% Evaluated Apr. 27, 2024, NVIDIA RTX 4090