---
license: other
language:
- en
pipeline_tag: text-generation
inference: false
tags:
- transformers
- gguf
- imatrix
- Chihiro-7B-v0.1
---
Quantizations of https://huggingface.co/yuuko-eth/Chihiro-7B-v0.1
### Inference Clients/UIs
* [llama.cpp](https://github.com/ggerganov/llama.cpp)
* [KoboldCPP](https://github.com/LostRuins/koboldcpp)
* [ollama](https://github.com/ollama/ollama)
* [text-generation-webui](https://github.com/oobabooga/text-generation-webui)
* [jan](https://github.com/janhq/jan)
* [GPT4All](https://github.com/nomic-ai/gpt4all)
---
# From original readme

This is an experimental Mistral-architecture SLERP merge with two brilliant base models. Zebrafish and Breeze were used together in this work.
Model configuration is as follows:
* [Breeze-7B-Instruct](https://huggingface.co/MediaTek-Research/Breeze-7B-Instruct-v0_1) as base.
* [Zebrafish-7B](https://huggingface.co/mlabonne/Zebrafish-7B) as model 1.
To use the model, please use either prompt templates suggested by the base models, or just slap the Mistral one on.
### Benchmarks
Evaluation suite: OpenLLM
| Model | ARC |HellaSwag| MMLU |TruthfulQA|Winogrande|GSM8K|
|-------------------------------------------------------------------|----:|--------:|--------------------------|---------:|---------:|----:|
|[Chihiro-7B-v0.1](https://huggingface.co/yuuko-eth/Chihiro-7B-v0.1)|68.52| 85.95| (not yet evaluated) | 63.81| 81.77|64.22|
Evaluation suite: Nous
| Model |AGIEval|GPT4All|TruthfulQA|Bigbench|Average|
|-------------------------------------------------------------------|------:|------:|---------:|-------:|------:|
|[Chihiro-7B-v0.1](https://huggingface.co/yuuko-eth/Chihiro-7B-v0.1)| 45.16| 75.26| 63.82| 47.38| 57.91|
Average: 47.38%
Average score: 57.91%
Evaluated Apr. 27, 2024, NVIDIA RTX 4090