File size: 1,354 Bytes
b34eea1 0b97d85 74b6cca b34eea1 74b6cca d795489 74b6cca b34eea1 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 |
---
license: apache-2.0
tags:
- merge
- mergekit
- CorticalStack/pastiche-crown-clown-7b-dare-dpo
- Equall/Saul-Instruct-v1
---
![image/webp](https://cdn.prod.arcee.ai/images/saul-calme.jpeg)
# Saul-Instruct-Clown-7b
Saul-Instruct-Clown-7b is a merge of the following models using [mergekit](https://github.com/cg123/mergekit):
* [CorticalStack/pastiche-crown-clown-7b-dare-dpo](https://huggingface.co/CorticalStack/pastiche-crown-clown-7b-dare-dpo)
* [Equall/Saul-Instruct-v1](https://huggingface.co/Equall/Saul-Instruct-v1)
## 🏆 Evaluation
### OpenLLM
Saul-Instruct-Clown-7b OpenLLM benchmark suite
| Model | Average | arc | HellaSwag | mmlu | TruthfulQA | gsm8k |
|---|---:|---:|---:|---:|---:|---:|
| [arcee-ai/Saul-Instruct-Clown-7b](https://huggingface.co/arcee-ai/Saul-Instruct-Clown-7b/) | 72.79 | 68.26 | 86.28 | 63.12 | 64.68 | 83.43 | 70.96 |
## 🧩 Configuration
```yaml
slices:
- sources:
- model: CorticalStack/pastiche-crown-clown-7b-dare-dpo
layer_range: [0, 32]
- model: Equall/Saul-Instruct-v1
layer_range: [0, 32]
merge_method: slerp
base_model: CorticalStack/pastiche-crown-clown-7b-dare-dpo
parameters:
t:
- filter: self_attn
value: [0, 0.5, 0.3, 0.7, 1]
- filter: mlp
value: [1, 0.5, 0.7, 0.3, 0]
- value: 0.5
dtype: bfloat16
``` |