Rose-20B / README.md
tavtav's picture
Update README.md
12f9fa3
|
raw
history blame
2.68 kB
---
language:
- en
pipeline_tag: text-generation
tags:
- text-generation-inference
- instruct
license: llama2
---
<h1 style="text-align: center">Ross-20B</h1>
<center><img src="https://files.catbox.moe/rze9c9.png" alt="roseimage" width="350" height="350"></center>
<center><i>Image sourced by Shinon</i></center>
<h2 style="text-align: center">Experimental Frankenmerge Model</h2>
## Other Formats
[GGUF](https://huggingface.co/TheBloke/Rose-20B-GGUF)
[GPTQ](https://huggingface.co/TheBloke/Rose-20B-GPTQ)
[AWQ](https://huggingface.co/TheBloke/Rose-20B-AWQ)
[exl2](https://huggingface.co/royallab/Rose-20B-exl2)
## Model Details
A Frankenmerge with [Thorns-13B](https://huggingface.co/CalderaAI/13B-Thorns-l2) by CalderaAI and [Noromaid-13-v0.1.1](https://huggingface.co/NeverSleep/Noromaid-13b-v0.1.1) by NeverSleep (IkariDev and Undi). This recipe was proposed by Trappu and the layer distribution recipe was made by Undi. I thank them for sharing their knowledge with me. This model should be very good at any roleplay scenarios. I called the model "Rose" because it was a fitting name for a "thorny maid".
The recommended format to use is Alpaca.
```
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
{prompt}
### Response:
```
Feel free to share any other prompts that works. This model is very robust.
**Warning: This model uses significantly more VRAM due to the KV cache increase resulting in more VRAM required for the context window.**
## Justification for its Existence
Potential base model for finetune experiments using our dataset to create Pygmalion-20B. Due to the already high capabilities, adding our dataset will mesh well with how the model performs.
Potential experimentation with merging with other 20B Frankenmerge models.
## Model Recipe
```
slices:
- sources:
- model: Thorns-13B
layer_range: [0, 16]
- sources:
- model: Noromaid-13B
layer_range: [8, 24]
- sources:
- model: Thorns-13B
layer_range: [17, 32]
- sources:
- model: Noromaid-13B
layer_range: [25, 40]
merge_method: passthrough
dtype: float16
```
Again, credits to [Undi](https://huggingface.co/Undi95) for the recipe.
## Reception
The model was given to a handful of members in the PygmalionAI Discord community for testing. A strong majority really enjoyed the model with only a couple giving the model a passing grade. Since our community has high standards for roleplaying models, I was surprised at the positive reception.
## Contact
Send a message to tav (tav) on Discord if you want to talk about the model to me. I'm always open to receive comments.