Text Generation
Transformers
PyTorch
mistral
Not-For-All-Audiences
nsfw
text-generation-inference
Inference Endpoints
File size: 2,290 Bytes
0abc348 63d3d4e 5545bf8 8803a5c 324652f 0abc348 f8231c7 8803a5c f8231c7 6f9eb65 0abc348 f692b96 0abc348 f692b96 03009b3 f692b96 9330b96 26d2f5d 0abc348 b755893 de556a6 512e211 930872f 0baa830 e3a73b4 512e211 3a93418 e3a73b4 dfb6449 e3a73b4 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 |
---
datasets:
- Norquinal/claude_multiround_chat_1k
- jondurbin/airoboros-gpt4-1.4
- Squish42/bluemoon-fandom-1-1-rp-cleaned
- totally-not-an-llm/EverythingLM-data-V2-sharegpt
- OpenLeecher/Teatime
- PygmalionAI/PIPPA
tags:
- not-for-all-audiences
- nsfw
license: cc-by-nc-4.0
---
## What is PetrolLM?
PetrolLM is [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) model fine-tune using QLoRA (4-bit precision) for the purposes of creative writing and roleplay.
The dataset consists of 5800 samples, with the composition as follows:
* AICG Logs (~17%)
* PygmalionAI/PIPPA (~17%)
* Squish42/bluemoon-fandom-1-1-rp-cleaned (~13%)
* OpenLeecher/Teatime (~2%)
* Norquinal/claude_multiround_chat_1k (~17%)
* jundurbin/airoboros-gpt4-1.4 (~17%)
* totally-not-an-llm/EverythingLM-data-V2-sharegpt (~17%)
These samples were then back-filled using gpt-4/gpt-3.5-turbo-16k or otherwise converted to fit the prompt format.
## Prompt Format
The model was finetuned with a prompt format similar to the original SuperHOT prototype:
```
---
style: roleplay
characters:
[char]: [description]
summary: [scenario]
---
<chat_history>
Format:
[char]: [message]
Human: [message]
```
## Use in Text Generation Web UI
Install the bleeding-edge version of `transformers` from source:
```
pip install git+https://github.com/huggingface/transformers
```
Or, alternatively, change `model_type` in `config.json` from `mistral` to `llama`.
## Use in SillyTavern UI
![](https://files.catbox.moe/2dkr28.png)
As an addendum, you can include one of the following as the `Last Output Sequence`:
```
Human: In your next reply, write at least two paragraphs. Be descriptive and immersive, providing vivid details about {{char}}'s actions, emotions, and the environment.
{{char}}:
```
```
{{char}} (2 paragraphs, engaging, natural, authentic, descriptive, creative):
```
```
[System note: Write at least two paragraphs. Be descriptive and immersive, providing vivid details about {{char}}'s actions, emotions, and the environment.]
{{char}}:
```
The third one seems to work the best. I would recommend experimenting with creating your own to best suit your needs.
## Finetuing Parameters
- LoRA Rank: 64
- LoRA Alpha: 16
- LoRA Dropout: 0.1
- BF16 Training
- Cutoff Length: 2048
- Training Epoch(s): 2 |