File size: 1,728 Bytes
a9846a4 4ba2bff a9846a4 6ef074e a9846a4 05d4591 848e64a ff899f5 05d4591 0bc522c 05d4591 a9846a4 05d4591 a9846a4 05d4591 6ef074e |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 |
---
language:
- en
license: apache-2.0
tags:
- text-generation-inference
- transformers
- unsloth
- trl
- sft
base_model: unsloth/Phi-3-mini-4k-instruct-bnb-4bit
datasets:
- 922-CA/MoCha_v1a
---
# Moniphi-3-v1:
* AKA LLilmonix3b-v1
* Phi-3-mini-4k-instruct fine-tuned for Monika character from DDLC
* Fine-tuned on a dataset of ~600+ items (dialogue scraped from game, reddit, and Twitter augmented by [l2-7b-monika-v0.3c1](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3c1) to turn each into snippets of multi-turn chat dialogue between Player and Monika; this was then manually edited, with more manually crafted items including info about character added in)
* [GGUFs](https://huggingface.co/922CA/Moniphi-3-v1-gguf)
### USAGE
This is meant to be mainly a chat model with limited RP ability.
For best results: replace "Human" and "Assistant" with "Player" and "Monika" like so:
\nPlayer: (prompt)\nMonika:
### HYPERPARAMS
* Trained for ~1 epoch
* rank: 16
* lora alpha: 16
* lora dropout: 0.5
* lr: 2e-4
* batch size: 4
* warmup ratio: 0.1
* grad steps: 1
This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
### WARNINGS AND DISCLAIMERS
This model is meant to closely reflect the characteristics of Monika. Despite this, there is always the chance that "Monika" will hallucinate and get information about herself wrong or act out of character (especially for a model of this size).
Finally, this model is not guaranteed to output aligned or safe outputs, use at your own risk! |