---
language:
- en
license: cc-by-nc-4.0
tags:
- text-generation-inference
- transformers
- unsloth
- mistral
- GEB
base_model: ammarali32/multi_verse_model
datasets:
- maldv/crabcanon
---
# SHRDFU-7b Δ
- **Developed by:** maldv
- **License:** cc-by-nc-4.0
- **Finetuned from model:** ammarali32/multi_verse_model
- **Methodology:** Peft to train; extending intelligence and problem solving w/ crabcanon
As I work on understanding how to layer information in to the model, this model used no conditioning and even with low LR's, had quite a sharp graph. It definitely inherited the style of the source.
I had been experimenting in this series with wrapping each paragraph or turn with `` bos/eos tokens. This may be semi-compatible with instruct, but is incompatible with alpaca and chatml. Good to know.
[](https://github.com/unslothai/unsloth)