metadata
language:
- en
license: cc-by-nc-4.0
tags:
- text-generation-inference
- transformers
- unsloth
- mistral
- GEB
base_model: ammarali32/multi_verse_model
datasets:
- maldv/crabcanon
SHRDFU-7b Δ
- Developed by: maldv
- License: cc-by-nc-4.0
- Finetuned from model: ammarali32/multi_verse_model
- Methodology: Peft to train; extending intelligence and problem solving w/ crabcanon
As I work on understanding how to layer information in to the model, this model used no conditioning and even with low LR's, had quite a sharp graph. It definitely inherited the style of the source.
I had been experimenting in this series with wrapping each paragraph or turn with <s></s>
bos/eos tokens. This may be semi-compatible with instruct, but is incompatible with alpaca and chatml. Good to know.