Venus 103b - version 1.2
Model Details
- A result of interleaving layers of Sao10K/Euryale-1.3-L2-70B and GOAT-AI/GOAT-70B-Storytelling
- The resulting model has 120 layers and 103 billion parameters.
- See mergekit-config.yml for details on the merge method used.
- See the
exl2-*
branches for exllama2 quantizations. The 5.65 bpw quant should fit in 80GB VRAM, and the 3.35/3.0 bpw quants should fit in 48GB VRAM. - Inspired by Goliath-120b
Warning: This model will produce NSFW content!
Results
- In my limited testing, I've found this model to be the most creative of the 103b merges I've made so far.
- Seems to tolerate higher temperatures than the previous Venus models
- Doesn't seem to suffer from any censorship issues
- Does not follow instructions as well as v1.1, but still does a bit better than v1.0
- Has some issues with formatting sometimes (i.e not closing asterisks or quotes)
Note that these are obviously just my personal observations, everyone will have their own unique experience based on their settings and specific scenarios they're using the model for.
- Downloads last month
- 16
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.