--- license: apache-2.0 language: - en tags: - not-for-all-audiences --- ## Silly RP 7B First time with merges really, basically experimenting and seeing what works. Seems solid in my tests so far, but no guarantees on quality – give it a shot and share your feedback! Eager to hear how others like it. Honestly, I'm still learning the ropes, like the benefits of different merge methods :P ## Extra info - Chat template: ChatML --- base_model: - tavtav/eros-7b-test - NousResearch/Nous-Hermes-2-Mistral-7B-DPO - maywell/Synatra-7B-v0.3-RP - NeverSleep/Noromaid-7B-0.4-DPO - cogbuji/Mr-Grammatology-clinical-problems-Mistral-7B-0.5 library_name: transformers tags: - mergekit - merge --- # output This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [NeverSleep/Noromaid-7B-0.4-DPO](https://huggingface.co/NeverSleep/Noromaid-7B-0.4-DPO) as a base. ### Models Merged The following models were included in the merge: * [tavtav/eros-7b-test](https://huggingface.co/tavtav/eros-7b-test) * [NousResearch/Nous-Hermes-2-Mistral-7B-DPO](https://huggingface.co/NousResearch/Nous-Hermes-2-Mistral-7B-DPO) * [maywell/Synatra-7B-v0.3-RP](https://huggingface.co/maywell/Synatra-7B-v0.3-RP) * [cogbuji/Mr-Grammatology-clinical-problems-Mistral-7B-0.5](https://huggingface.co/cogbuji/Mr-Grammatology-clinical-problems-Mistral-7B-0.5) ### Configuration The following YAML configuration was used to produce this model: ```yaml base_model: NeverSleep/Noromaid-7B-0.4-DPO models: - model: maywell/Synatra-7B-v0.3-RP parameters: weight: 0.2 - model: tavtav/eros-7b-test parameters: weight: 0.2 - model: cogbuji/Mr-Grammatology-clinical-problems-Mistral-7B-0.5 parameters: weight: 0.2 - model: NousResearch/Nous-Hermes-2-Mistral-7B-DPO parameters: weight: 0.2 merge_method: task_arithmetic parameters: weight: 0.17 dtype: float16 random_seed: 694201337567099116663322537 ```