ZeroXClem commited on
Commit
6849553
β€’
1 Parent(s): 72b1ee8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +157 -3
README.md CHANGED
@@ -4,13 +4,41 @@ tags:
4
  - merge
5
  - mergekit
6
  - lazymergekit
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  ---
8
 
9
  # ZeroXClem/Qwen2.5-7B-HomerCreative-Mix
10
 
11
- ZeroXClem/Qwen2.5-7B-HomerCreative-Mix is a merge of the following models using [mergekit](https://github.com/cg123/mergekit):
12
 
13
- ## 🧩 Configuration
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
 
15
  ```yaml
16
  # Merge configuration for ZeroXClem/Qwen2.5-7B-HomerCreative-Mix using Model Stock
@@ -24,5 +52,131 @@ base_model: newsbang/Homer-v0.5-Qwen2.5-7B
24
  normalize: false
25
  int8_mask: true
26
  dtype: bfloat16
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
27
 
28
- ```
 
4
  - merge
5
  - mergekit
6
  - lazymergekit
7
+ - creative
8
+ - roleplay
9
+ - instruct
10
+ - qwen
11
+ - model_stock
12
+ - bfloat16
13
+ base_model:
14
+ - newsbang/Homer-v0.5-Qwen2.5-7B
15
+ - allknowingroger/HomerSlerp1-7B
16
+ - bunnycore/Qwen2.5-7B-Instruct-Fusion
17
+ - bunnycore/Qandora-2.5-7B-Creative
18
+ language:
19
+ - en
20
+ library_name: transformers
21
  ---
22
 
23
  # ZeroXClem/Qwen2.5-7B-HomerCreative-Mix
24
 
25
+ **ZeroXClem/Qwen2.5-7B-HomerCreative-Mix** is an advanced language model meticulously crafted by merging four pre-trained models using the powerful [mergekit](https://github.com/cg123/mergekit) framework. This fusion leverages the **Model Stock** merge method to combine the creative prowess of **Qandora**, the instructive capabilities of **Qwen-Instruct-Fusion**, the sophisticated blending of **HomerSlerp1**, and the foundational conversational strengths of **Homer-v0.5-Qwen2.5-7B**. The resulting model excels in creative text generation, contextual understanding, and dynamic conversational interactions.
26
 
27
+ ## πŸš€ Merged Models
28
+
29
+ This model merge incorporates the following:
30
+
31
+ - [**bunnycore/Qandora-2.5-7B-Creative**](https://huggingface.co/bunnycore/Qandora-2.5-7B-Creative): Specializes in creative text generation, enhancing the model's ability to produce imaginative and diverse content.
32
+
33
+ - [**bunnycore/Qwen2.5-7B-Instruct-Fusion**](https://huggingface.co/bunnycore/Qwen2.5-7B-Instruct-Fusion): Focuses on instruction-following capabilities, improving the model's performance in understanding and executing user commands.
34
+
35
+ - [**allknowingroger/HomerSlerp1-7B**](https://huggingface.co/allknowingroger/HomerSlerp1-7B): Utilizes spherical linear interpolation (SLERP) to blend model weights smoothly, ensuring a harmonious integration of different model attributes.
36
+
37
+ - [**newsbang/Homer-v0.5-Qwen2.5-7B**](https://huggingface.co/newsbang/Homer-v0.5-Qwen2.5-7B): Acts as the foundational conversational model, providing robust language comprehension and generation capabilities.
38
+
39
+ ## 🧩 Merge Configuration
40
+
41
+ The configuration below outlines how the models are merged using the **Model Stock** method. This approach ensures a balanced and effective integration of the unique strengths from each source model.
42
 
43
  ```yaml
44
  # Merge configuration for ZeroXClem/Qwen2.5-7B-HomerCreative-Mix using Model Stock
 
52
  normalize: false
53
  int8_mask: true
54
  dtype: bfloat16
55
+ ```
56
+
57
+ ### Key Parameters
58
+
59
+ - **Merge Method (`merge_method`):** Utilizes the **Model Stock** method, as described in [Model Stock](https://arxiv.org/abs/2403.19522), to effectively combine multiple models by leveraging their strengths.
60
+
61
+ - **Models (`models`):** Specifies the list of models to be merged:
62
+ - **bunnycore/Qandora-2.5-7B-Creative:** Enhances creative text generation.
63
+ - **bunnycore/Qwen2.5-7B-Instruct-Fusion:** Improves instruction-following capabilities.
64
+ - **allknowingroger/HomerSlerp1-7B:** Facilitates smooth blending of model weights using SLERP.
65
+
66
+ - **Base Model (`base_model`):** Defines the foundational model for the merge, which is **newsbang/Homer-v0.5-Qwen2.5-7B** in this case.
67
+
68
+ - **Normalization (`normalize`):** Set to `false` to retain the original scaling of the model weights during the merge.
69
+
70
+ - **INT8 Mask (`int8_mask`):** Enabled (`true`) to apply INT8 quantization masking, optimizing the model for efficient inference without significant loss in precision.
71
+
72
+ - **Data Type (`dtype`):** Uses `bfloat16` to maintain computational efficiency while ensuring high precision.
73
+
74
+ ## πŸ† Performance Highlights
75
+
76
+ - **Creative Text Generation:** Enhanced ability to produce imaginative and diverse content suitable for creative writing, storytelling, and content creation.
77
+
78
+ - **Instruction Following:** Improved performance in understanding and executing user instructions, making the model more responsive and accurate in task execution.
79
+
80
+ - **Optimized Inference:** INT8 masking and `bfloat16` data type contribute to efficient computation, enabling faster response times without compromising quality.
81
+
82
+ ## 🎯 Use Case & Applications
83
+
84
+ **ZeroXClem/Qwen2.5-7B-HomerCreative-Mix** is designed to excel in environments that demand both creative generation and precise instruction following. Ideal applications include:
85
+
86
+ - **Creative Writing Assistance:** Aiding authors and content creators in generating imaginative narratives, dialogues, and descriptive text.
87
+
88
+ - **Interactive Storytelling and Role-Playing:** Enhancing dynamic and engaging interactions in role-playing games and interactive storytelling platforms.
89
+
90
+ - **Educational Tools and Tutoring Systems:** Providing detailed explanations, answering questions, and assisting in educational content creation with contextual understanding.
91
+
92
+ - **Technical Support and Customer Service:** Offering accurate and contextually relevant responses in technical support scenarios, improving user satisfaction.
93
+
94
+ - **Content Generation for Marketing:** Creating compelling and diverse marketing copy, social media posts, and promotional material with creative flair.
95
+
96
+ ## πŸ“ Usage
97
+
98
+ To utilize **ZeroXClem/Qwen2.5-7B-HomerCreative-Mix**, follow the steps below:
99
+
100
+ ### Installation
101
+
102
+ First, install the necessary libraries:
103
+
104
+ ```bash
105
+ pip install -qU transformers accelerate
106
+ ```
107
+
108
+ ### Example Code
109
+
110
+ Below is an example of how to load and use the model for text generation:
111
+
112
+ ```python
113
+ from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
114
+ import torch
115
+
116
+ # Define the model name
117
+ model_name = "ZeroXClem/Qwen2.5-7B-HomerCreative-Mix"
118
+
119
+ # Load the tokenizer
120
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
121
+
122
+ # Load the model
123
+ model = AutoModelForCausalLM.from_pretrained(
124
+ model_name,
125
+ torch_dtype=torch.bfloat16,
126
+ device_map="auto"
127
+ )
128
+
129
+ # Initialize the pipeline
130
+ text_generator = pipeline(
131
+ "text-generation",
132
+ model=model,
133
+ tokenizer=tokenizer,
134
+ torch_dtype=torch.bfloat16,
135
+ device_map="auto"
136
+ )
137
+
138
+ # Define the input prompt
139
+ prompt = "Once upon a time in a land far, far away,"
140
+
141
+ # Generate the output
142
+ outputs = text_generator(
143
+ prompt,
144
+ max_new_tokens=150,
145
+ do_sample=True,
146
+ temperature=0.7,
147
+ top_k=50,
148
+ top_p=0.95
149
+ )
150
+
151
+ # Print the generated text
152
+ print(outputs[0]["generated_text"])
153
+ ```
154
+
155
+ ### Notes
156
+
157
+ - **Fine-Tuning:** This merged model may require fine-tuning to optimize performance for specific applications or domains.
158
+
159
+ - **Resource Requirements:** Ensure that your environment has sufficient computational resources, especially GPU-enabled hardware, to handle the model efficiently during inference.
160
+
161
+ - **Customization:** Users can adjust parameters such as `temperature`, `top_k`, and `top_p` to control the creativity and diversity of the generated text.
162
+
163
+
164
+ ## πŸ“œ License
165
+
166
+ This model is open-sourced under the **Apache-2.0 License**.
167
+
168
+ ## πŸ’‘ Tags
169
+
170
+ - `merge`
171
+ - `mergekit`
172
+ - `model_stock`
173
+ - `Qwen`
174
+ - `Homer`
175
+ - `Creative`
176
+ - `ZeroXClem/Qwen2.5-7B-HomerCreative-Mix`
177
+ - `bunnycore/Qandora-2.5-7B-Creative`
178
+ - `bunnycore/Qwen2.5-7B-Instruct-Fusion`
179
+ - `allknowingroger/HomerSlerp1-7B`
180
+ - `newsbang/Homer-v0.5-Qwen2.5-7B`
181
 
182
+ ---