Update README.md
Browse files
README.md
CHANGED
@@ -6,15 +6,23 @@ tags:
|
|
6 |
- lazymergekit
|
7 |
- bunnycore/QandoraExp-7B
|
8 |
- trollek/Qwen2.5-7B-CySecButler-v0.1
|
|
|
|
|
|
|
|
|
9 |
---
|
|
|
|
|
|
|
10 |
|
11 |
-
|
12 |
|
13 |
-
|
14 |
-
|
15 |
-
* [trollek/Qwen2.5-7B-CySecButler-v0.1](https://huggingface.co/trollek/Qwen2.5-7B-CySecButler-v0.1)
|
16 |
|
17 |
-
## 𧩠Configuration
|
|
|
|
|
18 |
|
19 |
```yaml
|
20 |
slices:
|
@@ -33,5 +41,51 @@ parameters:
|
|
33 |
value: [1, 0.5, 0.7, 0.3, 0]
|
34 |
- value: 0.5
|
35 |
dtype: bfloat16
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
36 |
|
37 |
-
|
|
|
6 |
- lazymergekit
|
7 |
- bunnycore/QandoraExp-7B
|
8 |
- trollek/Qwen2.5-7B-CySecButler-v0.1
|
9 |
+
base_model:
|
10 |
+
- bunnycore/QandoraExp-7B
|
11 |
+
- trollek/Qwen2.5-7B-CySecButler-v0.1
|
12 |
+
library_name: transformers
|
13 |
---
|
14 |
+
# Qwen2.5-7B-Qandora-CySec
|
15 |
+
|
16 |
+
ZeroXClem/Qwen2.5-7B-Qandora-CySec is an advanced model merge combining Q&A capabilities and cybersecurity expertise using the mergekit framework. This model excels in both general question-answering tasks and specialized cybersecurity domains.
|
17 |
|
18 |
+
## π Model Components
|
19 |
|
20 |
+
- **[bunnycore/QandoraExp-7B](https://huggingface.co/bunnycore/QandoraExp-7B)**: Powerful Q&A capabilities
|
21 |
+
- **[trollek/Qwen2.5-7B-CySecButler-v0.1](https://huggingface.co/trollek/Qwen2.5-7B-CySecButler-v0.1)**: Specialized cybersecurity knowledge
|
|
|
22 |
|
23 |
+
## 𧩠Merge Configuration
|
24 |
+
|
25 |
+
The models are merged using spherical linear interpolation (SLERP) for optimal blending:
|
26 |
|
27 |
```yaml
|
28 |
slices:
|
|
|
41 |
value: [1, 0.5, 0.7, 0.3, 0]
|
42 |
- value: 0.5
|
43 |
dtype: bfloat16
|
44 |
+
```
|
45 |
+
|
46 |
+
### Key Parameters
|
47 |
+
|
48 |
+
- **Self-Attention (self_attn)**: Controls blending across self-attention layers
|
49 |
+
- **MLP**: Adjusts Multi-Layer Perceptron balance
|
50 |
+
- **Global Weight (t.value)**: 0.5 for equal contribution from both models
|
51 |
+
- **Data Type**: bfloat16 for efficiency and precision
|
52 |
+
|
53 |
+
## π― Applications
|
54 |
+
|
55 |
+
1. General Q&A Tasks
|
56 |
+
2. Cybersecurity Analysis
|
57 |
+
3. Hybrid Scenarios (general knowledge + cybersecurity)
|
58 |
+
|
59 |
+
## π Usage
|
60 |
+
|
61 |
+
```python
|
62 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
63 |
+
|
64 |
+
model_name = "ZeroXClem/Qwen2.5-7B-Qandora-CySec"
|
65 |
+
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
66 |
+
model = AutoModelForCausalLM.from_pretrained(model_name)
|
67 |
+
|
68 |
+
input_text = "What are the fundamentals of python programming?"
|
69 |
+
input_ids = tokenizer.encode(input_text, return_tensors="pt")
|
70 |
+
output = model.generate(input_ids, max_length=100)
|
71 |
+
response = tokenizer.decode(output[0], skip_special_tokens=True)
|
72 |
+
print(response)
|
73 |
+
```
|
74 |
+
|
75 |
+
## π License
|
76 |
+
|
77 |
+
This model inherits the licenses of its base models. Refer to bunnycore/QandoraExp-7B and trollek/Qwen2.5-7B-CySecButler-v0.1 for usage terms.
|
78 |
+
|
79 |
+
## π Acknowledgements
|
80 |
+
|
81 |
+
- bunnycore (QandoraExp-7B)
|
82 |
+
- trollek (Qwen2.5-7B-CySecButler-v0.1)
|
83 |
+
- mergekit project
|
84 |
+
|
85 |
+
## π Citation
|
86 |
+
|
87 |
+
If you use this model, please cite this repository and the original base models.
|
88 |
+
|
89 |
+
## π‘ Tags
|
90 |
|
91 |
+
merge, mergekit, lazymergekit, bunnycore/QandoraExp-7B, trollek/Qwen2.5-7B-CySecButler-v0.1, cybersecurity, Q&A
|