ZeroXClem commited on
Commit
6c8b513
Β·
verified Β·
1 Parent(s): 7bfa6d1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +60 -6
README.md CHANGED
@@ -6,15 +6,23 @@ tags:
6
  - lazymergekit
7
  - bunnycore/QandoraExp-7B
8
  - trollek/Qwen2.5-7B-CySecButler-v0.1
 
 
 
 
9
  ---
 
 
 
10
 
11
- # ZeroXClem/Qwen2.5-7B-Qandora-CySec
12
 
13
- ZeroXClem/Qwen2.5-7B-Qandora-CySec is a merge of the following models using [mergekit](https://github.com/cg123/mergekit):
14
- * [bunnycore/QandoraExp-7B](https://huggingface.co/bunnycore/QandoraExp-7B)
15
- * [trollek/Qwen2.5-7B-CySecButler-v0.1](https://huggingface.co/trollek/Qwen2.5-7B-CySecButler-v0.1)
16
 
17
- ## 🧩 Configuration
 
 
18
 
19
  ```yaml
20
  slices:
@@ -33,5 +41,51 @@ parameters:
33
  value: [1, 0.5, 0.7, 0.3, 0]
34
  - value: 0.5
35
  dtype: bfloat16
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
36
 
37
- ```
 
6
  - lazymergekit
7
  - bunnycore/QandoraExp-7B
8
  - trollek/Qwen2.5-7B-CySecButler-v0.1
9
+ base_model:
10
+ - bunnycore/QandoraExp-7B
11
+ - trollek/Qwen2.5-7B-CySecButler-v0.1
12
+ library_name: transformers
13
  ---
14
+ # Qwen2.5-7B-Qandora-CySec
15
+
16
+ ZeroXClem/Qwen2.5-7B-Qandora-CySec is an advanced model merge combining Q&A capabilities and cybersecurity expertise using the mergekit framework. This model excels in both general question-answering tasks and specialized cybersecurity domains.
17
 
18
+ ## πŸš€ Model Components
19
 
20
+ - **[bunnycore/QandoraExp-7B](https://huggingface.co/bunnycore/QandoraExp-7B)**: Powerful Q&A capabilities
21
+ - **[trollek/Qwen2.5-7B-CySecButler-v0.1](https://huggingface.co/trollek/Qwen2.5-7B-CySecButler-v0.1)**: Specialized cybersecurity knowledge
 
22
 
23
+ ## 🧩 Merge Configuration
24
+
25
+ The models are merged using spherical linear interpolation (SLERP) for optimal blending:
26
 
27
  ```yaml
28
  slices:
 
41
  value: [1, 0.5, 0.7, 0.3, 0]
42
  - value: 0.5
43
  dtype: bfloat16
44
+ ```
45
+
46
+ ### Key Parameters
47
+
48
+ - **Self-Attention (self_attn)**: Controls blending across self-attention layers
49
+ - **MLP**: Adjusts Multi-Layer Perceptron balance
50
+ - **Global Weight (t.value)**: 0.5 for equal contribution from both models
51
+ - **Data Type**: bfloat16 for efficiency and precision
52
+
53
+ ## 🎯 Applications
54
+
55
+ 1. General Q&A Tasks
56
+ 2. Cybersecurity Analysis
57
+ 3. Hybrid Scenarios (general knowledge + cybersecurity)
58
+
59
+ ## πŸ›  Usage
60
+
61
+ ```python
62
+ from transformers import AutoTokenizer, AutoModelForCausalLM
63
+
64
+ model_name = "ZeroXClem/Qwen2.5-7B-Qandora-CySec"
65
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
66
+ model = AutoModelForCausalLM.from_pretrained(model_name)
67
+
68
+ input_text = "What are the fundamentals of python programming?"
69
+ input_ids = tokenizer.encode(input_text, return_tensors="pt")
70
+ output = model.generate(input_ids, max_length=100)
71
+ response = tokenizer.decode(output[0], skip_special_tokens=True)
72
+ print(response)
73
+ ```
74
+
75
+ ## πŸ“œ License
76
+
77
+ This model inherits the licenses of its base models. Refer to bunnycore/QandoraExp-7B and trollek/Qwen2.5-7B-CySecButler-v0.1 for usage terms.
78
+
79
+ ## πŸ™ Acknowledgements
80
+
81
+ - bunnycore (QandoraExp-7B)
82
+ - trollek (Qwen2.5-7B-CySecButler-v0.1)
83
+ - mergekit project
84
+
85
+ ## πŸ“š Citation
86
+
87
+ If you use this model, please cite this repository and the original base models.
88
+
89
+ ## πŸ’‘ Tags
90
 
91
+ merge, mergekit, lazymergekit, bunnycore/QandoraExp-7B, trollek/Qwen2.5-7B-CySecButler-v0.1, cybersecurity, Q&A