Update README.md
Browse files
README.md
CHANGED
@@ -8,53 +8,18 @@ library_name: transformers
|
|
8 |
tags:
|
9 |
- mergekit
|
10 |
- merge
|
11 |
-
|
|
|
|
|
|
|
|
|
12 |
---
|
13 |
# merge
|
14 |
|
15 |
-
This is a merge of pre-trained language models
|
16 |
-
|
17 |
-
## Merge Details
|
18 |
-
### Merge Method
|
19 |
-
|
20 |
-
This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [Delta-Vector/Hamanasu-15B-Instruct](https://huggingface.co/Delta-Vector/Hamanasu-15B-Instruct) as a base.
|
21 |
-
|
22 |
-
### Models Merged
|
23 |
-
|
24 |
-
The following models were included in the merge:
|
25 |
-
* [prithivMLmods/Phi-4-o1](https://huggingface.co/prithivMLmods/Phi-4-o1)
|
26 |
-
* [prithivMLmods/Megatron-Opus-14B-2.0](https://huggingface.co/prithivMLmods/Megatron-Opus-14B-2.0)
|
27 |
-
* [pankajmathur/orca_mini_phi-4](https://huggingface.co/pankajmathur/orca_mini_phi-4)
|
28 |
|
29 |
-
|
30 |
|
31 |
-
|
32 |
|
33 |
-
|
34 |
-
models:
|
35 |
-
- model: prithivMLmods/Phi-4-o1
|
36 |
-
parameters:
|
37 |
-
density: 0.5
|
38 |
-
weight: 0.5
|
39 |
-
- model: pankajmathur/orca_mini_phi-4
|
40 |
-
parameters:
|
41 |
-
density: 0.5
|
42 |
-
weight: 0.5
|
43 |
-
- model: prithivMLmods/Megatron-Opus-14B-2.0
|
44 |
-
parameters:
|
45 |
-
density: 0.5
|
46 |
-
weight: 0.5
|
47 |
-
- model: Delta-Vector/Hamanasu-15B-Instruct
|
48 |
-
parameters:
|
49 |
-
density: 0.7
|
50 |
-
weight: 0.7
|
51 |
-
|
52 |
-
merge_method: ties
|
53 |
-
base_model: Delta-Vector/Hamanasu-15B-Instruct
|
54 |
-
parameters:
|
55 |
-
normalize: false
|
56 |
-
int8_mask: true
|
57 |
-
dtype: float16
|
58 |
-
tokenizer:
|
59 |
-
source: union
|
60 |
-
```
|
|
|
8 |
tags:
|
9 |
- mergekit
|
10 |
- merge
|
11 |
+
- roleplay
|
12 |
+
- experimental
|
13 |
+
language:
|
14 |
+
- ru
|
15 |
+
- en
|
16 |
---
|
17 |
# merge
|
18 |
|
19 |
+
This is a merge of pre-trained language models.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
20 |
|
21 |
+
This merge was made to test how phi4 reacts to merging.
|
22 |
|
23 |
+
Ii reacted good. Overall intellectual capabilities are good, rp language is rich, ERP is supported, but not as good as wanted. On ru it is better than it's parts, on eng it is good too.
|
24 |
|
25 |
+
Just remember, it is experimental, so tested on little amount of replies ~100
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|