OddTheGreat commited on
Commit
04330df
·
verified ·
1 Parent(s): a4dcb8e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -44
README.md CHANGED
@@ -8,53 +8,18 @@ library_name: transformers
8
  tags:
9
  - mergekit
10
  - merge
11
-
 
 
 
 
12
  ---
13
  # merge
14
 
15
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
16
-
17
- ## Merge Details
18
- ### Merge Method
19
-
20
- This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [Delta-Vector/Hamanasu-15B-Instruct](https://huggingface.co/Delta-Vector/Hamanasu-15B-Instruct) as a base.
21
-
22
- ### Models Merged
23
-
24
- The following models were included in the merge:
25
- * [prithivMLmods/Phi-4-o1](https://huggingface.co/prithivMLmods/Phi-4-o1)
26
- * [prithivMLmods/Megatron-Opus-14B-2.0](https://huggingface.co/prithivMLmods/Megatron-Opus-14B-2.0)
27
- * [pankajmathur/orca_mini_phi-4](https://huggingface.co/pankajmathur/orca_mini_phi-4)
28
 
29
- ### Configuration
30
 
31
- The following YAML configuration was used to produce this model:
32
 
33
- ```yaml
34
- models:
35
- - model: prithivMLmods/Phi-4-o1
36
- parameters:
37
- density: 0.5
38
- weight: 0.5
39
- - model: pankajmathur/orca_mini_phi-4
40
- parameters:
41
- density: 0.5
42
- weight: 0.5
43
- - model: prithivMLmods/Megatron-Opus-14B-2.0
44
- parameters:
45
- density: 0.5
46
- weight: 0.5
47
- - model: Delta-Vector/Hamanasu-15B-Instruct
48
- parameters:
49
- density: 0.7
50
- weight: 0.7
51
-
52
- merge_method: ties
53
- base_model: Delta-Vector/Hamanasu-15B-Instruct
54
- parameters:
55
- normalize: false
56
- int8_mask: true
57
- dtype: float16
58
- tokenizer:
59
- source: union
60
- ```
 
8
  tags:
9
  - mergekit
10
  - merge
11
+ - roleplay
12
+ - experimental
13
+ language:
14
+ - ru
15
+ - en
16
  ---
17
  # merge
18
 
19
+ This is a merge of pre-trained language models.
 
 
 
 
 
 
 
 
 
 
 
 
20
 
21
+ This merge was made to test how phi4 reacts to merging.
22
 
23
+ Ii reacted good. Overall intellectual capabilities are good, rp language is rich, ERP is supported, but not as good as wanted. On ru it is better than it's parts, on eng it is good too.
24
 
25
+ Just remember, it is experimental, so tested on little amount of replies ~100