bunnycore lbourdois commited on
Commit
e6d608f
·
verified ·
1 Parent(s): 75c3fb1

Improve language tag (#2)

Browse files

- Improve language tag (4065f247ada506dfe73e860e8ecb738bda117947)


Co-authored-by: Loïck BOURDOIS <[email protected]>

Files changed (1) hide show
  1. README.md +178 -164
README.md CHANGED
@@ -1,164 +1,178 @@
1
- ---
2
- library_name: transformers
3
- tags:
4
- - mergekit
5
- - merge
6
- base_model:
7
- - gz987/qwen2.5-7b-cabs-v0.3
8
- - ngxson/LoRA-Qwen2.5-7B-Instruct-abliterated-v3
9
- - bunnycore/Qwen2.5-7B-Instruct-Merge-Stock-v0.1
10
- - Xiaojian9992024/Qwen2.5-Dyanka-7B-Preview
11
- - Qwen/Qwen2.5-7B-Instruct
12
- - bunnycore/Qwen-2.5-7b-s1k-lora_model
13
- - gz987/qwen2.5-7b-cabs-v0.3
14
- - bunnycore/Qwen-2.5-7b-rp-lora
15
- - Qwen/Qwen2.5-7B-Instruct
16
- - ngxson/LoRA-Qwen2.5-7B-Instruct-abliterated-v3
17
- model-index:
18
- - name: Blabbertron-1.0
19
- results:
20
- - task:
21
- type: text-generation
22
- name: Text Generation
23
- dataset:
24
- name: IFEval (0-Shot)
25
- type: HuggingFaceH4/ifeval
26
- args:
27
- num_few_shot: 0
28
- metrics:
29
- - type: inst_level_strict_acc and prompt_level_strict_acc
30
- value: 74.33
31
- name: strict accuracy
32
- source:
33
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Blabbertron-1.0
34
- name: Open LLM Leaderboard
35
- - task:
36
- type: text-generation
37
- name: Text Generation
38
- dataset:
39
- name: BBH (3-Shot)
40
- type: BBH
41
- args:
42
- num_few_shot: 3
43
- metrics:
44
- - type: acc_norm
45
- value: 36.05
46
- name: normalized accuracy
47
- source:
48
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Blabbertron-1.0
49
- name: Open LLM Leaderboard
50
- - task:
51
- type: text-generation
52
- name: Text Generation
53
- dataset:
54
- name: MATH Lvl 5 (4-Shot)
55
- type: hendrycks/competition_math
56
- args:
57
- num_few_shot: 4
58
- metrics:
59
- - type: exact_match
60
- value: 49.24
61
- name: exact match
62
- source:
63
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Blabbertron-1.0
64
- name: Open LLM Leaderboard
65
- - task:
66
- type: text-generation
67
- name: Text Generation
68
- dataset:
69
- name: GPQA (0-shot)
70
- type: Idavidrein/gpqa
71
- args:
72
- num_few_shot: 0
73
- metrics:
74
- - type: acc_norm
75
- value: 6.94
76
- name: acc_norm
77
- source:
78
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Blabbertron-1.0
79
- name: Open LLM Leaderboard
80
- - task:
81
- type: text-generation
82
- name: Text Generation
83
- dataset:
84
- name: MuSR (0-shot)
85
- type: TAUR-Lab/MuSR
86
- args:
87
- num_few_shot: 0
88
- metrics:
89
- - type: acc_norm
90
- value: 13.51
91
- name: acc_norm
92
- source:
93
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Blabbertron-1.0
94
- name: Open LLM Leaderboard
95
- - task:
96
- type: text-generation
97
- name: Text Generation
98
- dataset:
99
- name: MMLU-PRO (5-shot)
100
- type: TIGER-Lab/MMLU-Pro
101
- config: main
102
- split: test
103
- args:
104
- num_few_shot: 5
105
- metrics:
106
- - type: acc
107
- value: 37.27
108
- name: accuracy
109
- source:
110
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Blabbertron-1.0
111
- name: Open LLM Leaderboard
112
- ---
113
- # merge
114
-
115
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
116
-
117
- ## Merge Details
118
- ### Merge Method
119
-
120
- This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) + [ngxson/LoRA-Qwen2.5-7B-Instruct-abliterated-v3](https://huggingface.co/ngxson/LoRA-Qwen2.5-7B-Instruct-abliterated-v3) as a base.
121
-
122
- ### Models Merged
123
-
124
- The following models were included in the merge:
125
- * [gz987/qwen2.5-7b-cabs-v0.3](https://huggingface.co/gz987/qwen2.5-7b-cabs-v0.3) + [ngxson/LoRA-Qwen2.5-7B-Instruct-abliterated-v3](https://huggingface.co/ngxson/LoRA-Qwen2.5-7B-Instruct-abliterated-v3)
126
- * [bunnycore/Qwen2.5-7B-Instruct-Merge-Stock-v0.1](https://huggingface.co/bunnycore/Qwen2.5-7B-Instruct-Merge-Stock-v0.1)
127
- * [Xiaojian9992024/Qwen2.5-Dyanka-7B-Preview](https://huggingface.co/Xiaojian9992024/Qwen2.5-Dyanka-7B-Preview)
128
- * [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) + [bunnycore/Qwen-2.5-7b-s1k-lora_model](https://huggingface.co/bunnycore/Qwen-2.5-7b-s1k-lora_model)
129
- * [gz987/qwen2.5-7b-cabs-v0.3](https://huggingface.co/gz987/qwen2.5-7b-cabs-v0.3) + [bunnycore/Qwen-2.5-7b-rp-lora](https://huggingface.co/bunnycore/Qwen-2.5-7b-rp-lora)
130
-
131
- ### Configuration
132
-
133
- The following YAML configuration was used to produce this model:
134
-
135
- ```yaml
136
- models:
137
- - model: Qwen/Qwen2.5-7B-Instruct+bunnycore/Qwen-2.5-7b-s1k-lora_model
138
- parameters:
139
- weight: 0.3
140
- - model: Xiaojian9992024/Qwen2.5-Dyanka-7B-Preview
141
- - model: bunnycore/Qwen2.5-7B-Instruct-Merge-Stock-v0.1
142
- - model: gz987/qwen2.5-7b-cabs-v0.3+ngxson/LoRA-Qwen2.5-7B-Instruct-abliterated-v3
143
- - model: gz987/qwen2.5-7b-cabs-v0.3+bunnycore/Qwen-2.5-7b-rp-lora
144
- base_model: Qwen/Qwen2.5-7B-Instruct+ngxson/LoRA-Qwen2.5-7B-Instruct-abliterated-v3
145
- merge_method: model_stock
146
- parameters:
147
- dtype: bfloat16
148
- tokenizer_source: Qwen/Qwen2.5-7B-Instruct
149
-
150
- ```
151
-
152
- # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
153
- Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__Blabbertron-1.0-details)
154
-
155
- | Metric |Value|
156
- |-------------------|----:|
157
- |Avg. |36.22|
158
- |IFEval (0-Shot) |74.33|
159
- |BBH (3-Shot) |36.05|
160
- |MATH Lvl 5 (4-Shot)|49.24|
161
- |GPQA (0-shot) | 6.94|
162
- |MuSR (0-shot) |13.51|
163
- |MMLU-PRO (5-shot) |37.27|
164
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - mergekit
5
+ - merge
6
+ base_model:
7
+ - gz987/qwen2.5-7b-cabs-v0.3
8
+ - ngxson/LoRA-Qwen2.5-7B-Instruct-abliterated-v3
9
+ - bunnycore/Qwen2.5-7B-Instruct-Merge-Stock-v0.1
10
+ - Xiaojian9992024/Qwen2.5-Dyanka-7B-Preview
11
+ - Qwen/Qwen2.5-7B-Instruct
12
+ - bunnycore/Qwen-2.5-7b-s1k-lora_model
13
+ - gz987/qwen2.5-7b-cabs-v0.3
14
+ - bunnycore/Qwen-2.5-7b-rp-lora
15
+ - Qwen/Qwen2.5-7B-Instruct
16
+ - ngxson/LoRA-Qwen2.5-7B-Instruct-abliterated-v3
17
+ language:
18
+ - zho
19
+ - eng
20
+ - fra
21
+ - spa
22
+ - por
23
+ - deu
24
+ - ita
25
+ - rus
26
+ - jpn
27
+ - kor
28
+ - vie
29
+ - tha
30
+ - ara
31
+ model-index:
32
+ - name: Blabbertron-1.0
33
+ results:
34
+ - task:
35
+ type: text-generation
36
+ name: Text Generation
37
+ dataset:
38
+ name: IFEval (0-Shot)
39
+ type: HuggingFaceH4/ifeval
40
+ args:
41
+ num_few_shot: 0
42
+ metrics:
43
+ - type: inst_level_strict_acc and prompt_level_strict_acc
44
+ value: 74.33
45
+ name: strict accuracy
46
+ source:
47
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Blabbertron-1.0
48
+ name: Open LLM Leaderboard
49
+ - task:
50
+ type: text-generation
51
+ name: Text Generation
52
+ dataset:
53
+ name: BBH (3-Shot)
54
+ type: BBH
55
+ args:
56
+ num_few_shot: 3
57
+ metrics:
58
+ - type: acc_norm
59
+ value: 36.05
60
+ name: normalized accuracy
61
+ source:
62
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Blabbertron-1.0
63
+ name: Open LLM Leaderboard
64
+ - task:
65
+ type: text-generation
66
+ name: Text Generation
67
+ dataset:
68
+ name: MATH Lvl 5 (4-Shot)
69
+ type: hendrycks/competition_math
70
+ args:
71
+ num_few_shot: 4
72
+ metrics:
73
+ - type: exact_match
74
+ value: 49.24
75
+ name: exact match
76
+ source:
77
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Blabbertron-1.0
78
+ name: Open LLM Leaderboard
79
+ - task:
80
+ type: text-generation
81
+ name: Text Generation
82
+ dataset:
83
+ name: GPQA (0-shot)
84
+ type: Idavidrein/gpqa
85
+ args:
86
+ num_few_shot: 0
87
+ metrics:
88
+ - type: acc_norm
89
+ value: 6.94
90
+ name: acc_norm
91
+ source:
92
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Blabbertron-1.0
93
+ name: Open LLM Leaderboard
94
+ - task:
95
+ type: text-generation
96
+ name: Text Generation
97
+ dataset:
98
+ name: MuSR (0-shot)
99
+ type: TAUR-Lab/MuSR
100
+ args:
101
+ num_few_shot: 0
102
+ metrics:
103
+ - type: acc_norm
104
+ value: 13.51
105
+ name: acc_norm
106
+ source:
107
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Blabbertron-1.0
108
+ name: Open LLM Leaderboard
109
+ - task:
110
+ type: text-generation
111
+ name: Text Generation
112
+ dataset:
113
+ name: MMLU-PRO (5-shot)
114
+ type: TIGER-Lab/MMLU-Pro
115
+ config: main
116
+ split: test
117
+ args:
118
+ num_few_shot: 5
119
+ metrics:
120
+ - type: acc
121
+ value: 37.27
122
+ name: accuracy
123
+ source:
124
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Blabbertron-1.0
125
+ name: Open LLM Leaderboard
126
+ ---
127
+ # merge
128
+
129
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
130
+
131
+ ## Merge Details
132
+ ### Merge Method
133
+
134
+ This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) + [ngxson/LoRA-Qwen2.5-7B-Instruct-abliterated-v3](https://huggingface.co/ngxson/LoRA-Qwen2.5-7B-Instruct-abliterated-v3) as a base.
135
+
136
+ ### Models Merged
137
+
138
+ The following models were included in the merge:
139
+ * [gz987/qwen2.5-7b-cabs-v0.3](https://huggingface.co/gz987/qwen2.5-7b-cabs-v0.3) + [ngxson/LoRA-Qwen2.5-7B-Instruct-abliterated-v3](https://huggingface.co/ngxson/LoRA-Qwen2.5-7B-Instruct-abliterated-v3)
140
+ * [bunnycore/Qwen2.5-7B-Instruct-Merge-Stock-v0.1](https://huggingface.co/bunnycore/Qwen2.5-7B-Instruct-Merge-Stock-v0.1)
141
+ * [Xiaojian9992024/Qwen2.5-Dyanka-7B-Preview](https://huggingface.co/Xiaojian9992024/Qwen2.5-Dyanka-7B-Preview)
142
+ * [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) + [bunnycore/Qwen-2.5-7b-s1k-lora_model](https://huggingface.co/bunnycore/Qwen-2.5-7b-s1k-lora_model)
143
+ * [gz987/qwen2.5-7b-cabs-v0.3](https://huggingface.co/gz987/qwen2.5-7b-cabs-v0.3) + [bunnycore/Qwen-2.5-7b-rp-lora](https://huggingface.co/bunnycore/Qwen-2.5-7b-rp-lora)
144
+
145
+ ### Configuration
146
+
147
+ The following YAML configuration was used to produce this model:
148
+
149
+ ```yaml
150
+ models:
151
+ - model: Qwen/Qwen2.5-7B-Instruct+bunnycore/Qwen-2.5-7b-s1k-lora_model
152
+ parameters:
153
+ weight: 0.3
154
+ - model: Xiaojian9992024/Qwen2.5-Dyanka-7B-Preview
155
+ - model: bunnycore/Qwen2.5-7B-Instruct-Merge-Stock-v0.1
156
+ - model: gz987/qwen2.5-7b-cabs-v0.3+ngxson/LoRA-Qwen2.5-7B-Instruct-abliterated-v3
157
+ - model: gz987/qwen2.5-7b-cabs-v0.3+bunnycore/Qwen-2.5-7b-rp-lora
158
+ base_model: Qwen/Qwen2.5-7B-Instruct+ngxson/LoRA-Qwen2.5-7B-Instruct-abliterated-v3
159
+ merge_method: model_stock
160
+ parameters:
161
+ dtype: bfloat16
162
+ tokenizer_source: Qwen/Qwen2.5-7B-Instruct
163
+
164
+ ```
165
+
166
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
167
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__Blabbertron-1.0-details)
168
+
169
+ | Metric |Value|
170
+ |-------------------|----:|
171
+ |Avg. |36.22|
172
+ |IFEval (0-Shot) |74.33|
173
+ |BBH (3-Shot) |36.05|
174
+ |MATH Lvl 5 (4-Shot)|49.24|
175
+ |GPQA (0-shot) | 6.94|
176
+ |MuSR (0-shot) |13.51|
177
+ |MMLU-PRO (5-shot) |37.27|
178
+