Files changed (1) hide show
  1. README.md +134 -24
README.md CHANGED
@@ -1,32 +1,28 @@
1
  ---
 
 
2
  license: apache-2.0
 
 
3
  datasets:
4
  - Locutusque/inst_mix_v2_top_100k
5
- language:
6
- - en
7
  pipeline_tag: text-generation
8
  widget:
9
- - text: >-
10
- <|USER|> Design a Neo4j database and Cypher function snippet to Display
11
- Extreme Dental hygiene: Using Mouthwash for Analysis for Beginners.
12
- Implement if/else or switch/case statements to handle different conditions
13
- related to the Consent. Provide detailed comments explaining your control
14
- flow and the reasoning behind each decision. <|ASSISTANT|>
15
  - text: '<|USER|> Write me a story about a magical place. <|ASSISTANT|> '
16
- - text: >-
17
- <|USER|> Write me an essay about the life of George Washington
18
- <|ASSISTANT|>
19
  - text: '<|USER|> Solve the following equation 2x + 10 = 20 <|ASSISTANT|> '
20
- - text: >-
21
- <|USER|> Craft me a list of some nice places to visit around the world.
22
- <|ASSISTANT|>
23
- - text: >-
24
- <|USER|> How to manage a lazy employee: Address the employee verbally. Don't
25
- allow an employee's laziness or lack of enthusiasm to become a recurring
26
- issue. Tell the employee you're hoping to speak with them about workplace
27
- expectations and performance, and schedule a time to sit down together.
28
- Question: To manage a lazy employee, it is suggested to talk to the
29
- employee. True, False, or Neither? <|ASSISTANT|>
30
  inference:
31
  parameters:
32
  temperature: 0.5
@@ -35,8 +31,109 @@ inference:
35
  top_k: 30
36
  max_new_tokens: 250
37
  repetition_penalty: 1.15
38
- tags:
39
- - merge
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
  ---
41
  # LocutusqueXFelladrin-TinyMistral248M-Instruct
42
  This model was created by merging Locutusque/TinyMistral-248M-Instruct and Felladrin/TinyMistral-248M-SFT-v4 using mergekit. After the two models were merged, the resulting model was further trained on ~20,000 examples on the Locutusque/inst_mix_v2_top_100k at a low learning rate to further normalize weights. The following is the YAML config used to merge:
@@ -56,4 +153,17 @@ dtype: float16
56
  The resulting model combines the best of both worlds. With Locutusque/TinyMistral-248M-Instruct's coding capabilities and reasoning skills, and Felladrin/TinyMistral-248M-SFT-v4's low hallucination and instruction-following capabilities. The resulting model has an incredible performance considering its size.
57
 
58
  ## Evaluation
59
- Found in the Open LLM Leaderboard.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - en
4
  license: apache-2.0
5
+ tags:
6
+ - merge
7
  datasets:
8
  - Locutusque/inst_mix_v2_top_100k
 
 
9
  pipeline_tag: text-generation
10
  widget:
11
+ - text: '<|USER|> Design a Neo4j database and Cypher function snippet to Display Extreme
12
+ Dental hygiene: Using Mouthwash for Analysis for Beginners. Implement if/else
13
+ or switch/case statements to handle different conditions related to the Consent.
14
+ Provide detailed comments explaining your control flow and the reasoning behind
15
+ each decision. <|ASSISTANT|> '
 
16
  - text: '<|USER|> Write me a story about a magical place. <|ASSISTANT|> '
17
+ - text: '<|USER|> Write me an essay about the life of George Washington <|ASSISTANT|> '
 
 
18
  - text: '<|USER|> Solve the following equation 2x + 10 = 20 <|ASSISTANT|> '
19
+ - text: '<|USER|> Craft me a list of some nice places to visit around the world. <|ASSISTANT|> '
20
+ - text: '<|USER|> How to manage a lazy employee: Address the employee verbally. Don''t
21
+ allow an employee''s laziness or lack of enthusiasm to become a recurring issue.
22
+ Tell the employee you''re hoping to speak with them about workplace expectations
23
+ and performance, and schedule a time to sit down together. Question: To manage
24
+ a lazy employee, it is suggested to talk to the employee. True, False, or Neither?
25
+ <|ASSISTANT|> '
 
 
 
26
  inference:
27
  parameters:
28
  temperature: 0.5
 
31
  top_k: 30
32
  max_new_tokens: 250
33
  repetition_penalty: 1.15
34
+ model-index:
35
+ - name: LocutusqueXFelladrin-TinyMistral248M-Instruct
36
+ results:
37
+ - task:
38
+ type: text-generation
39
+ name: Text Generation
40
+ dataset:
41
+ name: AI2 Reasoning Challenge (25-Shot)
42
+ type: ai2_arc
43
+ config: ARC-Challenge
44
+ split: test
45
+ args:
46
+ num_few_shot: 25
47
+ metrics:
48
+ - type: acc_norm
49
+ value: 24.74
50
+ name: normalized accuracy
51
+ source:
52
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Locutusque/LocutusqueXFelladrin-TinyMistral248M-Instruct
53
+ name: Open LLM Leaderboard
54
+ - task:
55
+ type: text-generation
56
+ name: Text Generation
57
+ dataset:
58
+ name: HellaSwag (10-Shot)
59
+ type: hellaswag
60
+ split: validation
61
+ args:
62
+ num_few_shot: 10
63
+ metrics:
64
+ - type: acc_norm
65
+ value: 27.79
66
+ name: normalized accuracy
67
+ source:
68
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Locutusque/LocutusqueXFelladrin-TinyMistral248M-Instruct
69
+ name: Open LLM Leaderboard
70
+ - task:
71
+ type: text-generation
72
+ name: Text Generation
73
+ dataset:
74
+ name: MMLU (5-Shot)
75
+ type: cais/mmlu
76
+ config: all
77
+ split: test
78
+ args:
79
+ num_few_shot: 5
80
+ metrics:
81
+ - type: acc
82
+ value: 26.12
83
+ name: accuracy
84
+ source:
85
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Locutusque/LocutusqueXFelladrin-TinyMistral248M-Instruct
86
+ name: Open LLM Leaderboard
87
+ - task:
88
+ type: text-generation
89
+ name: Text Generation
90
+ dataset:
91
+ name: TruthfulQA (0-shot)
92
+ type: truthful_qa
93
+ config: multiple_choice
94
+ split: validation
95
+ args:
96
+ num_few_shot: 0
97
+ metrics:
98
+ - type: mc2
99
+ value: 40.12
100
+ source:
101
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Locutusque/LocutusqueXFelladrin-TinyMistral248M-Instruct
102
+ name: Open LLM Leaderboard
103
+ - task:
104
+ type: text-generation
105
+ name: Text Generation
106
+ dataset:
107
+ name: Winogrande (5-shot)
108
+ type: winogrande
109
+ config: winogrande_xl
110
+ split: validation
111
+ args:
112
+ num_few_shot: 5
113
+ metrics:
114
+ - type: acc
115
+ value: 49.09
116
+ name: accuracy
117
+ source:
118
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Locutusque/LocutusqueXFelladrin-TinyMistral248M-Instruct
119
+ name: Open LLM Leaderboard
120
+ - task:
121
+ type: text-generation
122
+ name: Text Generation
123
+ dataset:
124
+ name: GSM8k (5-shot)
125
+ type: gsm8k
126
+ config: main
127
+ split: test
128
+ args:
129
+ num_few_shot: 5
130
+ metrics:
131
+ - type: acc
132
+ value: 0.0
133
+ name: accuracy
134
+ source:
135
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Locutusque/LocutusqueXFelladrin-TinyMistral248M-Instruct
136
+ name: Open LLM Leaderboard
137
  ---
138
  # LocutusqueXFelladrin-TinyMistral248M-Instruct
139
  This model was created by merging Locutusque/TinyMistral-248M-Instruct and Felladrin/TinyMistral-248M-SFT-v4 using mergekit. After the two models were merged, the resulting model was further trained on ~20,000 examples on the Locutusque/inst_mix_v2_top_100k at a low learning rate to further normalize weights. The following is the YAML config used to merge:
 
153
  The resulting model combines the best of both worlds. With Locutusque/TinyMistral-248M-Instruct's coding capabilities and reasoning skills, and Felladrin/TinyMistral-248M-SFT-v4's low hallucination and instruction-following capabilities. The resulting model has an incredible performance considering its size.
154
 
155
  ## Evaluation
156
+ Found in the Open LLM Leaderboard.
157
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
158
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Locutusque__LocutusqueXFelladrin-TinyMistral248M-Instruct)
159
+
160
+ | Metric |Value|
161
+ |---------------------------------|----:|
162
+ |Avg. |27.98|
163
+ |AI2 Reasoning Challenge (25-Shot)|24.74|
164
+ |HellaSwag (10-Shot) |27.79|
165
+ |MMLU (5-Shot) |26.12|
166
+ |TruthfulQA (0-shot) |40.12|
167
+ |Winogrande (5-shot) |49.09|
168
+ |GSM8k (5-shot) | 0.00|
169
+