mav23 commited on
Commit
8c83bed
1 Parent(s): 57fd4a2

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. .gitattributes +1 -0
  2. README.md +217 -0
  3. llama3.2-3b-enigma.Q4_0.gguf +3 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ llama3.2-3b-enigma.Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,217 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: llama3.2
5
+ tags:
6
+ - enigma
7
+ - valiant
8
+ - valiant-labs
9
+ - llama
10
+ - llama-3.2
11
+ - llama-3.2-instruct
12
+ - llama-3.2-instruct-3b
13
+ - llama-3
14
+ - llama-3-instruct
15
+ - llama-3-instruct-3b
16
+ - 3b
17
+ - code
18
+ - code-instruct
19
+ - python
20
+ - conversational
21
+ - chat
22
+ - instruct
23
+ base_model: meta-llama/Llama-3.2-3B-Instruct
24
+ datasets:
25
+ - sequelbox/Tachibana
26
+ - sequelbox/Supernova
27
+ pipeline_tag: text-generation
28
+ model_type: llama
29
+ model-index:
30
+ - name: Llama3.2-3B-Enigma
31
+ results:
32
+ - task:
33
+ type: text-generation
34
+ name: Text Generation
35
+ dataset:
36
+ name: Winogrande (5-Shot)
37
+ type: winogrande
38
+ args:
39
+ num_few_shot: 5
40
+ metrics:
41
+ - type: acc
42
+ value: 67.96
43
+ name: acc
44
+ - task:
45
+ type: text-generation
46
+ name: Text Generation
47
+ dataset:
48
+ name: ARC Challenge (25-Shot)
49
+ type: arc-challenge
50
+ args:
51
+ num_few_shot: 25
52
+ metrics:
53
+ - type: acc_norm
54
+ value: 47.18
55
+ name: normalized accuracy
56
+ - task:
57
+ type: text-generation
58
+ name: Text Generation
59
+ dataset:
60
+ name: IFEval (0-Shot)
61
+ type: HuggingFaceH4/ifeval
62
+ args:
63
+ num_few_shot: 0
64
+ metrics:
65
+ - type: inst_level_strict_acc and prompt_level_strict_acc
66
+ value: 47.75
67
+ name: strict accuracy
68
+ source:
69
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.2-3B-Enigma
70
+ name: Open LLM Leaderboard
71
+ - task:
72
+ type: text-generation
73
+ name: Text Generation
74
+ dataset:
75
+ name: BBH (3-Shot)
76
+ type: BBH
77
+ args:
78
+ num_few_shot: 3
79
+ metrics:
80
+ - type: acc_norm
81
+ value: 18.81
82
+ name: normalized accuracy
83
+ source:
84
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.2-3B-Enigma
85
+ name: Open LLM Leaderboard
86
+ - task:
87
+ type: text-generation
88
+ name: Text Generation
89
+ dataset:
90
+ name: MATH Lvl 5 (4-Shot)
91
+ type: hendrycks/competition_math
92
+ args:
93
+ num_few_shot: 4
94
+ metrics:
95
+ - type: exact_match
96
+ value: 6.65
97
+ name: exact match
98
+ source:
99
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.2-3B-Enigma
100
+ name: Open LLM Leaderboard
101
+ - task:
102
+ type: text-generation
103
+ name: Text Generation
104
+ dataset:
105
+ name: GPQA (0-shot)
106
+ type: Idavidrein/gpqa
107
+ args:
108
+ num_few_shot: 0
109
+ metrics:
110
+ - type: acc_norm
111
+ value: 1.45
112
+ name: acc_norm
113
+ source:
114
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.2-3B-Enigma
115
+ name: Open LLM Leaderboard
116
+ - task:
117
+ type: text-generation
118
+ name: Text Generation
119
+ dataset:
120
+ name: MuSR (0-shot)
121
+ type: TAUR-Lab/MuSR
122
+ args:
123
+ num_few_shot: 0
124
+ metrics:
125
+ - type: acc_norm
126
+ value: 4.54
127
+ name: acc_norm
128
+ source:
129
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.2-3B-Enigma
130
+ name: Open LLM Leaderboard
131
+ - task:
132
+ type: text-generation
133
+ name: Text Generation
134
+ dataset:
135
+ name: MMLU-PRO (5-shot)
136
+ type: TIGER-Lab/MMLU-Pro
137
+ config: main
138
+ split: test
139
+ args:
140
+ num_few_shot: 5
141
+ metrics:
142
+ - type: acc
143
+ value: 15.41
144
+ name: accuracy
145
+ source:
146
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.2-3B-Enigma
147
+ name: Open LLM Leaderboard
148
+ ---
149
+
150
+
151
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/64f267a8a4f79a118e0fcc89/it7MY5MyLCLpFQev5dUis.jpeg)
152
+
153
+
154
+ Enigma is a code-instruct model built on Llama 3.2 3b.
155
+ - High quality code instruct performance with the Llama 3.2 Instruct chat format
156
+ - Finetuned on synthetic code-instruct data generated with Llama 3.1 405b. [Find the current version of the dataset here!](https://huggingface.co/datasets/sequelbox/Tachibana)
157
+ - Overall chat performance supplemented with [generalist synthetic data.](https://huggingface.co/datasets/sequelbox/Supernova)
158
+
159
+
160
+ ## Version
161
+
162
+ This is the **2024-09-30** release of Enigma for Llama 3.2 3b, enhancing code-instruct and general chat capabilities.
163
+
164
+ Enigma is also available for [Llama 3.1 8b!](https://huggingface.co/ValiantLabs/Llama3.1-8B-Enigma)
165
+
166
+ Help us and recommend Enigma to your friends! We're excited for more Enigma releases in the future.
167
+
168
+
169
+ ## Prompting Guide
170
+ Enigma uses the [Llama 3.2 Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct) prompt format. The example script below can be used as a starting point for general chat:
171
+
172
+ ```python
173
+ import transformers
174
+ import torch
175
+
176
+ model_id = "ValiantLabs/Llama3.2-3B-Enigma"
177
+
178
+ pipeline = transformers.pipeline(
179
+ "text-generation",
180
+ model=model_id,
181
+ model_kwargs={"torch_dtype": torch.bfloat16},
182
+ device_map="auto",
183
+ )
184
+
185
+ messages = [
186
+ {"role": "system", "content": "You are Enigma, a highly capable code assistant."},
187
+ {"role": "user", "content": "Can you explain virtualization to me?"}
188
+ ]
189
+
190
+ outputs = pipeline(
191
+ messages,
192
+ max_new_tokens=1024,
193
+ )
194
+
195
+ print(outputs[0]["generated_text"][-1])
196
+ ```
197
+
198
+ ## The Model
199
+ Enigma is built on top of Llama 3.2 3b Instruct, using high quality code-instruct data and general chat data in Llama 3.2 Instruct prompt style to supplement overall performance.
200
+
201
+ Our current version of Enigma is trained on code-instruct data from [sequelbox/Tachibana](https://huggingface.co/datasets/sequelbox/Tachibana) and general chat data from [sequelbox/Supernova.](https://huggingface.co/datasets/sequelbox/Supernova)
202
+
203
+
204
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/63444f2687964b331809eb55/VCJ8Fmefd8cdVhXSSxJiD.jpeg)
205
+
206
+
207
+ Enigma is created by [Valiant Labs.](http://valiantlabs.ca/)
208
+
209
+ [Check out our HuggingFace page for Shining Valiant 2 and our other Build Tools models for creators!](https://huggingface.co/ValiantLabs)
210
+
211
+ [Follow us on X for updates on our models!](https://twitter.com/valiant_labs)
212
+
213
+ We care about open source.
214
+ For everyone to use.
215
+
216
+ We encourage others to finetune further from our models.
217
+
llama3.2-3b-enigma.Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6f719c097dfe9e43525b19dd586ed4afc46de287eaae249427fb214747024ec
3
+ size 1917191040