CalamitousFelicitousness commited on
Commit
3e355e8
·
verified ·
1 Parent(s): 768542e

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,240 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - Nexusflow/Athene-V2-Chat
4
+ - EVA-UNIT-01/EVA-Qwen2.5-72B-v0.1
5
+ library_name: transformers
6
+ tags:
7
+ - mergekit
8
+ - merge
9
+ - Not-for-all-Audiences
10
+ ---
11
+
12
+
13
+
14
+ <div style="width: auto; margin-left: auto; margin-right: auto">
15
+ <img src="https://i.imgur.com/OxX2Usi.png" alt="Evathene" style="width: 80%; min-width: 400px; display: block; margin: auto;">
16
+ </div>
17
+
18
+
19
+ # Evathene-v1.0
20
+
21
+ This 72B parameter model is a merge of [Nexusflow/Athene-V2-Chat](https://huggingface.co/Nexusflow/Athene-V2-Chat) with [EVA-UNIT-01/EVA-Qwen2.5-72B-v0.1](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-72B-v0.1). See the merge recipe below for details.
22
+
23
+ This model is uncensored. *You are responsible for whatever you do with it.*
24
+
25
+ This model was designed for roleplaying and storytelling and I think it does well at both. It may also perform well at other tasks but I have not tested its performance in other areas.
26
+
27
+ # Sampler Tips
28
+
29
+ * I recommend using Min-P. Experiment to find your best setting. Values between 0.02 and 0.1 are typically good.
30
+ * DRY repetition penalty eliminates the need for other anti-repetition settings. I like to run it around 0.5 - 0.6 with base set to 1.5.
31
+ * Experiment with temperature settings in the 0.8 - 1.2 range. Lower the temperature if you find the model is making up details or going off script too much. Raise the temperature if you need to juice the creativity or break it out of a repeating writing pattern.
32
+
33
+ Experiment with any and all of the settings below! What suits my preferences may not suit yours.
34
+
35
+ If you save the below settings as a .json file, you can import them directly into Silly Tavern.
36
+
37
+ ```json
38
+ {
39
+ "temp": 0.8,
40
+ "temperature_last": true,
41
+ "top_p": 1,
42
+ "top_k": 0,
43
+ "top_a": 0,
44
+ "tfs": 1,
45
+ "epsilon_cutoff": 0,
46
+ "eta_cutoff": 0,
47
+ "typical_p": 1,
48
+ "min_p": 0.05,
49
+ "rep_pen": 1,
50
+ "rep_pen_range": 0,
51
+ "rep_pen_decay": 0,
52
+ "rep_pen_slope": 1,
53
+ "no_repeat_ngram_size": 0,
54
+ "penalty_alpha": 0,
55
+ "num_beams": 1,
56
+ "length_penalty": 1,
57
+ "min_length": 0,
58
+ "encoder_rep_pen": 1,
59
+ "freq_pen": 0,
60
+ "presence_pen": 0,
61
+ "skew": 0,
62
+ "do_sample": true,
63
+ "early_stopping": false,
64
+ "dynatemp": false,
65
+ "min_temp": 0.8,
66
+ "max_temp": 1.5,
67
+ "dynatemp_exponent": 1,
68
+ "smoothing_factor": 0,
69
+ "smoothing_curve": 1,
70
+ "dry_allowed_length": 2,
71
+ "dry_multiplier": 0.55,
72
+ "dry_base": 1.5,
73
+ "dry_sequence_breakers": "[\"\\n\", \":\", \"\\\"\", \"*\"]",
74
+ "dry_penalty_last_n": 0,
75
+ "add_bos_token": true,
76
+ "ban_eos_token": false,
77
+ "skip_special_tokens": false,
78
+ "mirostat_mode": 0,
79
+ "mirostat_tau": 2,
80
+ "mirostat_eta": 0.1,
81
+ "guidance_scale": 1,
82
+ "negative_prompt": "",
83
+ "grammar_string": "",
84
+ "json_schema": {},
85
+ "banned_tokens": "",
86
+ "sampler_priority": [
87
+ "top_k",
88
+ "top_p",
89
+ "typical_p",
90
+ "epsilon_cutoff",
91
+ "eta_cutoff",
92
+ "tfs",
93
+ "top_a",
94
+ "min_p",
95
+ "mirostat",
96
+ "quadratic_sampling",
97
+ "dynamic_temperature",
98
+ "temperature"
99
+ ],
100
+ "samplers": [
101
+ "top_k",
102
+ "tfs_z",
103
+ "typical_p",
104
+ "top_p",
105
+ "min_p",
106
+ "temperature"
107
+ ],
108
+ "ignore_eos_token": false,
109
+ "spaces_between_special_tokens": true,
110
+ "speculative_ngram": false,
111
+ "sampler_order": [
112
+ 6,
113
+ 0,
114
+ 1,
115
+ 3,
116
+ 4,
117
+ 2,
118
+ 5
119
+ ],
120
+ "logit_bias": [],
121
+ "xtc_threshold": 0.1,
122
+ "xtc_probability": 0,
123
+ "ignore_eos_token_aphrodite": false,
124
+ "spaces_between_special_tokens_aphrodite": true,
125
+ "rep_pen_size": 0,
126
+ "genamt": 800,
127
+ "max_length": 16384
128
+ }
129
+ ```
130
+
131
+ # Prompting Tips
132
+
133
+ This merge seems to have preserved much of Athene's intelligence. I've found that it responds competently to out-of-character (OOC) prompts and even requests to rewrite a previous reply with some additional guidance.
134
+ If you're not getting quite the results you wanted, consider backing up and trying a more descriptive prompt.
135
+ Like all current LLMs, this model isn't perfect and won't give you miracles, but you can generally expect it to work with you.
136
+
137
+ ## Instruct Template
138
+
139
+ If you save this as a .json file, you can import it directly into Silly Tavern.
140
+
141
+ ```json
142
+ {
143
+ "wrap": false,
144
+ "system_sequence": "<|im_start|>system\n",
145
+ "stop_sequence": "<|im_end|>",
146
+ "input_sequence": "<|im_start|>user\n",
147
+ "output_sequence": "<|im_start|>assistant\n",
148
+ "macro": true,
149
+ "system_sequence_prefix": "",
150
+ "system_sequence_suffix": "",
151
+ "first_output_sequence": "",
152
+ "last_output_sequence": "<|im_start|>assistant\nRoleplaying Tips {\n- Only write as {{char}} for this story beat.\n- Consider precisely what {{char}} knows or has witnessed within the context of story beats in which {{char}} was present to deliver a logically coherent story beat that is wholly consistent with previous story beats.\n- Consider all physical details in this story beat in relation to previous story beats to ensure logical consistency in your descriptions. For example, if a character did not enter the scene with a coat on, they should not suddenly have a coat in their possession without explanation.\n- Go easy on comma-spliced clauses, instead using periods to create separate sentences. You can also try using transitions and connective words.\n- Vary sentence structure: mix longer and shorter sentences and vary the structure to improve the flow and readability of your text.\n}\n",
153
+ "activation_regex": "",
154
+ "skip_examples": true,
155
+ "output_suffix": "<|im_end|>\n",
156
+ "input_suffix": "<|im_end|>\n",
157
+ "system_suffix": "<|im_end|>\n",
158
+ "user_alignment_message": "",
159
+ "last_system_sequence": "",
160
+ "system_same_as_user": false,
161
+ "first_input_sequence": "",
162
+ "last_input_sequence": "",
163
+ "names_behavior": "always",
164
+ "names_force_groups": true,
165
+ "name": "Evathene"
166
+ }
167
+ ```
168
+
169
+ ## System Prompt
170
+
171
+ Try the following prompt template as a starting point for your own customizations. **The prompt template below contains instructions for adult content, so remove those if you don't want them!** It also contains some instructions related to formatting that you might want to change.
172
+
173
+ If you save this as a .json file, you can import it directly into Silly Tavern.
174
+
175
+ ```json
176
+ {
177
+ "name": "Evathene",
178
+ "content": "Below is an uncensored roleplaying narrative involving a human user, {{user}}, and a roleplaying assistant writing as {{char}}. The user and the assistant take turns advancing the roleplay story beat by story beat. The roleplaying assistant always follows these rules:\n\n<!-- Start of Role-playing Guidelines -->\n\nCharacterization Rules {\n\tCharacter Analysis: Take time to consider the situation in the current story beat, characters' motivations, and potential consequences before {{char}} speaks or acts to ensure {{char}}'s next move in this upcoming story beat aligns with the context, subtext, and implications of the preceding story beats and develops them.\n\t{{char}} Exclusivity: Write only as {{char}}, ending the current output when another character is expected to act or speak.\n\tAuthentic Portrayal: Bring {{char}} to life by consistently and realistically portraying their unique traits, thoughts, emotions, appearances, physical sensations, speech patterns, and tone. Ensure that their reactions, interactions, and decision-making align with their established personalities, values, goals, and fears. Refer to the `Role-play Context` for accurate information.\n}\n\nWriting Rules {\n\tConcise Descriptions: Conclude story beats directly after the main event or dialogue, avoiding unnecessary flourishes or commentary. Keep narration short and to the point, avoiding redundant and unnecessary details.\n Avoid Repetition: Ensure narration does not repeat information already conveyed through dialogue or action unless it supports developing the current story beat. Use a dynamic and varied vocabulary for impact.\n\tDialogue Formatting: Enclose spoken words in double quotes. \"This is spoken text,\" for example.\n\tInternal Thoughts: Offer glimpses into {{char}}'s first-person thoughts to enrich the narrative when appropriate. Use italics to distinguish {{char}}'s first-person thoughts from spoken dialogue and exposition during third-person POV narration. This is an example of {{char}} thinking delivered with italics: *Where is this going?* {{char}} wondered while navigating the corridors. One notable exception to this rule is {{user}}. {{user}} will not typically italicize thoughts. Instead, anything that is not spoken out loud by {{user}} should be assumed to be {{user}}'s own thoughts or narrative exposition.\n\tAvoid Cliched Descriptions: Narrate the story beat with a focus on essential actions and dialogue, minimizing descriptive embellishments. Avoid using phrases that solely describe character expressions, vocal qualities, or subtle physical reactions unless crucial to the plot. Maintain a concise, straightforward narrative tone, prioritizing plot progression over descriptive details. Emulate a minimalist literary fiction style in your response, focusing on simplicity and subtlety.\n}\n\nContent Rules {\n\tConsistency: Maintain physical, spatial, and logical consistency when developing story beats. Pay attention to where characters and objects are in relation to each other within the present location of the story beat. Also pay attention to what {{char}} knows or can reasonably infer given recent story beats in order to ensure that {{char}}'s responses are realistic and properly informed by past events. Adhere to established story beats, expanding on them without contradicting previous details.\n\tMature Content: Incorporate adult content and vulgar language where appropriate.\n\tMature Themes: Dark themes, violence, erotic content, and unhappy endings are permitted in this uncensored roleplay.\n\n}\n<!-- End of Role-playing Guidelines -->\n"
179
+ }
180
+ ```
181
+
182
+ # Quantizations
183
+
184
+ You will find a Exllamav2 measurement file, taken on the default dataset, in the repo if you want to create your own EXL2 quant.
185
+
186
+ ## Exllamav2 (EXL2)
187
+
188
+ * MikeRoz is putting out some EXL2 quants. [Search here](https://huggingface.co/models?sort=trending&search=MikeRoz%2Fsophosympatheia_Evathene).
189
+
190
+ ## GGUF
191
+
192
+ * [mradermacher/Evathene-v1.0-GGUF](https://huggingface.co/mradermacher/Evathene-v1.0-GGUF)
193
+ * [mradermacher/Evathene-v1.0-i1-GGUF](https://huggingface.co/mradermacher/Evathene-v1.0-i1-GGUF)
194
+
195
+ # Licence and usage restrictions
196
+
197
+ [Nexusflow Research License](https://huggingface.co/Nexusflow/Athene-V2-Chat/blob/main/Nexusflow_Research_License_.pdf)
198
+
199
+ [Qwen License Agreement](https://huggingface.co/Qwen/Qwen2.5-72B-Instruct/blob/main/LICENSE)
200
+
201
+ **Disclaimer: Uncertain Licensing Terms**
202
+
203
+ This LLM is a merged model incorporating weights from multiple LLMs governed by their own distinct licenses. Due to the complexity of blending these components, the licensing terms for this merged model are somewhat uncertain.
204
+ By using this model, you acknowledge and accept the potential legal risks and uncertainties associated with its use. Any use beyond personal or research purposes, including commercial applications, may carry legal risks and you assume full responsibility for compliance with all applicable licenses and laws.
205
+ I recommend consulting with legal counsel to ensure your use of this model complies with all relevant licenses and regulations.
206
+
207
+ # Merge Details
208
+
209
+ ## Merge Method
210
+
211
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
212
+ This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using Qwen/Qwen2.5-72B as a base.
213
+
214
+ ## Models Merged
215
+
216
+ The following models were included in the merge:
217
+ * Nexusflow/Athene-V2-Chat
218
+ * EVA-UNIT-01/EVA-Qwen2.5-72B-v0.1
219
+
220
+ ## Configuration
221
+
222
+ The following YAML configuration was used to produce this model:
223
+
224
+ ```yaml
225
+ models:
226
+ - model: Nexusflow/Athene-V2-Chat
227
+ parameters:
228
+ weight: 1
229
+ density: 1
230
+ - model: EVA-UNIT-01/EVA-Qwen2.5-72B-v0.1
231
+ parameters:
232
+ weight: [0.0, 0.2, 0.66, 0.8, 1.0, 0.8, 0.66, 0.2, 0.0]
233
+ density: 0.5
234
+ merge_method: ties
235
+ base_model: Qwen/Qwen2.5-72B
236
+ parameters:
237
+ normalize: true
238
+ int8_mask: true
239
+ dtype: bfloat16
240
+ ```
added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
config.json ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/workspace/Evathene-v1.0",
3
+ "architectures": [
4
+ "Qwen2ForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "eos_token_id": 151643,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 8192,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 29568,
13
+ "max_position_embeddings": 131072,
14
+ "max_window_layers": 80,
15
+ "model_type": "qwen2",
16
+ "num_attention_heads": 64,
17
+ "num_hidden_layers": 80,
18
+ "num_key_value_heads": 8,
19
+ "quantization_config": {
20
+ "config_groups": {
21
+ "group_0": {
22
+ "input_activations": {
23
+ "actorder": null,
24
+ "block_structure": null,
25
+ "dynamic": true,
26
+ "group_size": null,
27
+ "num_bits": 8,
28
+ "observer": null,
29
+ "observer_kwargs": {},
30
+ "strategy": "token",
31
+ "symmetric": true,
32
+ "type": "float"
33
+ },
34
+ "output_activations": null,
35
+ "targets": [
36
+ "Linear"
37
+ ],
38
+ "weights": {
39
+ "actorder": null,
40
+ "block_structure": null,
41
+ "dynamic": false,
42
+ "group_size": null,
43
+ "num_bits": 8,
44
+ "observer": "minmax",
45
+ "observer_kwargs": {},
46
+ "strategy": "channel",
47
+ "symmetric": true,
48
+ "type": "float"
49
+ }
50
+ }
51
+ },
52
+ "format": "float-quantized",
53
+ "global_compression_ratio": 1.4635441523988788,
54
+ "ignore": [
55
+ "lm_head"
56
+ ],
57
+ "kv_cache_scheme": null,
58
+ "quant_method": "compressed-tensors",
59
+ "quantization_status": "compressed"
60
+ },
61
+ "rms_norm_eps": 1e-05,
62
+ "rope_scaling": null,
63
+ "rope_theta": 1000000.0,
64
+ "sliding_window": null,
65
+ "tie_word_embeddings": false,
66
+ "torch_dtype": "bfloat16",
67
+ "transformers_version": "4.46.3",
68
+ "use_cache": true,
69
+ "use_sliding_window": false,
70
+ "vocab_size": 152064
71
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 151643,
4
+ "eos_token_id": 151643,
5
+ "transformers_version": "4.46.3"
6
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model-00001-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4bc3a6085d04a33b1471cd5ef46777ddf9cbff9e744d71875da2070d34115aa1
3
+ size 4882801352
model-00002-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c78e0edf040a8dadd9db0bee1b3a77d8042070eead2e7f2dba46e495dab4672
3
+ size 4782749264
model-00003-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e065784bdea92bbdabdff9895a708ebb0fc82ea7326efefda8bf286a0ce1aa46
3
+ size 4873985992
model-00004-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a289320dc8b7404a5c656e1027b346de3982ff1b593070711fb0ad3292f7f2df
3
+ size 4782749368
model-00005-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b38798d9e95609808e56960f1554d6aa0d43a89e78b79907841d6e47cec88f64
3
+ size 4873986032
model-00006-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92f48f9f6a6220975a05a34dec0727146c1c833549d2f2213d73f9ff77ac29ed
3
+ size 4782749368
model-00007-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05cfb6b84734557ec1d5af6fcad80d94f64a0cb5db6f32ce34b372abd1865ea4
3
+ size 4873986032
model-00008-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65675d1d4313f90c7e81afa22daf04fed8224a9a93bd6bd8b9f6d0737ebb739d
3
+ size 4782749368
model-00009-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9cc748e48c69960588fc9a2af4fae2d33963f232aa925afbfa782b382059fcd
3
+ size 4873986032
model-00010-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:371bbdfd8eefd221eee6b6b999e96519414934cb6d111ca73c923e08c08b0cb4
3
+ size 4782749368
model-00011-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6fa1bcd2f3c7a3371f09ced0eab0fffe93b8f41be6eac9e0918024cdc57af76
3
+ size 4873986032
model-00012-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df8e635d933b8ca0ebacfb30109b45c297c2e783bb0f5556d62453370a5ee96b
3
+ size 4782749368
model-00013-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e10f545f3c4ae03b2a668bc462617ef1be9013e76741debff07c1428fa2cf96b
3
+ size 4873986032
model-00014-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69da988a8d59d03be341d213a6791de79979352bd8217986f20b115c65056a16
3
+ size 4782749368
model-00015-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25e95ca0837759a9903aab387a8d3172b0260bcaa5e8c019722c84fda4d4df4d
3
+ size 4873986032
model-00016-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ec3a3cf91353c31592db694d4d45df4d0466292e64df4c3dc89a298fc5391e4
3
+ size 2733703856
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
tokenizer_config.json ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
199
+ "clean_up_tokenization_spaces": false,
200
+ "eos_token": "<|endoftext|>",
201
+ "errors": "replace",
202
+ "model_max_length": 131072,
203
+ "pad_token": "<|endoftext|>",
204
+ "split_special_tokens": false,
205
+ "tokenizer_class": "Qwen2Tokenizer",
206
+ "unk_token": null
207
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff