valoomba commited on
Commit
08c81e7
·
verified ·
1 Parent(s): 1f4c43f

Update config

Browse files
README.md CHANGED
@@ -1,8 +1,8 @@
1
  ---
2
  base_model:
3
- - rombodawg/Rombos-Coder-V2.5-Qwen-32b
4
  - FuseAI/FuseO1-DeepSeekR1-Qwen2.5-Coder-32B-Preview
5
  - Qwen/Qwen2.5-Coder-32B
 
6
  library_name: transformers
7
  tags:
8
  - mergekit
@@ -21,23 +21,25 @@ This model was merged using the sce merge method using [Qwen/Qwen2.5-Coder-32B](
21
  ### Models Merged
22
 
23
  The following models were included in the merge:
24
- * [rombodawg/Rombos-Coder-V2.5-Qwen-32b](https://huggingface.co/rombodawg/Rombos-Coder-V2.5-Qwen-32b)
25
  * [FuseAI/FuseO1-DeepSeekR1-Qwen2.5-Coder-32B-Preview](https://huggingface.co/FuseAI/FuseO1-DeepSeekR1-Qwen2.5-Coder-32B-Preview)
 
26
 
27
  ### Configuration
28
 
29
  The following YAML configuration was used to produce this model:
30
 
31
  ```yaml
32
- models:
33
- - model: rombodawg/Rombos-Coder-V2.5-Qwen-32b
34
- - model: FuseAI/FuseO1-DeepSeekR1-Qwen2.5-Coder-32B-Preview
35
-
36
- merge_method: sce
37
  base_model: Qwen/Qwen2.5-Coder-32B
38
-
39
- parameters:
40
- select_topk: 1.0 # or whatever fraction/threshold you prefer
41
  dtype: bfloat16
42
-
 
 
 
 
 
 
 
 
 
 
43
  ```
 
1
  ---
2
  base_model:
 
3
  - FuseAI/FuseO1-DeepSeekR1-Qwen2.5-Coder-32B-Preview
4
  - Qwen/Qwen2.5-Coder-32B
5
+ - rombodawg/Rombos-Coder-V2.5-Qwen-32b
6
  library_name: transformers
7
  tags:
8
  - mergekit
 
21
  ### Models Merged
22
 
23
  The following models were included in the merge:
 
24
  * [FuseAI/FuseO1-DeepSeekR1-Qwen2.5-Coder-32B-Preview](https://huggingface.co/FuseAI/FuseO1-DeepSeekR1-Qwen2.5-Coder-32B-Preview)
25
+ * [rombodawg/Rombos-Coder-V2.5-Qwen-32b](https://huggingface.co/rombodawg/Rombos-Coder-V2.5-Qwen-32b)
26
 
27
  ### Configuration
28
 
29
  The following YAML configuration was used to produce this model:
30
 
31
  ```yaml
 
 
 
 
 
32
  base_model: Qwen/Qwen2.5-Coder-32B
 
 
 
33
  dtype: bfloat16
34
+ merge_method: sce
35
+ parameters:
36
+ select_topk: 1.0
37
+ slices:
38
+ - sources:
39
+ - layer_range: [0, 64]
40
+ model: rombodawg/Rombos-Coder-V2.5-Qwen-32b
41
+ - layer_range: [0, 64]
42
+ model: FuseAI/FuseO1-DeepSeekR1-Qwen2.5-Coder-32B-Preview
43
+ - layer_range: [0, 64]
44
+ model: Qwen/Qwen2.5-Coder-32B
45
  ```
config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "_name_or_path": "Qwen/Qwen2.5-Coder-32B",
3
  "architectures": [
4
  "Qwen2ForCausalLM"
5
  ],
@@ -10,19 +9,18 @@
10
  "hidden_size": 5120,
11
  "initializer_range": 0.02,
12
  "intermediate_size": 27648,
13
- "max_position_embeddings": 32768,
14
  "max_window_layers": 64,
15
  "model_type": "qwen2",
16
  "num_attention_heads": 40,
17
  "num_hidden_layers": 64,
18
  "num_key_value_heads": 8,
19
- "rms_norm_eps": 1e-06,
20
- "rope_scaling": null,
21
  "rope_theta": 1000000.0,
22
- "sliding_window": null,
23
  "tie_word_embeddings": false,
24
  "torch_dtype": "bfloat16",
25
- "transformers_version": "4.48.2",
26
  "use_cache": true,
27
  "use_sliding_window": false,
28
  "vocab_size": 152064
 
1
  {
 
2
  "architectures": [
3
  "Qwen2ForCausalLM"
4
  ],
 
9
  "hidden_size": 5120,
10
  "initializer_range": 0.02,
11
  "intermediate_size": 27648,
12
+ "max_position_embeddings": 131072,
13
  "max_window_layers": 64,
14
  "model_type": "qwen2",
15
  "num_attention_heads": 40,
16
  "num_hidden_layers": 64,
17
  "num_key_value_heads": 8,
18
+ "rms_norm_eps": 1e-05,
 
19
  "rope_theta": 1000000.0,
20
+ "sliding_window": 131072,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
+ "transformers_version": "4.43.1",
24
  "use_cache": true,
25
  "use_sliding_window": false,
26
  "vocab_size": 152064
fusechat-sce.yml CHANGED
@@ -1,10 +1,13 @@
1
- models:
2
- - model: rombodawg/Rombos-Coder-V2.5-Qwen-32b
3
- - model: FuseAI/FuseO1-DeepSeekR1-Qwen2.5-Coder-32B-Preview
4
-
5
- merge_method: sce
6
  base_model: Qwen/Qwen2.5-Coder-32B
7
-
8
- parameters:
9
- select_topk: 1.0 # or whatever fraction/threshold you prefer
10
  dtype: bfloat16
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  base_model: Qwen/Qwen2.5-Coder-32B
 
 
 
2
  dtype: bfloat16
3
+ merge_method: sce
4
+ parameters:
5
+ select_topk: 1.0
6
+ slices:
7
+ - sources:
8
+ - layer_range: [0, 64]
9
+ model: rombodawg/Rombos-Coder-V2.5-Qwen-32b
10
+ - layer_range: [0, 64]
11
+ model: FuseAI/FuseO1-DeepSeekR1-Qwen2.5-Coder-32B-Preview
12
+ - layer_range: [0, 64]
13
+ model: Qwen/Qwen2.5-Coder-32B
generation_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 151646,
4
+ "eos_token_id": 151643,
5
+ "do_sample": true,
6
+ "temperature": 0.6,
7
+ "top_p": 0.95,
8
+ "transformers_version": "4.39.3"
9
+ }
model-00001-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5ba1c23eb69a3f3b3487b2bc6b87e4427dc160bcd4ee61fe3c114a5e4010de17
3
  size 4949338472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:881abc282596b0d4ed41c9b68fc30474d8cee6d1ebcb9545b38633d946b3fe82
3
  size 4949338472
model-00002-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:80f5bd38e7a52453bac081b7fa50b250d9d1b8a524028e48725015f681347550
3
  size 4991425976
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:399ef3f46a0eb39ee8964eb2ec89132b15a5ac0a3fbd1af66a09ed911bf0d41f
3
  size 4991425976
model-00003-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fbba1dfc5c20e5ee9e2a34f23da9964823142f9dcb6f15a73d410a1ca389c6da
3
  size 4876059408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:daa6fcba2548a945e10b52baff8b6ec5b6e1fba8ad5fe0acab2a7d1276c26c2c
3
  size 4876059408
model-00004-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:22ddefaf972d558135caf9c317d3e6b4a3a302f0dbcd5c56bd1ffa7b6f3ee665
3
  size 4876059400
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c922a2395c91941d46c84ce228e71b91bbb8044bcd65ffd0bdceeb191541e904
3
  size 4876059400
model-00005-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7bb27d554988b382131979fe3c2f05eccfa56525b0bda1a5451f2273ae079797
3
  size 4876059408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fde69d45e305fa8b02d9b28b1a131e1e02e8ce832d8c3d23f5c5b1e562aa7b83
3
  size 4876059408
model-00006-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3b2e19e490b9fe03807de588351668fa89196c15d56dda5fe2288fb77a4317e8
3
  size 4876059400
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebf633d81549a16c363ea1515a0f763232e35bdb665013cd441686ac447be788
3
  size 4876059400
model-00007-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:75fdff87f86d1ccf2e9e17a3b8af002692403800ff4da99b596e8d91758e3e91
3
  size 4876059408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56a06b6b218e1be01cae4c63d4cd0103ae9f7ee78b3a580407f6a2192e6d988b
3
  size 4876059408
model-00008-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c9c78d51ce56fe6386dbbcfd838a4049c6bcef6e767ebf6d503b93889682dc6e
3
  size 4876059400
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92204531cbc720f14e6391a871e90f41c508bfd4479cde3971224e2677310729
3
  size 4876059400
model-00009-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a2b2f9380c2249c19a79ccbbb77d993baa3142bcfc5159461346674b8ad8270d
3
  size 4876059408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a16a134bed3ea2ad8b43872f030a586f16a7a7d2f902cb28c850668bf1e2bc70
3
  size 4876059408
model-00010-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:403a5493e6c018ded33fe8812707be69bdc91da6fa69cffd5f60050f66dd406b
3
  size 4876059400
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4bc50c47244c7c74bc35ceed793673fceb1533c8e22dcf80e743ec913c95771f
3
  size 4876059400
model-00011-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:86da186ef5263051a0af2fcd9458b40b96a22248b788c56ad20e4d5f526e5d19
3
  size 4876059408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09c7b3fdd1dceeb950b4406a17d48308490d0c5e794dc62dab9a0bf5f45dd3d5
3
  size 4876059408
model-00012-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ce3ce5ce33cc9fad3b5ce1b9a4a21d28bb1f3775b4818705620f62260813a205
3
  size 4876059400
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7916d74ffd4f082bfc21bef028db82838e6f75cc44ca10bc729158346365f2a0
3
  size 4876059400
model-00013-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e4831189f6c08af1625184d09e7f9e885b1248ba5dfd2b6d1940bea2f4328ebf
3
  size 4876059400
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae2b8a869964fcf6281f68f98dc7820924645fbc369a5ebdf1101218b7d8a575
3
  size 4876059400
model-00014-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8cac904ae88e9958969d285bca1967b5e5fc408160aef401c5aa8e11bcddeaf4
3
  size 1950423728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0290ac48f7f0be2b9365ba34efd21d2e6ae160b604ad197421b8ace8650695de
3
  size 1950423728
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
- size 11421896
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88145e3c3249adc2546ede277e9819d6e405e19072456e4b521cbc724bd60773
3
+ size 7031660
tokenizer_config.json CHANGED
@@ -1,208 +1,35 @@
1
  {
2
  "add_bos_token": false,
3
- "add_prefix_space": false,
4
- "added_tokens_decoder": {
5
- "151643": {
6
- "content": "<|endoftext|>",
7
- "lstrip": false,
8
- "normalized": false,
9
- "rstrip": false,
10
- "single_word": false,
11
- "special": true
12
- },
13
- "151644": {
14
- "content": "<|im_start|>",
15
- "lstrip": false,
16
- "normalized": false,
17
- "rstrip": false,
18
- "single_word": false,
19
- "special": true
20
- },
21
- "151645": {
22
- "content": "<|im_end|>",
23
- "lstrip": false,
24
- "normalized": false,
25
- "rstrip": false,
26
- "single_word": false,
27
- "special": true
28
- },
29
- "151646": {
30
- "content": "<|object_ref_start|>",
31
- "lstrip": false,
32
- "normalized": false,
33
- "rstrip": false,
34
- "single_word": false,
35
- "special": true
36
- },
37
- "151647": {
38
- "content": "<|object_ref_end|>",
39
- "lstrip": false,
40
- "normalized": false,
41
- "rstrip": false,
42
- "single_word": false,
43
- "special": true
44
- },
45
- "151648": {
46
- "content": "<|box_start|>",
47
- "lstrip": false,
48
- "normalized": false,
49
- "rstrip": false,
50
- "single_word": false,
51
- "special": true
52
- },
53
- "151649": {
54
- "content": "<|box_end|>",
55
- "lstrip": false,
56
- "normalized": false,
57
- "rstrip": false,
58
- "single_word": false,
59
- "special": true
60
- },
61
- "151650": {
62
- "content": "<|quad_start|>",
63
- "lstrip": false,
64
- "normalized": false,
65
- "rstrip": false,
66
- "single_word": false,
67
- "special": true
68
- },
69
- "151651": {
70
- "content": "<|quad_end|>",
71
- "lstrip": false,
72
- "normalized": false,
73
- "rstrip": false,
74
- "single_word": false,
75
- "special": true
76
- },
77
- "151652": {
78
- "content": "<|vision_start|>",
79
- "lstrip": false,
80
- "normalized": false,
81
- "rstrip": false,
82
- "single_word": false,
83
- "special": true
84
- },
85
- "151653": {
86
- "content": "<|vision_end|>",
87
- "lstrip": false,
88
- "normalized": false,
89
- "rstrip": false,
90
- "single_word": false,
91
- "special": true
92
- },
93
- "151654": {
94
- "content": "<|vision_pad|>",
95
- "lstrip": false,
96
- "normalized": false,
97
- "rstrip": false,
98
- "single_word": false,
99
- "special": true
100
- },
101
- "151655": {
102
- "content": "<|image_pad|>",
103
- "lstrip": false,
104
- "normalized": false,
105
- "rstrip": false,
106
- "single_word": false,
107
- "special": true
108
- },
109
- "151656": {
110
- "content": "<|video_pad|>",
111
- "lstrip": false,
112
- "normalized": false,
113
- "rstrip": false,
114
- "single_word": false,
115
- "special": true
116
- },
117
- "151657": {
118
- "content": "<tool_call>",
119
- "lstrip": false,
120
- "normalized": false,
121
- "rstrip": false,
122
- "single_word": false,
123
- "special": false
124
- },
125
- "151658": {
126
- "content": "</tool_call>",
127
- "lstrip": false,
128
- "normalized": false,
129
- "rstrip": false,
130
- "single_word": false,
131
- "special": false
132
- },
133
- "151659": {
134
- "content": "<|fim_prefix|>",
135
- "lstrip": false,
136
- "normalized": false,
137
- "rstrip": false,
138
- "single_word": false,
139
- "special": false
140
- },
141
- "151660": {
142
- "content": "<|fim_middle|>",
143
- "lstrip": false,
144
- "normalized": false,
145
- "rstrip": false,
146
- "single_word": false,
147
- "special": false
148
- },
149
- "151661": {
150
- "content": "<|fim_suffix|>",
151
- "lstrip": false,
152
- "normalized": false,
153
- "rstrip": false,
154
- "single_word": false,
155
- "special": false
156
- },
157
- "151662": {
158
- "content": "<|fim_pad|>",
159
- "lstrip": false,
160
- "normalized": false,
161
- "rstrip": false,
162
- "single_word": false,
163
- "special": false
164
- },
165
- "151663": {
166
- "content": "<|repo_name|>",
167
- "lstrip": false,
168
- "normalized": false,
169
- "rstrip": false,
170
- "single_word": false,
171
- "special": false
172
- },
173
- "151664": {
174
- "content": "<|file_sep|>",
175
- "lstrip": false,
176
- "normalized": false,
177
- "rstrip": false,
178
- "single_word": false,
179
- "special": false
180
- }
181
  },
182
- "additional_special_tokens": [
183
- "<|im_start|>",
184
- "<|im_end|>",
185
- "<|object_ref_start|>",
186
- "<|object_ref_end|>",
187
- "<|box_start|>",
188
- "<|box_end|>",
189
- "<|quad_start|>",
190
- "<|quad_end|>",
191
- "<|vision_start|>",
192
- "<|vision_end|>",
193
- "<|vision_pad|>",
194
- "<|image_pad|>",
195
- "<|video_pad|>"
196
- ],
197
- "bos_token": null,
198
- "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
199
  "clean_up_tokenization_spaces": false,
200
- "eos_token": "<|endoftext|>",
201
- "errors": "replace",
202
- "extra_special_tokens": {},
203
- "model_max_length": 32768,
204
- "pad_token": "<|endoftext|>",
205
- "split_special_tokens": false,
206
- "tokenizer_class": "Qwen2Tokenizer",
207
- "unk_token": null
 
 
 
 
 
 
 
 
 
 
 
 
 
 
208
  }
 
1
  {
2
  "add_bos_token": false,
3
+ "add_eos_token": false,
4
+ "bos_token": {
5
+ "__type": "AddedToken",
6
+ "content": "<|begin▁of▁sentence|>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
  "clean_up_tokenization_spaces": false,
13
+ "eos_token": {
14
+ "__type": "AddedToken",
15
+ "content": "<|end▁of▁sentence|>",
16
+ "lstrip": false,
17
+ "normalized": true,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "legacy": true,
22
+ "model_max_length": 16384,
23
+ "pad_token": {
24
+ "__type": "AddedToken",
25
+ "content": "<|end▁of▁sentence|>",
26
+ "lstrip": false,
27
+ "normalized": true,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ },
31
+ "sp_model_kwargs": {},
32
+ "unk_token": null,
33
+ "tokenizer_class": "LlamaTokenizerFast",
34
+ "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set ns = namespace(is_first=false, is_tool=false, is_output_first=true, system_prompt='') %}{%- for message in messages %}{%- if message['role'] == 'system' %}{% set ns.system_prompt = message['content'] %}{%- endif %}{%- endfor %}{{bos_token}}{{ns.system_prompt}}{%- for message in messages %}{%- if message['role'] == 'user' %}{%- set ns.is_tool = false -%}{{'<|User|>' + message['content']}}{%- endif %}{%- if message['role'] == 'assistant' and message['content'] is none %}{%- set ns.is_tool = false -%}{%- for tool in message['tool_calls']%}{%- if not ns.is_first %}{{'<|Assistant|><|tool▁calls▁begin|><|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '\\n' + '```json' + '\\n' + tool['function']['arguments'] + '\\n' + '```' + '<|tool▁call▁end|>'}}{%- set ns.is_first = true -%}{%- else %}{{'\\n' + '<|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '\\n' + '```json' + '\\n' + tool['function']['arguments'] + '\\n' + '```' + '<|tool▁call▁end|>'}}{{'<|tool▁calls▁end|><|end▁of▁sentence|>'}}{%- endif %}{%- endfor %}{%- endif %}{%- if message['role'] == 'assistant' and message['content'] is not none %}{%- if ns.is_tool %}{{'<|tool▁outputs▁end|>' + message['content'] + '<|end▁of▁sentence|>'}}{%- set ns.is_tool = false -%}{%- else %}{% set content = message['content'] %}{% if '</think>' in content %}{% set content = content.split('</think>')[-1] %}{% endif %}{{'<|Assistant|>' + content + '<|end▁of▁sentence|>'}}{%- endif %}{%- endif %}{%- if message['role'] == 'tool' %}{%- set ns.is_tool = true -%}{%- if ns.is_output_first %}{{'<|tool▁outputs▁begin|><|tool▁output▁begin|>' + message['content'] + '<|tool▁output▁end|>'}}{%- set ns.is_output_first = false %}{%- else %}{{'\\n<|tool▁output▁begin|>' + message['content'] + '<|tool▁output▁end|>'}}{%- endif %}{%- endif %}{%- endfor -%}{% if ns.is_tool %}{{'<|tool▁outputs▁end|>'}}{% endif %}{% if add_generation_prompt and not ns.is_tool %}{{'<|Assistant|>'}}{% endif %}"
35
  }