Upload tokenizer
Browse files- tokenizer.json +8 -8
- tokenizer_config.json +4 -4
tokenizer.json
CHANGED
@@ -77,7 +77,7 @@
|
|
77 |
},
|
78 |
{
|
79 |
"id": 8,
|
80 |
-
"content": "
|
81 |
"single_word": false,
|
82 |
"lstrip": false,
|
83 |
"rstrip": false,
|
@@ -86,7 +86,7 @@
|
|
86 |
},
|
87 |
{
|
88 |
"id": 9,
|
89 |
-
"content": "
|
90 |
"single_word": false,
|
91 |
"lstrip": false,
|
92 |
"rstrip": false,
|
@@ -95,7 +95,7 @@
|
|
95 |
},
|
96 |
{
|
97 |
"id": 10,
|
98 |
-
"content": "
|
99 |
"single_word": false,
|
100 |
"lstrip": false,
|
101 |
"rstrip": false,
|
@@ -104,7 +104,7 @@
|
|
104 |
},
|
105 |
{
|
106 |
"id": 11,
|
107 |
-
"content": "
|
108 |
"single_word": false,
|
109 |
"lstrip": false,
|
110 |
"rstrip": false,
|
@@ -2417,10 +2417,10 @@
|
|
2417 |
"<|eom_id|>": 5,
|
2418 |
"<|eot_id|>": 6,
|
2419 |
"<|python_tag|>": 7,
|
2420 |
-
"
|
2421 |
-
"
|
2422 |
-
"
|
2423 |
-
"
|
2424 |
"<|reserved_special_token_0|>": 12,
|
2425 |
"<|reserved_special_token_1|>": 13,
|
2426 |
"<|reserved_special_token_2|>": 14,
|
|
|
77 |
},
|
78 |
{
|
79 |
"id": 8,
|
80 |
+
"content": "<|begin_of_thought|>",
|
81 |
"single_word": false,
|
82 |
"lstrip": false,
|
83 |
"rstrip": false,
|
|
|
86 |
},
|
87 |
{
|
88 |
"id": 9,
|
89 |
+
"content": "<|end_of_thought|>",
|
90 |
"single_word": false,
|
91 |
"lstrip": false,
|
92 |
"rstrip": false,
|
|
|
95 |
},
|
96 |
{
|
97 |
"id": 10,
|
98 |
+
"content": "<|begin_of_solution|>",
|
99 |
"single_word": false,
|
100 |
"lstrip": false,
|
101 |
"rstrip": false,
|
|
|
104 |
},
|
105 |
{
|
106 |
"id": 11,
|
107 |
+
"content": "<|end_of_solution|>",
|
108 |
"single_word": false,
|
109 |
"lstrip": false,
|
110 |
"rstrip": false,
|
|
|
2417 |
"<|eom_id|>": 5,
|
2418 |
"<|eot_id|>": 6,
|
2419 |
"<|python_tag|>": 7,
|
2420 |
+
"<|begin_of_thought|>": 8,
|
2421 |
+
"<|end_of_thought|>": 9,
|
2422 |
+
"<|begin_of_solution|>": 10,
|
2423 |
+
"<|end_of_solution|>": 11,
|
2424 |
"<|reserved_special_token_0|>": 12,
|
2425 |
"<|reserved_special_token_1|>": 13,
|
2426 |
"<|reserved_special_token_2|>": 14,
|
tokenizer_config.json
CHANGED
@@ -65,7 +65,7 @@
|
|
65 |
"special": true
|
66 |
},
|
67 |
"8": {
|
68 |
-
"content": "
|
69 |
"lstrip": false,
|
70 |
"normalized": false,
|
71 |
"rstrip": false,
|
@@ -73,7 +73,7 @@
|
|
73 |
"special": true
|
74 |
},
|
75 |
"9": {
|
76 |
-
"content": "
|
77 |
"lstrip": false,
|
78 |
"normalized": false,
|
79 |
"rstrip": false,
|
@@ -81,7 +81,7 @@
|
|
81 |
"special": true
|
82 |
},
|
83 |
"10": {
|
84 |
-
"content": "
|
85 |
"lstrip": false,
|
86 |
"normalized": false,
|
87 |
"rstrip": false,
|
@@ -89,7 +89,7 @@
|
|
89 |
"special": true
|
90 |
},
|
91 |
"11": {
|
92 |
-
"content": "
|
93 |
"lstrip": false,
|
94 |
"normalized": false,
|
95 |
"rstrip": false,
|
|
|
65 |
"special": true
|
66 |
},
|
67 |
"8": {
|
68 |
+
"content": "<|begin_of_thought|>",
|
69 |
"lstrip": false,
|
70 |
"normalized": false,
|
71 |
"rstrip": false,
|
|
|
73 |
"special": true
|
74 |
},
|
75 |
"9": {
|
76 |
+
"content": "<|end_of_thought|>",
|
77 |
"lstrip": false,
|
78 |
"normalized": false,
|
79 |
"rstrip": false,
|
|
|
81 |
"special": true
|
82 |
},
|
83 |
"10": {
|
84 |
+
"content": "<|begin_of_solution|>",
|
85 |
"lstrip": false,
|
86 |
"normalized": false,
|
87 |
"rstrip": false,
|
|
|
89 |
"special": true
|
90 |
},
|
91 |
"11": {
|
92 |
+
"content": "<|end_of_solution|>",
|
93 |
"lstrip": false,
|
94 |
"normalized": false,
|
95 |
"rstrip": false,
|