Spaces:
Running
Running
add xlm-roberta
Browse files- vocab/__init__.py +2 -0
- vocab/xlm_roberta/__init__.py +3 -0
vocab/__init__.py
CHANGED
@@ -116,6 +116,7 @@ all_tokenizers = [
|
|
116 |
("falcon_180b", "", "sentencepiece"),
|
117 |
("skywork_13b_base",),
|
118 |
("skywork_13b_math",),
|
|
|
119 |
# "goat",
|
120 |
|
121 |
# ##### glm系列
|
@@ -140,6 +141,7 @@ all_tokenizers = [
|
|
140 |
("mistral_7b",),
|
141 |
("mixtral_8_7b",),
|
142 |
|
|
|
143 |
("flan_t5_base",),
|
144 |
("fastchat_t5_3b",),
|
145 |
("pko_t5_large",),
|
|
|
116 |
("falcon_180b", "", "sentencepiece"),
|
117 |
("skywork_13b_base",),
|
118 |
("skywork_13b_math",),
|
119 |
+
("xlm_roberta", ),
|
120 |
# "goat",
|
121 |
|
122 |
# ##### glm系列
|
|
|
141 |
("mistral_7b",),
|
142 |
("mixtral_8_7b",),
|
143 |
|
144 |
+
|
145 |
("flan_t5_base",),
|
146 |
("fastchat_t5_3b",),
|
147 |
("pko_t5_large",),
|
vocab/xlm_roberta/__init__.py
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
from transformers import AutoTokenizer
|
2 |
+
|
3 |
+
tokenizer = AutoTokenizer.from_pretrained("FacebookAI/xlm-roberta-base", trust_remote_code=True)
|