Upload tokenizer
Browse files- README.md +4 -4
- tokenizer.json +2 -2
README.md
CHANGED
@@ -1,14 +1,14 @@
|
|
1 |
---
|
2 |
-
library_name: transformers
|
3 |
-
license: mit
|
4 |
base_model: microsoft/mdeberta-v3-base
|
5 |
-
tags:
|
6 |
-
- generated_from_trainer
|
7 |
datasets:
|
8 |
- tweet_sentiment_multilingual
|
|
|
|
|
9 |
metrics:
|
10 |
- accuracy
|
11 |
- f1
|
|
|
|
|
12 |
model-index:
|
13 |
- name: scenario-NON-KD-PR-COPY-CDF-ALL-D2_data-cardiffnlp_tweet_sentiment_multilingual_
|
14 |
results: []
|
|
|
1 |
---
|
|
|
|
|
2 |
base_model: microsoft/mdeberta-v3-base
|
|
|
|
|
3 |
datasets:
|
4 |
- tweet_sentiment_multilingual
|
5 |
+
library_name: transformers
|
6 |
+
license: mit
|
7 |
metrics:
|
8 |
- accuracy
|
9 |
- f1
|
10 |
+
tags:
|
11 |
+
- generated_from_trainer
|
12 |
model-index:
|
13 |
- name: scenario-NON-KD-PR-COPY-CDF-ALL-D2_data-cardiffnlp_tweet_sentiment_multilingual_
|
14 |
results: []
|
tokenizer.json
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9ddbe35b768b22d6cd0d61a0a88ea3a188b7c00bbb7b825f9f247cf0b79c2365
|
3 |
+
size 16350764
|