Update README.md
Browse files
README.md
CHANGED
@@ -3,11 +3,16 @@ license: mit
|
|
3 |
base_model: xlm-roberta-base
|
4 |
tags:
|
5 |
- generated_from_trainer
|
|
|
|
|
6 |
metrics:
|
7 |
- f1
|
8 |
model-index:
|
9 |
-
- name: xlm-roberta-base-finetuned-
|
10 |
results: []
|
|
|
|
|
|
|
11 |
---
|
12 |
|
13 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
@@ -22,16 +27,18 @@ It achieves the following results on the evaluation set:
|
|
22 |
|
23 |
## Model description
|
24 |
|
25 |
-
|
26 |
|
27 |
-
## Intended uses
|
|
|
28 |
|
29 |
-
More information needed
|
30 |
|
31 |
-
##
|
32 |
|
33 |
-
|
|
|
34 |
|
|
|
35 |
## Training procedure
|
36 |
|
37 |
### Training hyperparameters
|
|
|
3 |
base_model: xlm-roberta-base
|
4 |
tags:
|
5 |
- generated_from_trainer
|
6 |
+
- NERz
|
7 |
+
- crypto
|
8 |
metrics:
|
9 |
- f1
|
10 |
model-index:
|
11 |
+
- name: xlm-roberta-base-finetuned-ner-crypto
|
12 |
results: []
|
13 |
+
widget:
|
14 |
+
- text: "Didn't I tell you that that was a decent entry point on $PROPHET? If you are in - congrats, Prophet is up 90% in the last 2 weeks and 50% up in the last week alone"
|
15 |
+
|
16 |
---
|
17 |
|
18 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
|
27 |
|
28 |
## Model description
|
29 |
|
30 |
+
This model is a fine-tuned version of xlm-roberta-base, specializing in Named Entity Recognition (NER) within the cryptocurrency domain. It is optimized to recognize and classify entities such as cryptocurrency ticker symbols, names, and addresses within text.
|
31 |
|
32 |
+
## Intended uses
|
33 |
+
Designed primarily for NER tasks in the cryptocurrency sector, this model excels in identifying and categorizing ticker symbols, cryptocurrency names, and addresses in textual content.
|
34 |
|
|
|
35 |
|
36 |
+
## Limitations
|
37 |
|
38 |
+
Performance may be subpar when the model encounters entities outside its training data or infrequently occurring entities within the cryptocurrency domain. The model might also be susceptible to variations in entity presentation and context.
|
39 |
+
## Training and evaluation data
|
40 |
|
41 |
+
The model was trained using a diverse dataset, including artificially generated tweets and ERC20 token metadata fetched through the Covalent API (https://www.covalenthq.com/docs/unified-api/). GPT was employed to generate 500 synthetic tweets tailored for the cryptocurrency domain. The Covalent API was instrumental in obtaining a rich set of 20K+ unique ERC20 token metadata entries, enhancing the model's understanding and recognition of cryptocurrency entities.
|
42 |
## Training procedure
|
43 |
|
44 |
### Training hyperparameters
|