XLM-R-Large-Tweet

XLM-R-Large-Tweet is a version of the XLM-R-Large-Tweet-Base*, fine-tuned for sentiment analysis using 5,610 annotated Serbian COVID-19 vaccination-related tweets. Specifically, it is tailored for five-class sentiment analysis to capture finer sentiment nuances in the social media domain using the following scale: very negative, negative, neutral, positive, and very positive.

*XLM-R-Large-Tweet-Base is an additionally pretrained version of the XLM-RoBERTa large-sized model.

How to Use

To use the model, you can load it with the following code:

from transformers import AutoTokenizer, XLMRobertaForSequenceClassification

model_name = "DarijaM/XLM-R-Large-Tweet"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = XLMRobertaForSequenceClassification.from_pretrained(model_name)
Downloads last month
5
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.