|
--- |
|
license: apache-2.0 |
|
--- |
|
This BERT was fined-tuned on +400k nuclear energy data from twitter/X. The classification accuracy obtained is 96%. \ |
|
The number of labels is 3: {0: Negative, 1: Neutral, 2: Positive} |
|
|
|
This is an example to use it |
|
```bash |
|
from transformers import AutoTokenizer |
|
from transformers import pipeline |
|
from transformers import AutoModelForSequenceClassification |
|
import torch |
|
|
|
checkpoint = 'kumo24/bert-sentiment-nuclear' |
|
tokenizer=AutoTokenizer.from_pretrained(checkpoint) |
|
id2label = {0: "negative", 1: "neutral", 2: "positive"} |
|
label2id = {"negative": 0, "neutral": 1, "positive": 2} |
|
|
|
|
|
if tokenizer.pad_token is None: |
|
tokenizer.add_special_tokens({'pad_token': '[PAD]'}) |
|
|
|
model = AutoModelForSequenceClassification.from_pretrained(checkpoint, |
|
num_labels=3, |
|
id2label=id2label, |
|
label2id=label2id) |
|
|
|
device = torch.device("cuda" if torch.cuda.is_available() else "cpu") |
|
model.to(device) |
|
|
|
|
|
sentiment_task = pipeline("sentiment-analysis", |
|
model=model, |
|
tokenizer=tokenizer) |
|
|
|
print(sentiment_task("Michigan Wolverines are Champions, Go Blue!")) |
|
``` |