|
--- |
|
language: en |
|
license: apache-2.0 |
|
datasets: |
|
- empathic reactions to news stories |
|
model-index: |
|
- name: roberta-base-empathy |
|
results: |
|
- task: |
|
type: text-classification |
|
name: Text Classification |
|
dataset: |
|
name: Reaction to News Stories |
|
type: Reaction to News Stories |
|
config: sst2 |
|
split: validation |
|
metrics: |
|
- name: MSE loss |
|
type: MSE loss |
|
value: 7.07853364944458 |
|
- name: Pearson's R (empathy) |
|
type: Pearson's R (empathy) |
|
value: 0.4336383660597612 |
|
- name: Pearson's R (distress) |
|
type: Pearson's R (distress) |
|
value: 0.40006974689041663 |
|
|
|
--- |
|
|
|
# Roberta base finetuned on a dataset of empathic reactions to news stories (Buechel et al., 2018; Tafreshi et al., 2021, 2022) |
|
|
|
## Table of Contents |
|
- [Model Details](#model-details) |
|
- [How to Get Started With the Model](#how-to-get-started-with-the-model) |
|
- [Uses](#uses) |
|
- [Risks, Limitations and Biases](#risks-limitations-and-biases) |
|
- [Training](#training) |
|
|
|
## Model Details |
|
**Model Description:** This model is a fine-tuned checkpoint of [RoBERTA-base](https://huggingface.co/roberta-base), fine-tuned for Track 1 of the[WASSA 2022 Shared Task](https://aclanthology.org/2022.wassa-1.20.pdf) - predicting empathy and distress scores on a dataset of reactions to news stories. |
|
This model attained an average Pearson's correlation (r) of 0.416854 on the dev set (for comparison, the top team had an average r of .54 on the test set ). |
|
|
|
# Training |
|
|
|
#### Training Data |
|
An extended version of the [empathic reactions to news stories dataset](https://codalab.lisn.upsaclay.fr/competitions/834#learn_the_details-datasets) |
|
|
|
###### Fine-tuning hyper-parameters |
|
|
|
|
|
- learning_rate = 1e-5 |
|
- batch_size = 32 |
|
- warmup = 600 |
|
- max_seq_length = 128 |
|
- num_train_epochs = 3.0 |
|
|
|
|
|
|