sescore_english_webnlg / description.md
xu1998hz's picture
Update description.md
fd904ba
|
raw
history blame
3.07 kB
## Installation and usage
```bash
pip install -r requirements.txt
```
Minimal example (evaluating English text generation)
```python
import evaluate
sescore = evaluate.load("xu1998hz/sescore")
# for different versions of SEScore
# sescore = evaluate.load("xu1998hz/sescore_english_mt") -> for English at Machine Translation
# sescore = evaluate.load("xu1998hz/sescore_german_mt") -> for German at Machine Translation
# sescore = evaluate.load("xu1998hz/sescore_english_webnlg") -> for webnlg data-to-text
# sescore = evaluate.load("xu1998hz/sescore_english_coco") -> for image caption
score = sescore.compute(
references=['sescore is a simple but effective next-generation text evaluation metric'],
predictions=['sescore is simple effective text evaluation metric for next generation']
)
```
*SEScore* compares a list of references (gold translation/generated output examples) with a same-length list of candidate generated samples. Currently, the output range is learned and scores are most useful in relative ranking scenarios rather than absolute comparisons. We are producing a series of rescaling options to make absolute SEScore-based scaling more effective.
### Available pre-trained models
Currently, the following language/model pairs are available:
| Language | pretrained data | pretrained model link |
|----------|-----------------|-----------------------|
| English | MT | [xu1998hz/sescore_english_mt](https://huggingface.co/xu1998hz/sescore_english_mt) |
| German | MT | [xu1998hz/sescore_german_mt](https://huggingface.co/xu1998hz/sescore_german_mt) |
| English | webNLG17 | [xu1998hz/sescore_english_webnlg17](https://huggingface.co/xu1998hz/sescore_english_webnlg17) |
| English | CoCo captions | [xu1998hz/sescore_english_coco](https://huggingface.co/xu1998hz/sescore_english_coco) |
Please contact repo maintainer Wenda Xu to add your models!
## Limitations
*SEScore* is trained on synthetic data in-domain.
Although this data is generated to simulate user-relevant errors like deletion and spurious insertion, it may be limited in its ability to simulate humanlike errors.
Model applicability is domain-specific (e.g., CoCo caption-trained model will be better for captioning than MT-trained).
We are in the process of producing and benchmarking general language-level *SEScore* variants.
## Citation
If you find our work useful, please cite the following:
```bibtex
@inproceedings{xu-etal-2022-not,
title={Not All Errors are Equal: Learning Text Generation Metrics using Stratified Error Synthesis},
author={Xu, Wenda and Tuan, Yi-lin and Lu, Yujie and Saxon, Michael and Li, Lei and Wang, William Yang},
booktitle ={Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing},
month={dec},
year={2022},
url={https://arxiv.org/abs/2210.05035}
}
```
## Acknowledgements
The work of the [COMET](https://github.com/Unbabel/COMET) maintainers at [Unbabel](https://duckduckgo.com/?t=ffab&q=unbabel&ia=web) has been instrumental in producing SEScore.