aeth0r's picture
Update README.md
20f91c1 verified
---
license: mit
language:
- en
library_name: transformers
pipeline_tag: text-classification
widget:
- text: Prince Raoden went to Elantris. [SEP] Elantris is a great city.
---
# bert-base-cased-NER-reranker
A BERT model trained on the synthetic literary NER context retrieval dataset [Amalvy et. al, 2023](https://aclanthology.org/2023.emnlp-main.642/).
To use this model, construct a text of the form **NER-sentence [SEP] context-sentence**.
The model should predict the positive class if **context-sentence** is useful to predict **NER-sentence**, and the negative class otherwise.
# Performance Metrics
The model obtains 98.34 F1 on the synthetic test set.
See [Amalvy et. al, 2023](https://aclanthology.org/2023.emnlp-main.642/) for details about NER performance gains when using this retriever model to assit a NER model at inference.
# How to Reproduce Training
See the training script [here](https://github.com/CompNet/conivel/blob/gen/train_reranker.py).
# Citation
If you use this model in your research, please cite:
```bibtex
@InProceedings{Amalvy2023,
title = {Learning to Rank Context for Named Entity Recognition Using a Synthetic Dataset},
author = {Amalvy, A. and Labatut, V. and Dufour, R.},
booktitle = {2023 Conference on Empirical Methods in Natural Language Processing},
year = {2023},
doi = {10.18653/v1/2023.emnlp-main.642},
pages = {10372-10382},
}
```