|
--- |
|
language: fa |
|
license: apache-2.0 |
|
--- |
|
|
|
# ParsBERT (v3.0) |
|
A Transformer-based Model for Persian Language Understanding |
|
|
|
The new version of BERT v3.0 for Persian is available today and can tackle the zero-width non-joiner character for Persian writing. Also, the model was trained on new multi-types corpora with a new set of vocabulary. |
|
|
|
|
|
## Introduction |
|
|
|
ParsBERT is a monolingual language model based on Google’s BERT architecture. This model is pre-trained on large Persian corpora with various writing styles from numerous subjects (e.g., scientific, novels, news). |
|
|
|
Paper presenting ParsBERT: [arXiv:2005.12515](https://arxiv.org/abs/2005.12515) |
|
|
|
### BibTeX entry and citation info |
|
|
|
Please cite in publications as the following: |
|
|
|
```bibtex |
|
@article{ParsBERT, |
|
title={ParsBERT: Transformer-based Model for Persian Language Understanding}, |
|
author={Mehrdad Farahani, Mohammad Gharachorloo, Marzieh Farahani, Mohammad Manthouri}, |
|
journal={ArXiv}, |
|
year={2020}, |
|
volume={abs/2005.12515} |
|
} |
|
``` |
|
|
|
## Questions? |
|
Post a Github issue on the [ParsBERT Issues](https://github.com/hooshvare/parsbert/issues) repo. |