Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,69 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
language: fa
|
3 |
+
license: apache-2.0
|
4 |
+
---
|
5 |
+
|
6 |
+
# ALBERT Persian
|
7 |
+
|
8 |
+
A Lite BERT for Self-supervised Learning of Language Representations for the Persian Language
|
9 |
+
|
10 |
+
> میتونی بهش بگی برت_کوچولو
|
11 |
+
|
12 |
+
[ALBERT-Persian](https://github.com/m3hrdadfi/albert-persian) is the first attempt on ALBERT for the Persian Language. The model was trained based on Google's ALBERT BASE Version 2.0 over various writing styles from numerous subjects (e.g., scientific, novels, news) with more than 3.9M documents, 73M sentences, and 1.3B words, like the way we did for ParsBERT.
|
13 |
+
|
14 |
+
Please follow the [ALBERT-Persian](https://github.com/m3hrdadfi/albert-persian) repo for the latest information about previous and current models.
|
15 |
+
|
16 |
+
## Persian Sentiment [Digikala, SnappFood, DeepSentiPers]
|
17 |
+
|
18 |
+
It aims to classify text, such as comments, based on their emotional bias. We tested three well-known datasets for this task: `Digikala` user comments, `SnappFood` user comments, and `DeepSentiPers` in two binary-form and multi-form types.
|
19 |
+
|
20 |
+
### SnappFood
|
21 |
+
|
22 |
+
[Snappfood](https://snappfood.ir/) (an online food delivery company) user comments containing 70,000 comments with two labels (i.e. polarity classification):
|
23 |
+
|
24 |
+
1. Happy
|
25 |
+
2. Sad
|
26 |
+
|
27 |
+
| Label | # |
|
28 |
+
|:--------:|:-----:|
|
29 |
+
| Negative | 35000 |
|
30 |
+
| Positive | 35000 |
|
31 |
+
|
32 |
+
**Download**
|
33 |
+
You can download the dataset from [here](https://drive.google.com/uc?id=15J4zPN1BD7Q_ZIQ39VeFquwSoW8qTxgu)
|
34 |
+
|
35 |
+
|
36 |
+
## Results
|
37 |
+
|
38 |
+
The following table summarizes the F1 score obtained as compared to other models and architectures.
|
39 |
+
|
40 |
+
| Dataset | ALBERT-fa-base-v2 | ParsBERT-v1 | mBERT | DeepSentiPers |
|
41 |
+
|:------------------------:|:-----------------:|:-----------:|:-----:|:-------------:|
|
42 |
+
| SnappFood User Comments | 85.79 | 88.12 | 87.87 | - |
|
43 |
+
|
44 |
+
|
45 |
+
### BibTeX entry and citation info
|
46 |
+
|
47 |
+
Please cite in publications as the following:
|
48 |
+
|
49 |
+
```bibtex
|
50 |
+
@misc{ALBERTPersian,
|
51 |
+
author = {Mehrdad Farahani},
|
52 |
+
title = {ALBERT-Persian: A Lite BERT for Self-supervised Learning of Language Representations for the Persian Language},
|
53 |
+
year = {2020},
|
54 |
+
publisher = {GitHub},
|
55 |
+
journal = {GitHub repository},
|
56 |
+
howpublished = {\url{https://github.com/m3hrdadfi/albert-persian}},
|
57 |
+
}
|
58 |
+
|
59 |
+
@article{ParsBERT,
|
60 |
+
title={ParsBERT: Transformer-based Model for Persian Language Understanding},
|
61 |
+
author={Mehrdad Farahani, Mohammad Gharachorloo, Marzieh Farahani, Mohammad Manthouri},
|
62 |
+
journal={ArXiv},
|
63 |
+
year={2020},
|
64 |
+
volume={abs/2005.12515}
|
65 |
+
}
|
66 |
+
```
|
67 |
+
|
68 |
+
## Questions?
|
69 |
+
Post a Github issue on the [ALBERT-Persian](https://github.com/m3hrdadfi/albert-persian) repo.
|