ARBERTv2EL / README.md
AfnanTS's picture
Update README.md
96d0b82 verified
---
datasets:
- AfnanTS/Final_ArLAMA_DS_tokenized_for_ARBERTv2
language:
- ar
base_model:
- UBC-NLP/ARBERTv2
pipeline_tag: fill-mask
---
<img src="./Transparent.png" alt="Model Logo" width="30%" height="30%" align="right"/>
The **ArBERTV2_EL** model is a transformer-based Arabic language model fine-tuned using the Entity Linking (EL) task. This model leverages Knowledge Graphs (KGs) for intrinsic evaluation of Masked Language Modeling (MLM) models without directly evaluating the EL model. The EL task ensures that the model benefits from the incorporation of structured knowledge during pre-training.
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
Filling masked tokens in Arabic text, particularly in contexts enriched with knowledge from KGs.
### Downstream Use
Can be further fine-tuned for Arabic NLP tasks that require semantic understanding, such as text classification or question answering.
## How to Get Started with the Model
```python
from transformers import pipeline
fill_mask = pipeline("fill-mask", model="AfnanTS/ArBERTV2_EL")
fill_mask("اللغة [MASK] مهمة جدا."
```
## Training Details
### Training Data
Trained on the ArLAMA dataset, which is designed to represent Knowledge Graphs in natural language.
### Training Procedure
Continued pre-training of the ArBERTv2 model using Entity Linking (EL) task.