AmbatronBERTa / README.md
Peerawat2024's picture
Update README.md
aea6cc2 verified
---
license: unknown
base_model:
- airesearch/wangchanberta-base-att-spm-uncased
---
# AmbatronBERTa
AmbatronBERTa is a Thai language model fine-tuned specifically for text classification tasks, built upon the WangchanBERTa architecture.
## Model Description
AmbatronBERTa is designed to handle the complexities of the Thai language. It has been fine-tuned on a dataset of over 3,000 research papers to improve classification accuracy. Leveraging the transformer-based WangchanBERTa, it efficiently captures the nuances of Thai text, making it suitable for classifying documents across multiple fields.
## Developers
AmbatronBERTa was developed by students at **King Mongkut's University of Technology North Bangkok**:
- **Peerawat Banpahan**
- **Waris Thongpho**
## Use Cases
AmbatronBERTa can be applied to a wide range of tasks, such as:
- **Research Classification:** Categorizing academic papers into relevant topics.
- **Document Organization:** Classifying articles, blogs, and other documents by themes.
- **Sentiment Analysis:** Analyzing sentiment in Thai-language texts across various contexts.
## How to Use
To use AmbatronBERTa with the `transformers` library:
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
# Load the tokenizer and model
tokenizer = AutoTokenizer.from_pretrained("Peerawat2024/AmbatronBERTa")
model = AutoModelForSequenceClassification.from_pretrained("Peerawat2024/AmbatronBERTa")