|
--- |
|
license: unknown |
|
base_model: |
|
- airesearch/wangchanberta-base-att-spm-uncased |
|
--- |
|
# AmbatronBERTa |
|
|
|
AmbatronBERTa is a Thai language model fine-tuned specifically for text classification tasks, built upon the WangchanBERTa architecture. |
|
|
|
## Model Description |
|
AmbatronBERTa is designed to handle the complexities of the Thai language. It has been fine-tuned on a dataset of over 3,000 research papers to improve classification accuracy. Leveraging the transformer-based WangchanBERTa, it efficiently captures the nuances of Thai text, making it suitable for classifying documents across multiple fields. |
|
|
|
## Developers |
|
AmbatronBERTa was developed by students at **King Mongkut's University of Technology North Bangkok**: |
|
- **Peerawat Banpahan** |
|
- **Waris Thongpho** |
|
|
|
## Use Cases |
|
AmbatronBERTa can be applied to a wide range of tasks, such as: |
|
- **Research Classification:** Categorizing academic papers into relevant topics. |
|
- **Document Organization:** Classifying articles, blogs, and other documents by themes. |
|
- **Sentiment Analysis:** Analyzing sentiment in Thai-language texts across various contexts. |
|
|
|
## How to Use |
|
To use AmbatronBERTa with the `transformers` library: |
|
|
|
```python |
|
from transformers import AutoTokenizer, AutoModelForSequenceClassification |
|
|
|
# Load the tokenizer and model |
|
tokenizer = AutoTokenizer.from_pretrained("Peerawat2024/AmbatronBERTa") |
|
model = AutoModelForSequenceClassification.from_pretrained("Peerawat2024/AmbatronBERTa") |