|
--- |
|
license: mit |
|
language: |
|
- en |
|
pipeline_tag: text2text-generation |
|
--- |
|
# News2Topic-V2-Flan-T5-base |
|
|
|
## Model Details |
|
|
|
- **Model type:** Text-to-Text Generation |
|
- **Language(s) (NLP):** English |
|
- **License:** MIT License |
|
- **Finetuned from model:** FLAN-T5 Base Model (Google AI) |
|
|
|
## Uses |
|
|
|
The News2Topic Flan T5-base model is designed for automatic generation of topic names from news articles or news-like text. It can be integrated into news aggregation platforms, content management systems, or used for enhancing news browsing and searching experiences by providing concise topics. |
|
|
|
## How to Get Started with the Model |
|
|
|
``` |
|
from transformers import pipeline |
|
|
|
pipe = pipeline("text2text-generation", model="textgain/News2Topic-V2-Flan-T5-base") |
|
|
|
news_text = "Your news text here." |
|
print(pipe(news_text)) |
|
``` |
|
|
|
## Training Details |
|
|
|
The News2Topic V2 Flan T5-base model was trained on a 20K sample of the "Newsroom" dataset (https://lil.nlp.cornell.edu/newsroom/index.html), annotated with data generated by a fine-tuned GPT-3.5-turbo on synthetic curated data. |
|
|
|
The model was trained for 10 epochs, with a learning rate of 0.00001, a maximum sequence length of 512, and a training batch size of 12. |
|
|
|
## Citation |
|
|
|
**BibTeX:** |
|
``` |
|
@article{Kosar_DePauw_Daelemans_2024, |
|
title={Comparative Evaluation of Topic Detection: Humans vs. LLMs}, volume={13}, |
|
url={https://www.clinjournal.org/clinj/article/view/173}, journal={Computational Linguistics in the Netherlands Journal}, |
|
author={Kosar, Andriy and De Pauw, Guy and Daelemans, Walter}, |
|
year={2024}, |
|
month={Mar.}, |
|
pages={91–120} } |
|
``` |