GPT2 Fine-Tuned Banking 77

This is a fine-tuned version of the GPT2 model. It's best suited for text-generation.

Model Description

Kwaku/gpt2-finetuned-banking77 was fine tuned on the banking77 dataset, which is "composed of online banking queries annotated with their corresponding intents."

Intended Uses and Limitations

Given the magnitude of the Microsoft DialoGPT-large model, the author resorted to fine-tuning the gpt2 model for the creation of a chatbot. The intent was for the chatbot to emulate a banking customer agent, hence the use of the banking77 dataset. However, when the fine-tuned model was deployed in the chatbot, the results were undesirable. Its responses were inappropriate and unnecessarily long. The last word of its response is repeated numerously, a major glitch in it. The model performs better in text-generation but is prone to generating banking-related text because of the corpus it was trained on.

How to use

You can use this model directly with a pipeline for text generation:

>>>from transformers import pipeline

>>> model_name = "Kwaku/gpt2-finetuned-banking77"
>>> generator = pipeline("text-generation", model=model_name)
>>> result = generator("My money is", max_length=15, num_return_sequences=2)
>>> print(result)

[{'generated_text': 'My money is stuck in ATM pending. Please cancel this transaction and refund it'}, {'generated_text': 'My money is missing. How do I get a second card, and how'}]

Limitations and bias

For users who want a diverse text-generator, this model's tendency to generate mostly bank-related text will be a drawback. It also inherits the biases of its parent model, the GPT2.

Downloads last month
19
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train Kwaku/gpt2-finetuned-banking77