ZenGQ - BERT for Question Answering

This is a fine-tuned BERT model for question answering tasks, trained on a custom dataset.

Model Details

  • Model: BERT-base-uncased
  • Task: Question Answering
  • Dataset: Rep00Zon

Usage

Load the model

from transformers import AutoTokenizer, AutoModelForQuestionAnswering, pipeline

# Load the tokenizer and model from Hugging Face
tokenizer = AutoTokenizer.from_pretrained("prabinpanta0/ZenGQ")
model = AutoModelForQuestionAnswering.from_pretrained("prabinpanta0/ZenGQ")

# Create a pipeline for question answering
qa_pipeline = pipeline("question-answering", model=model, tokenizer=tokenizer)

# Define your context and questions
contexts = ["Berlin is the capital of Germany.",
          "Paris is the capital of France.",
          "Madrid is the capital of Spain."]
questions = [
    "What is the capital of Germany?",
    "Which city is the capital of France?",
    "What is the capital of Spain?"
]

# Get answers
for context, question in zip(contexts, questions):
    result = qa_pipeline(question=question, context=context)
    print(f"Question: {question}")
    print(f"Answer: {result['answer']}\n")

Training Details

  • Epochs: 3
  • Training Loss: 2.050335, 1.345047, 1.204442

Token

text = "Berlin is the capital of Germany. Paris is the capital of France. Madrid is the capital of Spain."
tokens = tokenizer.tokenize(text)
print(tokens)

Output: ['berlin', 'is', 'the', 'capital', 'of', 'germany', '.', 'paris', 'is', 'the', 'capital', 'of', 'france', '.', 'madrid', 'is', 'the', 'capital', 'of', 'spain', '.']

Dataset

The model was trained on the Rep00Zon dataset.

License

This model is licensed under the MIT License.

Downloads last month
12
Safetensors
Model size
109M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train prabinpanta0/ZenGQ