Fine-tuned DistilBERT for Named Entity Recognition (NER)

Model Description

This model is a fine-tuned version of DistilBERT for Named Entity Recognition (NER) tasks. It was trained on the CoNLL-2003 dataset, designed to identify entities such as persons, organizations, locations, and miscellaneous entities within English text.

  • Model Architecture: DistilBERT (pre-trained transformer-based model)
  • Task: Named Entity Recognition (NER)
  • Entity Types: PER (Person), ORG (Organization), LOC (Location), MISC (Miscellaneous)

Training Details

  • Dataset: CoNLL-2003 (standard dataset for NER tasks)
  • Training Data Size: 14,000 samples for training, 3,250 samples for evaluation
  • Epochs: 3
  • Batch Size: 16 (training), 64 (evaluation)
  • Learning Rate: 2e-5
  • Optimizer: AdamW with weight decay

Evaluation Metrics

The model was evaluated using the following metrics:

  • F1 Score: 0.928661
  • Accuracy: 0.983252
  • Precision: 0.918794
  • Recall: 0.938741

Example Usage

Here’s how to use this NER model with the Hugging Face Transformers library:

from transformers import pipeline

# Load the model from the Hugging Face Hub
ner_pipeline = pipeline("ner", model="Beehzod/smart-finetuned-ner")

# Example predictions
text = "Hugging Face Inc. is based in New York City, and its CEO is Clement Delangue."
results = ner_pipeline(text)

for entity in results:
    print(f"Entity: {entity['word']}, Label: {entity['entity']}, Score: {entity['score']:.4f}")
Downloads last month
22
Safetensors
Model size
65.8M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train Beehzod/smart-finetuned-ner

Space using Beehzod/smart-finetuned-ner 1