YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

PubMed 200k RCT DeBERTa v3 Model

This model is fine-tuned on the PubMed 200k RCT dataset using the DeBERTa v3 base model.

Model Details

  • Base model: microsoft/deberta-v3-base
  • Fine-tuned on: PubMed 200k RCT dataset
  • Task: Sequence Classification
  • Number of classes: 5
  • Max sequence length: 68

Usage

from transformers import AutoModelForSequenceClassification, AutoTokenizer

model = AutoModelForSequenceClassification.from_pretrained('Vedant101/bert-uncased-pubmed-200k')
tokenizer = AutoTokenizer.from_pretrained('Vedant101/bert-uncased-pubmed-200k')
Downloads last month
0
Safetensors
Model size
184M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support