Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

roberta-base fine-tuned on the SQuAD2 dataset for 2 epochs.

The fine-tuning process was performed on a single NVIDIA Tesla T4 GPU (15GB). The hyperparameters are:

max_seq_length=512
per_device_train_batch_size=8
gradient_accumulation_steps=4
total train batch size (w. parallel, distributed & accumulation) = 32
learning_rate=3e-5

Evaluation results

"eval_exact": 80.33352985766024,
"eval_f1": 83.38322909593009,

"eval_HasAns_exact": 77.81713900134953,
"eval_HasAns_f1": 83.925283241562,
"eval_HasAns_total": 5928,
"eval_NoAns_exact": 82.84272497897393,
"eval_NoAns_f1": 82.84272497897393,
"eval_NoAns_total": 5945,
"eval_best_exact": 80.33352985766024,
"eval_best_exact_thresh": 0.0,
"eval_best_f1": 83.38322909593005,
"eval_best_f1_thresh": 0.0,
"eval_samples": 11955,
"eval_total": 11873,

More information

Stanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by crowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span, from the corresponding reading passage, or the question might be unanswerable.

SQuAD2.0 combines the 100,000 questions in SQuAD1.1 with over 50,000 unanswerable questions written adversarially by crowdworkers to look similar to answerable ones. To do well on SQuAD2.0, systems must not only answer questions when possible, but also determine when no answer is supported by the paragraph and abstain from answering. (https://rajpurkar.github.io/SQuAD-explorer/)

Downloads last month
22
Safetensors
Model size
124M params
Tensor type
I64
F32
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.