FiD model trained on WebQ

-- This is the model checkpoint of FiD [2], based on the T5 large (with 770M parameters) and trained on the WebQ dataset [1].

-- Hyperparameters: 8 x 40GB A100 GPUs; batch size 8; AdamW; LR 3e-5; 30000 steps

References:

[1] Semantic parsing on freebase from question-answer pairs. EMNLP 2013.

[2] Leveraging Passage Retrieval with Generative Models for Open Domain Question Answering. EACL 2021.

Model performance

We evaluate it on the WebQ dataset, the EM score is 50.2 on the test set.

Downloads last month
32
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model’s pipeline type.