albert-base-v2-squad-fp16
This model is a fp16 quantized version of albert-base-v2-squad. It achieves the following results on the SQuAD 1.1 evaluation set (no model accuracy loss compared to fp32):
- Exact Match(EM): 84.68
- F1: 91.40
Inference API
You can test the model directly using the Hugging Face Inference API:
from transformers import pipeline
# Load the pipeline
qa_pipeline = pipeline("question-answering", model="xichenn/albert-base-v2-squad-fp16")
# Run inference
result = qa_pipeline(question="What is the capital of France?", context="France is a country in Europe. Its capital is Paris.")
print(result)
- Downloads last month
- 30
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.
Model tree for xichenn/albert-base-v2-squad-fp16
Base model
albert/albert-base-v2Datasets used to train xichenn/albert-base-v2-squad-fp16
Evaluation results
- Exact Match on squadvalidation set self-reported84.680
- F1 on squadvalidation set self-reported91.400