# Muril Large Squad2 This model is finetuned for QA task on Squad2 from [Muril Large checkpoint](https://huggingface.co/google/muril-large-cased). ## Hyperparameters ``` Batch Size: 4 Grad Accumulation Steps = 8 Total epochs = 3 MLM Checkpoint = google/muril-large-cased max_seq_len = 256 learning_rate = 1e-5 lr_schedule = LinearWarmup warmup_ratio = 0.1 doc_stride = 128 ``` ## Squad 2 Evaluation stats: Generated from [the official Squad2 evaluation script](https://worksheets.codalab.org/rest/bundles/0x6b567e1cf2e041ec80d7098f031c5c9e/contents/blob/) ```json { "exact": 82.0180240882675, "f1": 85.10110304685352, "total": 11873, "HasAns_exact": 81.6970310391363, "HasAns_f1": 87.87203044454981, "HasAns_total": 5928, "NoAns_exact": 82.3380992430614, "NoAns_f1": 82.3380992430614, "NoAns_total": 5945 } ``` ## Limitations MuRIL is specifically trained to work on 18 Indic languages and English. This model is not expected to perform well in any other languages. See the MuRIL checkpoint for further details. For any questions, you can reach out to me [on Twitter](https://twitter.com/batw0man)