You need to agree to share your contact information to access this dataset

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this dataset content.



Documentation

In this work, we introduce SPARROW, SPARROW is a evaluation benchmark for sociopragmatic meaning understanding. SPARROW comprises 169 datasets covering 13 task types across six primary categories (e.g., anti-social language detection, emotion recognition). SPARROW datasets encompass 64 different languages originating from 12 language families representing 16 writing scripts.

How to Use SPARROW

Request Access

To obtain access to the SPARROW benchmark on Huggingface, follow the following steps:

  • Login on your Haggingface account

  • Request access

    • Please fill in your actual full name and affiliation (e.g., the name of your research institute).
    • Please use your official email address if it is available.

Install Requirments

    pip install datasets transformers seqeval

Login with your Huggingface CLI

You can get/manage your access tokens in your settings.

    export HUGGINGFACE_TOKEN="" 
    huggingface-cli login --token $HUGGINGFACE_TOKEN

Submitting your results on SPARROW test

We design a public leaderboard for scoring PLMs on SPARRAW. Our leaderboard is interactive and offers rich meta-data about the various datasets involved as well as the language models we evaluate.

You can evalute your models using SPARROW leaderboard: https://sparrow.dlnlp.ai


Citation

If you use SPARROW for your scientific publication, or if you find the resources in this repository useful, please cite our paper as follows:

@inproceedings{zhang-etal-2023-skipped,
    title = "The Skipped Beat: A Study of Sociopragmatic Understanding in LLMs for 64 Languages",
    author = "Zhang, Chiyu  and
      Khai Duy Doan and,
      Qisheng Liao and,
      Abdul-Mageed, Muhammad",
    booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing (EMNLP)",
    year = "2023",
    publisher = "Association for Computational Linguistics",
}

Acknowledgments

We gratefully acknowledge support from the Natural Sciences and Engineering Research Council of Canada, the Social Sciences and Humanities Research Council of Canada, Canadian Foundation for Innovation, ComputeCanada and UBC ARC-Sockeye.

Downloads last month
35