Generated by GPT-5-mini| SQuAD | |
|---|---|
| Name | SQuAD |
| Released | 2016 |
| Creators | Stanford University |
| Domain | Machine reading comprehension |
| License | CC BY-SA 4.0 |
SQuAD SQuAD is a benchmark dataset for machine reading comprehension introduced by researchers at Stanford University designed to evaluate question answering systems trained on passages drawn from Wikipedia. The dataset influenced research at institutions including Google Research, Microsoft Research, Facebook AI Research, DeepMind, and companies such as OpenAI and IBM Research while aligning with evaluation standards used in shared tasks hosted at venues like ACL, NeurIPS, EMNLP, and NAACL. SQuAD catalyzed model development that involved architectures from groups working on BERT, ELMo, GPT, RoBERTa, and ALBERT.
SQuAD is a reading-comprehension corpus where models answer questions by extracting spans from paragraphs sourced from Wikipedia articles about subjects like Barack Obama, World War II, Shakespeare, Marie Curie, and Python (programming language). The dataset's creation drew on annotation practices related to projects at Stanford University and evaluation conventions used in competitions hosted by organizations including Kaggle, SemEval, TREC, SIGHAN, and IWSLT. SQuAD's public release generated follow-up datasets and extensions from teams at Allen Institute for AI, Carnegie Mellon University, University of Washington, Berkeley Artificial Intelligence Research, and MIT CSAIL.
The original SQuAD release was produced by crowdsourced annotators who selected passages from Wikipedia articles and authored question–answer pairs with span-based answers, a process inspired by annotation schemes from projects at Amazon Mechanical Turk, CrowdFlower, Prolific (company), Linguistic Data Consortium, and workflows practiced in corpora such as Penn Treebank and PropBank. The dataset comprises hundreds of paragraphs and tens of thousands of question–answer pairs drawn from articles on figures like Albert Einstein, Isaac Newton, Leonardo da Vinci, Ada Lovelace, and Nelson Mandela. Later versions expanded the task to include unanswerable questions, paralleling efforts by research groups at University of Oxford, University College London, ETH Zurich, Peking University, and Tsinghua University.
SQuAD evaluation uses exact match (EM) and F1 score metrics computed against human-annotated spans, metrics commonly reported in publications at ACL, EMNLP, NeurIPS, ICLR, and COLING. Leaderboards compare systems from teams affiliated with Google Research, Facebook AI Research, Microsoft Research, OpenAI, DeepMind, and academic labs at Stanford University, University of Washington, Carnegie Mellon University, UC Berkeley, and Massachusetts Institute of Technology. Tasks derived from SQuAD include extractive question answering, adversarial evaluation popularized by groups at NYU, Princeton University, Cornell University, University of Illinois Urbana-Champaign, and multilingual adaptations by teams at Google Translate and Microsoft Translator.
Initial baselines for SQuAD included feature-based systems developed in labs at Stanford University and neural architectures from groups at University of Washington and University of Michigan, followed by improvements from recurrent and attention-based models influenced by work at Facebook AI Research, Google Brain, DeepMind, and OpenAI. Transformer-based entries such as BERT and RoBERTa from Google Research and Facebook AI Research rapidly advanced leaderboard rankings, with subsequent compression and efficiency variants from Google Research (ALBERT), Hugging Face contributors, and research teams at Salesforce Research. The public leaderboard showcased contributions from academic teams at Carnegie Mellon University, University of California, Berkeley, University of Edinburgh, National University of Singapore, and corporate labs like IBM Research and Baidu Research.
SQuAD spurred progress in natural language processing research pursued by groups at Stanford University, Google Research, OpenAI, Facebook AI Research, and Microsoft Research, enabling downstream systems in question answering deployed by products from Google, Microsoft, Amazon (company), Apple Inc., and Facebook. It influenced the development of pretrained language models such as BERT, GPT-2, GPT-3, RoBERTa, XLNet, and ALBERT, and informed evaluation protocols in shared tasks run at EMNLP, ACL, NeurIPS, and ICLR. Applications inspired by SQuAD include search enhancements at Google Search, conversational agents from Amazon Alexa, customer support bots at companies like Zendesk and Salesforce, and question-answering interfaces in digital libraries curated by institutions such as Library of Congress and Europeana.
Researchers at University of Pennsylvania, ETH Zurich, University of Cambridge, Columbia University, and Johns Hopkins University have critiqued SQuAD for dataset biases, annotation artifacts, overfitting to span-extraction paradigms, and limited reasoning requirements compared with benchmarks such as Winograd Schema Challenge, DROP (dataset), Natural Questions, and HotpotQA. Analyses published in venues including ACL, EMNLP, NeurIPS, ICLR, and Transactions of the ACL highlighted issues like lexical overlap shortcuts, adversarial vulnerability demonstrated by papers from NYU and University of Maryland, and challenges in generalization observed by teams at Google Research and Facebook AI Research. These concerns motivated richer benchmarks and evaluation frameworks developed by entities such as Allen Institute for AI, Hugging Face, Carnegie Mellon University, and research groups across Europe and Asia.
Category:Datasets