LLMpediaThe first transparent, open encyclopedia generated by LLMs

SQuAD (versions 1.1 and 2.0)

Generated by GPT-5-mini
Note: This article was automatically generated by a large language model (LLM) from purely parametric knowledge (no retrieval). It may contain inaccuracies or hallucinations. This encyclopedia is part of a research project currently under review.
Article Genealogy
Parent: BERT Hop 4
Expansion Funnel Raw 77 → Dedup 0 → NER 0 → Enqueued 0
1. Extracted77
2. After dedup0 (None)
3. After NER0 ()
4. Enqueued0 ()
SQuAD (versions 1.1 and 2.0)
NameSQuAD (versions 1.1 and 2.0)
Released2016–2018
AuthorsRajpurkar, Jia, Liang
InstitutionStanford University
DomainQuestion answering
LicenseCC BY-SA 4.0 (original)

SQuAD (versions 1.1 and 2.0)

Stanford Question Answering Dataset (SQuAD) versions 1.1 and 2.0 are influential reading comprehension corpora used to train and evaluate machine reading systems, introduced from work at Stanford University and associated with researchers from the Stanford Natural Language Processing Group, the Allen Institute for Artificial Intelligence, and collaborators at institutions such as University of Washington and Carnegie Mellon University. The datasets catalyzed development of neural architectures exemplified by models from teams at Google Research, Facebook AI Research, and Microsoft Research and influenced benchmarks adopted in tracks at conferences including EMNLP, NAACL, and ACL. They are widely used alongside other corpora like CNN/Daily Mail (dataset), NarrativeQA, and HotpotQA in evaluations by groups at OpenAI, DeepMind, and university labs.

Background

SQuAD originated from efforts at Stanford University under guidance by faculty such as Christopher Manning and collaborators including authors affiliated with Princeton University and Columbia University, aiming to create a large-scale, high-quality reading comprehension benchmark comparable to resources like Penn Treebank and ImageNet. The first release, associated with workshops at NAACL 2016 and presentations at venues like EMNLP 2016, emphasized span-extraction task design inspired by prior corpora such as MS MARCO and evaluation traditions from TREC. Version 2.0 extended the corpus in a follow-up effort reported at venues such as ACL 2018 to introduce unanswerable questions, an idea informed by error analyses from teams at Google Brain and critiques published by researchers at University of Edinburgh and University of Oxford.

Dataset Composition

SQuAD 1.1 comprises question-answer pairs crowdsourced from passages drawn primarily from Wikipedia articles curated by editors and aligned with topics like biographies (e.g., entries on Barack Obama, Marie Curie, Albert Einstein), historical events (e.g., French Revolution, Battle of Waterloo, Apollo program), scientific topics (e.g., DNA, General relativity, Higgs boson), and cultural subjects (e.g., The Beatles, Mona Lisa, Hamlet). The collection process engaged annotators experienced in tasks of the kind used by groups at Amazon Mechanical Turk and quality-controlled using protocols akin to those in datasets from LDC. SQuAD 2.0 retained the answerable pairs from 1.1 and added adversarial unanswerable questions designed to resemble answerable ones, following approaches used in adversarial evaluation by teams at Google DeepMind and Facebook AI Research, and drawing on analysis techniques used in studies by researchers at MIT and Stanford University.

Evaluation Metrics and Benchmarks

Performance on both versions is measured primarily by Exact Match and F1-score, metrics adopted and popularized in evaluation suites used by competitions such as Kaggle and leaderboards hosted by organizations like Papers with Code, with leaderboards populated by models developed by labs at Google Research, Facebook AI Research, Microsoft Research, and startups such as Hugging Face. Baselines included feature-rich systems from teams at NYU and ensemble models incorporating representations from pretrained transformers developed by Google AI (e.g., BERT), OpenAI (e.g., GPT series), and model implementations from Hugging Face Transformers. SQuAD benchmarks informed highlights at conferences including NeurIPS and spurred comparisons in workshops organized by IEEE and AAAI.

Impact on NLP Research

SQuAD accelerated adoption of deep contextual representations exemplified by BERT (from Google AI), ELMo (from Allen Institute for AI and University of Washington), and transformer-based models from teams at Google Brain and OpenAI, leading to rapid improvements on leaderboards maintained by Stanford University and mirrored in evaluations by Microsoft Research. The dataset influenced curriculum in courses at MIT, Stanford University, and Carnegie Mellon University and shaped commercial NLP products at companies like Amazon (company), Google LLC, and Microsoft Corporation that integrated question answering features into digital assistants such as Amazon Alexa, Google Assistant, and Microsoft Cortana. Research threads catalyzed by SQuAD produced methods in transfer learning studied at UC Berkeley and multilingual extensions pursued by groups at Facebook AI Research and Google Translate teams, and inspired diagnostic datasets from labs at University of Pennsylvania and University of Cambridge.

Limitations and Criticisms

Scholars at University of Edinburgh, University of Oxford, and New York University highlighted that systems could exploit dataset artefacts, prompting critiques similar to those levied against earlier corpora like bAbI and CLEVR and methodological discussions at ICLR and ACL. Limitations include sensitivity to annotation biases documented by researchers at Princeton University and susceptibility to adversarial examples analyzed by teams at Google DeepMind and Facebook AI Research. Concerns about coverage and representativeness led to calls from ethicists at Harvard University and Yale University for broader domain diversity and bias auditing, echoing debates in panels at NeurIPS and policy discussions involving ACM and IEEE.

Category:Question answering datasets