LLMpediaThe first transparent, open encyclopedia generated by LLMs

Natural Questions

Generated by GPT-5-mini
Note: This article was automatically generated by a large language model (LLM) from purely parametric knowledge (no retrieval). It may contain inaccuracies or hallucinations. This encyclopedia is part of a research project currently under review.
Article Genealogy
Parent: SQuAD Hop 4
Expansion Funnel Raw 61 → Dedup 0 → NER 0 → Enqueued 0
1. Extracted61
2. After dedup0 (None)
3. After NER0 ()
4. Enqueued0 ()
Natural Questions
Natural Questions
Unknown authorUnknown author · Public domain · source
NameNatural Questions
DeveloperGoogle Research
Released2019
DomainQuestion answering, Information retrieval
LicenseCreative Commons
Size300000

Natural Questions

Natural Questions is a large-scale dataset for machine reading and open-domain question answering developed to bridge human information needs and machine comprehension. It was created to evaluate models on realistic queries derived from Google Search logs and grounded in full-text Wikipedia pages, supporting research in retrieval, reading comprehension, and answer verification. The dataset has influenced benchmark efforts across industry and academia, spurring comparisons with corpora such as SQuAD, TriviaQA, MS MARCO, and HotpotQA.

Overview

Natural Questions contains real anonymized queries from Google Search paired with corresponding long-form pages from Wikipedia annotated with short and long answers. The dataset emphasizes real-world query intent by sampling from query streams similar to those studied by PageRank researchers and teams at Google Research who investigate web-scale information access. Annotators with ties to projects at Stanford University and Carnegie Mellon University labeled spans and passages, aligning efforts with prior work tied to Allen Institute for AI initiatives. The corpus served as a testbed for models originating from groups at Facebook AI Research, DeepMind, Microsoft Research, and startups incubated at Y Combinator.

Dataset and Methodology

Natural Questions was constructed using anonymized queries, retrieval of candidate pages, and manual annotation of answer spans within Wikipedia articles. The methodology builds on retrieval pipelines similar to those in systems developed by Elastic NV and algorithms influenced by the work of researchers at University of Washington and University of Massachusetts Amherst. Annotation guidelines referenced crowd-sourcing practices used by Amazon Mechanical Turk and labeling protocols from teams at New York University and University of Edinburgh. The dataset includes short answer spans, long answer passages, and a “no answer” label for queries lacking explicit answers in the target page, following evaluation designs comparable to those used in datasets by University of Oxford researchers and the National Institute of Standards and Technology.

Applications and Benchmarks

Natural Questions has been applied to train models in open-domain question answering, retrieval-augmented generation, and passage ranking. It is commonly used alongside benchmarks from GLUE and SuperGLUE suites and compared with tasks from TREC and competitions hosted by Kaggle. Models developed by teams at Google DeepMind, OpenAI, Alibaba DAMO Academy, and Baidu Research have reported results on the dataset. The corpus enabled advances in architectures such as BERT, RoBERTa, ALBERT, ELECTRA, and transformer variants proposed at University of Toronto and University of California, Berkeley. It also guided improvements in retrieval systems exploiting techniques from BM25 implementations and neural retrievers inspired by work at Johns Hopkins University.

Evaluation Metrics and Results

Evaluation on Natural Questions uses metrics for both short and long answers, including exact match and F1 measures for span recovery, and passage-level metrics for long answer selection. These metrics align with scoring practices from shared tasks run by ACL and EMNLP and mirror evaluation pipelines from ICLR and NeurIPS workshops. Reported results show that transformer-based models from Google Research and ensembles by teams at Stanford University achieved substantial gains over baselines from earlier work at University of California, San Diego and University of Michigan, while retrieval-augmented approaches from Facebook AI Research and hybrid systems by Microsoft Research have pushed passage selection metrics higher. Leaderboards hosted by academic groups and industry labs document improvements over time against baselines established in the original release.

Limitations and Criticisms

Critics have pointed to annotation noise, query sampling biases, and language coverage limitations, echoing concerns raised in analyses from University of Pennsylvania and University of Cambridge. The dependence on Wikipedia restricts coverage for queries about current events and regional topics, a limitation also discussed in literature from Reuters Institute and teams at Columbia University. Privacy advocates and researchers affiliated with Electronic Frontier Foundation have raised questions about de-identification of search-derived queries, similar to debates surrounding datasets assembled by Twitter and archives curated by Internet Archive. Methodological critiques reference reproducibility discussions prominent at Harvard University and dataset audit frameworks used by groups at MIT.

History and Development

Natural Questions was released by researchers at Google Research following prior QA datasets such as SQuAD (Stanford), TriviaQA (University of Washington), and MS MARCO (Microsoft). Its development drew on annotation and retrieval lessons from projects at Carnegie Mellon University and Stanford University, and benefited from collaboration with crowdsourcing platforms used by groups at Princeton University and Yale University. Subsequent updates and analyses were published at venues including ACL, EMNLP, and NAACL, and inspired derivative datasets and tasks created by labs at ETH Zurich, Tsinghua University, and Peking University.

Category:Datasets