Generated by GPT-5-mini| CORD-19 | |
|---|---|
| Name | CORD-19 |
| Subject | Biomedical literature dataset on coronaviruses |
| Released | 2020 |
| Creators | Allen Institute for AI; National Institutes of Health; World Health Organization; Chan Zuckerberg Initiative |
| Language | English (primary), multilingual subsets |
| Format | Machine-readable text, metadata, JSON, PDF |
| License | Mixed (open-access, publisher agreements) |
| Website | (see project portals) |
CORD-19
CORD-19 is a large, machine-readable corpus of scientific literature assembled in response to the 2019–21 global COVID-19 pandemic and the emergence of SARS-CoV-2. Initially published in 2020, it aimed to accelerate text mining and computational analysis by researchers at institutions such as the Allen Institute for AI, the National Institutes of Health, the World Health Organization, and the Chan Zuckerberg Initiative. The dataset aggregated peer-reviewed articles, preprints, and historical papers on coronaviruses to support rapid discovery for clinical research, public health response, and bioinformatics.
The dataset emerged amid the declaration of a Public Health Emergency of International Concern by the World Health Organization and coordinated scientific mobilization including initiatives from the National Institutes of Health, the European Commission, and philanthropic actors like the Bill & Melinda Gates Foundation and the Chan Zuckerberg Initiative. It built on prior open-science efforts such as the Human Genome Project, the Protein Data Bank, and the PubMed literature ecosystem to enable computational epidemiology, natural language processing, and drug-repurposing pipelines. Contributors included publishers and repositories such as bioRxiv, medRxiv, PubMed Central, and major journals from publishers represented by Elsevier, Springer Nature, and the Nature Publishing Group.
CORD-19 combined tens of thousands of documents spanning peer-reviewed articles, preprints, and legacy coronavirus research. Primary sources included collections from PubMed Central, preprint servers bioRxiv and medRxiv, and publisher-provided content from houses like Elsevier, Wiley-Blackwell, and Oxford University Press. Historical repositories such as the WHO COVID-19 Database and legacy literature on SARS and MERS were incorporated to provide context from the outbreaks linked to SARS-CoV and MERS-CoV. Metadata fields paralleled bibliographic standards used by CrossRef, DOI registries, and indexing services such as Scopus and Web of Science; full-texts were provided in machine-readable formats including JSON and XML to support tools developed in environments like GitHub and cloud platforms such as Amazon Web Services.
The project coordination featured technical leadership from the Allen Institute for AI in partnership with the National Institutes of Health and the White House Office of Science and Technology Policy, with dataset stewardship influenced by academic groups at institutions including Stanford University, Massachusetts Institute of Technology, Harvard University, and University of Washington. Engineering workflows relied on open-source toolchains hosted on GitHub and continuous-updates pipelines to integrate contributions from publishers like Cell Press and The Lancet-affiliated titles. Governance and advisory inputs came from subject-matter experts at organizations such as the Centers for Disease Control and Prevention, the World Health Organization, and national research funders like the National Science Foundation.
Researchers applied the corpus across diverse computational tasks: natural language processing benchmarks, named-entity recognition for proteins, chemical entities, and pathogens, and knowledge graph construction linking genes, proteins, and therapeutics studied by groups at Broad Institute and Scripps Research. Drug-repurposing screens referenced compounds cataloged by institutions like European Medicines Agency and U.S. Food and Drug Administration, while epidemiological modeling teams at Imperial College London and Johns Hopkins University used literature-derived parameters for transmission and clinical severity. Text-mining efforts produced datasets for question-answering systems, citation-network analyses leveraging Google Scholar and bibliometric tools such as Altmetric, and topic-modeling used by computational linguistics groups at Carnegie Mellon University and University of Cambridge.
Access was enabled through public downloads and programmatic APIs with distribution coordinated under mixed licensing arrangements: many items were available under open-access policies consistent with Plan S principles and repositories like PubMed Central, while other components required publisher agreements or were made available under negotiated terms with houses such as Elsevier and Springer Nature. Data portals and mirrors were hosted by partners including the National Institutes of Health and cloud providers such as Amazon Web Services and Google Cloud Platform, enabling uptake by research teams at institutions like University of Oxford and industry labs at DeepMind and IBM Research.
Critiques addressed coverage biases, heterogeneity of licensing, and quality control challenges inherent in aggregating rapid preprints from bioRxiv and medRxiv alongside peer-reviewed literature in journals like The Lancet and New England Journal of Medicine. Concerns were raised by librarians and ethicists at institutions such as Columbia University and University of California, Berkeley about reproducibility, curation of retracted studies (e.g., high-profile retractions), and the potential for automated analyses to amplify flawed findings. Technical limitations included inconsistent metadata standards relative to CrossRef and variable text-extraction fidelity from PDF sources, prompting calls from the Open Science Framework and communities around arXiv for improved provenance, persistent identifiers, and better integration with curated databases like the Protein Data Bank and drug ontologies maintained by the National Library of Medicine.
Category:Datasets