Generated by GPT-5-mini| Stanford Large Network Dataset Collection | |
|---|---|
| Name | Stanford Large Network Dataset Collection |
| Established | 2009 |
| Curator | Stanford Network Analysis Project |
| Location | Stanford University, California |
| Disciplines | Computer Science; Network Science; Data Mining |
| Website | Stanford SNAP |
Stanford Large Network Dataset Collection
The Stanford Large Network Dataset Collection is a curated repository compiling large-scale network and graph datasets widely used in computer science, data mining, and network science research. It aggregates social networks, web graphs, collaboration networks, and biological interaction maps for benchmarking algorithms developed by researchers affiliated with institutions such as Stanford University, Massachusetts Institute of Technology, University of California, Berkeley, Carnegie Mellon University, and Princeton University. The collection supports reproducible evaluation and cross-comparison among work from groups like Google Research, Microsoft Research, Facebook AI Research, Amazon Web Services, and IBM Research.
The collection was initiated to address needs identified by communities around conferences and venues including NeurIPS, ICML, KDD, WWW, and SIGMOD. Early contributors and users include researchers from Stanford Network Analysis Project, faculty such as Jure Leskovec and collaborators from Leland Stanford Junior University departments who sought standardized datasets for studying processes described in papers presented at ACM SIGKDD Conference on Knowledge Discovery and Data Mining and IEEE INFOCOM. The repository emphasizes diverse graph types—directed, undirected, weighted, temporal—and provides metadata to enable methods compared in journals like Nature Communications, Science Advances, IEEE Transactions on Network Science and Engineering, and ACM Computing Surveys.
Notable entries comprise social networks sourced from platforms with ties to research from Twitter, Inc., Facebook, Inc., and datasets reflecting scholarly collaboration networks linked to arXiv and DBLP bibliographic records used in studies by teams at Cornell University, University of Illinois Urbana-Champaign, and Harvard University. Web graphs include crawls of domains cited in work by Yahoo! Research, Baidu Research, and datasets that echo collections used by Common Crawl. Biological and protein-interaction graphs relate to projects from National Institutes of Health-funded labs, European Bioinformatics Institute, and datasets analogous to those in BioGRID and STRING. Additional datasets capture infrastructure and communication patterns similar to those analyzed by researchers at AT&T Labs and Bell Labs Research.
Datasets are distributed in plain text adjacency lists, edge lists, compressed binary formats, and occasionally specialized matrix formats mirroring inputs used by software from GraphX, NetworkX, igraph, SNAP (network analysis) libraries and graph processing systems like Pregel and Apache Giraph. Access mechanisms parallel practices from digital repositories such as Zenodo, Figshare, and institutional archives at Stanford Libraries, with metadata fields that enable citation formats consistent with Digital Object Identifier assignments and identifiers used in CrossRef records. Licensing terms reflect precedents set by entities such as Creative Commons, Open Data Commons, and corporate research datasets redistributed by Microsoft Research Open Data.
Researchers apply these datasets to evaluate algorithms for community detection as developed in studies presented at ICDM and AISTATS, link prediction techniques referenced in KDD papers, and graph embedding methods popularized by teams at Facebook AI Research and Google DeepMind. Use cases span recommendation systems following approaches from Amazon.com and Netflix, Inc., contagion and diffusion modeling explored in work by Centers for Disease Control and Prevention-adjacent research, and infrastructure resilience studies inspired by analyses from National Aeronautics and Space Administration and Department of Energy laboratories. Educational courses at institutions like MIT, Stanford University, and University of Washington use the repository for laboratory assignments and student projects.
The collection serves as a standard benchmark for performance comparisons in algorithmic scalability, memory footprint, and runtime reported at venues such as SOSP, OSDI, and PODC. It underpins empirical evaluations in papers submitted to IEEE Big Data Conference and informs performance baselines used by open-source projects hosted on platforms like GitHub and GitLab. Comparative studies often reference metrics and tasks consistent with challenges organized by NeurIPS competitions and reproducibility tracks at ACM SIGCOMM and USENIX symposia.
Datasets carry a variety of licenses influenced by contributors from academic labs, industry research groups, and public repositories—examples include Creative Commons Attribution variants, permissive licenses similar to those advocated by Open Data Commons, and proprietary-use restrictions modeled by corporate research disclosures from Google LLC and Meta Platforms, Inc.. Users are asked to cite dataset providers and associated papers, following citation norms used by journals such as Journal of Machine Learning Research and conferences like KDD and NeurIPS.
Category:Datasets Category:Network science Category:Computer science datasets