Generated by GPT-5-mini| WLCG (Worldwide LHC Computing Grid) | |
|---|---|
| Name | WLCG (Worldwide LHC Computing Grid) |
| Abbreviation | WLCG |
| Formation | 2001 |
| Type | International scientific collaboration |
| Purpose | Distributed computing for particle physics |
| Headquarters | CERN |
| Region served | Worldwide |
| Parent organization | CERN |
WLCG (Worldwide LHC Computing Grid) is a global distributed computing infrastructure created to process, store, and analyze data from the Large Hadron Collider. It connects large research institutions, national laboratories, and university data centers to enable experiments such as ATLAS, CMS, ALICE, and LHCb. The project integrates technologies and organizations from projects and entities including CERN, the European Organization for Nuclear Research, national research agencies, and international grid initiatives.
WLCG was conceived to support the data demands of the Large Hadron Collider and its flagship experiments ATLAS experiment, CMS experiment, ALICE experiment, and LHCb experiment by federating resources across sites such as Fermilab, DESY, INFN, RAL, and TRIUMF. It builds on earlier grid projects like EGEE, GridPP, OpenScienceGrid, and LCG project to provide storage, compute, and network services for collaborations including CERN Council, European Commission, and research infrastructures such as GÉANT. WLCG interfaces with technologies promoted by standards bodies and consortia like Open Grid Forum, GEANT, and initiatives from National Science Foundation partners.
The architecture uses a tiered model with Tier-0 at CERN, regional Tier-1 centers at sites such as CCIN2P3, PIC, and NDGF, and numerous Tier-2 and Tier-3 compute centers at universities like University of Oxford, University of Manchester, MIT, and University of California, Berkeley. Core services include distributed data management systems derived from dCache, EOS (software), and CASTOR (storage), workload management systems influenced by HTCondor, PanDA, and DIRAC (project), and authentication/authorization integrated with technologies from Globus Toolkit and LCMAPS alongside identity federations such as eduGAIN and InCommon. Network connectivity is provided by research networks including GÉANT, ESnet, and national backbones like JANET and SURFnet with traffic engineered using protocols and equipment from Cisco Systems, Juniper Networks, and standards by IETF.
Operations follow coordinated schedules set by experiment collaborations like ATLAS collaboration and CMS Collaboration and are overseen by management bodies at CERN and national grid projects such as GridPP and OpenScienceGrid. Day-to-day operations use monitoring and incident response tools originating in projects like Nagios, Grafana Labs, and ELK Stack with service-level agreements influenced by funding agencies including European Research Council and national research councils such as the Science and Technology Facilities Council. Software deployment and configuration management often employ tools from Ansible, Puppet (software), and container platforms like Kubernetes for experiment-specific workloads coordinated with technical coordination groups tied to institute boards and collaboration boards.
WLCG handles exabytes of data annually produced by the Large Hadron Collider detectors during runs studied by teams including Nobel laureates and research groups across institutions such as University of Chicago, Princeton University, California Institute of Technology, and Max Planck Society. The grid aggregates hundreds of thousands of CPU cores and many petabytes of disk and tape storage at centers like Centre de Calcul de l'IN2P3 and KIT (Karlsruhe Institute of Technology), sustaining analysis campaigns that led to discoveries associated with the Higgs boson and precision measurements cited by collaborations in journals like Physical Review Letters and Journal of High Energy Physics. Performance metrics are benchmarked using tools and standards developed in consortia including W3C and measurement projects co‑ordinated with LHC Computing Grid Technical Design Report follow-ups.
Security architecture integrates certificate-based authentication from certificate authorities such as IGTF with authorization policies aligned to collaboration rules at CERN and national laboratories like Brookhaven National Laboratory and SLAC National Accelerator Laboratory. Data management practices employ replication, custodial storage, and metadata catalogs using systems developed by the experiments and sites, interoperating with storage technologies from vendors such as EMC Corporation and initiatives like Storage Resource Broker. Compliance, auditing, incident response, and continuity planning coordinate with national cyber response teams and standards bodies such as ISO while ensuring provenance and data preservation for long‑term reanalysis by institutions including INSPIRE-HEP and archives like CERN Document Server.
Development traces back to proposals and efforts in the late 1990s and early 2000s involving groups from CERN, Fermilab, SLAC National Accelerator Laboratory, and European grid endeavors such as DataGrid. Milestones include the formal LCG project launch, integration of Tier-1 centers, deployment of middleware from projects like EDG and gLite, and transitions to cloud-aware and containerized operations influenced by Amazon Web Services experiments and collaborations with national e‑infrastructure programs. The operational maturity that supported the 2012 discovery attributed to the ATLAS experiment and CMS experiment was followed by evolutions toward more flexible models embracing software-defined storage and compute orchestration with partners in academia and industry including Microsoft Research and Google.
Governance comprises representatives from CERN Council, national funding agencies such as National Science Foundation and Deutsche Forschungsgemeinschaft, and grid projects like GridPP, NDGF, and INFN CNAF. Participating institutions include major laboratories and universities such as CERN, Fermilab, DESY, TRIUMF, Brookhaven National Laboratory, Lawrence Berkeley National Laboratory, University of Tokyo, Tsinghua University, University of Melbourne, and research infrastructures coordinated by entities like European Grid Infrastructure. Strategic direction is set through technical boards, resource allocation committees, and collaboration boards drawing membership from experiments, national nodes, and agencies including STFC and CNRS.