Worldwide LHC Computing Grid

{{Short description|Grid computing project}}

{{cleanup|reason=Talk:Worldwide LHC Computing Grid#Inconsistences|date=March 2014}}

The Worldwide LHC Computing Grid (WLCG), formerly (until 2006){{cite news | url = http://www.isgtw.org/feature/happy-10th-birthday-wlcg| title = Happy 10th Birthday, WLCG!| work =International Grid Science This Week| first=Jacqui |last=Hayes | date=21 December 2011|access-date=2012-12-20}} the LHC Computing Grid (LCG), is an international collaborative project that consists of a grid-based computer network infrastructure incorporating over 170 computing centers in 42 countries, {{as of|lc=y|2017}}. It was designed by CERN to handle the prodigious volume of data produced by Large Hadron Collider (LHC) experiments.

{{Citation

| date =January 2017

| title =Worldwide LHC Computing Grid

| publisher =CERN

| url =http://wlcg-public.web.cern.ch/

| access-date =2017-08-04}}

{{Citation

| date =January 2017

| title =About

| publisher =CERN

| url =http://wlcg-public.web.cern.ch/about

| access-date =2017-08-04}}

File:Rack of Worldwide LHC Computing Grid.jpg

By 2012, data from over 300 trillion (3×1014) LHC proton-proton collisions had been analyzed,[https://web.archive.org/web/20120615043418/http://www.msnbc.msn.com/id/47783507/ns/technology_and_science-science/t/hunt-higgs-boson-hits-key-decision-point/ Hunt for Higgs boson hits key decision point] and LHC collision data was being produced at approximately 25 petabytes per year. {{As of|2017|post,}} the LHC Computing Grid is the world's largest computing grid comprising over 170 computing facilities in a worldwide network across 42 countries[http://wlcg-public.web.cern.ch/about Worldwide LHC Computing Grid about page] 2017: "Even after filtering out 99% of it, in 2017 we're expecting to gather around 50 petabytes of data. That's 50 million gigabytes, the equivalent to nearly 15 million high-definition (HD) movies."[http://wlcg-public.web.cern.ch/ Worldwide LHC Computing Grid main page] 2017: "WLCG is the world's largest computing grid. It is supported by many associated national and international grids across the world, such as European Grid Initiative (Europe-based) and Open Science Grid (US-based), as well as many other regional grids." scattered around the world that produce a massive distributed computing infrastructure with about 1,000,000 CPU cores, providing more than 10,000 physicists around the world with near real-time access to the LHC data, and the power to process it.

According to the [https://home.cern/science/computing/grid WLCG Website] as of 2024: "WLCG combines about 1.4 million computer cores and 1.5 exabytes of storage from over 170 sites in 42 countries [...] It runs over 2 million tasks per day and [...] global transfer rates exceeded 260 GB/s." Indicating substantial upgrades to WLCG over time beyond its initial release.

Background

The Large Hadron Collider at CERN was designed to test the existence of the Higgs boson, an important but elusive piece of knowledge that had been sought by particle physicists for over 40 years. A very powerful particle accelerator was needed, because Higgs bosons might not be seen in lower energy experiments, and because vast numbers of collisions would need to be studied. Such a collider would also produce unprecedented quantities of collision data requiring analysis. Therefore, advanced computing facilities were needed to process the data.

Description

A design report was published in 2005.{{Cite book |title= LHC Computing Grid: Technical Design Report |work= document LCG-TDR-001, CERN-LHCC-2005-024 |date= 20 June 2005 |publisher= The LCG TDR Editorial Board |isbn= 978-92-9083-253-9 |url= http://cdsweb.cern.ch/record/840543/files/lhcc-2005-024.pdf |access-date= 2 October 2011 }}

It was announced to be ready for data on 3 October 2008.{{cite web |url=http://lcg.web.cern.ch/LCG/lhcgridfest |title=LHC GridFest |year=2008 |publisher=CERN }}

A popular 2008 press article predicted "the internet could soon be made obsolete" by its technology.{{Cite news|title=Coming soon: superfast internet |work=The Times |location=London |author=Jonathan Leake |date=6 April 2008 |url=http://www.timesonline.co.uk/tol/news/science/article3689881.ece |access-date=25 January 2013 |url-status=dead |archive-url=https://web.archive.org/web/20110805105551/http://www.timesonline.co.uk/tol/news/science/article3689881.ece |archive-date=August 5, 2011 }}

CERN had to publish its own articles trying to clear up the confusion.{{Cite news |title= The Grid: separating fact from fiction |publisher= CERN |date= May 2008 |url= http://public.web.cern.ch/public/en/spotlight/SpotlightGridFactsAndFiction-en.html |access-date= 25 January 2013 }} Adapted from an article originally published in Symmetry Breaking.

It incorporates both private fiber-optic cable links and existing high-speed portions of the public Internet. At the end of 2010, the Grid consisted of some 200,000 processing cores and 150 petabytes of disk space, distributed across 34 countries.{{cite journal |title=High-energy physics: Down the petabyte highway |journal= Nature |date= 19 January 2011 |author= Geoff Brumfiel |doi= 10.1038/469282a |volume= 469 |issue= 7330 |pages= 282–283 |pmid=21248814|bibcode= 2011Natur.469..282B |doi-access= free }}

The data stream from the detectors provides approximately 300 GByte/s of data, which after filtering for "interesting events", results in a data stream of about 300 MByte/s. The CERN computer center, considered "Tier 0" of the LHC Computing Grid, has a dedicated 10 Gbit/s connection to the counting room.

The project was expected to generate multiple TB of raw data and event summary data, which represents the output of calculations done by the CPU farm at the CERN data center. This data is sent out from CERN to thirteen Tier 1 academic institutions in Europe, Asia, and North America,{{Cite web |title= The Grid: A system of tiers |publisher= CERN |url= https://home.cern/about/computing/grid-system-tiers |access-date= 2 October 2017 }} via dedicated links with 10 Gbit/s or higher of bandwidth. This is called the LHC Optical Private Network.{{Cite web |title= Network transfer architecture |publisher= CERN |url= http://lcg.web.cern.ch/LCG/public/data_transfer.htm |access-date= 2 October 2011 |archive-date= 1 October 2011 |archive-url= https://web.archive.org/web/20111001174103/http://lcg.web.cern.ch/LCG/public/data_transfer.htm |url-status= dead }}

More than 150 Tier 2 institutions are connected to the Tier 1 institutions by general-purpose national research and education networks.[http://gridcafe.web.cern.ch/gridcafe/animations/LHCdata/LHCdata.html final-draft-4-key] {{webarchive |url=https://web.archive.org/web/20080526104330/http://gridcafe.web.cern.ch/gridcafe/animations/LHCdata/LHCdata.html |date=May 26, 2008 }}

The data produced by the LHC on all of its distributed computing grid is expected to add up to 200 PB of data each year.{{cite news|date=28 April 2008|url=http://www.techworld.com/mobility/features/index.cfm?featureid=4074&pn=2|publisher=Techworld.com|title=Parallel Internet: Inside the Worldwide LHC computing grid|first=Jon|last=Brodkin}} In total, the four main detectors at the LHC produced 13 petabytes of data in 2010.

The Tier 1 institutions receive specific subsets of the raw data, for which they serve as a backup repository for CERN. They also perform reprocessing when recalibration is necessary. The primary configuration for the computers used in the grid is based on CentOS.{{Cite web|title = Linux @ CERN: /linux/nextversion.shtml|url = http://linux.web.cern.ch/linux/nextversion.shtml|website = linux.web.cern.ch|access-date = 2016-02-07|last = CERN|archive-date = 2015-04-29|archive-url = https://web.archive.org/web/20150429023826/http://linux.web.cern.ch/linux/nextversion.shtml|url-status = dead}} In 2015, CERN switched away from Scientific Linux to CentOS.

Distributed computing resources for analysis by end-user physicists are provided by multiple federations across the Europe, Asia Pacific and the Americas.

See also

{{Clear}}

References

{{Reflist}}