Frédéric Hemmer, CERN, IT DepartmentThe LHC Computing Grid – September 2007 Wolfgang von Rüden,...
-
Upload
dulcie-newman -
Category
Documents
-
view
218 -
download
0
Transcript of Frédéric Hemmer, CERN, IT DepartmentThe LHC Computing Grid – September 2007 Wolfgang von Rüden,...
Frédéric Hemmer, CERN, IT Department The LHC Computing Grid – September 2007 Wolfgang von Rüden, CERN, IT Department
The LHC Computing Grid
Frédéric Hemmer
Deputy Head, IT Department
25 October 2007
Visit of Dr Ronald Plasterk,
Netherlands Minister of Education,
Culture and Science
2
The LHC Computing Grid – October 2007 Frédéric Hemmer, CERN, IT Department
The LHC Data Challenge
• The accelerator will be completed in 2008 and run for 10-15 years
• Experiments will produce about15 Million Gigabytes of data each year (about 20 million CDs!)
• LHC data analysis requires a computing power equivalent to ~100,000 of today's fastest PC processors
• Requires many cooperating computer centres, as CERN can only provide ~20% of the capacity
3
The LHC Computing Grid – October 2007
CERN18%
All Tier-1s39%
All Tier-2s43%
CERN12%
All Tier-1s55%
All Tier-2s33%
CERN34%
All Tier-1s66%
CPU Disk Tape
Summary of Computing Resource RequirementsAll experiments - 2008From LCG TDR - June 2005
CERN All Tier-1s All Tier-2s TotalCPU (MSPECint2000s) 25 56 61 142Disk (PetaBytes) 7 31 19 57Tape (PetaBytes) 18 35 53
4
The LHC Computing Grid – October 2007 Frédéric Hemmer, CERN, IT Department
Solution: the Grid
• Use the Grid to unite computing resources of particle physics institutes around the world
The World Wide Web provides seamless access to information that is stored in many millions of different geographical locations
The Grid is an infrastructure that provides seamless access to computing power and data storage capacity distributed over the globe
5
The LHC Computing Grid – October 2007 Frédéric Hemmer, CERN, IT Department
How does the Grid work?
• It makes multiple computer centres look like a single system to the end-user
• Advanced software, called middleware, automatically finds the data the scientist needs, and the computing power to analyse it.
• Middleware balances the load on different resources.It also handles security, accounting, monitoring and much more.
6
The LHC Computing Grid – October 2007 Wolfgang von Rüden, CERN, IT Department
View of the ATLAS detector (under construction)
150 million sensors deliver data …
… 40 million times per second
7
The LHC Computing Grid – October 2007 Wolfgang von Rüden, CERN, IT Department
8
The LHC Computing Grid – October 2007 Wolfgang von Rüden, CERN, IT Department
9
The LHC Computing Grid – October 2007 Wolfgang von Rüden, CERN, IT Department
10
The LHC Computing Grid – October 2007
LHC Computing Grid project (LCG)
• More than 140 computing centres
• 12 large centres for primary data management: CERN (Tier-0) and eleven Tier-1s
• 38 federations of smaller Tier-2 centres
• 20 countries involved
SARA-NIKHEF
Frédéric Hemmer, CERN, IT Department
11
The LHC Computing Grid – October 2007
LCG Service HierarchyTier-0 – the accelerator centre• Data acquisition & initial processing• Long-term data curation• Distribution of data Tier-1 centres
Canada – Triumf (Vancouver)France – IN2P3 (Lyon)Germany – Forschunszentrum KarlsruheItaly – CNAF (Bologna)Netherlands – NIKHEF/SARA (Amsterdam)Nordic countries – distributed Tier-1
Spain – PIC (Barcelona)Taiwan – Academia SInica (Taipei)UK – CLRC (Oxford)US – FermiLab (Illinois) – Brookhaven (NY)
Tier-1 – “online” to the data acquisition process high availability
• Managed Mass Storage – grid-enabled data service
• Data-heavy analysis• National, regional support
Tier-2 – ~120 centres in ~29 countries• Simulation• End-user analysis – batch and interactive
12
The LHC Computing Grid – October 2007 Frédéric Hemmer, CERN, IT Department
14
The LHC Computing Grid – October 2007 Frédéric Hemmer, CERN, IT Department
WLCG Collaboration• The Collaboration
• 4 LHC experiments
• ~140 computing centres
• 12 large centres (Tier-0, Tier-1)
• 38 federations of smaller “Tier-2” centres
• ~35 countries
• Memorandum of Understanding• Agreed in October 2005, now being signed
• Resources• Focuses on the needs of the four LHC experiments
• Commits resources• each October for the coming year
• 5-year forward look
• Agrees on standards and procedures
• Relies on EGEE and OSG (and other regional efforts)
15
The LHC Computing Grid – October 2007 Frédéric Hemmer, CERN, IT Department
0
5
10
15
20
25
30
35
40
45
50
Apr 06
May 06
Jun 06
Jul 06
Aug 06
Sep 06
Oct 06
Nov 06
Dec 06
Jan 07
Feb 07
Mar 07
Apr 07
May 07
Jun 07
Jul 07
Aug 07
Sep 07
Oct 07
Nov 07
Dec 07
Jan 08
Feb 08
Mar 08
Apr 08
CERN + Tier-1s - Installed and Required CPU Capacity (MSI2K)
installed target
0
5
10
15
20
25
Apr 06
May 06
Jun 06
Jul 06
Aug 06
Sep 06
Oct 06
Nov 06
Dec 06
Jan 07
Feb 07
Mar 07
Apr 07
May 07
Jun 07
Jul 07
Aug 07
Sep 07
Oct 07
Nov 07
Dec 07
Jan 08
Feb 08
Mar 08
Apr 08
CERN + Tier-1s - Installed and Required DISK Capacity (PetaBytes)
installed target
Ramp-up to First Physics Run
4x 6x
Evolution of installed capacity from April 06 to June 07Target capacity from MoU pledges for 2007 (due July07) and 2008 (due April 08)
16
The LHC Computing Grid – October 2007 Frédéric Hemmer, CERN, IT Department
Impact of the LHC Computing Grid in Europe
• LCG has been the driving force for the European multi-science Grid EGEE (Enabling Grids for E-sciencE)
• EGEE is now a global effort, and the largest Grid infrastructure worldwide
• Co-funded by the European Commission (Cost: ~130 M€ over 4 years, funded by EU ~70M€)
• EGEE already used for >20 applications, including…
Medical ImagingEducation, TrainingBio-informatics
17
The EGEE project
• EGEE– Started in April 2004, now in 2nd phase with 91 partners in 32
countries– 3rd phrase (2008-2010) in preparation
• Objectives– Large-scale, production-quality
grid infrastructure for e-Science – Attracting new resources and
users from industry as well asscience
– Maintain and further improve“gLite” Grid middleware
Partners in the Netherlands: FOM/NIKHEF, SARA, UvA, Groningen
18
Applications on EGEE
• More than 25 applications from anincreasing number of domains– Astrophysics– Computational Chemistry– Earth Sciences– Financial Simulation– Fusion– Geophysics– High Energy Physics– Life Sciences– Multimedia– Material Sciences– …..
Summary of applications report: https://edms.cern.ch/document/722132
Increasing workloads
32%
Still expect factor 5 increase for LHC experiments over next yearStill expect factor 5 increase for LHC experiments over next year
19
20
Example: EGEE Attacks Avian Flu
• EGEE used to analyse 300,000 possible potential drug compounds against bird flu virus, H5N1.
• 2000 computers at 60 computer centres in Europe, Russia, Asia and Middle East ran during four weeks in April - the equivalent of 100 years on a single computer.
• Potential drug compounds now being identified and ranked.
Neuraminidase, one of the two major surface proteins of influenza viruses, facilitating the release of virions from infected cells. Image Courtesy Ying-Ta Wu, AcademiaSinica.
21
Example:Geocluster industrial application
• The first industrial application successfully running on EGEE
• Developed by the Compagnie Générale de Géophysique (CGG) in France, doing geophysical simulations for oil, gas, mining and environmental industries.
• EGEE technology helps CGG to federate its computing resources around the globe.
22
The LHC Computing Grid – October 2007 Frédéric Hemmer, CERN, IT Department
Sustainability• Need to prepare for permanent Grid infrastructure• Ensure a high quality of service for all user communities• Independent of short project funding cycles• Infrastructure managed in collaboration
with National Grid Initiatives (NGIs)• European Grid Initiative (EGI)
23
The LHC Computing Grid – October 2007 Frédéric Hemmer, CERN, IT Department
For more information about the Grid:
www.gridcafe.org
Thank you for your kind attention!
www.cern.ch/lcg
www.eu-egee.org
www.eu-egi.org/
24
The LHC Computing Grid – October 2007
High level services and middleware perspectivesScientific results using gridApplication porting and deployment Demos