[Leo Jason-Lloyd, Leonard Jason-Lloyd] an Introduc(BookSee.org)
Slide 1 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 GridPP and the Grid for Particle Physics...
-
Upload
isabella-wells -
Category
Documents
-
view
212 -
download
0
Transcript of Slide 1 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 GridPP and the Grid for Particle Physics...
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 1 of 24
GridPP and the Grid for Particle
PhysicsSteve Lloyd
Queen Mary, University of LondonNW Grid Seminar May 2006
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 2 of 24
19 UK Universities, CERN and
CCLRC (RAL & Daresbury) Funded by PPARC:
GridPP1 2001-2004 “From Web to Grid”
GridPP2 2004-2007 “From Prototype to Production”
GridPP3 2007-2011 (proposed)“From Production to
Exploitation”
Who are GridPP?
Developed a working, highly functional Grid
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 3 of 24
Why? - The CERN LHC
4 Large Experiments
The world’s most powerful particle accelerator – Starting 2007
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 4 of 24
ALICE- heavy ion collisions, to create quark-gluon plasmas- 50,000 particles in each collision
LHCb- to study the differences between matter and antimatter- detect over 100 million b and b-bar mesons each year
ATLAS- General purpose- Origin of mass- Supersymmetry- 2,000 scientists from 34 countries
CMS- General purpose- 1,800 scientists from over 150 institutes
The Experiments
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 5 of 24
7,000 tonnes 42m long22m wide22m high
2,000 Physicists150 Institutes34 Countries
The ATLAS Detector
(About the height of a 5 storey building)
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 6 of 24
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 7 of 24
Starting from this event…
We are looking for this “signature”Selectivity: 1 in 1013
Like looking for 1 person in a thousand world populations
Or for a needle in 20 million haystacks!
The LHC Data Challenge • ~100,000,000
electronic channels• 800,000,000 proton-
proton interactions per second
• 0.0002 Higgs per second
• 10 PBytes of data a year
• (10 Million GBytes = 14 Million CDs)
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 8 of 24
Solution – Build a Grid
• Share more than information• Efficient use of resources at many
institutes• Leverage over other sources of funding• Data, computing power, applications• Join local communities
Challenges:• share data between thousands of scientists with multiple interests• link major and minor computer centres• ensure all data accessible anywhere, anytime• grow rapidly, yet remain reliable for more than a decade• cope with different management policies of different centres• ensure data security• be up and running routinely by 2007
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 9 of 24
Middleware is Everything
MIDDLEWARE
CPUDisks, CPU etc
PROGRAMS
OPERATING SYSTEM
Word/Excel
Email/Web
Your Progra
mGames
CPUCluste
r
UserInterfac
eMachine
CPUCluste
r
CPUCluste
r
Resource Broker
Information Service
Single PC
Grid
DiskServer
Your Progra
m
Middleware is the Operating System of a distributed computing system
Replica CatalogueBookkeepin
g Service
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 10 of 24
GridPP Middleware Development
Workload Management
Storage Interfaces
Network Monitoring
SecurityInformation Services
Grid Data Management
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 11 of 24
International Context
LHC Computing Grid (LCG)Grid Deployment Project for
LHC
EU Enabling Grids for e-Science (EGEE) 2004-2008Grid Deployment Project for all disciplines
GridPP LCG
EGEE
GridPP is part of EGEE and LCG (currently the largest Grid in the world)
UK National Grid ServiceUK’s core production computational and data Grid
Open Science Grid (USA)Science applications from HEP to biochemistry
Nordugrid (Scandinavia)Grid Research and Development collaboration
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 12 of 24
What you need to use the Grid
1. Get a digital certificate (UK Certificate Authority)
2. Join a Virtual Organisation (VO)
3. Get access to a local User Interface Machine (UI) and copy your files and certificate there
Authentication – who you are
Authorisation – what you are allowed to do
4. Write some Job Description Language (JDL) and scripts to wrap your programs
############# HelloWorld.jdl #################Executable = "/bin/echo";Arguments = "Hello welcome to the Grid ";StdOutput = "hello.out";StdError = "hello.err";OutputSandbox = {"hello.out","hello.err"};#########################################
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 13 of 24
How it works
Job Descriptio
n
UI Machine
Resource Broker
Input SandboxScript you
want to run
Other files (Job
Options, Source...)
Storage Element
Compute Element
Storage Element
Grid
Proxy Certificate
Output SandboxOutput
files (Plots, Logs...)
Input Data
Output Data
Job
Output Sandbo
x
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 14 of 24
Tier StructureTier 0
Tier 1National centres
Tier 2Regional groups
Institutes
Workstations
Offline farm
Online system
CERN computer centre
RAL,UK
ScotGrid NorthGridSouthGrid London
FranceItalyGermanyUSA
Glasgow Edinburgh Durham
Useful model for Particle Physics but not necessary for others
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 15 of 24
UK Tier-1 Centre at RAL
• High quality data services• National and International Role• UK focus for International Grid
development•1000 Dual CPU•200 TB Disk•220 TB Tape
(Capacity 1PB)Grid Operations Centre
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 16 of 24
UK Tier-2 Centres
ScotGridDurham, Edinburgh, Glasgow NorthGridDaresbury, Lancaster, Liverpool,Manchester, Sheffield
SouthGridBirmingham, Bristol, Cambridge,Oxford, RAL PPD
LondonBrunel, Imperial, QMUL, RHUL, UCL
Mostly funded by HEFCE
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 17 of 24
The LCG Grid Status Worldwide
182 Sites
23,438 CPUs
9.2 PB Disk
2,200 Years of CPU time
UK
21 Sites
4,482 CPUs
180 TB Disk
593 Years of CPU time
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 18 of 24
Who else can use a Grid?• Astronom
y
• Healthcare
• Bioinformatics
• Gaming
• Engineering
• Commerce
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 19 of 24
“UK contributes to EGEE's battle with malaria”
BioMedSuccesses/Day 1107Success % 77%
WISDOM (Wide In Silico Docking On Malaria)
The first biomedical data challenge for drug discovery, which ran on the EGEE grid production service from 11 July 2005 until 19 August 2005.
GridPP resources in the UK contributed ~100,000 kSI2k-hours from 9 sites
Number of Biomedical jobs processed by country
Normalised CPU hours contributed to thebiomedical VO for UK sites, July-August 2005
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 20 of 24
"GridPP has been developed to help answer questions about the conditions in the Universe just after the Big Bang," said Professor Keith Mason, head of the Particle Physics and Astronomy Research Council (PPARC)."But the same resources and techniques can be exploited by other sciences with a more direct benefit to society."
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 21 of 24
GridPP and Industry• Use of our Grid• Security tools• GridSite Secure Web
Toolkit• R-GMA Information
System• APEL Accounting tools• Help getting PPARC
funding
What We Have to Offer:
Current Involvement:
• HP are sponsoring a joint project with GridPP at Bristol.
• GridPP has a collaboration with IBM through ScotGrid and R-GMA.
• Specific sites also have close relationships with various industrial suppliers.
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 22 of 24
Security Tools and Gridsite
• Set of extensions to Apache Webserver• Gridsite identifies users to websites by their digital
certificates
GridSite brings Grid Security to Websites
Security is Paramount• UK Certificate Authority • Users must have digital User Certificates (can be revoked)• Users must be part of a ‘Virtual Organisation’ (VOMS
Management)• All Grid computers must have Host Certificates (can be revoked)• Security policies agreed between sites
• Allows remote editing by authorised users without login access to host servers
• GridSiteWiki is an extension• GridSite is open source (http://www.gridsite.org/)
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 23 of 24
Future Challenges
(Ex-)Concorde(15 km)
CD stack with1 year LHC data(~ 20 km)
We are here(4 km)
• Scaling to full size ~20,000 → 100,000 CPUs
• Stability, Robustness etc• Security (Hackers Paradise!)• Sharing resources (in RAE
environment!)• International Collaboration• Increased Industrial take-up• Spread beyond Science• Continued funding beyond start of
LHC!
Steve Lloyd NW Grid Seminar - 11 May 2006 Slide 24 of 24
Further Info …
http://www.gridpp.ac.uk
RSS News feed