GridPP Building a UK Computing Grid for Particle Physics Professor Steve Lloyd, Queen Mary,...

Post on 13-Jan-2016

216 views 0 download

Tags:

Transcript of GridPP Building a UK Computing Grid for Particle Physics Professor Steve Lloyd, Queen Mary,...

GridPPBuilding a UK Computing Grid for Particle Physics

Professor Steve Lloyd, Queen Mary, University of London

Chair of the GridPP Collaboration Board

Steve Lloyd Public Service Summit - 22 September 2004 Slide 2

Outline

• Why? – The CERN LHC and the Data Deluge• What? - GridPP and the Grid

– What is the Grid?– Applications and Middleware– Tier-1 and Tier-2 Regional Centres

• How? - GridPP Management• Summary and Challenges

Steve Lloyd Public Service Summit - 22 September 2004 Slide 3

What is GridPP?

19 UK Universities, CCLRC (RAL & Daresbury) and CERN

Funded by the Particle Physics and Astronomy Research Council (PPARC)

GridPP1 - 2001-2004 £17m "From Web to Grid"

GridPP2 - 2004-2007 £15m "From Prototype to Production"

Steve Lloyd Public Service Summit - 22 September 2004 Slide 4

The CERN LHC

4 Large Experiments

The world’s most powerful particle accelerator - 2007

Steve Lloyd Public Service Summit - 22 September 2004 Slide 5

LHC Experiments

• > 108 electronic channels• 8x108 proton-proton collisions/sec• 2x10-4 Higgs per sec• 10 Petabytes of data a year • (10 Million GBytes = 14 Million CDs)

Searching for the Higgs Particle and exciting new Physics

Starting from this event

Looking for this ‘signature’

e.g. ATLAS

Steve Lloyd Public Service Summit - 22 September 2004 Slide 6

What is the Grid?

MIDDLEWARE

CPUDisks, CPU etc

PROGRAMS

OPERATING SYSTEM

Word/Excel

Email/Web

Your Progra

mGames

CPUCluste

r

UserInterfac

eMachine

CPUCluste

r

CPUCluste

r

Resource Broker

Information Service

Single PC

Grid

DiskCluste

r

Your Progra

m

Middleware is the Operating System of a distributed computing system

Steve Lloyd Public Service Summit - 22 September 2004 Slide 7

What is the Grid?

From this:

To this:

Steve Lloyd Public Service Summit - 22 September 2004 Slide 8

International Collaboration

• EU DataGrid (EDG) 2001-2004– Middleware Development

Project

• US and other Grid projects– Interoperability

• LHC Computing Grid (LCG)– Grid Deployment Project for

LHC • EU Enabling Grids for e-Science

in Europe (EGEE) 2004-2006– Grid Deployment Project for all

disciplines

Steve Lloyd Public Service Summit - 22 September 2004 Slide 9

The LCG Grid

Steve Lloyd Public Service Summit - 22 September 2004 Slide 10

Grid Snapshot

Steve Lloyd Public Service Summit - 22 September 2004 Slide 11

GridPP1 Areas

6/Feb/2004

£3.57m

£5.67m

£3.74m

£2.08m£1.84m

CERN

DataGrid

Tier - 1/A

ApplicationsOperations

LHC Computing Grid Project (LCG)Applications, Fabrics, Technology and Deployment

European DataGrid (EDG)Middleware Development

UK Tier-1/A Regional CentreHardware and Manpower

Grid Application DevelopmentLHC and US Experiments + Lattice QCD

Management Travel etc

Steve Lloyd Public Service Summit - 22 September 2004 Slide 12

GridPP2 Areas

July 2004

£0.75m

£2.71m

£2.88m

£0.87m

£0.69m

£2.71m

£2.89m

£1.00m

£2.40m

Tier-1/AHardware

Tier-2Operations

Applications

M/S/N

LCG-2

MgrTravel

Ops

Tier-1/AOperations

Management Travel etc

UK Tier-1/A Regional Centre Hardware

UK Tier-2 Regional Centres Manpower

UK Tier-1/A Manpower

LHC Computing Grid Project (LCG)Manpower

Middleware, Security and Networking Manpower

Grid Application Development LHC and US Experiments + Lattice QCD, Phenomenology and Generic Portal

Steve Lloyd Public Service Summit - 22 September 2004 Slide 13

Application Development

Fabric

TapeStorage

Elements

RequestFormulator and

Planner

Client Applications

ComputeElements

Indicates component that w ill be replaced

DiskStorage

Elements

LANs andWANs

Resource andServices Catalog

ReplicaCatalog

Meta-dataCatalog

Authentication and SecurityGSISAM-specific user, group , node, st at ion regis tration B bftp ‘cookie’

Connectivity and Resource

CORBA UDP File transfer protocol s - ftp, b bftp, rcp GridFTP

Mass Storage s ystems protocol se.g. encp, hp ss

Collective Services

C atalogproto co ls

Signi fi cant Event Log ger Naming Service Database ManagerC atalog Manager

SAM R es ource M an ag em entB atch Sys tems - LSF, FB S, PB S,

C ondorData Mov erJob Services

Storage ManagerJob ManagerCache ManagerRequest Manager

“Dataset Editor” “File Storage Server”“Project Master” “Station M aster” “Station M aster”

Web Python codes, Java codesCom mand line D0 Fram ework C++ codes

“Stager”“Optim iser”

CodeRepostory

Name in “quotes” is SAM-given software component name

or addedenhanced using PPDG and Grid tools

GANGA

SAMGridLattice QCD

AliEn → ARDA

CMS

BaBar

Steve Lloyd Public Service Summit - 22 September 2004 Slide 14

Middleware Development

Configuration Management

Storage Interfaces

Network Monitoring

Security

Information Services

Grid Data Management

Steve Lloyd Public Service Summit - 22 September 2004 Slide 15

UK Tier-1/A Centre

• High quality data services• National and International

Role• UK focus for International

Grid development

•700 Dual CPU•80 TB Disk•60 TB Tape (Capacity 1PB)

Grid Operations Centre

Steve Lloyd Public Service Summit - 22 September 2004 Slide 16

UK Tier-2 Centres

ScotGridDurham, Edinburgh, Glasgow NorthGridDaresbury, Lancaster, Liverpool,Manchester, Sheffield

SouthGridBirmingham, Bristol, Cambridge,Oxford, RAL PPD, Warwick

LondonGridBrunel, Imperial, QMUL, RHUL, UCL

Mostly funded by HEFCE

Steve Lloyd Public Service Summit - 22 September 2004 Slide 17

UK Core e-Science

Programme

Institutes

Tier-2 Centres

CERNLCG

EGEE

GridPP

GridPP in Context

Tier-1/A

Middleware, Security,

Networking

Experiments

GridSupportCentre

Not to scale!

Apps Dev

AppsInt

GridPP

Steve Lloyd Public Service Summit - 22 September 2004 Slide 18

Management

Collaboration Board

Project ManagementBoard

Project Leader

Project Manager

DeploymentBoard

UserBoard

Production Manager

Dissemination Officer

EGEE Leader

Tier-1 Board

Project Map

Risk Register

Tier-2 Board

CERN LCG Liaison

Deployment Team

Steve Lloyd Public Service Summit - 22 September 2004 Slide 19

Summary

BaBar

D0CDF

ATLAS

CMS

LHCb

ALICE

19 UK Institutes

RAL Computer Centre

CERN ComputerCentre

SAMGrid

BaBarGrid

LCG

EDGGANGA

EGEE

UK PrototypeTier-1/A Centre

CERN PrototypeTier-0 Centre

4 UK Tier-2 Centres

LCG

UK Tier-1/ACentre

CERN Tier-0Centre

200720042001

4 UK Prototype Tier-2 Centres

ARDA

Separate Experiments,

Resources, Multiple Accounts

'One' Production GridPrototype Grids

Steve Lloyd Public Service Summit - 22 September 2004 Slide 20

Challenges

Concorde(15 km)

CD stack with1 year LHC data(~ 20 km)

We are here(1 km)

• Scaling to full size ~10,000 → 100,000 CPUs

• Stability, Robustness etc• Security• Sharing resources (in RAE

environment!)• International Collaboration• Continued funding beyond

start of LHC!

Steve Lloyd Public Service Summit - 22 September 2004 Slide 21

Further Info

http://www.gridpp.ac.uk