ATLAS DC2 and Rome Production Experience

32
ATLAS DC2 and Rome Production Experience LCG Database Deployment and Persistency Workshop October 19, 2005 CERN, Geneva, Switzerland Alexandre Vaniachine, David Malon (ANL)

description

ATLAS DC2 and Rome Production Experience. LCG Database Deployment and Persistency Workshop October 19, 2005 CERN, Geneva, Switzerland Alexandre Vaniachine, David Malon (ANL). Outline and Introduction. ATLAS Experience: Increased fluctuations in server load on the grid - PowerPoint PPT Presentation

Transcript of ATLAS DC2 and Rome Production Experience

Page 1: ATLAS DC2 and  Rome Production Experience

                                                                                                                                                              

ATLAS DC2 and Rome Production Experience

LCG Database Deployment and Persistency Workshop

October 19, 2005

CERN, Geneva, Switzerland

Alexandre Vaniachine, David Malon (ANL)

Page 2: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

Outline and Introduction

ATLAS Experience:Increased fluctuations in server load on the grid

Requires on-demand DB services deploymentATLAS Solutions:

Client-side: database client libraryServer-side: on-demand grid-enabled DB services

IntroductionATLAS productions run on a world-wide

federation of computational grids harnessing the power of more than twenty thousand processors

Page 3: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

ATLAS Federation of Grids

14 kCPU140 sites

5.7 kCPU49 sites

3.5 kCPU30 sites

July 20:

15 kCPU50 sites

Page 4: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

Data Workflow on the Grids

An emerging hyperinfrastructure of databases on the grid plays the dual role: both as a built-in part of the middleware (monitoring, catalogs, etc.) and as a distributed production system infrastructure orchestrating scatter-gather workflow of applications and data on the grid

To further detail the database-resident data flow on the grids ATLAS Data Challenges exercise the Computing Model to identify potential bottlenecks

Page 5: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

Databases and the GridsProduction System

LCG NorduGrid Grid3/OSG

File Transport Production DB

Non-LHC Sites ATLAS Sites

VDC Database

Sites

Sites

Cluster

Head Node Head Node

Worker Node Worker Node Worker Node

CMS Sites

RFT Database

RLS Database

TAG Database

RLS Database RLS Database

GridWorkflow

Orchestration

World-Wide Federation ofComputational

Grids

AMI Database

RB Database

Conditions DBs

Production System

LCG NorduGrid Grid3/OSG

File Transport Production DB

Non-LHC Sites ATLAS Sites

VDC Database

Sites

Sites

Cluster

Head Node Head Node

Worker Node Worker Node Worker Node

CMS Sites

RFT Database

RLS Database

TAG Database

RLS Database RLS Database

GridWorkflow

Orchestration

World-Wide Federation ofComputational

Grids

AMI Database

RB Database

Conditions DBs

Page 6: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

Expectations and Realities

Expectations Scalability achieved through replica servers deployment Database replicas will be deployed down to each Tier1 center To ease an administration burden the database server replica

installation was reduced to a one line command The needed database services level is known in advance

Realities Only the centers that has problems with access to the central

databases (because of firewalls or geographical remoteness resulting in low data throughput) deployed the replica servers

Difficulties in switching to the replica servers deployed Difficult to know the needed database services level in advance

Page 7: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

0

2000

4000

6000

8000

10000

12000

14000

Jul Aug Sep Oct Nov Dec Jan Feb Mar Apr May

Job

s/d

ay

LCG/CondorG

LCG/Original

NorduGrid

Grid3

Data Challenge 2 (long jobs period)

Data Challenge 2(short jobs period)

Rome Production (mix of jobs)

0

2000

4000

6000

8000

10000

12000

14000

Jul Aug Sep Oct Nov Dec Jan Feb Mar Apr May

Job

s/d

ay

LCG/CondorG

LCG/Original

NorduGrid

Grid3

Data Challenge 2 (long jobs period)

Data Challenge 2(short jobs period)

Rome Production (mix of jobs)

Production Rate Growth

CERNDatabase

Capacities Bottleneck

Page 8: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

First Lessons Learned

Most of the jobs request the same dataserver-side in-memory query caching eliminates hardware load

Database workload depends on the mix of jobsWith short jobs (reconstruction) the load increased

remote clients failed to connect because of their local TCP/IP socket timeouts settings

Similar to the Denial-of-Service attack condition the world-wide spread of production (see next slide)

Makes impractical to the TCP/IP socket timeouts adjustmentsThe only solution is to deploy even more servers

Page 9: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

Number of J obs

Grid3

LCG

LCG-CG

NorduGrid

Grid3

LCG

LCG-CG

NorduGrid

Spread of Grid Production

ATLAS Rome Production: 84 sites in 22 countries uibk.ac.at triumf.ca

umomtreal.ca utoronto.ca

cern.ch unibe.ch

csvs.ch golias.cz

skurut.cz gridka.fzk.de

atlas.fzk.de lcg-gridka.fzk.de

benedict.dk nbi.dk

morpheus.dk ific.uv.es

ft.uam.es ifae.es

marseille.fr cclcgcdli.in2p3.fr

clrece.in2p3.fr cea.fr

isabella.gr kfki.hu

cnaf.it lnl.it

roma1.it mi.it

ba.it pd.it

lnf.it na.it

to.it fi.it

ct.it ca.it

fe.it pd.it

roma2.it bo.it

pi.it sara.nl

nikhef.nl uio.no

hypatia.no zeus.pl

lip.pt msu.ru

hagrid.se bluesmoke.se

sigrid.se pdc.se

chalmers.se brenta.si

savka.sk ihep.su

sinica.tw ral.uk

shef.uk ox.uk

ucl.uk ic.uk

lancs.uk man.uk

ed.uk UTA.us

BNL.us BU.us

UC_ATLAS.us PDSF.us

FNAL.us IU.us

OU.us PSU.us

Hamptom.us UNM.us

UCSanDiego.us Uflorida.su

SMU.us CalTech.us

ANL.us UWMadison.us

UC.us Rice.us

Unknown

Number of jobs: 573315

U.S.

CERN

Page 10: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

Data Mining the Experience

The data-mining of the collected operations data reveals a striking feature – a very high degree of correlations between the failures: if the job submitted to some cluster failed, there is a high

probability that a next job submitted to the cluster would fail too if the submit host failed, all the jobs scattered over different

clusters will fail tooTaking these correlations into account is not yet

automated by the grid middlewareThat is why production databases and grid monitoring

data that are providing immediate feedback on the Data Challenge operations to the production operators is very important for efficient utilization of the grid capacities

Page 11: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

Increased Fluctuations

Among the database experience issues encountered is the increase in fluctuations in database servers’ workloads due to the chaotic nature of grid computations

The observed fluctuations in database access patterns are of general nature and must be addressed by grid-enabling database services

Page 12: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

Scalability Challenge

Database services capacities should be adequate for peak demand

The chaotic natureof Grid computing increases fluctuations in demand for database services: 14×statistical

0

50

100

150

200

6/20 7/4 7/18 8/1 8/15 8/29 9/12 9/26

DC2 Week

Que

ries

per

sec

ond

(dai

ly a

vera

ge)

Page 13: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

Jobs Load

No apparent correlations between jobs and database server load fluctuations

Job failures?0

100

200

300

400

11/04/04 11/14/04 11/24/04 12/04/04 12/14/04Qu

eri

es

pe

r s

ec

on

d (

da

ily a

ve

rag

e)

0

1000

2000

3000

4000

5000

6000

11/5/04 11/12/04 11/19/04 11/26/04 12/3/04 12/10/04 12/17/04

Jo

bs p

er

day

Any correlations?

0

100

200

300

0 1000 2000 3000 4000 5000 6000

Jobs per day

Qu

erie

s p

er s

eco

nd

(d

aily

avr

erag

e)

Page 14: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

ATLAS Efficiency on LCG

Efficiency fluctuates because clusters are “large error amplifiers”?

0.00%

25.00%

50.00%

75.00%

100.00%

12-Jul-04 31-Aug-04 20-Oct-04 9-Dec-04 28-Jan-05 19-Mar-05 8-May-05

Page 15: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

Power Users

The organized Rome Production was for the ATLAS biannual Physics Workshop in RomePreparations for the Rome Workshop expedited large growth

in ATLAS software use by physicistsAn increasing number of groups and individual physicists -

the Power Users - engage in a medium scale production on their local production facilities and world-wide grids

Empowered by grids, occasionally, ATLAS Power Users created bottlenecks on ATLAS and/or CERN/IT shared database resources

To coordinate database services workload ATLAS created the Power Users Club: http://uimon.cern.ch/twiki/bin/view/Atlas/PowerUsersClub

Is it only in ATLAS we got the Power Users – how do other LHC experiments deal with chaotic workload from physicists’ activities?

Page 16: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

Server Indirection

One of lessons learnt in ATLAS Data Challenges is that the database server address should NOT be hardwired in data processing transformations

The logical-physical indirection for database servers is now introduced in ATLASSimilar to the logical-physical file Replica Location

Service indirection of the grid file catalogsLogical-Physical resolution can be done against the local

file updated from Catalogue on configurable time period or directly against the Catalogue residing on Oracle DB

Page 17: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

To improve robustness of database access in a data grid environment ATLAS developed the application-side solution: database client library indirection, etc

ConnectionManagement module (by Yulia Shapiro) is adopted by POOL/CORAL

Client Library

Page 18: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

Deployment Experience

Central Deployment:Most of ATLAS current experience in production

Advanced planning for capacities requiredRemote site firewall problem

Replica deployment on Worker Node:Extensive experience in ATLAS Data Challenge 1

Replica update problemReplica deployment on the Head Node:

Use of grid commands to deploy database server replicaRequires dedicated Head Node allocation

The new most promising option: on-demand Edge Services Framework from OSG (see next slide)

Page 19: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

Joint OSG and EGEE Operations Workshop RAL UK September 27 2005

Abhishek Singh Rana and Kate Keahey OSG Edge Services Framework www.opensciencegrid.org

The Open Science Grid Consortium

ESF - Phase 1

CDFCMS ATLAS

GuestVO

ESF

SECE

Site

GT4 Workspace Service & VMM

Dynamically deployed ES Wafers for each VO

Wafer imagesstored in SE

Compute nodes and Storage nodes

Snapshot ofES Wafers

implemented asVirtual Workspaces

Page 20: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

Edge Services Framework

ATLAS is involved in ESF OSG Activity:http://osg.ivdgl.org/twiki/bin/view/EdgeServicesThe objective for the next month is to deploy a prototype of

ESF, dynamically deploying a workspace configured with a real ESF service and have it run in a scenario in which the Edge Service will normally be used

The first ATLAS grid-enabled application scheduled for deployment:mysql-gsi-0.1.0-5.0.10-beta-linux-i686 database

release built by the DASH project:http://www.piocon.com/ANL_SBIR.php

Page 21: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

Replication Experience

Since the 3D Kick-off Workshop ATLAS is using JDBC Octopus replication tools with increasing confidencehttp://uimon.cern.ch/twiki/bin/view/Atlas/DatabaseReplication

With ATLAS patches Octopus Replicator now reached the production level Used routinely for Geometry DB replication from Oracle to MySQL

Replication to SQLite was done recently with ease Octopus allows easy configuration of the datatype mapping between

different technologies which is important for the POOL/RAL applications Last month Octopus was used for the large-scale Event TAGS replication of

the Rome Physics Workshop data within CERN and Europe Replication from CERN Tier0 to US ATLAS Tier1 and BNL was also

accomplished successfully but was slower We have found that replication of the data on a large-scale and over large

distances will benefit from the increased throughput achieved by improvements on the transport-level of MySQL JDBC drivers

Page 22: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

POOL Collections: Setup

Data 38 datasets, 2.3M events from ATLAS Rome Workshop ROOT collections (1000 events each) generated from AOD ROOT collections merged into 38 collections, 1 per dataset Merged collections created as tables in Oracle and MySQL 38 collections merged into 2 master collections (1 indexed) Total size of 40 collections in Oracle ~ 3*(7.2 GB) ~ 22 GB

Databases Oracle collections used RAL collection schema and dedicated

COOLPROD integration server (configured by CERN IT/PDS). MySQL collections used non-RAL schema and dedicated

lxfs6021 server configured by ATLAS.Slide by Kristo Karr

Page 23: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

POOL Collections: ReplicationTools All collection merging and replication performed via the POOL collections utility:

CollAppend

Rates Merge of ROOT Collections from Castor (CERN) ROOT ->Oracle: 49 HzROOT->MySQL: 85 Hz. Replication of Collection to Same Database (CERN) Oracle->Oracle: 79 HzMySQL->MySQL: 127 Hz. Replication of Collection to Different Database (CERN) Oracle->MySQL: 67 Hz MySQL->Oracle: 62 Hz.Also measured replication to MySQL server at US ATLAS Tier1 at BNL Replication of Collection to Different Database (CERN to BNL) MySQL->MySQL: 8 Hz Replication of Collection to Different Database (CERN from BNL) MySQL->MySQL: 89 HzNotes: Timings are without bulk insertion, which improves rates by a factor of 3-4 according to

preliminary tests (no cache size optimization). Internal MySQLMySQL rates are 713 Hz, internal OracleOracle 800 Hz. Octopus and CollAppend rates agree.

Slide by Kristo Karr

Page 24: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

Feedback on the Current Services at CERN Tier 0

Almost a year ago the IT DB Services for Physics section was combined with Persistency Framework POOL A very positive experience through the whole period starting

with [email protected] mechanisms and now using the [email protected]

Built strong liaisons through Dirk GeppertVarious Oracle issues were resolved successfullyAll LHC experiments should be serviced on equal footingA long-in-advance planning for services is unrealistic

Oracle RAC should deliver the on-demand servicesA once-per-year workshop is not enough for efficient

information sharing between the experiments:An LCG technical board of liaisons from IT/PDS and experiments?

Page 25: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

Summary

The chaotic nature Grid computing increases fluctuations in database services demand

Grid Database Access requires changes on theClient-side:

ATLAS Client Library

Server-side: On-demand deployment: ESFScalable technology: DASH (and/or FroNTier)

Page 26: ATLAS DC2 and  Rome Production Experience

                                                                                                                                                              

Backup Slides

Page 27: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

ATLAS Efficiency on NorduGrid

Fluctuations are limited when efficiency is close to the 100% boundary

0.00%

25.00%

50.00%

75.00%

100.00%

12-Jul-04 31-Aug-04 20-Oct-04 9-Dec-04 28-Jan-05 19-Mar-05 8-May-05

Page 28: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

ATLAS Efficiency on Grid3

ATLAS achieved highest overall production efficiency on Grid3 (now OSG)

0.00%

25.00%

50.00%

75.00%

100.00%

22-Jun-04 11-Aug-04 30-Sep-04 19-Nov-04 8-Jan-05 27-Feb-05 18-Apr-05 7-Jun-05

Page 29: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

CERN Tier 0 Services FeedbackSeptember-December 2004

A spectrum of Oracle issues was addressed in the period: Continued Oracle prodDB monitoring, performance, troubleshooting

Yulia Shapiro is now helping Luc Goossens Best practices for pioneering ATLAS database application - Geometry DB:

Completed review of publishing procedures with the help of CERN IT/DB experts Another ATLAS Oracle database application DB application in development:

established CollectionsT0 DB on devdb (for T0 exercise + LCG3D testbed) Oracle client is now in the software distribution and production:

Thousands of distributed reconstruction jobs used new Oracle GeometryDB A very positive experience with [email protected] mechanisms, e.g.

emergency restoration of GeometryDB on development server very helpful in resolving Oracle “features” encountered by GeometryDB users:

vSize (V. Tsulaia), valgrind (David Rousseau, Ed Moyse), CLOB (S. Baranov) IT DB Services for Physics section is combined with Persistency Framework (POOL) Dirk Geppert is a newly established ATLAS Liaison

Page 30: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

CERN Tier 0 Services Feedback January – May 2005

Many thanks to CERN/IT FIO/DS for their excellent MySQL servers supportatlasdbdev server was serviced by vendor following disk alertatlasdbpro server is continuously up and running for 336 days

Thanks to CERN/IT ADC/PDS many Oracle performance issues with ProductionDB/ATLAS_PRODSYS on atlassg and (GeometryDB/ATLASDD ) on pdb01/atlas were investigated and improved indexing, caching, internal OCI error, deadlocks, etcservice interruption rescheduled to minimize effect on production

Page 31: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

CERN Tier 0 Services Feedback January – May 2005

Suggested improvements in the policies for the service interruptions announcements

Suggested improved information sharing on Oracle database usage experience between the LHC experiments

Requested notification when the tnsnames.ora file at CERN is changedsetup a monitoring system at BNL (Alex Undrus)

Requested support for the tnsnames.ora independence in POOL/RAL

Page 32: ATLAS DC2 and  Rome Production Experience

LCG Database Deployment and Persistency Workshop, October 17-19, 2005

Alexandre Vaniachine (ANL)

CERN Tier 0 Services Feedback June - September 2005

Many thanks to CERN/IT FIO/DS group for their excellent support of ATLAS MySQL serversIn July atlasdbpro server was serviced by vendor after

379 days of uninterrupted operations that contributed to the success of DC2 and Rome production

Later in the reporting period both operating systems and the server versions were upgraded

Thanks to CERN/IT ADC/PDS group various ATLAS Oracle database applications issues and security patches were quickly addressed and resolved: Dirk Geppert et al.