Center for Computational Sciences O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY...

15
Center for Computational Sciences OAK RIDGE NATIONAL LABORATORY U. S. DEPARTMENT OF ENERGY Vision for OSC Computing and Computational Sciences http:// www.ccs.ornl.gov/ Thomas Zacharia Associate Laboratory Director Computing and Computational Sciences Oak Ridge National Laboratory Earth Simulator Rapid Response Meeting May 15-16, 2002

Transcript of Center for Computational Sciences O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY...

Page 1: Center for Computational Sciences O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Vision for OSC Computing and Computational Sciences

Cen

ter

for

Co

mp

uta

tio

nal

Sci

ence

s

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Vision for OSC Computing and Computational Sciences

http://www.ccs.ornl.gov/

Thomas ZachariaAssociate Laboratory Director

Computing and Computational SciencesOak Ridge National Laboratory

Earth Simulator Rapid Response Meeting May 15-16, 2002

Page 2: Center for Computational Sciences O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Vision for OSC Computing and Computational Sciences

Cen

ter

for

Co

mp

uta

tio

nal

Sci

ence

s

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Charge from Dr. Orbach

• Review “. . . current state of the national computer vendor community relative to high performance computing”

• “. . . Vision for what realistically should be accomplished in the next five years within the Office of Science in high performance computing”

Page 3: Center for Computational Sciences O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Vision for OSC Computing and Computational Sciences

Cen

ter

for

Co

mp

uta

tio

nal

Sci

ence

s

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Dr. Orbach’s Vision for OSC ComputingStatement to ASCAC Committee, May 8, 2002

• “… there is a centrality of computation in everything that we do”

• “… large scale computation is the future of every program in the Office of Science”

• “… we want to have our own computing program in non-defense computational science”

Page 4: Center for Computational Sciences O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Vision for OSC Computing and Computational Sciences

Cen

ter

for

Co

mp

uta

tio

nal

Sci

ence

s

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

FY 03 Budget Request for OSC Computing Considerably Lower

than Required to Meet Goals

0

100,000,000

200,000,000

300,000,000

400,000,000

500,000,000

600,000,000

700,000,000

800,000,000

900,000,000

DOE-SC NNSA NSF

Fiscal Year

Bu

dg

et D

oll

ars

Page 5: Center for Computational Sciences O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Vision for OSC Computing and Computational Sciences

Cen

ter

for

Co

mp

uta

tio

nal

Sci

ence

s

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

As Fraction of Total Budget, OSC is Half NNSA and NSF and Needs Significant

Increase to Meet Goals

0

2

4

6

8

10

12

DOE-SC NNSA NSF

Co

mp

uti

ng

Bu

dg

et /

To

tal

Bu

dg

et (

%)

Page 6: Center for Computational Sciences O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Vision for OSC Computing and Computational Sciences

Cen

ter

for

Co

mp

uta

tio

nal

Sci

ence

s

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Earth Simulator has Heightened Urgency for Infrastructure Strategy for

Scientific ComputingCritical Steps:• Invest in critical software

with integrated science, and computer science development teams

• Deploy scientific computing hardware infrastructure in support of “large scale computation”– Cray, HP, IBM, SGI– IBM is the largest US

installation• Develop new initiative to

support advanced architecture research

Top 500 SupercomputersUS has been #1 in 12 of 19 lists

A concerted effort will be required to regain US leadership in high performancecomputing. The LINPACK benchmark generally overestimates the effectivenessof an architecture for applications such as climate by a substantial factor. Stability and reliability are also important system properties.

Page 7: Center for Computational Sciences O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Vision for OSC Computing and Computational Sciences

Cen

ter

for

Co

mp

uta

tio

nal

Sci

ence

s

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Invest in Critical Software with Integrated Science and Computer Science

Development Teams SciDAC: a Good Start Towards Scientific Computing Software

• Scientific Applications– Climate Simulation– Computational Chemistry– Fusion – 5 Topics– High Energy Nuclear

Physics – 5 Topics• Collaboratories

– Four Projects• Middleware & Network

Research– Six Projects

• Computer Science

– Scalable Systems Software

– Common Component Architecture

– Performance Science and Engineering

– Scientific Data Management

• Applied Mathematics

– PDE Linear/Nonlinear Solvers and Libraries

– Structured Grids/AMR

– Unstructured Grids

Dave Bader, SciDAC PI Meeting, Jan 15, 2002, Washington DC

Page 8: Center for Computational Sciences O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Vision for OSC Computing and Computational Sciences

Cen

ter

for

Co

mp

uta

tio

nal

Sci

ence

s

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Deploy Scientific Computing Hardware Infrastructure to Support

“Large-Scale Computation”• Provide most effective and efficient computing resources for a set

of scientific applications• Serve as focal point for scientific research community as it adapts

to new computing technologies• Provide organizational framework needed for multidisciplinary

activities– Addressing software challenges requires strong, long term

collaborations among disciplinary computational scientists, computer scientists, and applied mathematicians

• Provide organizational framework needed for development of community codes– Implementing many scientific codes requires wide range of disciplinary

expertise

• Organizational needs will continue to grow as computers advance to petaflops scale

Dave Bader, SciDAC PI Meeting, Jan. 15, 2002, Washington, DC

Page 9: Center for Computational Sciences O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Vision for OSC Computing and Computational Sciences

Cen

ter

for

Co

mp

uta

tio

nal

Sci

ence

s

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Earth Simulator has Widened Gap with DOE Scientific Computing Hardware Infrastructure

Top left: comparison between ES and SC resources – highlights widening gap between SC capabilities and others

Top right: comparison between ES and US resources of comparable peak performance – highlights architectural difference and need for new initiative to close the gap

Right: comparison between ES and US resources of comparable cost

1,000

3,000

5,000

7,000

Earth Simulator SEABORG CHEETAH

Sim

ulat

ions

yea

rs/d

ay

WideningGap

1,000

3,000

5,000

7,000

Earth Simulator POWER4 H+ (40TFlops) Power5 (50 TFlops)

Sim

ulat

ions

yea

rs/d

ay TechnologyGap

0

2,000

4,000

6,000

8,000

10,000

Earth Simulator POWER4 H+ (3*40TFlops) Power5 (3*50 TFlops)

Sim

ulat

ions

yea

rs/d

ay

Page 10: Center for Computational Sciences O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Vision for OSC Computing and Computational Sciences

Cen

ter

for

Co

mp

uta

tio

nal

Sci

ence

s

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Possible U.S. Response in the Near Term for Increased Computing Capacity

• 40 TFlops Peak• 5120 Vector Processors• 8 GFlops Processor• 8 Processors per Node• $500 M Procurement• $50M/yr Maintenance• Limited Software Investment to date• Significant Ancillary Impact on

Biology, Nanoscience, Astrophysics, HENP, Fusion

• 40 TFlops Peak• 5120 Power5 Processors• 8 GFlops Processor• 64 Processors per Node• $100 M Procurement• $10M/yr Maintenance• SciDAC Investment in Computational

Science and related ISICs• Significant Ancillary Impact on

Biology, Nanoscience, Astrophysics, HENP, Fusion

Earth Simulator US Alternative

Page 11: Center for Computational Sciences O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Vision for OSC Computing and Computational Sciences

Cen

ter

for

Co

mp

uta

tio

nal

Sci

ence

s

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Best Performance of High Resolution Atmospheric Model

Inter-node bandwidth (Mb/s)

Performance of Hi-Resolution Atmospheric Model

105

104

103

102

101

102 103 104 105

Earth Simulator (2560)

AlphaES45 (2048)

AlphaES40 (256)

SP3 WHII (512)T3E (512)

GF

lop

s

Page 12: Center for Computational Sciences O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Vision for OSC Computing and Computational Sciences

Cen

ter

for

Co

mp

uta

tio

nal

Sci

ence

s

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

1996 1997 1998 1999 2000 2001 2002 2003 2004 2005

Year

100

1000

10000

100000

1000000D

olla

rs/G

Flo

ps

C/L/D

ASCI

T3EBeowulfsCOTSJPL

QCDSPColumbia QCDOC

Columbia/IBM

Blue Gene/L

ASCI Blue

ASCI White ASCI Compaq

Develop New Initiative to Support Advanced Architecture:

BlueGene Offers Possible Option

Page 13: Center for Computational Sciences O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Vision for OSC Computing and Computational Sciences

Cen

ter

for

Co

mp

uta

tio

nal

Sci

ence

s

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

QCDSP (600GF based on Texas Instruments DSP C31)– Gordon Bell Prize for Most Cost Effective Supercomputer in '98– Columbia University Designed and Built– Optimized for Quantum Chromodynamics (QCD)– 12,000 50MF Processors– Commodity 2MB DRAM

QCDOC (20TF based on IBM System-on-a-Chip)– Collaboration between Columbia University and IBM Research– Optimized for QCD– IBM 7SF Technology (ASIC Foundry Technology)– 20,000 1GF processors (nominal)– 4MB Embedded DRAM + External Commodity DDR/SDR SDRAM

BlueGene L/D (180TF based on IBM System-on-a-Chip)– Designed by IBM Research in IBM CMOS 8SF Technology– 64,000 2.8GF processors (nominal)– 4MB Embedded DRAM + External Commodity DDR SDRAM

BlueGene Architecture is a (more) General Purpose Machine that builds on QCDOC

Page 14: Center for Computational Sciences O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Vision for OSC Computing and Computational Sciences

Cen

ter

for

Co

mp

uta

tio

nal

Sci

ence

s

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Host System:– Diagnostics, booting, archive– Application dependent requirements

System Organization (conceptual)

File Server Array– ~ 500 RAID PC servers– Gb Ethernet and/or Infiniband– Application dependent requirements

BlueGene/L Processing Nodes81920 Nodes– Two major partitions– 65536 nodes production– Platform (256 TFlops peak)– 16384 nodes partitioned into code

development platforms

Host Computer

Top View of system

50 Feet

Page 15: Center for Computational Sciences O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Vision for OSC Computing and Computational Sciences

Cen

ter

for

Co

mp

uta

tio

nal

Sci

ence

s

OAK RIDGE NATIONAL LABORATORYU. S. DEPARTMENT OF ENERGY

Summary• Continue investment in critical software with integrated

science, and computer science development teams

• Deploy scientific computing hardware infrastructure in support of “large scale computation”

• Develop new initiative to support advanced architecture research

• Develop a bold new facilities strategy for OSC computing

• Increase OSC computing budget to support outlined strategy

Without sustained commitment to scientific computing,

key computing and computational sciences capabilities,

including personnel, will erode beyond recovery.