TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 ·...

151
Trinity Pla+orm Introduc2on and Usage Model ACES Team August 2015 LAUR1526834

Transcript of TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 ·...

Page 1: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Pla+orm  Introduc2on  and  Usage  Model  

ACES Team  

August  2015  

LA-­‐UR-­‐15-­‐26834  

Page 2: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Topics  

•  Introduc9on  to  Sierra  (LLNL)  -­‐  NDA  •  Trinity  status  and  9me-­‐line  •  Trinity  plaGorm  update  •  Introduc9on  to  the  Trinity  Usage  Model  •  User  ques9ons/feedback      

2  

Page 3: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Detailed  Agenda  

Time   Topic  

5   Mee9ng  introduc9on  &  goals  (Manuel  Vigil)  

20   Sierra  Introduc9on,  NDA-­‐Lab  MOWs  only  (Robin  Goldstone,  LLNL)  

Trinity  Pla+orm  &  Project  Status  

10   Trinity  project  9meline  w/  ATCC  process  (Manuel  Vigil)  

10   Trinity  plaGorm  overview  &  status  including  classified/unclassified  resources  (ScoZ  Hemmert)  

10   File  systems  (BreZ  KeZering,  Ruth  Klundt)  

Trinity  Usage  Model  

5   Overview  of  Trinity  Usage  Model  (Jeff  Johnson,  Karen  Haskell)  

10   User  support  and  Geang  Started  –  Accessing,  help,  logging  in  (Ben  Santos,  Rob  Cunningham,  Lisa  Ice)  

10   Programming  environment  (Jennifer  Green,  David  Shrader,  Kevin  Pedrea,  Mahesh  Rajan)  

15   Burst  Buffer  (Cornell  Wright)   3  

Time   Topic  

10   Break  

10   Applica9on  Readiness  (Cornell  Wright,  Joel  Stevenson)  

15   Center  of  Excellence  (Tim  Kelley  (LANL),  Hai  Ah  Nam  (LANL),  Rob  Hoekstra(SNL),  Mike  Glass  (SNL),  (LLNL))  

10   Campaign  storage  (Kyle  Lamb,  BreZ  KeZering)  

10   Visualiza9on  &  Data  analysis  (Laura  Monroe,  David  Karelitz,  Warren  Hunt)  

10   Data  Movement  &  Archive  (BreZ  Hollander,  Susie  McRee)  

30   User  forum  &  Trinity  ques2ons  (all  speakers)  

5   Wrap  up  (Manuel  Vigil)  

Page 4: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Usage  Model  -­‐  Timeline  

Manuel Vigil August 2015

 

LA-­‐UR-­‐15-­‐26470  

Page 5: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

ASC  PlaGorm  Timeline  A

dvan

ced

Tech

nolo

gy

Sys

tem

s (A

TS)

Fiscal  Year  

‘12   ‘13   ‘14   ‘15   ‘16   ‘17  

Use  Re9re  

‘19  ‘18   ‘20  

Com

mod

ity

Tech

nolo

gy

Sys

tem

s (C

TS)

Procure&  Deploy  

Cielo  (LANL/SNL)  

Sequoia    (LLNL)  

ATS  1  –  Trinity    (LANL/SNL)  

ATS  2  –    Sierra  (LLNL)  

ATS  3  –    (LANL/SNL)  

Tri-­‐lab  Linux  Capacity  Cluster  II  (TLCC  II)  

CTS  1  

CTS  2  

‘21  

System  Delivery  

Jan.  2015  

ATS  bars  (start  represents  CD-­‐1  signed  off,  RFP  released,  and  program  budget  alloca9ons)  CTS  bars  (start  represents  RFP  released  and  program  budget  alloca9ons)  

Page 6: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Overview  of  Trinity  Award  

•  Subcontractor  –  Cray,  Inc.    

•  Firm  Fixed  Price  Subcontract:    –  Trinity  PlaGorm  (including  File  System)  –  Burst  Buffer  –  2  Applica9on  Regression  Test  Systems  –  1  System  Development  Test  System  –  On-­‐site  System  and  Applica9on  Analysts  –  Center  of  Excellence  for  Applica9on  Transi9on  Support  –  Advanced  Power  Management  –  Trinity  System  Maintenance  

•  Acquired  and  managed  by  the  New  Mexico  Alliance  for  Compu9ng  at  Extreme  Scale  

6  

Page 7: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Project  Drivers    

•  Sa9sfy  the  mission  need  for  more  capable  plaGorms  –  Trinity  is  designed  to  support  the  largest,  most  demanding  ASC  applica9ons  –  Increases  in  geometric  and  physics  fideli9es  while  sa9sfying  analysts’  9me-­‐to-­‐  

solu9on  expecta9ons  –  Foster  a  compe99ve  environment  and  influence  next  genera9on  architectures  

in  the  HPC  industry  •  Trinity  is  enabling  new  architecture  features  in  a  produc9on  compu9ng  

environment  –  Trinity’s  architecture  will  introduce  new  challenges  for  code  teams:  transi9on  

from  mul9-­‐core  to  many-­‐core,  high-­‐speed  on-­‐chip  memory  subsystem,  wider  SIMD/vector  units  

–  Tightly  coupled  solid  state  storage  serves  as  a  “burst  buffer”  for  checkpoint/restart  file  I/O  &  data  analy9cs,  enabling  improved  9me-­‐to-­‐  solu9on  efficiencies  

–  Advanced  power  management  features  enable  measurement  and  control  at  the  system,  node,  and  component  levels,  allowing  explora9on  of  applica9on  performance/waZ  and  reducing  total  cost  of  ownership  

•  Mission  Need  Requirements  are  primarily  driving  memory  capacity  –  Over  2  PB  of  aggregate  main  memory  

7  

Page 8: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Pla+orm  

•  Trinity  is  a  single  system  that  contains  both  Intel  Haswell  and  Knights  Landing  processors  –  Haswell  par99on  sa9sfies  FY16  mission  needs  (well  suited  to  exis9ng  

codes).  –  KNL  par99on  delivered  in  FY16  results  in  a  system  significantly  more  

capable  than  current  plaGorms  and  provides  the  applica9on  developers  with  an  aZrac9ve  next-­‐genera9on  target  (and  significant  challenges)  

–  Aries  interconnect  with  the  Dragonfly  network  topology    

•  Based  on  mature  Cray  XC30  architecture  with  Trinity  introducing  new  architectural  features  –  Intel  Knights  Landing  (KNL)  processors  –  Burst  Buffer  storage  nodes  –  Advanced  power  management  system  sosware  enhancements   8  

Page 9: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Advanced  Technology  System  Scheduling  Governance  Model  Advanced  Technology  Compu9ng  Campaigns  (ATCCs)  

9  

•  The  goal  of  this  model  is  to  effec9vely  allocate  and  schedule  AT  compu9ng  resources  among  all  three  NNSA  laboratories  for  weapons  deliverables  that  merit  priority  on  this  class  of  resource  

•  Objec9ves  of  this  governance  model  are  to:  –  Ensure  that  AT  system  resources  are  allocated  on  a  priority-­‐driven  basis  

according  to  SSP  Program  requirements;  –  U9lize  ASC  AT  systems  for  the  most  demanding  workload  categories,  for  which  they  

were  designed  and  procured;  and  –  Support  the  role  of  AT  systems  to  prepare  ASC  resources  (including  our  people,  our  

applica9ons,  and  our  compu9ng  environments),  for  significant  changes  in  future  computer  and  system  architectures.  

•  Alloca9ons  –  Each  Lab  has  1/3  alloca9on  on  each  AT  system  and  will  hold  an  internal  process  for  each  

six-­‐month  campaign  to  review/priori9ze  proposed  projects.    

•  Repor9ng  and  Accountability    –  Host  labs  will  prepare  ATCC  u9liza9on  reports  on  a  monthly  basis.    –  Once  a  year  the  host  lab  briefs  NNSA  HQ  staff  on  the  simula9ons  that  were  carried  out  

in  the  previous  two  ATCCs  on  the  AT  system.    

Page 10: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

July  1  April  1   Oct  1  Jan  1  

Security  Accredita2on  

2015  

ASC  L2  Milestone  

Trinity  Phase  1  ATCC-­‐1  

Phase  1  System  Build  and  Delivery  

July  1  April  1   Oct  1  Jan  1  2016  

CD4  Approval  

Applica9on  code  transi9on  

Physical  Infrastructure  Integra2on  

Phase  1  Acceptance  

Phase  2  System  Build  and  Delivery  

Final  Acceptance  

ACES  System  Integra2on  and  System  tes2ng  with  Science  

codes  

Trinity  Final  System  in    Classified  Network  

Classified  network  integra2on  and  cyber  

tes2ng  

ASC  L2  Milestone  

Final  System  Integra2on  

and  Acceptance  Tes2ng  

Vendor  Integra2on  

Trinity  PlaGorm  Schedule  Highlights  2015-­‐2016    

Slide  10  

Page 11: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Update  

Slide  11  

•  Trinity  is  the  first  instan9a9on  of  the  ASC’s  ATS  plaGorm  strategy  •  SCC  Physical  Infrastructure  Integra9on  completed  •  Trinity  Testbeds  delivered  –  Available  June  2015  •  Trinity  Phase  1  delivered  •  Trinity  Center  of  Excellence  ac9vi9es  ongoing  at  SNL,  LLNL,  and  LANL  •  Trinity  System  on  schedule    •  Next  6+  months  

–  Trinity  vendor  integra9on,  Aug-­‐Oct  2015  –  Trinity  Usage  Model  presenta9on  to  LANL,  SNL,  LLNL  –  August  2015  –  Acceptance  Tes9ng  scheduled  for  December  2015  

•  Cray’s  Rhine/Redwood  so\ware  is  in  schedule  cri2cal  path  –  Unclassified  Science  Runs  for  system  stabiliza9on,  Jan-­‐Feb,  2016  

•  Rhine-­‐Redwood  system  sosware  •  Burst  Buffer  Usage  by  applica9ons  

Page 12: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Pla+orm  Overview  

Scott Hemmert  

August  2015  

SAND2015-­‐6938  PE  

Page 13: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Architecture  

13  

Compute  (Intel  “Haswell”)  9436  Nodes  (~11  PF)  

Compute  (Intel  Xeon  Phi)  >9500  Nodes  

~40  PF  Total  Performance  and  2.1PiB  of  Total  Memory  

Gateway  Nodes   Lustre  Routers  (222  total,  114  Haswell)  

Burst  Buffer  (576  total,  300  Haswell)  

2x  648  Port  IB  Switches  

39  PB  File  System  

39  PB  File  System  

78  PB  Usable  ~1.6  TB/sec  –  2  Filesystems  

Cray  Sonexion©  Storage  System  

Cray  Development  &  Login  Nodes  

40  GigE  Network  

GigE  Network  

GigE  

40  GigE  

FDR  IB  

3.69  PB  Raw  3.28  TB/s  BW  

Page 14: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

14  

x  00   01   02   03   04   05   06   07  

10   11   12   13   14   15   16   17  

20   21   22   23   24   25   26   27  

30   31   32   33   34   35   36   37  

40   41   42   43   44   45   46   47  

50   51   52   53   54   55   56   57  

Green  Links  (15x1)  To  15  Other  

Blades  in  Chassis,  1  Tile  Each  Link  

(5.25  GB/s  per  link)  

Black  Links  (5x3)  To  5  Other  

Chassis  in  Group,  3  Tiles  Each  Link  

(15.75  GB/s  per  link)  

Blue  Links  (10x1)  To  Other  Groups,  10  Global  Links  

(4.7  GB/s  per  link)  

Cray  Aries  Blade   1.  Chassis  

16  Blades  Per  Chassis  16  Aries,  64  Nodes  

All-­‐to-­‐all  Electrical  Backplane  

2.  Group  

6  Chassis  Per  Group  96  Aries,  384  Nodes  

Electrical  Cables,  2-­‐D  All-­‐to-­‐All  

3.  Global  

G0   G1   G2   G3   G4  

Up  to  241  Groups  Up  to  23136  Aries,  92544  Nodes  

Op2cal  Cables,  All-­‐to-­‐All  between  Groups    

Gemini:    2  nodes,      62.9  GB/s  rou9ng  bw  Aries  4  nodes,  204.5  GB/s  rou9ng  bw    Aries  has  advanced  adap9ve  rou9ng  

x  

NIC  A  

Host  A  

NIC  B  

Host  B  

NIC  C  

Host  C  

NIC  D  

Host  D  

Cray  Aries  Interconnect  

Page 15: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Haswell  Compute  Node  

Haswell  16  Core  588  GF  

PCIe-­‐3  x16

 

Node  

DDR4  DIMM  16GBs  

DDR4  DIMM  16GBs  

DDR4  DIMM  16GBs  

DDR4  DIMM  16GBs  

 QPI  

Southbridge  Chip  

 QPI  

Haswell  16  Core  588  GF  

DDR4  DIMM  16GBs  

DDR4  DIMM  16GBs  

DDR4  DIMM  16GBs  

DDR4  DIMM  16GBs  

128  GB,  2133  

15  

Page 16: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  KNL  Compute  Node  Single  Socket  -­‐  Self  Hosted  Node  

PCIe-­‐3  x16

   Node  

Single  Socket  -­‐  Self  Hosted  Node  

96  GB  DDR4  2400  –Memory  

Southbridge  Chip  

DMI2  

~PCIe-­‐2    x4  

16  GB  On  Pkg  Mem  

>3  TF  KNL  (72  cores)  

16  

Page 17: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Test  Bed  Systems  

•  Gadget  –  Sosware  Development  Testbed  •  Applica9on  Regression  Testbeds  

– Configura9on  •  100  Haswell  Compute  Nodes  •  720  TB  /  15  GB/s  Sonexion  2000  Filesystem  •  6  Burst  Buffer  Nodes  

– Trini9te  •  LANL  Yellow  Network  

– Mutrino  •  Sandia  SRN  Network  

17  

Page 18: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

ACES File System Team August 2015

LA-­‐UR-­‐15-­‐24093  

Trinity  File  Systems  Introduc2on  and  Usage  Model  

Page 19: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Thanks  to  all  those  who  helped  

•  LANL:  Kyle  Lamb,  Mike  Mason,  Terri  Morris,  &  Alfred  Torrez  

•  SNL:  Ruth  Klundt  •  Cray:  Chris  Brady,  Kim  Schumann,  &  Mark  Swan  

Page 20: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  file  systems  &  their  purposes  

•  Home  space  –  NFSv3  for  user’s  own  files  –  Use  judiciously  for  wri9ng  applica9on  results  –  Backed-­‐up  and  quotas  

•  Project  space  –  NFSv3  to  share  such  files  as  source  code,  common  data  sets,  binaries,  etc.  –  Backed-­‐up  and  quotas  

•  Packages  space  –  NFSv3  to  share  tools  and  applica9ons,  especially  3rd  Party,  that  are  useful  to  many  code  teams  but  made  

available  by  one  person  –  Backed-­‐up  and  quotas  

•  Netscratch  space  –  NFSv3  for  faster  serial  access  to  restart,  graphics,  9me-­‐step  data,  etc.  –  Not  backed-­‐up  and  quotas  

•  Scratch  space  –  Lustre  for  fast  parallel  access  to  restart,  graphics,  9me-­‐step  data,  etc.    –  Not  backed-­‐up,  but  purged  

•  UDSL  (User  Dynamic  Shared  Library)  space  –  NFSv3  for  user/code-­‐specific  libraries  read  by  applica9ons  at  launch  –  Read-­‐only  from  computes;  write  from  login  nodes  

•  System  library  and  image  space  –  NFSv3  for  OS  images  and  OS/installed  product  libraries  –  Cfengine-­‐managed  

Page 21: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Things  to  know  about  Trinity’s  Lustre  PFSes  

•  Capaci9es  (usable)  –  Trinity:  39  PB  per  PFS  –  Trini9te  &  Mutrino:  720  TB  –  Gadget:  360  TB  

•  Sustained  bandwidth  (7.5  GB/s/SSU)  –  Trinity:  810  GB/s  per  PFS  (108  SSUs/PFS)  –  Trini9te  &  Mutrino:  15  GB/s  (2  SSUs)  –  Gadget:  7.5  GB/s  (1  SSU)  

•  OST  counts  (2  per  SSU)  –  Trinity:  216  per  PFS  –  Trini9te  &  Mutrino:  4  –  Gadget:  2  

•  Use  GridRAID  

Page 22: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Cray    XC-­‐LC  

Trinity  file  systems  

Boot  RAID  

UDSL  RAID  

Customer  NFS  Filers  

System  Libs  &  Image  

User  Libs  

6  Login  Nodes  

HDDs  

Home,  Projects,  Netscratch  

40  GigE  Fibre  Channel  

2  UDSL  Nodes  

2  Boot  Nodes  

Haswell  Computes  

32  DSL/DVS  Nodes  

Read  /  Write  

ssh  

6  NFS  Nodes  

Read  Only  

KNL  Computes  

Page 23: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  parallel  file  systems  

FDR  IB  

Lustre1  (39  PB  Usable)  

6  Login  Nodes   ssh  

Cray    XC-­‐LC  Haswell  Computes   KNL  Computes  

Lustre2  (39  PB  Usable)  

222  LNet  Routers  

Page 24: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Cray    XC-­‐LC  

Trini9te  file  systems  

Boot  RAID  

Customer  NFS  Filers  

System  Libs  &  Image  

1  Login  Node  

HDDs  

Home,  Projects,  Netscratch  

40  GigE  Fibre  Channel  

2  Boot  Nodes  

Haswell  Computes  

ssh  

2  NFS  Nodes  

KNL  Computes  

Page 25: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trini9te  parallel  file  systems  

FDR  IB  

Lustre  (720  TB  Usable)  

1  Login  Node   ssh  

Cray    XC-­‐LC  Haswell  Computes   KNL  Computes  

3  LNet  Routers  

Page 26: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Mutrino  file  systems  Cray    XC-­‐LC  

Boot  RAID  

Customer  NFS  Filers  

System  Libs  &  Image  

1  Login  Node  

HDDs  

Home,  Projects  

40  GigE  Fibre  Channel  

2  Boot  Nodes  

Haswell  Computes  

ssh  

2  NFS  Nodes  

KNL  Computes  

Page 27: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Mutrino  parallel  file  systems  

FDR  IB  

Lustre  (720  TB  Usable)  

1  Login  Node   ssh  

Cray    XC-­‐LC  Haswell  Computes   KNL  Computes  

3  LNet  Routers  

Page 28: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Cray    XC-­‐LC  

Gadget  file  systems  

Boot  RAID  

UDSL  RAID  

Customer  NFS  Filers  

System  Libs  &  Image  

User  Libs  

1  Login  Node  

HDDs  

Home,  Projects,  Netscratch  

40  GigE  Fibre  Channel  

2  UDSL  Nodes  

2  Boot  Nodes  

Haswell  Computes  

4  DSL/DVS  Nodes  

Read  /  Write  

ssh  

1  NFS  Node  

Read  Only  

KNL  Computes  

Page 29: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Gadget  parallel  file  systems  

FDR  IB  

Lustre  (360  TB  Usable)  

1  Login  Node   ssh  

Cray    XC-­‐LC  Haswell  Computes   KNL  Computes  

2  LNet  Routers  

Page 30: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Measurements  suggest  5-­‐16  I/O  streams/OST  for  op9mal  write  performance  

4344  

8550  

12772  

13565  

13594  

13620  

13634  

13746  

13793  

13762  

13699  

13758  

13727  

13728  

13970  

13880  

3607  

7030  

10545  

12549   13493  

13452  

13381  

13657  

13582  

13532  

13543  

13426  

13419  

13448  

13670  

13327  

0  

2000  

4000  

6000  

8000  

10000  

12000  

14000  

16000  

1   2   3   4   5   6   7   8   9   10   11   12   13   14   15   16  

MiB/s  

Streams/OST  

IOR  Op2mal  Write  

Trini9te  Op9mal  Write  1PPN  

Trini9te  Op9mal  Write  2  PPN  

Page 31: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Measurements  suggest  10-­‐14  I/O  streams/OST  for  op9mal  read  performance  

6800.55  

10044.2   12

886.8  

14503   16

600.7  

17245.4  

18440.8  

19282.1  

19336.2  

19696.2  

19928.9  

20492.1  

20613.5  

20850.5  

17607.2  

15216.7  

4992.16  

8740.68   11094   13

055.7  

16304.4  

17261.8  

17852.3  

18403.4  

19581  

20561.4  

20678.1  

20271.3  

20677.1  

17661.6  

19035.8  

13965.7  

0  

5000  

10000  

15000  

20000  

25000  

1   2   3   4   5   6   7   8   9   10   11   12   13   14   15   16  

MiB/s  

Streams/OST  

IOR  Op2mal  Read  

Trini9te  Op9mal  Read  1PPN  

Trini9te  Op9mal  Read  2  PPN  

Page 32: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Gadget  op9mal  N-­‐N  IOR  write  

8220  

8407  

8100  

8150  

8200  

8250  

8300  

8350  

8400  

8450  

MiB/s  

Op2mal  IOR  Write  

Post  Network  Fixes  

Post  DNE  

Page 33: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Gadget  op9mal  N-­‐N  IOR  read  

10882.1  

8397.96  

0  

2000  

4000  

6000  

8000  

10000  

12000  

MiB/s  

Op2mal  IOR  Read  

Post  Network  Fixes  

Post  DNE  

Page 34: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Gadget  N-­‐1  strided  IOR  write  1097.68   18

69.03  

3755.17  

4270.91  

5253.4   5844.59  

6032.68  

1112.16   18

44.17  

3879.23  

4107.73  

5487.07  

4127.79  

5877.23  

1069.46   17

80.43  

3505.61  

3986.28  

3660.6  

3891.77  

3963.52  

1071.39   1704.16  

3565.93   42

86.29  

4426.31  

4285.22  

4083.46  

0  

1000  

2000  

3000  

4000  

5000  

6000  

7000  

32   64   128   256   320   512   640  

MiB/s  

Processor  Count  

I/O  Performance  N-­‐1  Max  Write  

N-­‐1  Strided  MPIIO  Baseline  

N-­‐1  Strided  POSIX  Baseline  

N-­‐1  Strided  MPIIO  Post  DNE  

N-­‐1  Strided  POSIX  Post  DNE  

Page 35: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Gadget  N-­‐1  strided  IOR  read  11138.7  

509.59  

972.03  

1870.91  

2285.84   35

32.43  

4070.18  

10649.3  

519.14  

1009.9   1954.34  

2322.33  

3192.81  

3988.86  

10960  

512.98  

965.02  

1884.93  

2303.49   35

89.74  

4004.91  

10573.6  

524.92  

950.73  

1872.08  

2282.53   35

58.28  

3947.86  

0  

2000  

4000  

6000  

8000  

10000  

12000  

32   64   128   256   320   512   640  

MiB/s  

Processor  Count  

I/O  Performance  N-­‐1  Max  Read  

N-­‐1  Strided  MPIIO  Baseline  

N-­‐1  Strided  POSIX  Baseline  

N-­‐1  Strided  MPIIO  Post  DNE    

N-­‐1  Strided  POSIX  Post  DNE  

Page 36: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trini9te  &  Mutrino  op9mal  N-­‐N  IOR  write  

12145  

13836  

11219  

15217  

18242   18641  

0  

2000  

4000  

6000  

8000  

10000  

12000  

14000  

16000  

18000  

20000  

1  

Op2mal  IOR  Write  Comparisons  

Mutrino  Inital  

Mutrino  Network  Fix  

Trini9te  Ini9al  

Trini9te  Network  Fix  

Mutrino  Post  DNE  Upgrade  

Trini9te  Post  DNE  

Page 37: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trini9te  &  Mutrino  op9mal  N-­‐N  IOR  read  

10059.5  

4616.7  

11354.6  

20304.9  

11112.8  

17523.4  

0  

5000  

10000  

15000  

20000  

25000  

1  

Op2mal  IOR  Read  Comparisons  

Mutrino  Ini9al  

Mutrino  Network  Fix  

Trini9te  Ini9al  

Trini9te  Network  Fix  

Mutrino  Post  DNE  Upgrade  

Triniate  Post  DNE  

Page 38: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trini9te  &  Mutrino  N-­‐1  strided  IOR  write  

1074.7  

7012.7  

9003.4  

9337.51  

8925.78  

8036.04  

1044.98  

7023.45  

8455.84  

8714.99  

8405.16  

8342.56  

7586.43  

1066.86  

7404.32  

8749.33  

7946.82  

7338.09  

7320.34  

6459.44  

1066.73  

6643.77  

8317.73  

5322  

6820  

6379  

1058.39  

7002.04  

8393.25  

7148.1  

6655  

6057  

0  

1000  

2000  

3000  

4000  

5000  

6000  

7000  

8000  

9000  

10000  

32   256   512   1024   1600   2048   3200  

MiB/s  

Processor  Count  

N-­‐1  I/O  Performance  Max  Write  

Mutrino  N-­‐1  Strided  MPIIO  Baseline  

Mutrino  N-­‐1  Strided  POSIX  Baseline  

Mutrino  N-­‐1  Strided  MPIIO  Post  DNE  

Trini9te  N-­‐1  Strided  MPIIO  Post  DNE  

Trini9te  N-­‐1  Strided  POSIX  Post  DNE  

Page 39: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trini9te  &  Mutrino  N-­‐1  strided  IOR  read  9101.91  

4852.8  

5305.22  

4957.73  

4855.37  

4432.31  

10383.5  

5064.86  

5054.23  

4734.79  

4666.5  

4712.35  

4633.13  

11991.4  

4863.35   6143.4  

5749.37  

5412.93  

5402.68  

5376.09  

8861.59  

3236.46  

3010.7  

2727  

5426  

5422  

7513.6  

10296.9   12

122.2  

13322.5  

13520.8  

13375  

0  

2000  

4000  

6000  

8000  

10000  

12000  

14000  

16000  

32   256   512   1024   1600   2048   3200  

MiB/s  

Processor  Count  

N-­‐1  I/O  Performance  Max  Read  

Mutrino  N-­‐1  Strided  MPIIO  Baseline  

Mutrino  N-­‐1  Strided  POSIX  Baseline  

Mutrino  N-­‐1  Strided  MPIIO  Post  DNE  

Trini9te  N-­‐1  Strided  MPIIO  Post  DNE  

Trinii9te  N-­‐1  Strided  POSIX  Post  DNE  

Page 40: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Usage  Model  

Karen Haskell Jeff Johnson August  2015  

LA-­‐UR  15-­‐xxxxx  

Page 41: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

ASC  Usage  Models  

•  The  Trinity  Usage  Model  is  our  “contract”  with  users  for  how  to  best  use  the  machine.  

•  The  support  model,  form  and  func9onality  described  should  be  familiar  to  users  of  Cielo.  

•  The  Usage  Model  will  form  baseline  requirements  for  the  FY17  ASC  L2  ATCC  Produc9on  Readiness  Milestone.  

•  Like  Cielo,  Trinity  will  be  jointly  managed  by  LANL  and  SNL;  management  by  two  labs  rather  than  a  single  laboratory.      

•  As  an  ACES  tri-­‐lab  ASC  plaGorm,  policies  and  opera9ons  may  differ  from  those  of  local  laboratory  plaGorms.  

41  

Page 42: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Availability  

•  Final  phase  1  document  will  be  available  by  ATCC-­‐1  

•  Not  intended  to  be  a  living  document,  though  it  will  be  updated  for  Trinity  phase  2  

•  Widely  distributable,  documen9ng  the  Trinity  system  at  the  9me  of  publica9on  

•  Intended  as  an  agreement  with  the  users  of  the  system  as  to  how  the  machine  will  be  configured  and  managed  

 42  

Page 43: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Feedback  

•  We  will  incorporate  feedback  from  this  series  of  mee9ngs  – Email  tum-­‐[email protected]  with  ques9ons  or  comments  

•  Today’s  presenta9ons  and  the  Usage  Model  will  be  available  at  hZp://trinity.lanl.gov/  and  on  the  ACES  web  pages  at  hZp://aces.sandia.gov    

43  

Page 44: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Pla+orms  User  Support  

ACES User Support Team  

August  2015  

LA-­‐UR  15-­‐xxxxx  

Page 45: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

ACES  User  Support  

45  

Similar  Contact  Informa9on  to  Cielo  •  Help  available  8am  to  5pm  Mountain  Time  M-­‐F  •  Phone:  1-­‐877-­‐378-­‐1110  •  email:  trinity-­‐[email protected]  

Page 46: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

ACES  User  Support  

46  

 Documenta2on  •  hZps://aces.sandia.gov  •  Web  pages  also  accessible  on  Trinity  filesystems:    /usr/projects/hpcdoc/hpc.lanl.gov  

Page 47: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Alloca2ons  and  Accounts  

47  

•  Alloca9ons  –  Apply  for  an  ATS  Campaign  –  Alloca9ons  follow  the  ATCC  process:    hZps://aces.sandia.gov/atcc.html    

•  Accounts  requests  •  Trinity  

o  LLNL  &  Sandia  users:  through  SARAPE  hZp://sarape.sandia.gov  o  LANL  users:  hZps://hpcaccounts.lanl.gov/  

•  Trini9te  o  LLNL  &  Sandia  users:  through  SARAPE  hZp://sarape.sandia.gov  o  LANL  users:  hZps://hpcaccounts.lanl.gov/  

•  Mutrino  o  LLNL  &  LANL  users:  through  SARAPE  hZp://sarape.sandia.gov  o  Sandia  users:  through  hZps://webcars.sandia.gov  

Page 48: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Accessing  Trinity  

48  

•  Start  from  a  local  gateway  (e.g.  cslic.llnl.gov,  lynx‑s.sandia.gov)  

•  Get  9cket:    kinit -f    •  Use  ssh  to  reach  LANL  gateway:    red‑wtrw.lanl.gov  

•  From  there,  ssh  into  Trinity  front-­‐end:    tr‑fe  or  File  Transfer  Agents  (FTAs):    rfta‑fe  

Page 49: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Accessing  Trinity  Applica2on  Regression  Testbed  (ART)  Systems  

49  

Trini9te   Mutrino  

Diagram will be provided in published slides and on web pages

Log  in  to  Mutrino  via    srngate.sandia.gov    

Page 50: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Scheduling  and  Jobs  

50  

Job  Scheduling  •  Moab/Torque/ALPS  

λ  Same  commands  on  all  ASC  plaGorms  at  all  three  Labs:    msub,  showq,  checkjob,  etc.  

•  Fairshare-­‐based  scheduling,  32-­‐day  history,  no  “bank”,  no  upper-­‐limit  on  total  CPU  9me  per  Campaign  

Running  Jobs  •  aprun -n  …    (for  default  MPI)  •  mpirun -np  …  (for  Open  MPI  apps)  •  Modulefiles  sosware  environment  

Page 51: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Road  Show    Programming  Environment  and  Tools  

Page 52: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Programming  Environment  

52  

q  Programming  Languages  Ø  Fortran  Ø  C  Ø  C++  Ø  Python  Ø  Chapel  

q  Compilers  Ø  Cray  Compiling  Environment  (CCE)  Ø  Intel  Ø  GNU  

q  Programming  Models  (Distributed  Memory)  Ø  MPI:  Cray  Mpich  &  OpenMPI  Ø  SHMEM  Ø  OpenMP4.0  Ø  PGAS  &  Global  View  Ø  UPC  Ø  CAF  Ø  Chapel  

q  I/O  libraries  Ø  NetCDF  Ø  HDF5  

q  Tools  Ø  Use  modules  to  setup  sosware  env.  

q  Debuggers  Ø  DDT  Ø  TotalView  Ø  Abnormal  Termina9on  Processing  (ATP)  Ø  STAT  Ø  Cray  Compara9ve  Debugger  

q  Performance  Analysis  Ø  Vtune  and  Vector  Advisor  Ø  CrayPat  and  Cray  Appren9ce2  

q  Op9mized  Scien9fic  Libraries  Ø  Libsci  Ø  MKL  Ø  LAPACK  Ø  ScalAPACK  Ø  BLAS  Ø  Cray  Adap9ve  FFTs  (CRAFFT)  Ø  Cray  PETSc    Ø  Cray  Trilinos  Ø  FFTW  

Compute  Node  Linux  (CNL)  with  LWK  on  Compute  Nodes  and  Full  SUSE  Linux  on  Login  Nodes  

Page 53: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Modules  and  Compiling  

q Modules  tool  used  to  control  needed  environment  variable  and  directories  in  your  path    Ø module  list  shows  default  

§   OS  modules,  PrgEnv  Modules,  and  support  modules  Ø Use  module  avail  to  look  for  choices  Ø Use  “unload”,  “load”,  “swap”  module  commands  to  set  it  to  your  

preference  q  A  special  “PrgEnv”  module  controls  compiler,  debugger,  and  misc  

library  setup.  Select  from:    Ø  PrgEnv-­‐intel;  PrgEnv-­‐cray;  PrgEnv-­‐gnu;  PrgEnv-­‐pgi;  

q  Default  compute  node  compiler  wrapper  commands  for  all  compilers  (same  as  Cielo):  Ø  In,  cc,  and  CC  

53  

Page 54: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Running Applications MOAB/TORQUE/ALPS ( same as Cielo)  

q ALPS  is  the  allocator  provided  by  Cray  for  the  compute  nodes.  Ø  Works  in  concert  with  Moab  and  Torque  Ø  Provide  batch  job  submission  and  job  script  launch  facili9es  Ø  Commands  within  ALPS:  basic  commands  

§  running  (aprun)  §  status  enquiry  (apstat)  §  termina9on  (apkill)  

q Useful  man  pages  Ø  intro_alps  Ø  aprun  

q  Use  the  aprun  command  and  the  $OMP_NUM_THREADS  environment  variable  to  run  a  hybrid  program.  You  may  need  aprun  op9ons  "-­‐n",  "-­‐N",  and  "-­‐d"  to  get  the  desired  combina9on  of  MPI  processes,  nodes,  and  cores.  Ø  aprun  -­‐n  mpi_procs  -­‐N  mpi_procs_per_node  -­‐d  threads_per_mpi_proc  mpi_program  

q  Very  good  Cray  documenta9on  at  hZp://docs.cray.com  Ø  Cray  Applica9on  Developer’s  Environment  User’s  Guide  

54  

Page 55: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Major  Differences  between  Cielo  and  Trinity  

55  

Cielo   Trinity  (  Phase-­‐1)  

Processor   AMD  Opteron  6100  (MagnyCours)  

Intel  Xeon  E5-­‐2698V3  (Haswell)  

nodes,  cores/node,  memory/node   8518  compute,  16cores/node,  32GB/node  

9436  compute,  32cores/node,  128GB/node  

Interconnect   Cray  Gemini  (3D  Torus)   Cray  Aries(Dragonfly)  

Debuggers   TotalView   DDT  and  TotalView  

Compiler  (Default)   PGI   Intel    

Scien9fic  Library   libsci,  ACML,  MKL   libsci,  MKL  

SIMD     SSE4a   AVX2  with  FMA  

Peak  Node  GFLOPS   153.6   972.8  at  base  Turbo  Freq.  of  1.9GHz  

Page 56: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Compilers  and  flags  

56  

Intel   Cray   GNU   PGI  

Op9miza9on   -­‐fast**  –no-­‐ipo   -­‐O3  –h  scalar3  

-­‐O3  –ffast-­‐math  

-­‐fastsse  

AVX   -­‐xcore-­‐avx2   -­‐h  vector3   -­‐mavx   -­‐Mvect=simd:256  

OpenMP   -­‐qopenmp*   -­‐h  omp   -­‐f  openmp   -­‐mp=nonuma  

Fortran  Format   -­‐free   -­‐ffree   -­‐ffree-­‐form   -­‐Mfree  

Array  bounds  check  

-­‐check  bounds   -­‐h  bounds   -­‐�ounds-­‐check  

-­‐Mbounds  

Report   -­‐Wextra  -­‐Wall   -­‐h  negmsgs   -­‐fopt-­‐info-­‐vec   -­‐Minfo=all  

*Intel  v15  and  greater,  otherwise  –openmp  **-­‐fast  includes  –fp-­‐model  fast=2:  for  9ghter  accuracy  add  –fp-­‐model  fast=1  or  fp-­‐model  precise  

Page 57: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Going from Cielo PGI to Trinity Intel

PGI Intel Description-­‐fast   -­‐fast  -­‐no-­‐ipo   Standard optimization

-­‐mp=nonuma   -­‐qopenmp*   OpenMP support

-­‐Mfixed   -­‐fixed   Fortran fixed format support

-­‐Mfree   -­‐free   Fortran free format support

-­‐byteswapio   -­‐convert  big_endian  

Read and write Fortran unformatted data files as big

endian.Explicit  link  to  mkl  libs   -­‐mkl   Link to Intel MKL

-­‐V   -­‐-­‐version   Show compiler version

57  *Intel  v15  and  greater,  otherwise  -­‐openmp  

Page 58: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

PGAS (Partitioned Global Address Space) languages: Unified Parallel C (UPC) and Coarray Fortran (CAF)

q PrgEnv-­‐cray  module  provides  na9ve  support  for  Coarray  Fortran  and  Unified  Parallel  C:  

q CAF  example:  Ø  sn  -­‐h  caf  -­‐o  CAFhello  CAFhello.f90  

q UPC  example:  Ø  cc  -­‐h  upc  -­‐o  UPCProg  UPCProg.c  

58  

Chapel  (an  emerging  parallel  programming  language  being  developed  by  Cray  Inc.)  is  also  available  through  a  module    

Page 59: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

RUR  -­‐  Resource  U9liza9on  Repor9ng  (New  in  Trinity)    

59  

q Measures  resources  used  by  each  applica9on  q Users  op9onally  receive  report  wriZen  to  their  home  directory,  

appended  to  ~/rur.out  q Available  Metrics  

Ø  Timestamps:  Start  and  end  9me  of  the  applica9on  Ø  Taskstats:  User  CPU  9me,  max  memory  used,  exit  code,  …  Ø  Energy:  Total  energy  used  by  the  applica9on  

q Example:  uid: 20983, apid: 175244, jobid: 349.mutrino-moab, cmdname: ./bin/mpicth_244, plugin: energy {"nodes_throttled": 2, "nodes_with_changed_power_cap": 0, "max_power_cap_count": 0, "energy_used": 25015991, "max_power_cap": 0, "nodes_accel_power_capped": 0, "min_power_cap": 0, "min_power_cap_count": 0, "nodes_power_capped": 0, "nodes": 96}

Page 60: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Cornell Wright Nathan Hjelm August 2015

Trinity  Burst  Buffer  Hardware,  Sosware  and  Usage  

UNCLASSIFIED - LA-UR-15-26608

Page 61: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED - LA-UR-15-26608

Slide 61

Burst Buffers will improve Productivity and Enable Memory Hierarchy Research

•  Burst Buffer will improve operational efficiency by reducing defensive IO time

•  Burst Buffer fills a gap in the Memory and Storage Hierarchy and enables research into related programming models

§  Technology Drivers: –  Solid State Disk (SSD) cost decreasing –  Lower cost of bandwidth than hard disk drive

§  Trinity Operational Plans:

–  SSD based 3 PB Burst Buffer –  3.28 TB/Sec (2x speed of Parallel File

System)

Page 62: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED - LA-UR-15-26608

Burst Buffer – more than checkpoint

§  Use Cases: –  Checkpoint

•  In-job drain, pre-job stage, post-job drain

–  Data analysis and visualization -  In-transit -  Post-processing -  Ensembles of data

–  Data Cache •  Demand load •  Data staged

–  Out of core data •  Data intensive workloads that exceed memory capacity

Slide 62

Page 63: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED - LA-UR-15-26608

Slide 63

Trinity System Overview

Page 64: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED - LA-UR-15-26608

Slide 64

Trinity Burst Buffer Hardware §  576 Burst Buffer Nodes

–  Announced as Cray DataWarp™ –  On high speed interconnect - globally accessible –  Trinity IO Node + PCIe SSD Cards –  Distributed throughout cabinets

Metric Phase I Full Trinity

Burst Buffer PFS Burst

Buffer PFS

Nodes 300 BB Nodes

114 LNET Routers

576 BB Nodes

234 LNET Routers

Bandwidth 1.7 TB/S 0.71 TB/S 3.3 TB/S 1.45 TB/S Capacity 1.9 PB 82 PB 3.7 PB 82 PB

Memory Multiple 1.6 X 67 X 1.75 X 39 X Full system

checkpoint cost 12% 21%

Page 65: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED - LA-UR-15-26608

Slide 65

Cray XC40 DataWarp Blade ( 2 Burst buffer Nodes )

Page 66: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED - LA-UR-15-26608

Slide 66

Cray DataWarp Blade

é Service Blade

(2 nodes)

ç SSD Cards

Page 67: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED - LA-UR-15-26608

Slide 67

Burst Buffer Operating Modes

Mode Description

Private Scratch Per node burst buffer (BB) space

Shared Scratch Shared space, files may be striped across all BB nodes. à Used for Trinity Checkpoints ß

Shared Cache Parallel File System (PFS) cache. Transparent and explicit options

Load Balanced Read Only Cache

PFS files replicated into multiple BB nodes to speed up widely read files

Page 68: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED - LA-UR-15-26608

Slide 68

Burst Buffer System Software §  Burst Buffer SSD partitioned into allocation units

–  Allocation units belong to LVM volume group

§  Workload manager –  Job submission requests BB capacity –  Starts job when capacity available

§  DataWarp registration service –  Selects allocation units –  Creates XFS logical volumes on SSD –  Mounts via DVS on compute nodes

§  Multiple Access Modes –  Scratch / Cache –  Striped / Private / Load Balanced (RO)

§  Trinity BB Checkpoints will use Striped Scratch

§  Automated stage/drain of specified directories from/to PFS

§  Per job write limits (endurance management)

§  Administrative Functions – configuration, monitoring, repair

Page 69: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED - LA-UR-15-26608

DataWarp Job Script – Access §  Jobdw

– Create and configure a DataWarp job instance –  #DW jobdw [access_mode=mode] [capacity=n]

[max_mds=n] [modified_threshold=bytes] [pfs=path] [read_ahead=bytes:rasize] [sync_on_close=yes|no] [sync_to_pfs=yes|no] [ type=cache|scratch]

§  Persistentdw – Configure access to an existing persistent

DataWarp instance –  #DW persistentdw name=piname

Slide 69

Page 70: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED - LA-UR-15-26608

DataWarp Job Script – Staging

§  stage_in – Stage files into a DataWarp instance – #DW stage_in destination=dpath

source=spath [tolerate_errors=yes|no] type=type

§  stage_out – Stage files from a DataWarp instance – #DW stage_out destination=dpath

source=spath type=type

Slide 70

Page 71: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED - LA-UR-15-26608

DataWarp Job Script – Swap

§  Swap – Configure swap space per compute node – #DW swap nGiB

Slide 71

Page 72: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED - LA-UR-15-26608

Single-Job Burst Buffer

Slide 72

Job Phases A B C

Compute

Burst Buffer

PFS

Job

Create, Stage Compute Drain,

Destroy

Page 73: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED - LA-UR-15-26608

Multi-Job Burst Buffer

Slide 73

Job 0

Compute

Burst Buffer

PFS

Job 1 Job N Job N+1

Job Depend

… Create, Stage Compute Compute Drain,

Destroy

Perm Perm Perm Perm

Page 74: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED - LA-UR-15-26608

Compute + Analysis Burst Buffer

Slide 74

Job 0

Compute

Burst Buffer

PFS

Job A Job B Job N+1

Job Depend

Create, Stage Compute Analyze Drain,

Destroy

Perm Perm Perm Perm

Page 75: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED - LA-UR-15-26608

Slide 75

Applications have Options

§  Hierarchical Input Output (HIO) Library –  Hides vendor specific interface –  Provides additional performance, reliability,

management and diagnostic functions –  Recommended approach

§  Scalable Checkpoint Restart Library (SCR) –  Adam Moody (LLNL) planning to support Trinity BB

with SCR §  Direct POSIX calls from application

–  Will require Cray DataWarp specific ioctl calls to exploit striping and stage / drain functions

Page 76: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED - LA-UR-15-26608

Slide 76

Acknowledgements

§  Many people and groups have contributed to the concept, design and development of burst buffer, Trinity and the materials in this talk.

§  Including: –  Gary Grider –  Josip Loncaric –  Doug Doerfler (SNL) –  Nathan Hjelm –  Nick Wright (NERSC)

–  Dave Henseler (Cray) –  Bob Pearson (Cray) –  Bronis de Supinski (LLNL) –  Adam Moody (LLNL) –  John Bent (EMC)

Page 77: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Pla+orm  Applica2on  Readiness  

Application Readiness Cornell Wright, Joel Stevenson

 August  2015  

Page 78: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Applica9on  Readiness  Objec9ves  

•  Understand  applica9on/user  requirements,  usage  paZerns,  and  workflows  

•  Be  an  advocate  for  customer  needs  •  Work  with  applica9on  teams  to  ensure  produc9on-­‐run  

readiness  on  current  and  incoming  systems  •  Engage  with  Trinity  Early  Access  codes  to  achieve  rapid,  

produc9ve  use  •  Iden9fy  causes  of  unexpected  behavior;  submit  bugs  and  

track  fixes  •  Collaborate  with  subsystem  teams  such  as  Systems  

Management,  File  Systems,  I/O,  Archive  and  Tools  •  Stress  and  regression  test  of  produc9on  and  new  systems  

and  system  configura9ons  

78  

Page 79: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Applica9on  Readiness  Approach  

•  Each  par9cipa9ng  site  (LANL,  LLNL,  and  SNL)  has  some  form  of  a  local  applica9on  readiness  team.  

•  Local  applica9on  readiness  teams  will  work  with  code  developers  at  their  sites  to  port  codes  from  Cielo  to  Trinity.  

•  ACES  applica9on/user  support  personnel  have  the  responsibility  for  working  with  site  local  applica9on  readiness  teams  and  their  code  developers  to  port  codes  to  Trinity  Phase  1  (Haswell).  We  call  this  partnership  the  “ACES  Applica9on  Readiness  Team”.  

•  The  Trinity  Applica9on  Center  of  Excellence  (COE)  has  the  responsibility  for  working  with  site  local  applica9on  readiness  teams  and  their  code  developers  to  port  selected  codes  to  Trinity  Phase  2  (KNL).  

79  

Page 80: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

ACES  Applica9on  Readiness  Team  

•  ACES  applica9on/user  support  experts:  –  Consult  at  LANL  –  OneStop  at  SNL  

•  LANL  AR  team:  –  Cornell  Wright  –  Dave  Nystrom  –  David  Gunter  –  Nathan  Hjelm  –  Howard  Pritchard  – Mike  Berry  (Cray)  

•  SNL  AR  team:  –  Joel  Stevenson  –  Karen  Haskell  –  Dennis  Dinge  – Mike  Davis  (Cray)  

•  LLNL  Development  Environment  group:  –  ScoZ  Futral  

80  

Page 81: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Phase  1  (Haswell)    

•  Codes  running  on  Cielo  need  to  run  well  on  Trinity  Phase  1  from  the  beginning  of  General  Availability.    

•  The  ACES  Applica9on  Readiness  Team  will  work  with  code  teams  to  port/test  the  codes  and  evaluate  their  produc9on  readiness.    

•  The  ACES  Applica9on  Readiness  Team  will  also  work  closely  with  the  Campaign  users  to  exercise  the  codes  with  user  workflows  and  evaluate  the  performance  of  the  plaGorm/codes  on  the  user  Campaign  workload.  

81  

Page 82: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Phase  2  (KNL)    

•  The  first  applica9on  codes  to  be  ported  to  Trinity  Phase  2  have  been  iden9fied  and  will  be  managed  by  the  Trinity  Applica9on  Center  of  Excellence  (COE)  ac9vi9es.    Three  codes  have  been  selected,  one  from  each  lab.  

•  Addi9onal  codes  will  be  selected  for  COE  efforts  based  on  ASC  programma9c  needs.  ASC  resources  are  limited,  so  aZen9on  will  focus  on  a  limited  number  of  applica9ons  in  the  beginning.  

•  Before  becoming  Trinity  Phase  2,  applica9on  codes  will  need  to  demonstrate  success  on  the  KNL  architecture.  Success  on  the  KNL  architecture  may  be  demonstrated  in  several  ways:  –  Work  with  Trinity  COE  to  insure  the  code  is  ready  for  Trinity  KNL  usage.  –  Demonstrate  on  Trinity  KNL  architecture  testbed(s)  that  the  code  uses  the  

hardware  efficiently  and  is  ready  for  Trinity  KNL  usage.  Then  the  code  may  move  to  the  ACES  Applica9on  Readiness  Team  for  adop9on  in  Trinity  Phase  2.  

–  Demonstrate  on  Trinity  KNL  that  the  code  uses  the  hardware  efficiently  and  is  ready  for  Trinity  KNL  usage.  Then  the  code  may  move  to  the  ACES  Applica9on  Readiness  Team  for  adop9on  in  Trinity  Phase  2.  

82  

Page 83: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Applica9on  Readiness  Process  

•  Engage  consultants,  web  documents,  etc.  for  normal  needs.    When  this  is  not  sufficient  or  not  working  .  .  .  –  Los  Alamos  –  contact  Cornell  Wright,  [email protected]  –  Sandia  –  contact  Joel  Stevenson,  [email protected]  –  Livermore  –  contact  ScoZ  Futral,  [email protected]  

•  Trinity  Applica9ons  Technical  Concall  – Wednesdays  3  MT  /  2  PT  –  LANL  6-­‐1220  or  866-­‐880-­‐6515  PC  043534  #            

•  Issues/problems  that  are  primarily  system  related  can  go  directly  to  the  Los  Alamos  AR  team  

83  

Page 84: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Campaign  Storage  and  Usage  Model  

Kyle Lamb  

August  2015  

Page 85: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Campaign  Storage  Topics  

•  Who  •  Why  •  How  •  Where  •  When      

85  

Page 86: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Campaign  Storage  Thanks  

•  Many  Thanks  go  to:  – HPC-­‐3  Storage    – HPC-­‐5  Storage  – Vendors  

86  

Page 87: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

LANL  Archive  Performance  

87  

•  Restore  of  a  25TB  file  =  30Hours  •  File  sizes  expected  in  the  100s  of  TB  for  Trinity  •  Need  to  filter  data  going  to  archive,  archive  is  for  forever  data  (code  repos,  input  decks,  some  visualiza9ons)  not  necessarily  checkpoints  

•  Need  to  start  planning  for  1PB  sized  checkpoints,  with  longer  residence  with  high  performance  access  

Page 88: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Campaign  Storage  Why  

88  

•  Long  term  high  performance  data  storage  repository  for  large  data  sets  

•  Store  data  for  the  dura9on  of  a  compu9ng  campaign  on  a  storage  9er  that  offers  higher  resilience  than  Scratch  and  higher  performance  than  Archive  

•  At  comple9on  of  campaign,  user  selec9on  is  made  to  determine  what  needs  to  be  archived  

Page 89: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Campaign  Storage  Specifica2ons  

89  

•  Capacity  ~25PB  (with  future  expansion  expected)  

•  Performance  ~20-­‐25GB/s  (expansion  will  increase  performance)  

•  Quotas  enabled  •  Full  file  system  view  (ls  and  grep  work)  •  No  update  in  place  

Page 90: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Campaign  Storage  Technology  

90  

•  Posix  file  system  with  Object  storage  backend  •  U9lizes  “cloud  driven”  erasure  code  to  gain  highly  reliable  disk  based  storage  (20+4,  30+6,  40+8)  

•  Higher  performance  gained  through  large  scale  parallel  access  to  archive  grade  hard  drives  

•  Not  intended  for  high  duty  cycle  workloads  e.g.  Scratch  

Page 91: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Campaign  Storage  Access  

91  

•  Available  only  via  the  FTAs  not  mounted  to  any  cluster  

•  Mounted  on  interac9ve  FTAs  but  performance  available  via  batch  FTAs  and  PFTOOL  *new*  

•  FTAs  will  mount  all  produc9on  file  systems  and  u9lize  HSI  for  interface  with  HPSS  

Page 92: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

FTA  Cluster  

Long  Term  Storage/Archive  

Campaign  Storage  Diagram  

92  

Lustre1  (39  PB  Usable)  

2-­‐4  Interac9ve  FTAs  

ssh  

Campaign  Storage  

Lustre2  (39  PB  Usable)  

20-­‐25  Batch  FTAs  

HPSS  

NFS  Home/Projects  

Page 93: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Campaign  Storage  Timeline  

93  

•  Small  3PB  version  will  be  released  with  Trinity  for  Open  Science  

•  Larger  25PB  version  available  when  Trinity  goes  into  produc9on  

•  Capacity  and  Performance  increases  will  occur  with  need  and  available  budget  

Page 94: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Visualiza2on  on  Trinity  

Visualization on Trinity Laura  Monroe,  David  Modl,  Robert  Kares,  John  PatcheZ  (LANL)  

David  Karelitz,  Ken  Moreland,  Warren  Hunt  )SNL)  Eric  Brugger  (LLNL)  

August  2015  

Page 95: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

§  Interactive post-processing with off-platform rendering –  Dedicated viz nodes located within the platform run data servers

that read data from the global PFS and extract visible geometry. –  The visible geometry is sent through the GigE backbone over to

local clients for hardware rendering. –  Standard approach used with Cielo and previous ASC platforms. –  A low risk option: EnSight data servers will run out of the box on

the dedicated viz nodes on the Haswell side of Trinity.

Viz Usage Scenarios on Trinity: Post-processing, off-platform rendering

Slide 95

Page 96: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

§  Interactive post-processing with on-platform rendering –  Dedicated viz nodes located within the platform run data servers that

read data from the global PFS and extract visible geometry. –  The visible geometry is passed to clients running on the dedicated

viz nodes that render on the node CPU’s using a Mesa 3D interface to Intel’s OpenSWR renderer.

–  New option available with Trinity. –  Requires successful development by Intel of their OpenSWR

renderer for Haswell and KNL and deployment of the renderer into mesa 3D.

–  OpenSWR development for Haswell is relatively complete. –  KNL support for OpenSWR is under development and is higher risk. –  EnSight, ParaView and VisIt all plan to support this option.

Viz Usage Scenarios on Trinity Post-processing on-platform rendering

Slide 96

Page 97: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

§  Co-processing for in-transit viz with the burst buffer –  In this scenario dedicated viz nodes located within the platform

share access with the compute nodes of a running simulation job to a burst buffer filesystem.

–  The simulation job writes viz data to the burst buffer and data servers running on the viz nodes read data from the burst buffer and extract visible geometry.

–  The visible geometry can be rendered either on or off platform. –  This type of co-processing activity is termed in-transit viz. –  Can be used for both interactive and batch visualization.

Viz Usage Scenarios on Trinity In-transit visualization using burst buffer

Slide 97

Page 98: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

§  In-Situ viz using ParaView/Catalyst or VisIt/LibSim in simulation codes –  Simulation code is linked with the ParaView/Catalyst or VisIt/

LibSim libraries to allow complete visualization pipelines to be executed at runtime in the simulation memory space.

–  Visualization pipelines are defined by Python scripts read at runtime by ParaView/Catalyst or VisIt/LibSim to produce data products such as imagery, vis dumps, or derived data products

–  Requires integration with the simulation code source to pass control to ParaView/Catalyst or VisIt/LibSim and to link with the analysis libraries.

–  This approach has recently been successfully tested on Cielo using ParaView/Catalyst on the RAGE code.

Viz Usage Scenarios on Trinity: In-situ visualization

Slide 98

Page 99: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

§  For Trinity we intend to continue the viz model very successfully utilized for Cielo and most previous generations of ASC machines.

§  2% of Trinity nodes will be dedicated exclusively for visualization and analysis. –  The partition of dedicated viz nodes will be on both the Haswell

and KNL sides of the machine. –  Percentage based on observed Cielo use

§  Percentage reviewed every ATCC cycle, and readjusted if needed §  Can request extra resources to handle special needs

–  Nodes should support dynamic linking to shared object libraries,TCP/IP sockets for internal and external communication, DISCOM

Dedicated Viz Nodes for Trinity

Slide 99

Page 100: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

§  Rendering issue on Intel Xeon Phi’s –  OpenGL libraries not updated on this platform –  OpenGL runs slowly

§  Impact to Trinity and Cori platforms –  In particular impacts in-situ rendering

§  Intel plans: –  Development on Intel’s open source OpenSWR and deployment of the core

renderer into Mesa 3D, providing a Mesa interface to the codes –  Provide analysis/proposal/commitment to address this –  Develop and deploy software onto KNL

§  Lab plans: working group –  Review proposal –  Work with Intel to develop and deploy Mesa/OpenSWR on Trinity and Cori –  Representatives from LANL/SNL/LLNL/LBL –  Lab people are EnSight/ParaView/VisIt developers and systems people

Rendering on KNLs

Slide 100

Page 101: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

§  Ensight, ParaView, VisIt §  All three support:

–  OpenSWR/Mesa 3D support for on-platform rendering using Haswell and eventually KNL.

§  EnSight and ParaView support: –  Co-processing with the burst buffer for the in-transit viz

usage scenario. (VisIt may support this) §  ParaView and VisIt support:

–  Batch production of images via in-situ viz for LANL, SNL and LLNL codes.

Viz Packages

Slide 101

Page 102: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Brett Hollander – LANL Susie McRee - SNL

Archive  and  Data  Transfer  

Page 103: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Archive  in  the  Trinity  Era  

•  HPSS  will  be  the  “forever”  storage  system.  – Checkpoints  will  not  be  stored  in  HPSS  under  new  storage  model  

– HPC-­‐3  recommends  that  single  file  store  does  not  exceed  25TB  

–  If  user  believes  they  have  a  need  to  archive  large  checkpoint  files,  mee9ng  between  HPC/Program  management  necessary.    Excep9ons  will  be  few  for  files  >  25TB  

– To  ensure  annual  storage  does  not  exceed  budgeted  capacity,  quotas  on  LANL  HPSS  will  be  enforced.  

Page 104: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Archive  Quotas  

•  Why  LANL  HPSS  quotas?  – Must  curb  growth  to  sustain  economically  feasible  Archive.  

– Out  of  190  ac9ve  users,  the  top  10  take  up  45%  of  archive  capacity  ($2.5M  worth  of  tape)  

– Only  5%  recall  rate  – LANL  HPC  approached  by  ASC  to  implement    

Page 105: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Archive  Quotas  Details  

•  Status:  Proposed  HPSS  quotas  in  discussion  with  ASC  •  “Sos”  quotas,  no  automa9c  storage  shutoff  •  Quota  reset  each  FY  •  Amount  of  storage  granted  per  user  TBD  •  User  will  receive  no9fica9ons  at  70%,  90%,  100%  •  HPC+ASC  management  discussion  takes  place  at  100%    •  Management  can  request  more  storage  in  HPSS/Campaign  or  make  determina9on  that  user  must  delete  files  or  cease  Archival  storage  

•  Without  an  excep9on,  if  user  con9nues  to  store  past  100%  their  job  priority  will  fall  and  HPC  cluster  accounts  may  get  disabled  

 

Page 106: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Archive  Performance  

•  HPSS  will  provide  less  performance  than  Campaign  Storage  – HPSS  not  designed  for  fast  retrieval  back  to  File  system.  

– Single  file/single  node  (300MB/s)  – Mul9-­‐file/mul9-­‐node:  (2-­‐3GB/s)  – Expect  modest  increase  in  performance  by  Trinity  9meframe  

Page 107: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Tri-­‐lab  Classified  Data  Transfer  

107  

Page 108: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

LANL  Internal  Data  Movement  

•  Use  PFTool  to…  –  Move  data  between  Campaign  Storage  and  Lustre    

•  Use  HSI  to…  –  Move  data  from  Lustre/HPSS  to  SNL,  LLNL  –  Move  data  between  HPSS  and  Campaign  –  Move  between  Campaign  and  Lustre    

Page 109: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Data  Transfer  Best  Prac2ces  

•  Use  of  HTAR  is  highly  recommended  for  transfers  of  1,000+  small  files  (<  1GB  each).  

 •  Use  of  HSI  Transfer  Agent  will  improve  performance  via  parallel  file  transfer.    It  is  recommended  that  files  1TB  or  larger  be  stored  with  HSI  TA.  

 •  Given  that  current  DISCOM  bandwidth  maximum  throughput  is  approximately  1TB/hour,  transferring  files  >  10TB  is  not  suggested.  

Page 110: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Ques2ons  on  any  of  the  topics  presented?  

 •  For  general  Trinity  informa9on  ,  see  trinity.lanl.gov  (public  website)  

•  For  detailed  informa9on  about  Trinity  plaGorms,  visit  hZps://aces.sandia.gov/trinity  (authen9ca9on  required)  

110  

Page 111: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

   

Backup Slides

111  

Page 112: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

Burst Buffer Backup

Slide 112

Page 113: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

Slide 113

HIO Design Goals

HIO, short for Hierarchical Input Output, is the burst buffer enablement library for Trinity and other systems. Goals:

–  Support Trinity and future Burst Buffer implementations •  Isolate application from BB technology shifts

–  Easy to incorporate into existing applications •  Lightweight, simple to configure

–  Improve checkpoint performance –  Improve job reliability and efficiency –  Support checkpoint and analysis scenarios (e.g., viz) –  Extend to bridge to future IO

Page 114: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

Slide 114

Why implement HIO as a library ? §  Simplest packaging and delivery available

§  Self contained, minimal external or system dependencies

§  Easiest for applications to adopt

§  Library approach facilitates rapid prototyping and deployment, responsiveness to application needs

§  Library also provides a vehicle to provide (at no cost to applications): –  Checkpointing best practices –  Performance and functional diagnostics –  Mitigation for system problems

§  Why not provide via extensions to MPI-IO now ? –  Existing implementation perform poorly –  Unloved by users

§  Integration with other IO packages will be investigated in future –  HDF5 ? SCR ? ADIOS ? –  HIO library should be largely reusable in that environment

Page 115: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

HIO Project Features §  Full thread safety §  C/C++/Fortran support §  Configurable diagnostic and performance monitoring §  Header / Library packaging §  Open-source intention §  Support tri-lab ATS and CTS systems (more than Trinity) §  On-site (at LANL) prototype on Gadget §  Intent to prototype EAP support for HIO as POC and test

vehicle §  PFS-only version available before Trinity

Slide 115

Page 116: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

Primary HIO API Calls

§  Init / Term / CP Mgmt: –  hio_init_mpi() –  hio_config_set()

–  hio_should_checkpoint() –  hio_fini()

Slide 116

§  Open / Close: –  hio_dataset_open() –  hio_element_open()

–  hio_element_close() –  hio_dataset_close()

§  Read / Write: –  hio_element_write

{strided}{nb}()

–  hio_element_read {strided}{nb}()

–  hio_wait()

Page 117: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

hio – API calls Categorized § Context Management

– hio_init_single – hio_init_mpi – hio_fini

§ Checkpoint Management – hio_dataset_should_checkpt

§ Data Collection Control – hio_dataset_open – hio_dataset_close – hio_dataset_get_id – hio_dataset_unlink – hio_element_open – hio_element_close – hio_element_size – hio_dataset_construct

§ Data Movement – hio_element_write – hio_element_write_nb – hio_element_write_strided – hio_element_write_strided_nb – hio_element_flush – hio_element_read – hio_element_read_nb – hio_element_read_strided – hio_element_read_strided_nb – hio_dataset_flush

§ Request Control – hio_complete – hio_request_test – hio_request_wait

§ Configuration – hio_config_set_value – hio_config_get_value – hio_config_get_count – hio_config_get_info

§ Performance Reporting – hio_perf_get_count – hio_perf_get_info – hio_perf_get_value

§ Error Reporting – hio_err_get_last – hio_err_print_last – hio_err_print_all

Slide 117

Page 118: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

DataWarp Administrative Interfaces

§  Commands for setup and initial stage in –  Issued by pre-compute system job based on

#dw commands in job script §  Commands for final stage out and takedown

–  Issued by post-compute system job §  Monitoring for errors, endurance §  Commands for management – online / offline for

repair, re-configuration

Slide 118

Page 119: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

End of Burst Buffer

Backup Slide 119

Page 120: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Center  of  Excellence  SNL  

Rob Hoekstra, Mike Glass, Ron Green(Intel) et al.  

August  2015  

SAND2015-­‐6916  PE  

Page 121: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Priori2es  for  CoE  

•  Proxy  Applica9ons  •  Solvers  •  Kokkos  •  Codes  •  Conclusions      

121  

Page 122: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Proxy  Applica2ons  

122  

•  miniAero  – Explicit  Aerodynamics  

•  miniFE/HPCG  –  Implicit  Finite  Element  

•  (mini)FENL  –  In  progress,  “miniDriver”  more  representa9ve  of  real  code  data  structures  

•  BDDC    Domain  Decomp.  Solver  Proxy  

Page 123: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trilinos  NextGen  Solver  Stack  

123  

•  Teko:    Block  Segregated  Precondi9oning  •  Anasazi:    Parallel  Eigensolvers  •  Belos:    Parallel  Block  Itera9ve  Solvers  (FT/CA)  •  MueLU:    Mul9-­‐Level  Precondi9oning  •  Ifpack2:    ILU  Subdomain  Precondi9oning  •  ShyLU:    Hybrid  MPI/Threaded  Linear  Solver  •  Tpetra:    Distributed  Linear  Algebra  •  Zoltan2  Load  Balancing  &  Par99oning  •  Kokkos:    Heterogeneous  Node  Parallel  Kernels  

Applica2ons  &  Advanced  Analysis  Components  

Back-­‐ends:  OpenMP,  pthreads,  Cuda,  Qthreads,  ...  

Anasazi:  Eigensolvers   Belos:  Linear  Sovlers  

Muelu:  Mul2Grid  Precondi2oners  

ShyLU/Basker:  Direct  Solvers  Ifpack2:  Subdomain  Precondi2oners  

Tpetra:  Scalable  Linear  Algebra  

Kokkos  Sparse  Linear  Algebra  

Kokkos  Containers  Zoltan2:  Load  Balance/Par22oning  

Kokkos  Core  

Page 124: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Kokkos  

124  

–  Many-­‐core  challenge:  “Refactor  no  more  than  1+ɛ  9mes!”  –  In  Use:  Trilinos  Solvers,  Proxies,  Science  Applica9ons  

Applica2on  and  Domain  Specific  Library  Layer(s)  

Back-­‐ends:  OpenMP,  pthreads,  Cuda,  Qthreads,  ...  

Kokkos  Sparse  Linear  Algebra  Kokkos  Containers  Kokkos  Core  

0 4 8

12 16 20 24

K20X IvyBridge SandyBridge XeonPhi B0 XeonPhi C0 IBM Power7+

MiniFE  Proxy  -­‐  2me  for  200  itera2ons  

NVIDIA ELL NVIDIA CuSparse Kokkos OpenMP

Time  (sec)  

Page 125: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Codes  

•  SIERRA  – Nalu  (Science  Code  &  Acceptance  Test)  – Solid  Mechanics/Structural  Dynamics  – Aero  

•  RAMSES  – Solvers,  etc.  

•  ATDM  – Dharma/Kokkos  – Agile  Components  

125  

Page 126: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

SIERRA  

•  Solvers  – Mul9-­‐grid  (Nalu,  full  scale  Cielo/Sequoia)  – Domain  Decomposi9on  (vendor  &  custom  subdomain  solver  

•  Threading  and  Vectoriza9on  –  (mini)FENL  in  progress  –  Intrinsics  &  Improved  Intel  Compiler  

•  Speciality  Algorithms  – Contact  (exploring  threading  and  novel  new  algorithmic  approaches)  

126  

Page 127: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Agile  Components  

127  

•  AgileComponents  infrastructure  –  Defining  a  strategy  for  impact  

•  Deployed  through  algorithms  products  –  Trilinos  –  general  algorithms  toolkit  –  Dakota  –  UQ  and  op9miza9on  

•  Applica2ons  –  Sierra  –  RAMSES  –  Science  Codes  –  ASCR  

AgileComponents

R&D  100  Award  1000s  of  Users  Worldwide  Used  at  all  DOE  labs  

Page 128: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Exploring  Tools  (Ron  Green  et  al.)  

•  Intel  Studio  XE  2016  Tools  

•  Vtune  

•  MKL  

•  Vector  Advisor  

128  

Page 129: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

KNL  Developer  Challenges  

•  HBM  –  How  apps  take  advantage?  –  Open  ques9on,  answer  will  vary  by  applica9on  –  Tools  support  Fortran:  

•  !dir$  aZributes  fastmem  ::  <array  or  data  object>  –  Tools  Support  C/C++:  memkind  API  

•  hZps://github.com/memkind    –  Tools  Support  tracing  –  Intel  Vtune  support  (coming)  

•  Hybrid  Model  MPI  +  X  –  Where  X  ==  Threading  (OpenMP  4.x,  pthreads,  TBB,  Cilk  Plus)  

•  Vectoriza9on  –  absolutely  essen9al  –  -­‐xmic-­‐avx512  +  Improved  –qopt-­‐report  in  Intel  compilers  v15  and  v16  –  Intel  Advisor  XE  –  Vectoriza9on  Advisor  tool  

•  Premiers  in  the  Intel  Parallel  Studio  2016  Aug  2015  •  Requires  Intel  compilers  v16.0    

129  

Page 130: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Needs  

•  HBM  Explora9on  

•  Hackathons!  (with  real  codes)  

•  White  Boxes  

•  ???  

130  

Page 131: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Center  of  Excellence  LANL  

Hai Ah Nam  August  2015  

Page 132: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

Trinity Center of What?

Center of Excellence refers to a team, a shared facility or an entity that provides leadership, best practices, research, support and/or training for a focus area.

Slide 132

PURPOSE: “Support the transition of key applications to the Trinity System”

Page 133: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

Trinity Center of Who?

Center of Excellence refers to a team, a shared facility or an entity that provides leadership, best practices, research, support and/or training for a focus area.

Slide 133

PURPOSE: “Support the transition of key applications to the Trinity System” Team: Cray – John Levesque (Senior Level Analyst); Mike Berry (on site); Jim Schwarzmeier; performance team

Intel – Ron Green (on site); monthly interaction with SMEs, Technical Advocates, and others as needed

Site: CoE Site Lead Focus Area/Application LANL: Tim Kelley xRage (rad-hydro, multi-physics, multi-material, AMR)

SNL: Rob Hoekstra Sierra (Fluid, Therm, Aero, Struct. Dynamics, Solid Mech.)

LLNL: Shawn Dawson Mercury (MC neutron transport)

Page 134: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

Trinity Center of Who?

Center of Excellence refers to a team, a shared facility or an entity that provides leadership, best practices, research, support and/or training for a focus area.

Slide 134

PURPOSE: “Support the transition of key applications to the Trinity System” Team: Cray – John Levesque (Senior Analyst); Mike Berry (on site); Jim Schwarzmeier; performance team

Intel – Ron Green (on site); monthly interaction with SMEs, Technical Advocates, and others as needed LANL – Tim Kelley (CCS-7), xRage development team, Hai Ah Nam (CCS-2), Ben Santos (HPC-3), Cornell Wright (HPC-5), Jorge Roman (HPC-3), Peter Lamborn (HPC-3)

Bi-weekly meeting (Tues @ 11AM, 200-116)

Page 135: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

Trinity Center of Excellence

Center of Excellence refers to a team, a shared facility or an entity that provides leadership, best practices,

research, support and/or training for a focus area.

Slide 135

PURPOSE: “Support the transition of key applications to the Trinity System” Leadership, Early access to systems, software, tools Best Practices, The first to bleed… uncover “features” that are in fact bugs Research:

Engagement with other CoEs (Sierra, NERSC, ORNL, ANL) to bring back information to Trinity CoE & LANL

Share knowledge 1. hpc.lanl.gov à Platforms à Trinitite (Trinity) 2. ic-wiki.lanl.gov à Trinity Application Readiness (CoE, Open Science)

Page 136: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

Trinity Center of Excellence

Center of Excellence refers to a team, a shared facility or an entity that provides leadership, best practices, research, support and/or training for a focus area.

Slide 136

PURPOSE: “Support the transition of key applications to the Trinity System” Support & Cray Workshops & Tutorials Training: - 12 Steps to Performance Portability 6/16/15 (slides available)

- 1st 2 weeks December 2015 – Comprehensive Trinity Phase I Workshop - 4-5 codes teams doing hands-on performance tutorials (contact if interested) - 2 more KNL Workshops, dates TBD

Intel Discovery and Dungeon Sessions - xRage Discovery Workshop (6/3) - KNL Whitebox (early release) - Deep dive on applications kernels

Save the date.

Agenda coming

Page 137: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

xRage on Trinity

Slide 137

§  Major modernization efforts (cleaning house) –  Break into packages

•  Clarify data & control flow, reduce complexity, separate concerns

–  Clean up within packages •  Interfaces for packages, refactoring data structures •  Long term goal: physics code is a stateless transformer

–  Migrate packages to target architecture •  Prototyping ideas (tiled approach)

§  Address identified bottlenecks & inhibitors of vectorization (e.g. solver, data structures)

Page 138: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

Trinity Resources

§  [email protected] –  Access to Cray & Intel SMEs

§  LANL Help Resources –  HPC ([email protected]) –  Institutional Computing ([email protected])

§  Everyone –  Sharing best practices (IC-wiki, HPC.lanl.gov)

Slide 138

Page 139: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Campaign  Storage  Backup  Material  

139  

•  Managing  Files  We  provide  the  following  special  parallel  copy  and  list  commands  to  efficiently  manipulate  and  navigate  Campaign  Storage  

•  pfcp:  Copy  file(s)  from  sourcePath  to  desNnaNonPath  in  parallel    gpfst1-­‐fe$  pfcp  fileA  fileB  gpfst1-­‐fe$  pfcp  -­‐v  fileA*  subdir  

gpfst1-­‐fe$  pfcp  -­‐R  subdirA  subdirB          •  pfls:  Lists  file(s)  based  on  sourcePath  in  parallel  

 gpfst1-­‐fe$  pfls  fileA    gpfst1-­‐fe$  pfls  -­‐v  fileA*  gpfst1-­‐fe$  pfls  -­‐R  subdirA                

Page 140: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

Visualization Backup

Slide 140

Page 141: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Viz  Tools/Sosware  

•  EnSight  –  Commercial  tool  supported  by  CEI  

•  ParaView  – Open-­‐source  tool  supported  by  Kitware  Inc.  

•  VisIt  – Open-­‐source  tool  supported  by  LLNL    

•  All  tools  supported  for  Phase  1  •  All  tools  support  OpenSWR/Mesa  for  on-­‐plaGorm  rendering  (Haswell)  

141  

Page 142: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Viz  Scenarios  

•  Interac9ve  Visualiza9on/Post-­‐Processing  •  In-­‐situ  Visualiza9on  

–  Tightly  coupled  simula9on  code  and  in-­‐situ  viz  tool  for  run-­‐9me,  in-­‐memory  visualiza9on  and  analysis  

–  Sierra/Catalyst  (ParaView)  &  VisIt/LibSim  supported  •  In-­‐transit  Visualiza9on  

–  Visualiza9on/analysis  proceeds  concurrently  on          other  compute  nodes    

–  Poten9al  use  case  for  burst  buffer  – Ac9ve  research  area  

142  

Page 143: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Viz  Models  

•  Off-­‐plaGorm  rendering/Geometry  Transfer  –  Interac9ve  – Geometry/data  is  manipulated  on  Trinity  viz  nodes  but  rendered  by  viz  tool  on  non-­‐cluster  hardware  

 

•  On-­‐plaGorm  rendering/Image  Transfer  –  Interac9ve  –  Images  composited  &  rendered  on  Trinity  viz  nodes  and  sent  to  wai9ng  viz  tool  

– KNL  rendering  (with  OpenSWR)  is  uncertain  143  

Page 144: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Viz  Nodes  &  Queue  

•  2%  of  Haswell  and  Knights  Landing  par99ons  to  be  dedicated  visualiza9on/analysis  queue  

•  Con9nues  previous  Cielo  &  ASC  viz  models  •  Viz  node  alloca9on  reviewed  every  ATCC  cycle  •  Addi9onal  visualiza9on  and  analysis  resources  can  be  requested  for  special  needs  

•  Nodes  should  support  dynamic  linking  to  shared  object  libraries  and  TCP/IP  sockets  for  communica9on/DISCOM  

144  

Page 145: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Viz  on  Knights  Landing  

•  Xeon  Phi’s  have  severe  performance  handicap  when  rendering  through  tradi9onal  OpenGL/Mesa  –  Large  impact  on  in-­‐situ  visualiza9on    

•  Intel  ac9vely  developing  OpenSWR  as  a  sosware  rendering  solu9on  – OpenSWR  integra9on  with  Mesa  –  First  OpenSWR  deployment  to  Haswell  – Hopeful  OpenSWR  deployment  for  KNL  –  Tri-­‐Lab,  LBL  and  CEI/Kitware/VisIt  development  involvement  in  OpenSWR  and  KNL  rendering  

145  

Page 146: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Sandia-­‐specific  viz  

•  EnSight  – SNL:  “ensight2cluster  -­‐m  trinity”  

•  ParaView  – SNL:  “paraview2trinity”  

•  Sierra-­‐Catalyst  in-­‐situ  visualiza9on  – Exis9ng  capability  supported  as  part  of  phase  1  – Work  is  ongoing  to  support  Aero  

146  

Page 147: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Operated by Los Alamos National Security, LLC for the U.S. Department of Energy's NNSA

UNCLASSIFIED

End of Visualization

Backup Slide 147

Page 148: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Filesystems  Begin  Backup  slides  

Rob Hoekstra, Mike Glass, Ron Green(Intel) et al.  

August  2015  

SAND2015-­‐6916  PE  

Page 149: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

RAID  construc9on  differs  enabling  different  rebuild  strategies  

•  8  Data  +  2  Parity  over  10  HDDs  –  Each  8+2  exists  on  10  fixed  HDDs  

•  One  Disk  Fails  –  Read  9  other  parts  from  9  HDDs  

–  Write  reconstructed  part  to  one  HDD  

•  Failed  HDD  must  be  replaced  before  reconstruc9on  begins  

GridRAID  •  8  Data  +  2  Parity  over  41  HDDs  

–  2  spares  distributed  over  41  HDDs  

•  One  Disk  Fails  –  Read  9  other  parts  from  40  

HDDs  –  Write  reconstructed  part  over  

40  HDDs  •  Failed  HDD  must  be  replaced  

before  chance  of  losing  two  more  HDDs  –  Free-­‐up  spare  space  by  

wri9ng  spare  parts  from  40  HDDs  to  1  HDD  

Tradi9onal  RAID6  

Page 150: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

GridRAID  reconstructs  ~4x  faster  than  Tradi9onal  RAID  

•  Reconstruc9on  9me  dominated  by  read  – GridRAID  reads  from  40  HDDs  vs.  9  

•  GridRAID  replacement  HDD  reconstruc9on  uses  I/O  load  parameter  –  If  PFS  I/O  load  light,  reconstruc9on  uses  more  than  minimum  bandwidth  seang  

–  If  PFS  I/O  load  heavy,  reconstruc9on  won’t  exceed  minimum  bandwidth  seang  

•  Some  performance  degrada9on  is  possible  

Page 151: TrinityPlaormIntroducon’ and’Usage’Model · 2020-06-02 · Trinity’Usage’Model’G’Timeline’ Manuel Vigil August 2015! LA,UR15,26470 !

Trinity  Filesystems  End  Backup  slides  

Rob Hoekstra, Mike Glass, Ron Green(Intel) et al.  

August  2015  

SAND2015-­‐6916  PE