1 / 18

NERSC 5 Status NERSC 5 Team Bill Kramer Ernest Orlando Lawrence Berkeley National Laboratory

NERSC 5 Status NERSC 5 Team Bill Kramer Ernest Orlando Lawrence Berkeley National Laboratory. Here I Am Negotiating with Our New Vendor -NOT. Outline. Goals and Outline for NERSC 5 RFP Technology Observations Projections of NERSC-5 Characteristics. NERSC-5 Firsts.

rhea
Download Presentation

NERSC 5 Status NERSC 5 Team Bill Kramer Ernest Orlando Lawrence Berkeley National Laboratory

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. NERSC 5 Status NERSC 5 Team Bill Kramer Ernest Orlando Lawrence Berkeley National Laboratory

  2. Here I Am Negotiating with Our New Vendor -NOT

  3. Outline • Goals and Outline for NERSC 5 RFP • Technology Observations • Projections of NERSC-5 Characteristics

  4. NERSC-5 Firsts • The HECRTF and NRC Reports recommend coordination of Government procurements • NERSC-5 is possibly the first procurement to coordinate with other agencies • Joint application and kernel benchmarks with DOD HPCMP TI-06 • Gamess, Membench, CPUbench, SPIObench • Joint application and kernel benchmarks with NSF • Gamess, Milc, Paratec, Membench, CPUbench, SPIObench • NERSC-5 evaluation had four organizations observing the BVSS process • Larry Davis – DOD HPCMP, Rob Pennington – NCSA, Jim Kasdorf – PSC, Steve Brandt – LSU • At the end – all observers agreed with the decision and all complimented our process – and some may replicate it • NSF is using a number of things NERSC does

  5. NERSC 5 is probably the most reviewed and watched RFP • Determined to be a strategic IT project by DOE • Subject to formal Project Management methods and Earned Value Management • Monthly reporting • Quarterly IPT meetings • First IT procurement designated–administrative pathfinding for all of Office of Science (PNNL and ORNL follow) • Part of the NERSC OMB 300 submission • Quarterly reporting • Part of SC “Joule Metrics” to OMB • Must be no more than 10% above of cost and 10% behind schedule • Selected by Lab as the only IT project for best business practice review • Required unexpected acquisition plan review • RFP and Contract reviews • Site Office, Chicago Office, HQ Contract, MICS

  6. NERSC-5 Goals • Sustained System Performance over 3 years • 7.5 to 10 Sustained Teraflop/s averaged over 3 years • System Balance • Aggregate memory • Users have to be able to use at least 80% of the available memory for user code and data. • Global usable disk storage • At least 300 TB with an option for 150 TB more a year later • Integrate with the NERSC Global Filesystem (NGF) • Expected to significantly increase computational time for NERSC users in the 2007 Allocation Year • Dec 1, 2006 – November 30, 2007 • Have full impact for AY 2008 • Can arrive in FY 2006

  7. Greenbook and Other Plans • Coordinated requirements with the NUG Greenbook • Aligned with the NERSC plan for 2006-2010

  8. NUG Greenbook 2005General Recommendations • …The upcoming procurement must ensure a large increase in compute cycles available, as well as an appropriate balance of cache memory, processor memory, memory bandwidth, internode communication speed, intranode communication speed,… • …minimize the time-to-completion of large jobs, as well as maximize the overall efficiency of the hardware. …"large" can refer to jobs requiring long running times, a large number of processors, exceptionally large memory, or any combination of these • Significantly strengthen the computational science "infrastructure" at NERSC…local disk and archival storage, networking between NERSC's supercomputers and local storage and between NERSC and the WAN, and capabilities for local and remote data analysis and visualization must all be developed… • …current and potential future scientific applications are especially data or I/O intensive. These requirements should be carefully evaluated in order to support as wide a range of science as possible while also realizing significant benefits in both performance and cost in the computer configuration.

  9. NERSC-5 Benchmarks • Selection of benchmarks - several considerations • Representative of the workload • Represent different algorithms and methods • Are portable to likely candidate architectures with limited effort • Work in a repeatable and testable manner • Are tractable for a non-expert to understand • Can be instrumented • Authors agree we can use and distribute it • NERSC-5 started with approximately 20 candidates – settled on 7

  10. NERSC 5 Benchmarks • Application Benchmarks • CAM3 - Climate model, NCAR • GAMESS - Computational chemistry, Iowa State, Ames Lab • GTC - Fusion, PPPL • MADbench - Astrophysics (CMB analysis), LBL • Milc - QCD, multi-site collaboration • Paratec - Materials science,developed LBL and UC Berkeley • PMEMD – Life Science, University of North Carolina-Chapel Hill • Micro benchmarks test specific system features • Processor, Memory, Interconnect, I/O, Networking • Composite Benchmarks • Sustained System Performance Test (SSP), Effective System Performance Test (ESP), Full Configuration Test, Throughput Test and Variability Tests

  11. Application Summary

  12. Application Benchmarks represent 85% of the Workload

  13. Comments • Applications tests represent over 85% of the NERSC workload by discipline area. • Cover most frequently used programming libraries and programming languages. • For each benchmark, at least two test cases were prepared and validated on two or more architectures prior to the release of the RFP. • The two tests comprised medium, run on 64 processors, and large, run on 256 processors. • For technical reasons, the CAM3 benchmark was run on 56 and 240 processors • The GAMESS L benchmark was run on 384 processors to be compatible with the DOD HPCMP TI-06 benchmark • Extra-large benchmarks were prepared for MADbench (1024 processors) and MILC (2048 processors).

  14. Technology Observations • All bids were for multi core chips • Clock speed increasing at a much slower rate • The performance penalty is not as bad as we thought it might be • Power and cooling continue to increase • Flop/s per $ improving faster than Flop/s per Watt or Flop/s per sf • All proposals • Hybrid systems with Proprietary interconnects • High processor counts • One phase delivery - Influence of Sarbanes-Oxley? • Ran most to all benchmarks • Vertically integrated SW • NERSC Performance predictions were accurate • NERSC predicted systems would be between $0.75 - $1 per peak MFlop/s and $6-8 per SSP MFlop/s. • Bids and selection were better than expected.

  15. Technology Observations • Variety of topologies – between and within nodes • No vector or CPU accelerated systems proposed • Non commodity memory is very expensive • External storage cost getting cheaper for capacity • Delivery dates all at last moment • All proposers can move disk drives off the single system • It means they all use standards compliant storage • Declining viable bidders interested for full system and support of this size • Is SW risk getting better? Maybe • Efficiencies were stable and better than projected • ESP got much better commitments • No new technology for computer security • No innovative technology offered

  16. High Level NERSC 5 Features • >15 TF Sustained System Performance • Geometric Mean • Seaborg = .89 TF • Bassi ~ .8 TF • >300 TB of usable disk • Multiple 10 GigE connections • Many 1 GigE connections • O(50) FibreChannel Connections • CPUs’ > 4.5 GF per core • Multi-core sockets • Nodes are small SMPs • O(10,000) nodes • ~2 GB of memory per core • Expect programming model to remain MPI • Linux based user environment • Many reliability metrics • Access to NGF via GPFS

  17. The Phasing of NERSC-5 • Small Test System • Summer 2006 – user access not planned • Fall of 2006 - Phase 1 • 1/3 of compute resources • ~80% of I/O infrastructure • Winter 2007 – Phase 2 • 2/3 more compute nodes • Remaining disks and controllers • Winter 2008 – option to upgrade to at least double the sustained performance • Summer 2008 – Major software upgrade • Winter/Spring 2009 – option for a 1 Petaflop/s peak system – not currently in the NERSC budget

  18. Summary • The procurement team did a great job creating the ability to acquire an extraordinary NERSC-5 system • The negotiating team did a fantastic job reaching a fair but firm agreement that has all the features our users are accustomed • We now have a lot of work to do By the way, we are not done yet – stay tuned.

More Related