1 / 18

Nuclear Physics Greenbook Presentation (Astro,Theory, Expt)

Nuclear Physics Greenbook Presentation (Astro,Theory, Expt). Doug Olson, LBNL NUG Business Meeting 25 June 2004 Berkeley. Reminding you what Nuclear Physics is.

colin
Download Presentation

Nuclear Physics Greenbook Presentation (Astro,Theory, Expt)

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Nuclear Physics Greenbook Presentation(Astro,Theory, Expt) Doug Olson, LBNL NUG Business Meeting 25 June 2004 Berkeley

  2. Reminding you what Nuclear Physics is. • The mission of the Nuclear Physics (NP) program is to advance our knowledge of the properties and interactions of atomic nuclei and nuclear matter and the fundamental forces and particles of nature. • The program seeks to understand how quarks bind together to form nucleons and nuclei, to create and study the quark-gluon plasma that is thought to have been the primordial state of the early universe, and to understand energy production and element synthesis in stars and stellar explosions. D. Olson, NP Reqs, NUG Mtg

  3. D. Olson, NP Reqs, NUG Mtg

  4. Contents • Questions • Answers • Some Observations D. Olson, NP Reqs, NUG Mtg

  5. Questions about current needsand 3-4 years out • 1. What are your most important processing needs?   (e.g. single CPU speed, # of parallel CPU's, memory, ...) 2. What are your most important storage needs?   (e.g. I/O bandwidth to disk, disk space, HPSS bandwidth, space, ...) 3. What are your most important network needs?   (e.g. wide-area bandwidth, bandwidth between NERSC resources, ...) 4. What are your most important remote access needs?   (e.g. remote login, remote visualization, data transfer,    single sign-on to automate work across multiple sites, ...) 5. What are your most important user services needs?   (e.g. general helpdesk questions, tutorials, debugging help, ...) 6. Do you have special software requirements, if so what? 7. Do you have special visualization requirements, if so what? 8. Is automating your work across multiple sites important?   Called distributed workflow.  Sites could be other large   centers, a cluster, your desktop, etc. 9. Anything else important to your project? • Asked to all PI’s with NP awards • 9 responders & good cross section D. Olson, NP Reqs, NUG Mtg

  6. Responses • Astronomy • Swesty, SUNY SB, TSI Collaboration • Nugent, LBNL • Nuclear Theory • Ji, NCSU, QCD • Pieper, ANL, Light nuclei • Dean, ORNL, Nuclear many body problem • Vary, Iowa State, Nuclear reactions • Lee, Kentucky, Lattice QCD • Experiment • Klein, LBNL, IceCube • Olson, LBNL, STAR D. Olson, NP Reqs, NUG Mtg

  7. 1. What are your most important processing needs?   (e.g. single CPU speed, # of parallel CPU's, memory, ...) • Theory • Small cluster of nodes (1-16 nodes) and long run duration (12-24h or more) • implementation of execution of single-processor tasks would be welcomed. • Faster processors are always a good thing! • Up to 2048 parallel (for SMMC) • Need faster interprocessor b/w (for two other codes) • Eq. 256 CPU Altrix is 7X faster than Seaborg • total number of cycles obtained over a range of processors (50 to 500). generally .5 - 2 Gbytes/processor are needed on processors with 1-4x seaborg speed. • Memory per CPU - by far the most important to our project (saves I/O to disk and/or cuts down on inter-node communication) D. Olson, NP Reqs, NUG Mtg

  8. 1. What are your most important processing needs?   (e.g. single CPU speed, # of parallel CPU's, memory, ...) • Astro • Getting a little more CPU speed is ok, but BY FAR we need faster bandwidth between processors and between nodes. • 1024-2048 processors now • Lower latency communications • Expt • Not parallel algorithms, compute at PDSF & other linux clusters D. Olson, NP Reqs, NUG Mtg

  9. 2. What are your most important storage needs?   (e.g. I/O bandwidth to disk, disk space, HPSS bandwidth, space, ...) • Theory • Bandwidth to disk and gpfs disk space (increase by 100X for disk space). • Inode limit is a persistant pain, but can be lived with. • HPSS bandwidth, space • Single file system across machines (seaborg, newton) • Astro • Fine now, may change as we move more to 3-D. • improved parallel I/O throughput to disk for > 1024 processor jobs • increased scratch disk capacity • Expt • Database, MySQL in use now • Disk - scalable size & I/O performance, >100TB, > 1 GB/sec • HPSS – size & I/O • Automated caching, replication & I/O load balancing D. Olson, NP Reqs, NUG Mtg

  10. 3. What are your most important network needs?   (e.g. wide-area bandwidth, bandwidth between NERSC resources, ...) • Theory • Moving 20 GB datasets today ORNL-NERSC-MSU • Moving 0.5 TB datasets in 2 years ORNL-NERSC-MSU-LLNL-PNNL • bandwidth between NERSC resources • Astro • Improved throughput between NERSC, ORNL, and Stony Brook • Expt • WAN bandwidth end-to-end (means endpoints or other LAN effects are often the problem), labs & universities D. Olson, NP Reqs, NUG Mtg

  11. 4. What are your most important remote access needs?   (e.g. remote login, remote visualization, data transfer,    single sign-on to automate work across multiple sites, ...) • Theory • X-windowed system • Data transfer is becoming an increasingly important need. • ssh/scp with authorized keys is fine. one-time passwords would severely handicap my use of a local emacs and tramp to edit, view, and transfer files. • Astro • Single sign-on to allow process automation is very important right now. Of CRITICAL importance is avoidance of one-time authentication methods which would kill any hopes of scientific workflow automation. • Some remote viz. • Expt • Data transfer • Single sign-on across sites for automated workflow D. Olson, NP Reqs, NUG Mtg

  12. 5. What are your most important user services needs?   (e.g. general helpdesk questions, tutorials, debugging help, ...) • Theory • support/online-help for Windows-based X-servers • Programming languages online-references or links to online-references would be great. • General helpdesk and sometimes tutorials. • Online tutorials (stored and indexed) • Astro • Biggest problems are dealing with new compiler bugs. • Performance optimization, requires help people who have access to the IBM compiler group to code kernels tuned. • Expt • General user support and collaboration software installation is very good. • Need troubleshooting across sites and WAN D. Olson, NP Reqs, NUG Mtg

  13. 6. Do you have special software requirements, if so what? • Theory • Part of my plans involve solving a large sparse eigenvalue problem. Software like Aztec is going to be useful for this. • Astro • We continue to rely on the availability of HDF5 v1.4.5 for our I/O needs on seaborg. HDF5 1.6.x will not suffice as we have uncovered show-stopping bugs in this release. • Expt • Community & collaboration software (CERN, ROOT, …) • Current install/maintenance procedures work well D. Olson, NP Reqs, NUG Mtg

  14. 7. Do you have special visualization requirements,if so what? • Theory • We would welcome introduction of visual debugging tools for fortran/C++, especially for MPI or HPF programs, if possible of course. • Astro • Some, but most have been covered by the viz group. • We continue to rely heavily on the NERSC viz group to help us address our viz needs. D. Olson, NP Reqs, NUG Mtg

  15. 8. Is automating your work across multiple sites important?   Called distributed workflow.  Sites could be other large   centers, a cluster, your desktop, etc. • Theory • Yes. We are considering how to develop common component software for nuclear physics problems. The low-energy nuclear theory community will increasingly move towards integrated code environments. This includes data movement, and workflow across several sites. (We do this now with NERSC/ORNL/MSU). • I do a fair amount of post processing with Speakeasy on my workstation. This involves mixing results from NERSC, Argonne's parallel machines, and Los Alamos' qmc at present. D. Olson, NP Reqs, NUG Mtg

  16. 8. Is automating your work across multiple sites important?   Called distributed workflow.  Sites could be other large   centers, a cluster, your desktop, etc. • Astro • Yes! We are currently working with the SPA (Scientific Process Automation) team from the SciDAC Scientific Data Managment ISIC on automatic our workflow between NERSC and our home computing site at Stony Brook. • Expt • Yes. Experiment collaboration computing is spread across large & small sites and desktops. Need more integration with security & tools for a more seamless environment. D. Olson, NP Reqs, NUG Mtg

  17. 9. Anything else important to your project? • Theory • Nersc is a great help, keep up a great work! • My biggest concern with NERSC at the present time is that it has fallen behind the curve on the machine front. While I still consider NERSC a valuable resource to my research, I have diversified significantly during this FY. • Any performance tools, such as POE, that help diagnose the bottlenecks in a code and help  suggest routes to improvements. • Astro • Memory bandwidth and latency. • Expt • User management across site & national boundaries. Separate user registration & accounts across many sites will become too burdensome. Think single sign-on & seamless! D. Olson, NP Reqs, NUG Mtg

  18. Observations • A strong need for greater inter-processor bandwidth • Faster processors, more memory • Single file system view across NERSC • Greater parallel FS performance (>1024) • More space • More/better data management tools • Single sign-on across sites • Help with Inter-site (WAN) issues • Much scientific computing now has workflow across several sites D. Olson, NP Reqs, NUG Mtg

More Related