Download
physics on grids in the u s n.
Skip this Video
Loading SlideShow in 5 Seconds..
physics on grids in the U.S. PowerPoint Presentation
Download Presentation
physics on grids in the U.S.

physics on grids in the U.S.

113 Views Download Presentation
Download Presentation

physics on grids in the U.S.

- - - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - - -
Presentation Transcript

  1. physics on grids in the U.S.

  2. grids in the U.S. span projects, domains, countries, disciplines, oceans: they are a cultural phenomenon. Mainly funded by the US Department of Energy and the National Science Foundation, NASA and NIH. Particle Physics Data Grid. teams working together across a set of funded projects, research collaborations, long-term funded institutions and ad-hoc efforts Ruth Pordes

  3. Physicists are using these grids for data distribution and processing D0 Distributed Computing (Grid) proposed from get-go: Automated data requesting, multi-hop transport, and caching Across ~20 sites. 30% non-ATLAS resources. CMS simulations taking cycles “now and then”. Efficiency increased from 60-75% over the year through instrumentation and analysis. US ATLAS DC2 on Grid3: raw # jobs V. White, Chep2000 DZero reprocessing, summer 2005: Single site submission reprocessing on 9 sites in 7 countries; initial sharing of resources with LCG, OSG. Ruth Pordes

  4. Science results are emerging that rely on using grids:TeraGrid Simulation of 14 Billion Years of Universal Evolution Using the optimal (super)computer for each stage of the jobs. ..Simulated how a cube of the universe 250 million light years on each side developed over 14 billion years taking a snapshot of the universe every seven million years… http://www.interactions.org/sgtw/ Ruth Pordes

  5. Individual researchers do take advantage of spare compute cycles on the grid SDSS Southern Coadd project combined images from a ~300 square degree region along the southern celestial equator, imaged an average of about ten times to allow scientists to detect very faint and distant astronomical objects. ~15 sites used in an “opportunistic” more to coadd all available data. >44,000 computing jobs requiring >two terabytes of input data transferred to remote grid nodes were processed. Elongated blue object shows strong-lensing arc system discovered. The arc is a background galaxy whose image has been distorted by the gravitational strong lensing effect due to the foreground cluster of galaxies (orange-yellow objects). Ruth Pordes

  6. Cyberinfrastructure NMI PACI Terascale ITR Cyberinfrastructure and (LHC) Physics on Grids • Physics distributed computing, in particular LHC data analysis, needs more capability, scaling and performance than Grids in the US offer today. • x2~3 in end to end data transfer; ~x10 in job complexity x throughput; • improvement from 65->99% efficiency; • added data, job, error management services. • NSF Cyberinfrastructure (S. Kim, Oct 2004): • Science Frontiers as the Drivers • Balance capability and capacity • Massive data generated at the periphery; • E-Science (D. Atkins, Oct 2004): • Accelerate of cycles of discovery. • Reduce constraints of time and space. • Enhance sharing, re-use, multi-use of resources. • Enable rapid response to the unexpected. • Capture of process; not just end results • New “and-and” organizational forms. • Data Curation and Administration S. Kim, NSF, 2004 2000 2005 2010 Ruth Pordes

  7. E-labs : experiencial, continuous, and open forums for scholarly collaboration. Ruth Pordes

  8. F. Berman Internet 2004 Data Movement and Administration • Networks provisioning. Management of data transport for performance and efficiency. • Data Administration, history, location transparency, replication: • (non-conflicting) namespaces and identifiers, • directory and registry services - enterprise, interrealm; • Curation and Access • Multi-petabyte storage needs becoming “commonplace”. • Data lifecycles require technology transitions. • Intelligent common replication and placement services. All sciences and middleware groups focused on these issues. Ruth Pordes

  9. Security, Policy and Trust Beyond Authentication: accurately revoke, control, audit. Inter-domain negotiation: agreements and acceptance of “identities”, authorizations, policies. • Roles, authorities, responsibilities of Virtual Organizations Permissions and access controls: role definition, rights & delegations, privacy management, etc. • Technology: Extended X509 certificate attributes;SAML policy descriptions; Callouts from services to map attributes to rights. • Business model in support of roles e.g. CMSproductionsubmitter; SDSSserendipityworkinggroup etc. Research groups administer and prioritize access for their organizations. technologies from many CS & Physics Croups INFN/EGEE BNL Globus Virginia Tech UCSD/FNAL DESY/FNAL Ruth Pordes

  10. Errors, and Knowing what’s going on Reporting errors such they are intelligible and intelligent can be incrementally addressed. End to end Auditingto provide parallel, global, diagnosis of what went wrong in the complex sequence of events across many sites; e.g. The Fault Tolerant Shell ?a small language for that makes failures a first class concept. Aims to combine the ease of scripting with very precise error semantics, Providing for timeouts, retry, and alternation. You might think of this as exception handling for scripts. D. Thain, University of Notre Dame. try for 30 minutes cd /tmp rm -f data forany host in xxx yyy zzz wget http://${host}/fresh.data data end end EGEE Operations Workshop S. Belforte: “Experiments need to to control the progress of their application to take proper actions..helping the Grid to work by having it expose much of its status to the users Ruth Pordes

  11. Capabilities and User Environment User/Organizations need - easy, complete interfaces to the grid resources. - access to experiment databases, software releases, dataset merging and management; job splitting etc at the grid sites. Virtual Machine technologies are an active area of research as wrappers to dynamically glide in and isolate extra-site resident services. Ruth Pordes

  12. Grids in the U.S. Laboratories, Universities and Experiments allow their Farms and Storage Systems to be accessed across local and wide area networks and shared through the use grid middleware (as well as supporting direct use locally). Many times they make these resources accessible to multiple Grid domains. Users and Organizations expect transparent execution and storage environments that span across any boundaries between the resources they are using. There are thus many “Grids” that intersect and overlap. Each infrastructure is defined by the interfaces and administrative services that it provides. TeraGrid and Open Science Grid are 2 national grid infrastructures in the U.S. that are providing and augmenting their capabilities in support of physics. Ruth Pordes

  13. TeraGrid NSF funded large facilities: at 10 sites where the core set are the US national supercomputers. TeraGrid wide area network for high speed data serving between them. Diverse: hardware operating systems results in most application running at most at 1 or 2 sites. Deep Software stack (CTSS): includes compiler and libraries. Allocations for Use: through agreements: “exchangeable computing units” which can be spent at any site. Science Gateways: portals between a community and TeraGrid. Pretty broad definition is what each research group needs. US LHC simulations run on TeraGrid - issues of data handling and validation remain. Ruth Pordes

  14. Open Science Grid Community Built by application, middleware and site, research, university and laboratory contributors. • funded by DOE, NSF, (NIH) funded projects and institutions. • ~50 sites from Grid3, OSG, OSG Integration grid. Grid of Grids made coherent by a set of Interfaces, core Services and Policies. Autonomous Sites: Minimal impact on facilities existing resources and procedures; heterogeneous model but in practice all sites today are Linux. Integration Grid: Important structure for testing new services, new types of application, integration Extensible middleware: Core components packaged and integration by Virtual Data Toolkit (condor, globus, voms etc) with flexible process for additions and support. Interoperating: Core services interoperate with LCG-2, TeraGrid etc. + site in Brazil Ruth Pordes

  15. Open Science Grid Consortium “Open Science” like the “Open Source” in “Open Source” Software. Open to the sciences to contribute; Provide Common Infrastructure for many different sciences. US LHC S&C, Lab facilities, Trillium are lead contributors in experiment style collaboration: • US LHC is building the systems for LHC data analysis as part of the Open Science Grid. • The DOE Labs are making their resources available. Middlewareteams are lead contributors: Condor and Globus teams are energetically engaged. Market Place resource allocation allows for opportunistic use of resources, and internal management across sub-groups. working model – governance – of OSG is somewhere between the open source model for software and that of a particle physics experiment collaboration. Ruth Pordes

  16. physics & grids - a 2 way street Manage the Resources. Physicist - Computer Scientist teams optimally result in seamless merging of results into the mainstream of grid middleware to benefit others and adiabatic integration into experiment infrastructures to their required schedules. On an open road seeing a rainbow does not necessarily mean there is a pot of gold at the end. You can get bogged down without the right tools It’s the overall effectiveness that counts, not the secs to go from 0 to 100 Ruth Pordes