1 / 28

Outline

Outline. ENES The PRISM project: partners goals model components standard physical interfaces architecture and GUI PRISM first coupler: Oasis3 (08/2003): configuration communication interpolations/transformations PRISM final coupler (12/2004) configuration, communication,

Download Presentation

Outline

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Outline • ENES • The PRISM project: • partners • goals • model components • standard physical interfaces • architecture and GUI • PRISM first coupler: Oasis3 (08/2003): • configuration • communication • interpolations/transformations • PRISM final coupler (12/2004) • configuration, • communication, • interpolations/transformations

  2. ENES • Climate research in Europe: • Societal/political needs in Europe high (IPCC, mitigation,…) • Recognised excellence; scientific diversity (models, approaches,…) • How to organise Earth System modelling in Europe ? • The « one-big-centre-does-it-all » not suitable: • - expertise lies within national centres • diversity is key to research • Need for shared infrastructures: • - software (PRISM) • - hardware

  3. ENES • ENES: European Network for Earth System modelling • «Think tank » to organize, plan and seek funding for efficient distributed Earth System modelling in Europe • Follows a EuroClivar recommendation • Open to any institute/industry (MoU) • Coordinated by Guy Brasseur (MPI, Hamburg) • 50 members so far (http://enes.org)

  4. PRISM ENES A long term strategy: • Jointly develop shared software infrastructure for Earth System modelling • Provide European integrated service to access and use this infrastructure • Provide and manage hyper-computing access by 2010

  5. The PRISM project • PRISM: PRogram for Integrated Earth System Modelling • A European project, started December 2001, funded for 3 years by the European Commission (4.8 M€) • Coordinators: • Guy Brasseur (MPI, Hamburg) • Gerbrand Komen (KNMI, Amsterdam) • PRISM Director: Reinhard Budich (MPI)

  6. PRISM partners => 22 partners: leading climate research institutes and computer vendors • MPG-IMET, Germany • KNMI, Netherlands • MPI-MAD, Germany • Met-Office, UK • UREADMY, UK • IPSL, France • Météo-France, France • CERFACS, France • DMI, Denmark • SHMI, Sweden • NERSC, Norway • ETH Zurich, Switzerland • ING, Italy • MPI-BGC, Germany • PIK, Germany • ECMWF, Europe • UCL-ASTR, Belgium • NEC Europe • FECIT/Fujitsu • SGI Europe • SUN Europe

  7. PRISM goals Help climate modellers spend more time on science: • Provide software infrastructure • to easily assemble Earth system coupled models based on existing state-of-art European component models • launch/monitor complex/ensemble Earth system simulations

  8. PRISM goals • Define and promote technical and scientific standards for ESM: • Scientific standards: • Physical interfaces between model components • Global Earth System parameters • Technical standards: • Compiling, running, post-processing environment • Architecture and Graphical User Interface • Coupler and I/O software • Data and grid format • Coding and quality Interaction with other groups (ESMF, ESG/NOMADS, CF, RPN?,...)

  9. Atmosphere: Météo-France (ARPEGE), MPG-IMET(ECHAM), IPSL (LMDZ), MetOffice (Unified Model), UREADMY, INGV Atmospheric Chemistry: MPG-IMET, UREADMY, IPSL, MetOffice, Météo-France, KNMI Land Surface: IPSL (Orchidée), MetOffice, MPG-IMET, UREADMY, Météo-France (ISBA) Coupler: CERFACS, NEC, CCRLE, FECIT, SGI, MPI-MAD Regional Climate: SHMI, DMI, MetOffice Sea Ice: NERSC, UCL-ASTR, MetOffice, IPSL, MPG-IMET Ocean Biogeochemistry: MPI-BGC, IPSL, MPG-IMET, MetOffice Ocean: UREADMY, MetOffice (FOAM), MPI-M (HOPE), IPSL (OPA/ORCA) PRISM model components

  10. A proposal for PRISM standard O-A-SI physical interfaces: Atmosphere model 1- Surface pressure 2-4 Air temperature, humidity and wind 5- Wind module 6- Height of these 4 variables 1- Rainfall + int. energy 2- Snowfall + int. energy 3- Incoming solar radiat. 4- Solar zenith angle 5- Fraction of diffuse solar radiation 6- Downward infrared radiation 7- Sensitivity of atmos temp. & humidity to surf. fluxes 1*- Sensible heat flux 2*- Surf. emissivity 3*- Albedo, direct 4*- Albedo, diffuse 5*- Surf. radiative temp. 6*- Evaporation + int. energy [+ Qlat] 7*- Wind stress 8- Subgrid fractions 3 1 1 2 Surface layer turbulence 2 1*- rCd 2*- rCe 3*- rCh 1*- Surf. Temp 2*- Surf. Roughness 3*- Displacement height 4x- Surface velocity 4 4 5 +3 Ocean surface module Land surface model Sea ice model wave model 1x- Non solar heat flux 2x- Solar radiation 3x- Fresh water flux 4x- Salt flux 5x- Wind stress 6x- U^3 7x- Mass of snow and ice 8-Subgrid fractions 1-2 Temp./Salinity at sea-ice base 3- Sea surface temperature 4- Surf. radiative temp. 5- Surface ocean current 6- Sea surface salinity 7- Surface height 8- Absorbed solar radiation (in first oceanic layer) 1- Continental runoff + internal Energy 8 Note on subgrid fraction dependance: <>x- Sea Ice categories (incl. open ocean) <>*- Sea Ice or Land Surf. categories 6 7 Iceberg parameters Ocean model

  11. Prism central server Web User/ developer Web Prism local site PRISM architecture and GUI: • PRISM central server + PRISM local sites • GUI: adaptation of ECMWF prepIFS and SMS scheduler

  12. PRISM first coupler: Oasis3 • Based on Oasis developed since 1991 in CERFACS • to couple existing GCMs developed independently • at the time: • Models at relatively low resolution (~10000-20000 pts) • Small number of 2D coupling fields (~10) • Low coupling frequency (~once/day) • flexibility was very important, efficiency not so much! • performs: • synchronisation of the component models • coupling fields exchange and interpolation • I/O actions • tested on VPP5000, NEC SX5, SGI Octane and O3000, Compaq Alpha cluster, Linux PC cluster (MPI-Globus)

  13. PRISM project first coupler: Oasis3 • …and : • AWI (Germany) • PIK (Germany) • Met Office (UK) • UGAMP (UK) • KNMI (Netherlands) • CSIRO (Australia) • FSU/COAPS (USA) • LASG (China) • JAMSTEC (Japan) • …? Oasis regular users: • CERFACS • METEO-FRANCE (France) • IPSL- LODYC, LMD, LSCE (France) • ECMWF (UK) • UCL (Belgium) • MPI - M&D (Germany) • SMHI (Sweden) • BMRC (Australia) • IRI (USA)

  14. PRISM project first coupler: Oasis3 • Oasis3 configuration: • In text file namcouple read by Oasis3 at the beginning of the run, e.g. • total run time • number and names of component models • number and names of coupling fields; for each field: • source and target symbolic name • coupling and/or I/O status, • coupling or I/O period • transformations/interpolations • … • Component model grid (longitudes, latitudes, masks, mesh surfaces, mesh corner locations) must be available in binary or NetCDF files.

  15. O A O A Oasis3 O A O B A B A B A A file A A PRISM project first coupler: Oasis3 Oasis3 communication: • New PRISM System model interface (PSMILe) based on MPI1 or MPI2 message passing • Parallel communication between parallel models and Oasis3 interpolation process • Direct communication between models with same grid and partitioning • I/O functionality (automatic switch between coupled and forced mode) • Modularity: at each model time step, exchange is performed or not depending on user’s specifications in namcouple. • Automatic time integration depending on user’s specification

  16. Oasis3 O A O A Oasis3 O A O PRISM project first coupler: Oasis3 • Oasis3 interpolations/transformations => performed by separate sequential process • => on 2D scalar fields only • Interfacing with RPN Fast Scalar INTerpolator package • nearest-neighbour, bilinear, bicubic for regular Lat-Lon grids • Interfacing with SCRIP1.4 library (Los Alamos Software Release LACC 98-45): • nearest-neighbour, 1st and 2nd order conservative remapping for all grids • bilinear and bicubic interpolation for «logically-rectangular» grids • Bilinear and bicubic interpolation for reduced atmospheric grids • Other spatial transformations: flux correction, merging, etc. • General algebraic operations

  17. Need to optimise and parallelise the coupler PRISM project final coupler • Higher resolution, parallel and scalable models • Higher coupling frequencies desirable • Higher number of models and (3D) coupling fields • The final PRISM coupler will be composed of: • a Driver • a Transformer • a new PRISM System Model Interface Library

  18. PRISM project final coupler • Final coupler configuration (XML files): • The user chooses the models through the GUI. • Each component model comes with: • an Application Description (AD) • a Potential Model Input and Output Description (PMIOD). • The user configures his particular coupled run through the GUI : • total run time, etc. • for each field described in the PMIOD: • source or target • coupling or I/O status, • coupling or I/O period • transformations/interpolations, etc. • Based on the user’s choice, the GUI produces the XML configuration files. • At run-time the Driver reads and distributes configuration information. • The PSMILes and Transformer act accordingly to the user’s specifications.

  19. OB O1 OB C O1 C C C OB C PRISM project final coupler Final coupler communication: • More elaborate PSMILe based on MPI1 or MPI2 (grid definition transferred through the PSMILe API) • Modularity as for Oasis3: at each model time step, exchange is performed or not depending on user’s specifications. • As for Oasis3, automatic time integration • As for Oasis3, I/O functionality (automatic switch between coupled and forced mode) • Parallel communication: as for Oasis3 • + repartitioning. • Parallel calculation of interpolation weights and addresses in the source PSMILe • Extraction of useful part of source field only.

  20. The PRISM project final coupler • Final coupler interpolations/transformations • => as for Oasis3 + • Support of vector fields • Support of 3D fields • More flexibility for field combination/merging, etc.

  21. Conclusions • ENES and PRISM • PRISM first coupler: Oasis3, now available • PRISM final coupler prototype due 11/2003 • PRISM final coupler due 12/2004 • … and after PRISM ? • Follow-on project re-submitted at next EU-call in 2004 (CAPRI rejected) • International interaction and collaboration essential in all cases! • http://www.enes.org ; http//www.cerfacs.fr/PRISM/prism.html • valcke@cerfacs.fr

  22. PRISM project first coupler: Oasis3 • Oasis3 communication; PSMILe API: : • Initialization: • call prism_init_comp(…) • Retrieval of component model local communicator • call prism_get_localcomm (…) • Coupling or I/O field declarations (name, type, shape, local partition, …) • call prism_def_var(field_idx, …) • End of definition • call prism_enddef(…) • In model time stepping loop, coupling or I/O field exchange • call prism_put(field_id1, time, field_array1, ierror), • call prism_get(field_id2, time, field_array2, ierror) • => Automatic averaging/accumulation, coupling exchange, and/or I/O depending on time argument and user’s specifications in namcouple • Termination: • call prism_terminate(…)

  23. PRISM project final coupler Final coupler communication; PSMILe API: : • As for Oasis3 PSMILe + • Definition of grid (1D, 2D, 3D) • call prism_def_grid(…) • call prism_set_corners(…) • call prism_set_mask(…) • Definition of grid for vector and bundle fields • call prism_set_vector(…) • Call prism_set_subgrid(…) • Coupling or I/O field declarations support vector, bundles, 1D, 2D and 3D fields • Extraction of SCC and SMIOC information: • call prism_get_persist(…)

  24. ATM AD Driver ATM PMIOD V1: in, metadata V1 V2: out, metadata V2 V3: out, metadata V3 user user OCE AD • SCC • ATM:... • OCE:... • LAND:... user OCE PMIOD V1: out, metadata V1 V2: in, metadata V2 user ATM OCE OCE SMIOC V1 : to ATM, T1 V2 : from ATM, T2 ATM SMIOC V1 : from OCE, T1 V2: to OCE, T2 V3 : to LAND V1 V2 T user V1 V3 V2 LAND AD user LAND V4 LAND PMIOD V3: in, metadata V3 V4: in, metadata V4 Definition Phase fileV4 Composition Phase Deployment Phase LAND SMIOC V3 : from ATM V4 : from fileV4

  25. Driver Mi PMIOD V1: out, metadata V1 V2: out, metadata V2 V3: in, metadata V3 Mj PMIOD V1: in, metadata V1 V4: out, metadata V4 user user user SCC V1 : Mi -> Mj, Tli, Tnlij V2 : Mi -> Mj, Tij (+ V6) V4 : Mj -> Mk user Mj Mi Mj SMIOC V1 : cf SCC V4 : cf SCC Mi SMIOC V1 : cf SCC V2 : cf SCC V3 : in, fileV3, Tli V2 V7 V2 T V1 V1 V3 user V4 Mk PMIOD V4: in, metadata V4 V5: in, metadata V5 V7 fileV3 V6 V5 V5 user Mk fileV6 Mk SMIOC V4 : cf SCC V5 : in, fileV5, TnlV5k fileV5 Definition Phase Composition Phase Mi: Model i T: Transformer PMIOD: Potential Model Input and Output Description Deployment Phase SMIOC: Specific Model Input and Output Config. SCC: Specific Coupling Configuration

  26. Share Software structure of an Earth System Model Running environment Coupling infrastructure Scientific code Supporting software

  27. PRISM ESMF On going PRISM / ESMF collaboration Earth System Model Running environment Coupling infrastructure User code Supporting software

More Related