1 / 28

Kuai Fang 1 , Chaopeng Shen 1 , & Daniel Kifer 2

Deep learning for estimating land surface response with uncertainty: soil moisture and other opportunities . Kuai Fang 1 , Chaopeng Shen 1 , & Daniel Kifer 2 1 Civil and Environmental Engineering, Penn State University 2 Computer Science and Engineering, Penn State University

iblevins
Download Presentation

Kuai Fang 1 , Chaopeng Shen 1 , & Daniel Kifer 2

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Deep learning for estimating land surface response with uncertainty: soil moisture and other opportunities Kuai Fang1, Chaopeng Shen1, & Daniel Kifer2 1Civil and Environmental Engineering, Penn State University 2Computer Science and Engineering, Penn State University cshen@engr.psu.edu https://github.com/mhpi @ChaopengShen

  2. Content • Deep learning models for soil moisture • Separating uncertainties with LSTM models • Perspectives on incubating DL-powered research

  3. Background • Land surface provides important lower boundary conditions (vapor, heat, momentum, CO2) for the atmosphere • Assimilating soil moisture bias improves weather forecasts • Land surface models often produce biases • Is this idea “oversold”? We need to see...

  4. Background • Soil Moisture Active Passive (SMAP) • Launched recently (2015/04) • 2~3 days revisit time • Senses moisture-dependent top surface soil • Use cases • Weather forecasting (land feedback) • Runoff prediction • Vegetation water stress • Agriculture • How does soil moisture really work? • Inferring precipitation?

  5. Use DL to for moisture projection Self-learned Memory system • Time Series Deep Learning (DL) • Long-short Term Memory (LSTM) • Use DL as a dynamical model! • (hopefully) less bias for atm. models • Free from structural assumptions • May be used to correct feedbacks to atmosphere (y)

  6. Soil moisture prolongation LSTM model (y) SMAP L3_P Atm. Forcing (NLDAS) train (x) Surface soil moisture (optional) Land Surface Model (Noah) solutions Soil texture, slope, land cover, irrigation, depth to water table, etc

  7. Using LSTM, we prolonged SMAP to … • Spatio-temporally seamless coverage

  8. Using LSTM, we prolonged SMAP to … • Examined comparison with in-situ data & long-term projections • Long-term projections DL models are powerful. Combined DL-SMAP could help improve soil moisture estimates But can we estimate the uncertainty from those DL models? – important for DA

  9. Deep learning models for soil moisture • Separating uncertainties with LSTM models • Perspectives on incubating DL-powered research

  10. How to estimate the uncertainty? MCD+A Input-dependent Aleatoric uncertainty • Cannot be reduced by more data • Magnitude maybe predictable Epistemic uncertainty • Comes from uncertain model param. • Captured by Monte Carlo dropout (MCD) To be predicted by LSTM along with yi W: weight Ŵ: dropout weight x: input f: deep network Shen 2018 WRR Kendall and Gal 2017

  11. Uncertainty estimation with deep network? Cited 807!! • Gal 2016 argued that NN+Monte Carlo dropout=deep Gaussian process http://mlg.eng.cam.ac.uk/yarin/blog_3d801aa532c1ce.html • Issues: • (Very) approximate derivation between neural network and Gaussian Process (prior & variation distribution). Does this really work? • Requires hyper-parameter (dropout). Conditioned by training data • Does it really measure similarity? • Accurately reflect uncertainty?

  12. Probabilistic DL Model Soil Moisture Prediction (y) Forcing Input DL Model Aleatoric Uncertainty (σx) ŷ1 Dropout Model 1 Monte Carlo Dropout ŷ2 Dropout Model 2 …… σx ŷn Dropout Model n Combine of aleatoric and epistemic uncertainty: + where is the dropout rate σmc:Std of MCD ensemble

  13. Can error magnitude? • Training • 2015/04 - 2016/04 • Validation • 2016/04 - 2017/04 • hyper-parameter ( is a function of the dropout rate) • Temporal test • 2017/04 - 2018/04 • same pixels as training set

  14. Do those two uncertainties behave as intended?  Does uncertainty respond to aleatoric noise (but Gaussian noise added to SMAP observations: σnoise

  15. Do those two uncertainties behave as claimed?  Does epistemic uncertainty behave like GP? Training Basin First time it was reported that MCD responds to similarity!

  16. What if training data is not representative? Representative training set Biased training set

  17. Conclusions • An automatically estimated error model! (disruptive to UQ?) • We showed that & mostly works as intended. Through spatial autocorrelation (in input attributes), does measure similarity. • When training data is representative, MCD+A is good at predicting error , but requires a hyper-parameter. If training data is biased, will be misled. Often this evaluation was not thorough • Better prior error model helps

  18. Recent enhancement: closed-loop predictions

  19. Also applying LSTM to streamflow predictions

  20. Deep learning models for soil moisture • Separating uncertainties with LSTM models • Perspectives on incubating DL-powered research

  21. Process-based modeling vs machine learning ===== BDML strength ===== • Built from the top-down, directly from observations  accurate • Less biased • Identify things we don’t know? • Highly efficient in computation ==== Limitations ====== • Can’t observe everything! • May be difficult to interpret • May not fully respectphysical laws • Does not understandcausal relationships ===== PBM strength ==== • Built from the bottom-up to observe emergent patterns • We know what we put in • We can do experiments & identify causal relationships ===== Limitations ==== • Human biases • Parameter calibration • What we don’t know? • Errors compound? Synergy? Top down Bottom up

  22. Potential hydrologic DL applications New DL-based method • Challenging-to-model problems • Prepare inputs • Dynamical predictions • Scaling • Parameterization • Measure information content of predictors • …… • Extract knowledge?

  23. How can we incubate progress as a community? • Open competitions as an organizing event • Community organization, shared resources (data, models, infrastructure…)

  24. Interpretive/interrogative • Visualization • Relevance backpropagation • Trained explanation network

  25. Thank you! Email: Chaopeng Shen cshen@engr.psu.edu @ChaopengShen https://github.com/mhpi

  26. Hyper-parameter – dropout rate

  27. How does the prior error model impact uncertainty estimation? Apply Inverse-Gamma distribution as prior to σx: Temporal generation test result

More Related