Loading in 2 Seconds...
Loading in 2 Seconds...
GEWEX Aerosol Panel: A critical review of the efficacy of commonly used aerosol optical depth retrieval. GEWEX Panel: Sundar Christopher, Richard Ferrare, Paul Ginoux, Stefan Kinne, Gregory G. Leptoukh, Jeffrey Reid, Paul Stackhouse
Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.
GEWEX Panel: Sundar Christopher, Richard Ferrare, Paul Ginoux, Stefan Kinne, Gregory G. Leptoukh, Jeffrey Reid, Paul Stackhouse
Programmatic support: Hal Maring, Charles Ichoku, Bill Rossow
Comments on this presentation: email@example.com
By late October we expect to send sections to the instrument science teams for “fact checking.”
Report will layout the nature of the aerosol problem, with a synopsis of the literature and commentary on verification methods and findings.
Phase 2-detailed independent evaluation, will not start until MODIS collection 6 and MISR v 23 is officially released.bottom line up front
But, most products are in the twilight zones of “research,” “development” and “production.”
This is reinforced with the funding situation where money for product development, maintenance, and verification is limited. Developers spend more time “using” than “supporting” their products.
By the time the wider community figures out how a product is doing, a new version is released.
Situation: Confusion and some rancor in the community as to the actual efficacy and appropriate application of these data sets
Response: Reform the GEWEX Aerosol Panel (GAP)the aerosol problem
Team is to be small and well rounded. Team size is to be no larger than what is necessary (e.g., no large committees).
Members are to be ‘jurors drawn from accomplished peers’ to examine the 7 most common global products: AVHRR (GACP and NOAA), MISR, MODIS (Standard & Deep Blue), OMI, and POLDER.the GAP mandate
Phase 2: Based on Phase1 examine in detail specific issues in the generation of retrieval and gridded products.
Peer review of the peer review: Findings given to teams for comment before release. After an iteration, team rebuttals can be made public record.
Where are we now? Nearing the end of phase 1. Report is 100+ pages and growing…the GAP mandate
However, all of the world’s major Numerical Weather Prediction (NWP) have aerosol assimilation programs and aerosol data has worked its way into numerous applications.
There are aerosol observability issues
reliable and timely delivery
bias (contextual, sampling) understanding / removal
error characterization (essential for assimilations)customers and issues
data product evaluation, validation, and verification.
Reliable and timely delivery of satellite aerosol and fire products is only half the challenge. If products are to be integrated, then biases need to be removed through careful product evaluation and verification. Contextual & sampling biases need to be understood. Because of possible degradation in model performance through data assimilation, aerosol product error characterization has been emphasized more in the operations than climate communities. Indeed,despite popular misconceptions, operational data characterization requirements are often more strict than what is commonly used in the climate research community.
Reid, J. S., Benedetti, A., Colarco, P. R., Hansen, J.A., 2011. International operational aerosol observability workshop, Bulletin of the American Meteorological Society, 92, Issue 6, pp. ES21-ES24 doi: 10.1175/2010BAMS3183.1
Richard Ferrare (NASA LaRC): lidar, field work, multi-sensor products
Paul Ginoux (NOAA GFDL): Global modeling and aerosol sources
Stefan Kinne (Max Plank): GEWEX Cloud, AEROCOM
Gregory Leptoukh (NASA GSFC): Level 3 product development and distribution
Jeffrey Reid (NRL): co-chair, aerosol observability, field work, verification, operational development
Paul Stackhouse: GEWEX radiation, atmospheric radiation and energetics.panel membersHQ: Hal Maring and Charles Ichoku
“Advantage of Human Eye”
Basically want to know were stuff is. Can do one-up corrections
Model Aps, V&V, Inventory
Have stronger time constraints and need spatial bias elimination.
Quantify bias & uncertainty everywhere and correct where you can.
Parametric Modeling and Lower Order Process Studies
Correlations de-emphasize bias
Need to de-trend biases in retrieval and in sampling
Higher Order Process Study
Push multi-product and satellite data
V&V statistics must speak to these applications!
Hence, there is no “one size fits all” error parameter. Sorry….
differences are a mix of radiometric bias
tend to overestimate
AOD (at low AOD)
over oceans especially MISR
Zhang and Reid, 2010
global AOD difference between sensors
Mischenko et al., 2007
ASO clear sky bias, Zhang and Reid 2009
- slope -
If we knew AOD then we would not need MODIS. All we have is MODIS’s own estimate of AOD….
From Shi et al., 2010, ACP
There is no shortage of validation studies. But, they tend to be direct regression based, have important details missing, and are conducted over limited periods of times and/or locations. Hence, they tend to be of limited utility.
While there are many cases of satellite cal-val components from field missions, analyses are usually not repeated for new product versions.
Even well designed third party studies are generally not utilized or cited by the production teams.verification (1)
Over land, there is strong regional and temporally correlated biases across both algorithms and sensors, largely due to the lower boundary condition.
Radiance calibration is a significant problem and pops up in indirect ways. NASA is working on it.
Demonstration of diversity in aerosol products has little barring on relative product efficacy.
bottom line: The difference between “face value statistics” and an error bar for an individual retrieval is vast. How does this effect you? Depends on what you do with the data.verification (2)
Better strategies for “level 3 products” need to be devised and supported. One size fits none…..
One size fits all verification does not work either. But, there is a total lack of agreement on key verification metrics. The USER community needs to agree on what they think is important.
It should be a programmatic requirement of the science teams to develop prognostic error models as part of any mass produced and distributed product. Program offices need to fund this.key recommendations (1)
Developers and outside entities to work more together in verification studies.
Field work needs to be better utilized. After a first round of verification studies, next generation algorithms do not typically make use of older studies. Field work should focus more on verifying higher level products.key recommendations (2)