1 / 74

Data Mining and Gated Expert Neural Networks for Prognostic of Systems Health Monitoring

Data Mining and Gated Expert Neural Networks for Prognostic of Systems Health Monitoring. Mo Jamshidi , Ph.D., DEgr., Dr. H.C. F-IEEE, F-ASME, F-AAAS, F-NYAS, F-HAE, F-TWAS Regents Professor, Electrical and Computer Engr. Department & Director, Autonomous Control Engineering (ACE) Center

ludlow
Download Presentation

Data Mining and Gated Expert Neural Networks for Prognostic of Systems Health Monitoring

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Data Mining and Gated Expert Neural Networks for Prognostic of Systems Health Monitoring MoJamshidi, Ph.D., DEgr., Dr. H.C. F-IEEE, F-ASME, F-AAAS, F-NYAS, F-HAE, F-TWAS Regents Professor, Electrical and Computer Engr. Department & Director, Autonomous Control Engineering (ACE) Center University of New Mexico, Albuquerque, NM, USA Advisor, NASA JPL (1991-93), Headquarters (1996-2003) Sr. Research Advisor, US AF Research Lab. (1984-90,2001-present) Consultant, US DOE Oak Ridge NL (1988-92), Office of Renewable Energy (2001-2003) Vice President, IEEE Systems, Man and Cybernetics Society http://ace.unm.edu www.vlab.unm.edumoj@wacong.org.org Fairbanks, Alaska, USA May 24 2005

  2. OUTLINE Definition of Prognostics History of Prognostics Approaches of Prognostics Principle Component Analysis – PCA PCA via Neural Network Architecture Prognostics via Neural Networks Gated Approach to Hardware Prognostics Applications – Health and Industry Conclusion and Future Efforts

  3. Prognostics vs. Diagnostics vs. Health Monitoring – Are They the Same? • Health Monitor: “ v: to keep track of [current status] systematically with a view to collect information.” • Diagnosis: “n: identifying the nature or cause of some phenomenon.” • Prognosis: “n: a prediction about how something (as the weather) will develop, forecasting.” • Conclusion: they are not the same… • The Webster’s New World Dictionary.

  4. So How Are They Related? • Health monitoring uses instrumentation to collect information about the subject system. • Diagnostics uses the information in real time to detect abnormal operation or outright faults. • Prognosticsuses the information to predict the onset of abnormal conditions and faults prior the actual failure to allow the operators to gracefully plan for shutdown or, if required, operate the system in a degraded but safe-to-use mode until a shutdown and maintenance can be accomplished.

  5. A Brief History of Automated Diagnostics and Prognostics • Before the advent of inexpensive computing, diagnosis was ad-hoc, manual, and depended on human experts. • With the advent of accessible digital computers, early expert systems attempt diesel locomotive engine diagnostics based on oil analysis. Humans still required for prognostics. • 1970’s saw the start of equipment health monitoring for high-value systems (i.e. nuclear power plants) and on-line diagnostics using minicomputers. Human interpretation was still required. • 1980’s saw the use of personal computers and digital analyzers to do equipment health monitoring. Some automatic shut-down on extreme exception was included, but human involvement was still required.

  6. A Brief History (Contd.) • 1990’s saw built-in test and real-time diagnostics added to military electronics and high-value civilian systems. Health monitoring/diagnostics at this point were evolving into decision support systems for the operator. • NOW – Diagnostics pervasive • Automobiles (On Star ™, OBD II, heavy equipment, trucks, etc.) • Electronics/electro-mechanical devices (copiers, complex manufacturing equipment, etc.)

  7. A Brief History (Contd.) • Aviation (Boeing-777, Air Bus, etc.) • Prognostics at the component/ subsystem level start to appear for the first time. • Still no system-wide prognostics! By and large, prognostics are still done by the human operators deciding how much further they can go before stopping.

  8. Literature Survey … • Diagnostics are well developed. • Prognostics are not! • Logical next step … Intelligent System Level Prognostics

  9. Approaches to Diagnostics and Prognostics • Data Driven Methods • Analytical Methods • Knowledge based Methods

  10. Data Signatures • Library of predictive algorithms based on a number of advanced pattern recognition techniques - such as multivariate statistics, neural networks, signal analysis • Identify the partitions that separate the early signatures of functioning systems from those signatures of malfunctioning systems

  11. Predictive indicators of failures • A viable prognostic system should be able to provide an accurate picture of faults, component degradation, and predictive indicators of failures • Allowing our operators to take preventive maintenance actions to avoid costly damage on critical parts and to maintain availability/readiness rates for the system.

  12. Data Driven Methods • The huge amount of data has to be reducedintelligently for any careful fault diagnosis. • Reduce the superficial dimensionality of data to intrinsic dimensionality (i.e., number of independent variables with significant contributions to nonrandom variations in the observations).

  13. Data Driven Methods • Feature extraction: • Partial Least Square (PLS) • Fisher Discriminant Analysis • Canonical Variate Analysis • Principal Component Analysis • We will only focus on PCA and its non-linear relative (NLPCA).

  14. Principal Component Analysis • What is PCA? • It is a way of identifying patterns in data, and expressing the data in such a way as to highlight their similarities and differences. Since patterns in data can be hard to find in data of high dimension, where the luxury of graphical representation is not available.

  15. Principal Component Analysis • PCA is a powerful tool for analyzing data. • The other main advantage of PCA is that once you have found these patterns in the data, and you compress the data, i.e. by reducing the number of dimensions, you have not much loss of information.

  16. PCA … • The feature variables in PCA (also referred to as factors) are linear combinations of the original problem variables.

  17. Classical Statistics based PCA steps… • Get Data • Subtract the mean • Calculate the covariance matrix • Calculate eigenvalues and eigenvectors of covariance matrix • Choose feature vector (data compression begins from here) • Derive the new data set (reduced)

  18. Principal Component Analysis (PCA) • Assuming a data set of containingn observations andmvariables (i.e., a n x mmatrix), PCA divides into two matrices or the scores dimension (n x f) and which is the loading matrix dimension (m x f) plus a matrix of residualsof dimension (n x m).

  19. Principal Component Analysis (PCA) • It is known that PCA optimizes the process by minimizing the Euclidean norm of the residual matrix . • To satisfy this condition, it is known that columns of are the eigenvectors corresponding to the f largest eigenvalues of the covariance matrix of .

  20. Principal Component Analysis (PCA) • In other words, PCA transforms our data from m to f dimension by providing a linear mapping: • where represents a row of the original data set and represents the corresponding row of .

  21. Non-Linear PCA (NLPCA) • In Kramer’s NLPCA, the linear transformation in PCA is generalized to any nonlinear function such that • where is a nonlinear vector function composed of f individual nonlinear functions analogous to the columns of .

  22. Non-Linear PCA (NLPCA)

  23. Analytical Methods • The analytical methods generate features using detailed mathematical models. • Based on the measured input and output , it is common to generate residuals , parameter estimates , and state estimates . • The residuals are the outcomes of consistency checks between the plant observations and a mathematical model.

  24. Integrated Method for Fault Diagnostics and Prognostics (IFDP) • Based on • NLPCA for dimensionality reduction • Society of experts (E-AANN, KSOM, RBFC) • Gated Experts • All developed in Matlab with Simulink for model simulations

  25. Extended Auto-Associative Neural Networks (E-AANN)

  26. Kohonen Self-Organizing Maps (KSOM) • KSOM defines a mapping from the input data space n onto a regular two-dimensional array of nodes. • In the System, a KSOM input is a vector combining both inputs and outputs of a certain the System component. • Every node i is defined by a prototype vector min. Input vector xn is compared with every mi and the best match mb is selected.

  27. Kohonen Self-Organizing Maps (KSOM) Three-dimensional input data in which each sample vector x consists of the RGB (red-green-blue) values of a color vector.

  28. Radial Basis Function based Clustering (RBFC) • The RBF rulebase is identified by our clustering algorithm. • We will consider a specific case of a rulebase with n inputs and a single output. The inputs to the rulebase are assumed to be normalized to fall within the range [0,1].

  29. Gated Experts for Combining Predictions of Different Methods • The Gated Experts (GE) architecture [Weigened et al, 1995] was developed as a method for adaptively combining predictions of multiple experts operating in an environment with changing hidden regimes. • The predictions are combined using a gate block, which dynamically assigns probabilities to the forecast of each expert being correct based on how close the current regime in the data fits the area of expertise for that expert.

  30. Gated Experts for Combining Predictions of Different Methods • The training process for the GE architecture uses the expectation-maximization (EM) algorithm, which combines both supervised and unsupervised learning. • The supervised component in experts learns to predict the conditional mean for the next observed value, and the unsupervised component in the gate learns to discover hidden regimes and assign the probabilities to experts’ forecasts accordingly.

  31. Gated Experts for Combining Predictions of Different Methods • The unsupervised component is also present in experts in the form of a variance parameter, which each expert adjusts to match the variance of the data for which it was found most responsible by the gate.

  32. Prototype Hardware Implementations • A Chiller at Texas A&M University with (Langari and his team) • A laser pointing system prototype at the University of New Mexico (Jamshidi and ACE team) • A COIL laser at AFRL - USAF (Jamshidi & Stone) • A flash memory line at Intel Corp. (Jamshidi & Stone)

  33. Input 1 Input System Boundary Vs 3 Input 2 Input Chiller Model at Texas A&M University

  34. Training Data and Test Data Whole data with 1000 samples

  35. Training Data and Test Data Normalized training data with 2% noise (sorted)

  36. Training Data and Test Data Normalized test data with 2% noise (sorted)

  37. One Sensor with Drift Error Test data with 2% noise, sensor 3 has drift error

  38. One Sensor with Drift Error Drift error and sensor 3 data

  39. One Sensor with Shift Error Test data with 2% noise, sensor 3 has shift error

  40. One Sensor with Shift Error E-AANN output, the input data had 2% noise and shift error

  41. One Sensor with Shift Error Shift error and sensor 3 data

  42. One Sensor with Shift Error The difference between E-AANN input and output, the input data had 2% noise and shift error

  43. PCA Application to Cardiac Output • Cardiac output is defined by two factors. • Stroke volume • Heart Rate • Cardiac Output = Heart rate X Stroke volume (ml/min) (beats/min) (ml/beat) CO for basal metabolic rate is about 5.5L/min

  44. The human heart

  45. Prognostics of CO using PCA Analysis • PCA is used in identifying patterns in data, and expressing the data in such a way to highlight their similarities and differences. • PCA assists us in making an accurate prognostic analysis of a patients Cardiac output performance and hence predict possible heart failures.

  46. Good data representation

  47. By taking several measurements of CO, one is able to predict the possibilities of heart failure, and this allows for PCA to be very useful in the prognostics of Cardiac output. • PCA takes these millions of output measurements and crunches them into a graph representation, from which we can easily visualize CO defects.

  48. Why prognostics ? • In medicine, the cheapest way to cure disease is to prevent it. This is done with early diagnostics, medicines, vaccines, etc.. • However with an accurate prognostics approach, conditions like heart attack and heart failure can be greatly minimized. • PCA enables us to arrive at prognostics.

More Related