1 / 36

Kalman Filters

Kalman Filters. Introduction. Mathematical formulation is described by state space concepts . Solution is computed recursively . Both stationary and also non-stationary environments. Each updated estimate of the state computed from The previous estimate, and

azize
Download Presentation

Kalman Filters

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Kalman Filters ELE 774 - Adaptive Signal Processing

  2. Introduction • Mathematical formulation is described by state space concepts. • Solution is computed recursively. • Both stationary and also non-stationary environments. • Each updated estimate of the state computed from • The previous estimate, and • The new input data (innovation). • A unifying framework for the family of recursive least-squares (RLS) filters. ELE 774 - Adaptive Signal Processing

  3. Recursive MMS Estimation for Scalar RVs • Assume a complete set of observed r.v.s upto time n-1 y(1), y(2), ..., y(n-1) • Let the minimum mean-square estimate of the zero mean x(n-1) be where is the space spanned by the observations y(1) ... y(n-1). • Let there be a new observation y(n), • We estimate x(n) using the observations y(1), y(2),...,y(n-1), y(n) • Do this either by storing y(1), y(2),...,y(n-1) and redo the whole problem, or • Exploit and use the new observation y(n), i.e. use a recursive estimation procedure. ELE 774 - Adaptive Signal Processing

  4. Recursive MMS Estimation for Scalar RVs • What is new in the new observation y(n)? Innovations! • Define the forward prediction error • Prediction order (n-1) increases linearly with n. • According to principle of orthogonality, • fn-1(n) is orthogonal to y(1), y(2), ..., y(n-1). • fn-1(n)is a measure of the new information in y(n) ═>innovations! • Information provided by y(n) is composed of two parts • One that in not new, contained in • One that is new, contained in one-step prediction of y(n) ELE 774 - Adaptive Signal Processing

  5. Recursive MMS Estimation for Scalar RVs • Refer to the prediction error as innovations, and define • Properties of the innovation a(n): • Property 1: a(n) is orthogonal to the observations y(1), y(2), ...,y(n-1) Follows from the principle of orthogonality. • Property 2: a(1), a(2), ..., a(n) are orthogonal to each other Innovations process is white. Follow from ELE 774 - Adaptive Signal Processing

  6. Recursive MMS Estimation for Scalar RVs • Property 3: There is one to one correspondence between One sequence may be obtained from the other by means of a causal and causally invertible filter without any loss of information. To show this use Gram-Schmidt orthogonalization procedure: ELE 774 - Adaptive Signal Processing

  7. Recursive MMS Estimation for Scalar RVs • Collecting all terms together • kth row of the matrix gives the coefficients of the forward prediction-error filter of order k-1. • The innovations can be calculated from the observations, or • The observations can be calculated from the innovations • There is no loss of information in this transformation. ELE 774 - Adaptive Signal Processing

  8. Recursive MMS Estimation for Scalar RVs means or, equivalently ELE 774 - Adaptive Signal Processing

  9. Recursive MMS Estimation for Scalar RVs • Clearly, • Recalling that innovations are orthogonal to each other, and choosing bk to minimize the mean-square estimation error we get • Now rewrite • Adding a correction term bna(n) to the previous estimate gives the updated estimate , can be calculated recursively. ELE 774 - Adaptive Signal Processing

  10. Recursive MMS Estimation for Scalar RVs • Predictor – Corrector structure • The use of observations to compute a forward prediction error – innovations, • The use of the innovation to update (correct) the minimum mean-square estimate of a r.v. related linearly to the observations ELE 774 - Adaptive Signal Processing

  11. Process Equation Measurement Equation Discrete-Time Dynamical System • A linear discrete-time dynamical system can be characterized by ELE 774 - Adaptive Signal Processing

  12. Discrete-Time Dynamical System • The state vector, x(n), is the minimal set of data that is sufficient to uniquely describe the unforced dynamical behaviour of the system. • fewest data on the past behaviour needed to predict the future one. • Dimension M. • The observation vector, y(n), is the set of observed data. • Dimension N. ELE 774 - Adaptive Signal Processing

  13. Discrete-Time Dynamical System Transition matrix • The process equation: • models an unknown physical stochastic phenomenon denoted by the state x(n) as the output of a linear dynamical systemexcited by the white noise n1(n). • Properties of the transition matrix • 1. Product rule • 2. Inverse rule • Corollary ELE 774 - Adaptive Signal Processing

  14. Discrete-Time Dynamical System • The measurement equation • gives the relation between the state x(n) and the output y(n), with zero-mean white measurement noise (disturbance) n2(n) • Initial value of the state : x(0) • uncorrelated with both n1(n) and n2(n), • Noise vectors n1(n) and n2(n) are statistically independent ELE 774 - Adaptive Signal Processing

  15. Process Equation Measurement Equation Kalman Filtering • We need to solve these eqn.s jointly to find the state x(n) • Use the entire observed data, • consisting of the observations y(1), y(2), ..., y(n) • to find the minimum mean-square estimate of the statex(i), n ≥ 1 • i = n, filtering, • i > n, prediction, • i < n, smoothing. ELE 774 - Adaptive Signal Processing

  16. The Innovations Process • Let the MMS estimate of y(n) be • Span of the vector space is y(1), y(2), ..., y(n-1) • The innovations process associated with y(n) • (Similar to the scalar case) • the Mx1 vector a(n) represents the new information in the observed data y(n). ELE 774 - Adaptive Signal Processing

  17. The Innovations Process • Properties of the innovations process: • Property I: • a(n) is orthogonal to all past observationsy(1), ..., y(n-1) • Property II: • The innovations process consists of a sequence of vector random variables that are orthogonal to each other • Property III: • There is a one-to-one correspondence between the observations and the observation process. One sequence may be obtained from the other by means of linear stable operators, without loss of information. ELE 774 - Adaptive Signal Processing

  18. The Innovations Process • Correlation Matrix of the Innovations Process • Starting from initial state (n=0), we write • i.e. x(k) is a linear combination of x(0), n1(1), ..., n1(k-1) • We know that , then • 1. • 2. • 3. ELE 774 - Adaptive Signal Processing

  19. The Innovations Process • Recall that • Then, given the past decisions , i.e. , the MMS estimate of y(n) • Hence • Predicted state-error vector at time n using data up to time n-1. ELE 774 - Adaptive Signal Processing

  20. The Innovations Process • Autocorrelation of the innovation processa(n) where the predicted state-error correlation matrix is • statistical description of the error in the predicted estimate ELE 774 - Adaptive Signal Processing

  21. Estimation of the State • Minimum mean-square estimation of the state x(n) • Estimate may be expressed as a linear combination of the sequence of innovations process: • Using principle of orthogonality ELE 774 - Adaptive Signal Processing

  22. Estimation of the State • Minimum mean-square estimate of x(n) • Let i=n+1 ELE 774 - Adaptive Signal Processing

  23. Estimation of the State • Summation in is • Then • Kalman Gain • Define • Then ELE 774 - Adaptive Signal Processing

  24. Estimation of the State • Convenient way to calculate the Kalman gain • Hence,Kalman gain becomes • Rewriting the estimate have to be calculated at each iteration! Let’s make it recursive. ELE 774 - Adaptive Signal Processing

  25. Estimation of the State • Kalman gain computer ELE 774 - Adaptive Signal Processing

  26. Estimation of the State • Riccati Equation: • Predicted state-error vector: • After substitution and manipulations ELE 774 - Adaptive Signal Processing

  27. Estimation of the State • We want to find K(n,n-1), hence from previous slide • Then using (1) and (2), we get the Riccati difference equation where ELE 774 - Adaptive Signal Processing

  28. Estimation of the State • Riccati equation solver ELE 774 - Adaptive Signal Processing

  29. Estimation of the State • Kalman’s one-step prediction algorithm: ELE 774 - Adaptive Signal Processing

  30. Estimation of the State • One step prediction algorithm ELE 774 - Adaptive Signal Processing

  31. Filtering • Compute the filtered estimate by using the one-step prediction algorithm. • The state x(n) and the process noise n1(n) are independent of each other. • The MMSE estimate x(n+1) given the observations upto time n is where second line follows from the fact that y(n) and n1(n) are independent of each other. ELE 774 - Adaptive Signal Processing

  32. Filtering • Filtered Estimation Error and Conversion Factor • Define the filtered estimation error vector • We know that • Then Conversion factor ELE 774 - Adaptive Signal Processing

  33. Filtering • Substituting gives • Filtered State-Error Correlation Matrix • Definefiltered state-error vector ELE 774 - Adaptive Signal Processing

  34. Filtering • Manipulations give us • Initial Conditions: • Initial state of the process equation is not precisely known • In the absence of any observed data at n=0, let and if x is zero mean and ELE 774 - Adaptive Signal Processing

  35. Filtering • Kalman filter based on one-step prediction ELE 774 - Adaptive Signal Processing

  36. Summary ELE 774 - Adaptive Signal Processing

More Related