1 / 22

Exploiting Sparse Markov and Covariance Structure in Multiresolution Models

Reading Group. Exploiting Sparse Markov and Covariance Structure in Multiresolution Models. ( MIT EECS: Myung Jin Choi , Venkat Chandrasekaran , Alan S. Willsky ). Presenter: Zhe Chen ECE / CMR Tennessee Technological University October 22, 2010. Outline. Overview Preliminaries

Download Presentation

Exploiting Sparse Markov and Covariance Structure in Multiresolution Models

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Reading Group Exploiting Sparse Markov and Covariance Structure in Multiresolution Models (MIT EECS: Myung Jin Choi, VenkatChandrasekaran, Alan S. Willsky) Presenter:Zhe Chen ECE / CMR Tennessee Technological University October 22, 2010

  2. Outline Overview Preliminaries Multiresolution Models with Sparse In-scale Conditional Covariance Inference Exploiting Sparsity in Markov and Covariance Structure Learning MR Models with Sparse In-scale Conditional Covariance Experimental Results Conclusion

  3. Overview In this paper, the authors consider Gaussian multiresolution (MR) models in which coarser, hidden variables serve to capture statistical dependencies among the finest scale variables. The authors propose a new class of Gaussian MR models that capture the residual correlations within each scale using sparse covariance structure. The authors use the sparse tree structure between scales, to propagate information from scale-to-scale, and then perform residual filtering within each scale using the sparse conditional covariance structure. In addition, the authors develop methods for learning such models given data at the finest scale. The structure optimization within each scale can be formulated as a convex optimization problem.

  4. Preliminaries (0) Multiresolution (MR) models provide compact representations for encoding statistical dependencies among a large collection of random variables. In MR models, variables at coarser resolutions serve as common factors for explaining statistical dependencies among finer scale variables. For example, suppose that we would like to discover the dependency structure of the monthly returns of 100 different stocks by looking at pairwise correlations. It is likely that the covariance matrix will be full, i.e., the monthly return of one stock is correlated to all other 99 stocks, because stock prices tend to move together driven by the market situation. Therefore, it is more informative to introduce a hidden variable corresponding to the market.

  5. Preliminaries (1) One approach in MR modeling is to use tree-structured graphical models in which nodes at any scale are connected to each other only through nodes at other scales. Examples of MR tree models for a one-dimensional process (left) and for a two-dimensional process (right). Shaded nodes represent original variables at the finest scale and white nodes represent hidden variables at coarser scales.

  6. Preliminaries (2) A sparse graphical model The sparsity pattern of the corresponding information matrix Let be a jointly Gaussian random vector with a mean vector and a positive-definite covariance matrix . If the variables are Markov with respect to a graph , the inverse of the covariance matrix (also called information matrix) is sparse with respect to .

  7. Preliminaries (3) Sparsity pattern of a covariance matrix The corresponding graphical model Conjugate graph encoding the sparsity structure of the covariance matrix Conjugate graph – two graphs with covariance matrices that are inverses of one another. Specifically, in the conjugate graph, when two nodes are not connected with a conjugate edge, they are uncorrelated with each other.

  8. Multiresolution Models with Sparse In-scale Conditional Covariance (1) The authors propose a class of MR models with tree-structured connections between different scales and sparse conditional covariance structure at each scale. The authors define in-scale conditional covariance as the conditional covariance between two variables (in the same scale) when conditioned on variables at other scales. The authors’ model has a sparse graphical model for inter-scale structure and a sparse conjugate graph for in-scale structure. The authors refer to such an MR model as a Sparse In-scale Conditional Covariance Multiresolution (SIM) model.

  9. Multiresolution Models with Sparse In-scale Conditional Covariance (2) (a) An MR model with a sparse graphical structure (b) A SIM model with sparse conjugate graph within each scale (c) A graphical model corresponding to the model in (b) Examples of MR models

  10. Multiresolution Models with Sparse In-scale Conditional Covariance (3) (Shaded matrices are dense, and non-shaded matrices are sparse.) Partition the information matrix of a SIM model by scale: The matrix can be decomposed as a sum of , corresponding to the hierarchical inter-scale tree structure, and , corresponding to the conditional in-scale structure. Let .

  11. Inference Exploiting Sparsity in Markov and Covariance Structure (1) Let be a collection of random variables with a prior distribution , and be a set of noisy measurements: where is a selection matrix, and is a zero-mean Gaussian noise vector with a diagonal covariance matrix . Then, the MAP estimate is given as follows where is a diagonal matrix, and . How to solve it?

  12. Inference Exploiting Sparsity in Markov and Covariance Structure (2) The authors use the matrix splitting idea in developing an efficient inference method for the SIM model. The goal is to solve the equation where , , and are all sparse matrices. The authors alternate between two inference steps corresponding to inter-scale computation (also called tree inference) and in-scale computation in the MR model matrices.

  13. Inference Exploiting Sparsity in Markov and Covariance Structure (3) • Tree inference: • Set where D is a diagonal matrix added to ensure that M is positive-definite. • With the right-hand side vector fixed, solving the above equation is efficient since M has a tree structure. • In-scale inference: • Set • Evaluating the right-hand side only involves multiplications of a sparse matrix and a vector, so can be computed efficiently.

  14. Learning MR Models with Sparse In-scale Conditional Covariance (1) Suppose that we are given a target covariance and wish to learn a sparse graphical model that best approximates the covariance. Standard approaches in Gaussian graphical model selection solve the following log-determinant optimization problem to find an approximate covariance matrix: The authors again use the log-determinant problem, but now in the information matrix domain: where is a target information matrix.

  15. Learning MR Models with Sparse In-scale Conditional Covariance (2) • Given a target covariance of the variables at the finest scale, the objective is to introduce hidden variables at coarser scales and learn a SIM model. • The authors’ learning procedure consists of three steps • 1. Learning the inter-scale model • 2. Finding the target information matrix • 3. Obtaining sparse in-scale conditional covariance. Find a SIM model that approximates by solving the following problem: where is the in-scale information matrix at scale m and and are the set of all possible in-scale and inter-scale edges, respectively.

  16. Experimental Results (1) The following results are compared with a single-scale approximate model, a tree-structured MR model, and a sparse MR model. 1. Stock returns - modeling the dependency structure of monthly stock returns.

  17. Experimental Results (2) • the single-scale • (Divergence: 122.48) (b) sparse MR approximation (Divergence: 28.34) (c) Sparsity pattern of the in-scale conditional covariance of the SIM approximation (Divergence: 16.36) (divergence between the approximate and the empirical distribution) The SIM model structure is a more natural representation for capturing in-scale statistics. Sparsity pattern of the information matrix of:

  18. Experimental Results (3) Single-scale approximation Original model Tree approximation SIM model SIM approximation is close to the original covariance. 2. Fractional Brownian Motion - consider fractional Brownian motion (fBm) with Hurst parameter defined on the time interval with the covariance function: Covariance realized by each model using 64 time samples:

  19. Experimental Results (4) Scale 3 (4 x 4) Scale 4 (8 x 8) Scale 5 (16 x 16) • 3. Polynomially Decaying Covariance for a 2-D Field • Consider a collection of 256 Gaussian random variables arranged spatially on a 16 x 16 grid. The variance of each variable is given by and the covariance between each pair of variables is given by , where d(s, t) is the spatial distance between nodes s and t. • Conjugate graph at each scale of the SIM model for polynomially decaying covariance approximation:

  20. Experimental Results (5) The SIM modeling approach provides a significant gain in convergence rate over other model. Residual error vs. computation time to solve the inference problem in the polynomially decaying covariance example:

  21. Conclusion The authors have proposed a method to learn a Gaussian MR model with sparse in-scale conditional covariance at each scale and sparse inter-scale graphical structure connecting variables across scales. And an efficient inference algorithm has been given.

  22. Thank you!

More Related