1 / 32

Large Scale Metabolic Network Alignments by Compression

Large Scale Metabolic Network Alignments by Compression Michael Dang, Ferhat Ay , Tamer Kahveci ACM-BCB 2011. Bioinformatics Lab. University of Florida. Network Alignment. Bayati et al. ICDM 2009. Metabolic Network Alignment. Alignment with Heterogeneous Entities.

cyrus-ramos
Download Presentation

Large Scale Metabolic Network Alignments by Compression

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Large Scale Metabolic Network • Alignments by Compression • Michael Dang, FerhatAy, Tamer Kahveci • ACM-BCB 2011 Bioinformatics Lab. University of Florida

  2. NetworkAlignment Bayati et al. ICDM 2009 Ferhat Ay

  3. Metabolic Network Alignment Alignment with Heterogeneous Entities Network Alignment Subnetwork Mappings • Querying Network • databases • Functional Similarity • of Reactions Ferhat Ay

  4. Existing Work • Heymans et al. (2003) – Undirected, Hierarchical Enzyme Similarity • Pinter et al. (2005) – Directed,Only Multi-Source Trees • Singh et al. (2007) – PPI Networks, Sequence Similarity • Dost et al. (2007) – QNET, Color Coding, Tree queries of size at most 9 • Kuchaiev et al. (2010) – GRAAL,Solely Based on Network Topology • Ay et al. (2011) – SubMAP, Considers Subnetwork Mappings • Shih et al. – Next Talk! • Clustering of input networks is necessary for aligning large networks Ferhat Ay

  5. Alignment Phase - SubMAP Ay et. al., RECOMB 2010, JCB 2011 Ferhat Ay

  6. Performance Bottleneck 2 Gigabytes 30 minutes Ferhat Ay

  7. Alignment with Compression Compress Align Refine Ferhat Ay

  8. Alignment with/without compression • Without Compression • With Compression Ferhat Ay

  9. Outline of the method • Compression Phase • Minimum Degree Selection (MDS) method • Optimality analysis • Alignment Phase • Refinement Phase • Overall Complexity • How Much Should We Compress? • Experimental Results Ferhat Ay

  10. Compression Phase Original Network Compressed Network Encapsulated View What Alignment Algorithm Sees Ferhat Ay

  11. MDS – Minimum Degree Selection After After Before Ferhat Ay

  12. Overall Compression Step i Step 1 Step 2 Level 1 ………… Level 2 ………… ………… Level c …….… Ferhat Ay

  13. Optimality Condition for MDS Minimum Degree Node Optimal? Ferhat Ay

  14. Optimality Condition for MDS Optimal? - Can be optimal - At most one node away Minimum Degree Node Ferhat Ay

  15. How Far Away We Are from Optimal Compression? Number of compression steps for the optimal compression and MDS Sizes of the compressed networks for the optimal compression and MDS By the inequality “How far is our compression method from the optimal compression?” Ferhat Ay

  16. Alignment Phase Network 1 Network 2 Compressed Network 1 Compressed Network 2 Network Alignment Algorithm Alignment Ferhat Ay

  17. Refinement Phase Alignment Algorithm Refined Alignment Refine Ferhat Ay

  18. Complexity Analysis Compression Phase: Alignment Phase: Refinement Phase: Overall Complexity (with compression): Complexity of SubMAP (without compression): k = largest subnetwork size c = compression level n, m = sizes of networks Ferhat Ay

  19. How much should we compress? • Examples n=20, m=20, k=2  c ~ 1.37 n=20, m=80, k=1  c ~ 2.11 n=80, m=80, k=2  c ~ 2.15 n=200, m=400, k=1  c ~ 3.11 Ferhat Ay

  20. Experimental Results Ferhat Ay

  21. To compress or not to compress? Ferhat Ay

  22. Compression Rates in Practice KEGG Metabolic Networks with sizes ranging from 10 to 279 Ferhat Ay

  23. What Do We Gain by Compression? Subnetwork size k=1 Subnetwork size k=2 Ferhat Ay

  24. What Do We Lose by Compression? Correlation of mappings scores found by compressed alignment with the ones found by SubMAP Ferhat Ay

  25. Conclusions • We developed a scalable compression technique with optimality bounds. • Our method respects network topology while aligning the networks unlike clustering-based methods. • It provides significant improvement on resource utilization of existing network alignment algorithms. Ferhat Ay

  26. Future Directions • Improving the scale of alignment to genome-wide metabolic networks (without initial clustering). • Evaluating the performance of our compression technique on PPI networks. • Improving the accuracy of compressed alignment w.r.t original alignment for larger levels of compression. • Integrating our compression framework with other existing network alignment methods. Ferhat Ay

  27. Acknowledements Michael Dang NSF IIS-0845439 NSF CCF-0829867 Tamer Kahveci Ferhat Ay

  28. A bit of advertisement Computing Innovation Fellow University of Washington Department of Genome Sciences http://cifellows.org/ http://www.gs.washington.edu/ http://noble.gs.washington.edu/~wnoble/

  29. THANK YOU. QUESTIONS? Ferhat Ay

  30. APPENDIX Ferhat Ay

  31. Ferhat Ay

  32. Ferhat Ay

More Related