1 / 58

HMM for multiple sequences

HMM for multiple sequences. Pair HMM. HMM for pairwise sequence alignment, which incorporates affine gap scores. “Hidden” States Match (M) Insertion in x (X) insertion in y (Y) Observation Symbols Match (M): {( a,b )| a,b in ∑ }. Insertion in x (X): {( a,- )| a in ∑ } .

tristana
Download Presentation

HMM for multiple sequences

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. HMM for multiple sequences

  2. Pair HMM HMM for pairwise sequence alignment, which incorporates affine gap scores. “Hidden” States • Match (M) • Insertion in x (X) • insertion in y (Y) Observation Symbols • Match (M): {(a,b)| a,b in∑ }. • Insertion in x (X): {(a,-)| a in ∑ }. • Insertion in y (Y): {(-,a)| a in ∑ }.

  3.   End Begin  Pair HMMs   1- X 1--2  1-2 M   Y 1-

  4. A T C G T A C w 1 2 3 4 5 6 7 0 0 v A 1 T 2 G 3 T 4 T 5 A 6 T 7 Alignment: a path  a hidden state sequence A T - G T T A T A T C G T - A C M M Y M M X M M

  5. Multiple sequence alignment(Globin family)

  6. Profile model (PSSM) • A natural probabilistic model for a conserved region would be to specify independent probabilities ei(a) of observing nucleotide (amino acid) a in position i • The probability of a new sequence x according to this model is

  7. Profile / PSSM LTMTRGDIGNYLGLTVETISRLLGRFQKSGML LTMTRGDIGNYLGLTIETISRLLGRFQKSGMI LTMTRGDIGNYLGLTVETISRLLGRFQKSEIL LTMTRGDIGNYLGLTVETISRLLGRLQKMGIL LAMSRNEIGNYLGLAVETVSRVFSRFQQNELI LAMSRNEIGNYLGLAVETVSRVFTRFQQNGLI LPMSRNEIGNYLGLAVETVSRVFTRFQQNGLL VRMSREEIGNYLGLTLETVSRLFSRFGREGLI LRMSREEIGSYLGLKLETVSRTLSKFHQEGLI LPMCRRDIGDYLGLTLETVSRALSQLHTQGIL LPMSRRDIADYLGLTVETVSRAVSQLHTDGVL LPMSRQDIADYLGLTIETVSRTFTKLERHGAI • DNA / proteins Segments of the same length L; • Often represented as Positional frequency matrix;

  8. Searching profiles: inference • Give a sequence S of length L, compute the likelihood ratio of being generated from this profile vs. from background model: • R(S|P)= • Searching motifs in a sequence: sliding window approach

  9. Match states for profile HMMs • Match states • Emission probabilities .... .... Begin Mj End

  10. Components of profile HMMs • Insert states • Emission prob. • Usually back ground distribution qa. • Transition prob. • Mi to Ii, Ii to itself, Ii to Mi+1 • Log-odds score for a gap of length k (no logg-odds from emission) Ij Begin Mj End

  11. Components of profile HMMs • Delete states • No emission prob. • Cost of a deletion • M→D, D→D, D→M • Each D→D might be different Dj Begin Mj End

  12. Full structure of profile HMMs Dj Ij Begin Mj End

  13. Deriving HMMs from multiple alignments • Key idea behind profile HMMs • Model representing the consensus for the alignment of sequence from the same family • Not the sequence of any particular member HBA_HUMAN ...VGA--HAGEY... HBB_HUMAN ...V----NVDEV... MYG_PHYCA ...VEA--DVAGH... GLB3_CHITP ...VKG------D... GLB5_PETMA ...VYS--TYETS... LGB2_LUPLU ...FNA--NIPKH... GLB1_GLYDI ...IAGADNGAGV... *** *****

  14. Deriving HMMs from multiple alignments • Basic profile HMM parameterization • Aim: making the higher probability for sequences from the family • Parameters • the probabilities values : trivial if many of independent alignment sequences are given. • length of the model: heuristics or systematic way

  15. Sequence conservation: entropy profile of the emission probability distributions

  16. Searching with profile HMMs • Main usage of profile HMMs • Detecting potential sequences in a family • Matching a sequence to the profile HMMs • Viterbi algorithm or forward algorithm • Comparing the resulting probability with random model

  17. Searching with profile HMMs • Viterbi algorithm (optimal log-odd alignment)

  18. Searching with profile HMMs • Forward algorithm: summing over all potent alignments

  19. Variants for non-global alignments • Local alignments (flanking model) • Emission prob. in flanking states use background values qa. • Looping prob. close to 1, e.g. (1- ) for some small . Dj Ij Mj Begin End Q Q

  20. Variants for non-global alignments • Overlap alignments • Only transitions to the first model state are allowed. • When expecting to find either present as a whole or absent • Transition to first delete state allows missing first residue Dj Q Q Ij Begin Mj End

  21. Variants for non-global alignments • Repeat alignments • Transition from right flanking state back to random model • Can find multiple matching segments in query string Dj Ij Mj Q Begin End

  22. Estimation of prob. • Maximum likelihood (ML) estimation • given observed freq. cja of residue a in position j. • Simple pseudocounts • qa: background distribution • A: weight factor

  23. Optimal model construction: mark columns (a) Multiple alignment: (c) Observed emission/transition counts x x . . . x bat A G - - - C rat A - A G - C cat A G - A A - gnat - - A A A C goat A G - - - C 1 2 . . . 3 0 1 2 3 A - 4 0 0 C - 0 0 4 G - 0 3 0 T - 0 0 0 A 0 0 6 0 C 0 0 0 0 G 0 0 1 0 T 0 0 0 0 M-M 4 3 2 4 M-D 1 1 0 0 M-I 0 0 1 0 I-M 0 0 2 0 I-D 0 0 1 0 I-I 0 0 4 0 D-M - 0 0 1 D-D - 1 0 0 D-I - 0 2 0 match emissions insert emissions (b) Profile-HMM architecture: D D D state transitions I I I I beg M M M end 1 2 3 4 0

  24. Optimal model construction • MAP (match-insert assignment) • Recursive calculation of a number Sj • Sj: log prob. of the optimal model for alignment up to and including column j, assuming j is marked. • Sj is calculated from Si and summed log prob. between i and j. • Tij: summed log prob. of all the state transitions between marked i and j. • cxy are obtained from partial state paths implied by marking i and j.

  25. Optimal model construction • Algorithm: MAP model construction • Initialization: • S0 = 0, ML+1 = 0. • Recurrence: for j = 1,..., L+1: • Traceback: from j = L+1, while j > 0: • Mark column j as a match column • j = j.

  26. Weighting training sequences • Input sequences are random? • “Assumption: all examples are independent samples” might be incorrect • Solutions • Weight sequences based on similarity

  27. Weighting training sequences • Simple weighting schemes derived from a tree • Phylogenetic tree is given. • [Thompson, Higgins & Gibson 1994b] • [Gerstein, Sonnhammer & Chothia 1994]

  28. Weighting training sequences 7 V7 t6 = 3 I1+I2+I3 6 V6 t5 = 3 I1+I2 t4 = 8 I4 t3 = 5 V5 5 t2 = 2 I3 t1 = 2 I1 I2 1 2 3 4 w1:w2:w3:w4 = 35:35:50:64 I1:I2:I3:I4 = 20:20:32:47

  29. Multiple alignment by training profile HMM • Sequence profiles could be represented as probabilistic models like profile HMMs. • Profile HMMs could simply be used in place of standard profiles in progressive or iterative alignment methods. • ML methods for building (training) profile HMM (described previously) are based on multiple sequence alignment. • Profile HMMs can also be trained from initially unaligned sequences using the Baum-Welch (EM) algorithm

  30. Multiple alignment by profile HMM training- Multiple alignment with a known profile HMM • Before we estimate a model and a multiple alignment simultaneously, we consider as simpler problem: derive a multiple alignment from a known profile HMM model. • This can be applied to align a large member of sequences from the same family based on the HMM model built from the (seed) multiple alignment of a small representative set of sequences in the family.

  31. Multiple alignment with a known profile HMM • Align a sequence to a profile HMMViterbi algorithm • Construction a multiple alignment just requires calculating a Viterbi alignment for each individual sequence. • Residues aligned to the same match state in the profile HMM should be aligned in the same columns.

  32. Multiple alignment with a known profile HMM • Given a preliminary alignment, HMM can align additional sequences.

  33. Multiple alignment with a known profile HMM

  34. Multiple alignment with a known profile HMM • Important difference with other MSA programs • Viterbi path through HMM identifies inserts • Profile HMM does not align inserts • Other multiple alignment algorithms align the whole sequences.

  35. Profile HMM training from unaligned sequences • Harder problem • estimating both a model and a multiple alignment from initially unaligned sequences. • Initialization: Choose the length of the profile HMM and initialize parameters. • Training: estimate the model using the Baum-Welch algorithm (iteratively). • Multiple Alignment: Align all sequences to the final model using the Viterbi algorithm and build a multiple alignment as described in the previous section.

  36. Profile HMM training from unaligned sequences • Initial Model • The only decision that must be made in choosing an initial structure for Baum-Welch estimation is the length of the model M. • A commonly used rule is to set M be the average length of the training sequence. • We need some randomness in initial parameters to avoid local maxima.

  37. Multiple alignment by profile HMM training • Avoiding Local maxima • Baum-Welch algorithm is guaranteed to find a LOCAL maxima. • Models are usually quite long and there are many opportunities to get stuck in a wrong solution. • Solution • Start many times from different initial models. • Use some form of stochastic search algorithm, e.g. simulated annealing.

  38. Multiple alignment by profile HMM -similar to Gibbs sampling • The ‘Gibbs sampler’ algorithm described by Lawrence et al.[1993] has substantial similarities. • The problem was to simultaneously find the motif positions and to estimate the parameters for a consensus statistical model of them. • The statistical model used is essentially a profile HMM with no insert or delete states.

  39. Multiple alignment by profile HMM training-Model surgery • We can modify the model after (or during) training a model by manually checking the alignment produced from the model. • Some of the match states are redundant • Some insert states absorb too many sequences • Model surgery • If a match state is used by less than ½ of training sequences, delete its module (match-insert-delete states) • If more than ½ of training sequences use a certain insert state, expand it into n new modules, where n is the average length of insertions • ad hoc, but works well

  40. Phylo-HMMs: model multiple alignments of syntenic sequences • A phylo-HMM is a probabilistic machine that generates a multiple alignment, column by column, such that each column is defined by a phylogenetic model • Unlike single-sequence HMMs, the emission probabilities of phylo-HMMs are complex distributions defined by phylogenetic models

  41. Applications of Phylo-HMMs • Improving phylogenetic modeling that allow for variation among sites in the rate of substitution (Felsenstein & Churchill, 1996; Yang, 1995) • Protein secondary structure prediction (Goldman et al., 1996; Thorne et al., 1996) • Detection of recombination from DNA multiple alignments (Husmeier & Wright, 2001) • Recently, comparative genomics (Siepel, et. al. Haussler, 2005)

  42. Phylo-HMMs: combining phylogeny and HMMs • Molecular evolution can be viewed as a combination of two Markov processes • One that operates in the dimension of space (along a genome) • One that operates in the dimension of time (along the branches of a phylogenetic tree) • Phylo-HMMs model this combination

  43. Single-sequence HMM Phylo-HMM

  44. Phylogenetic models • Stochastic process of substitution that operates independently at each site in a genome • A character is first drawn at random from the background distribution and assigned to the root of the tree; character substitutions then occur randomly along the tree branches, from root to leaves • The characters at the leaves define an alignment column

  45. Phylogenetic Models • The different phylogenetic models associated with the states of a phylo-HMM may reflect different overall rates of substitution (e.g. in conserved and non-conserved regions), different patterns of substitution or background distributions, or even different tree topologies (as with recombination)

  46. Phylo-HMMs: Formal Definition • A phylo-HMM is a 4-tuple : • : set of hidden states • : set of associated phylogenetic models • : transition probabilities • : initial probabilities

  47. The Phylogenetic Model • : • : substitution rate matrix • : background frequencies • : binary tree • : branch lengths

  48. The Phylogenetic Model • The model is defined with respect to an alphabet  whose size is denoted d • The substitution rate matrix has dimension dxd • The background frequencies vector has dimension d • The tree has n leaves, corresponding to n extant taxa • The branch lengths are associated with the tree

  49. Probability of the Data • Let X be an alignment consisting of L columns and n rows, with the ith column denoted Xi • The probability that column Xi is emitted by state sjis simply the probability of Xi under the corresponding phylogenetic model, • This is the likelihood of the column given the tree, which can be computed efficiently using Felsenstein’s “pruning” algorithm (which we will describe in later lectures)

  50. Substitution Probabilities • Felsenstein’s algorithm requires the conditional probabilities of substitution for all bases a,b and branch lengths tj • The probability of substitution of a base b for a base a along a branch of length t, denoted is based on a continuous-time Markov model of substitution, defined by the rate matrix Qj

More Related