1 / 39

Defining Scoring Functions, Multiple Sequence Alignment Lecture #4

Defining Scoring Functions, Multiple Sequence Alignment Lecture #4. Background Readings : Chapter 6 in Biological Sequence Analysis , Durbin et al., 2001. Chapter 3.4, 3.5 in Introduction to Computational Molecular Biology , Setubal and Meidanis, 1997.

wweaver
Download Presentation

Defining Scoring Functions, Multiple Sequence Alignment Lecture #4

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Defining Scoring Functions, Multiple Sequence AlignmentLecture #4 Background Readings: Chapter 6 in Biological Sequence Analysis, Durbin et al., 2001. Chapter 3.4, 3.5 in Introduction to Computational Molecular Biology, Setubal and Meidanis, 1997. This class has been edited from Nir Friedman’s lecture. Changes made by Dan Geiger and Shlomo Moran. .

  2. Reminder: A Probabilistic Model for Defining the Scoring Function • Assume alignment without indels. • Assume each position (nucleotide /amino-acid) is independent of other positions. • Consider two options: M: the sequences are Matched(related) R: the sequences are Random (unrelated)

  3. R: Unrelated Sequences • Our random model of unrelated sequences is simple • Each position is sampled independently from a distribution over the alphabet  • We assume there is a distribution q() that describes the probability of letters in such positions. • Then:

  4. M: Related Sequences • We assume that each pair of aligned positions (s[i],t[i]) evolved from a common ancestor • Let p(a,b) be a distribution over pairs of letters. • p(a,b) is the probability that some ancestral letter evolved into this particular pair of letters.

  5. Odds-Ratio Test for Alignment If Q > 1, then the two strings s and t are more likely to be related (M) than unrelated (R). If Q < 1, then the two strings s and t are more likely to be unrelated (R) than related (M).

  6. Log Odds-Ratio Test for Alignment Score(s[i],t[i]) Taking logarithm of Q yields If log Q > 0, then s and t are more likely to be related. If log Q < 0, then they are more likely to be unrelated. Next we estimate Pr((a,b)|M)=P(a,b) for amino acids

  7. Estimating p(·,·) for proteins(Setubel et al, p. 81-84) We need to define for each pair (a,b) of amino acids (possibly a=b) the probability p(a,b) that a is aligned with bgiven that the sequences are related. p(a,b) should be determined by the data we have. The next slides present one way to define p(a,b) .

  8. Estimating p(·,·) for proteins First, we define for each letter a a probability qa, which is the relative frequency of the amino-acid a in a collection of proteins which is considered reliable for this purpose. Specifically, qa = na/n where na is the number of occurrences of letter a and n is the total number of letters in the collection, so n = ana.

  9. Estimating p(·,·) for proteins For ab, we use as a data a collection of accepted mutations: Formally, it is a multiset of the form {(a,b): a,b are amino acids and ab}. We consider the mutations as unordered pairs. The collection of accepted mutations contains mutations in alignments of closely related protein sequences. For example, Hemoglobin alpha chain in humans and other organisms (homologous proteins).

  10. Estimating p(·,·) for proteins Using the collection of accepted mutations, which we consider as unordered pairs of amino acids. Since the pairs are unordered, we denote them by {a,b}.Some parameters: Mutation counts fab=fba is the number of mutations {a, b}: ab in the above collection. f is the number of mutations (unordered pairs) in the collection. fab /f is the conditional probability:

  11. Estimating p(·,·) for proteins We still have to define the relative mutability, Pr(xy|M). This isthe probability that in the “match” model M, a random pair (x,y) is a mutation. The relative mutability is a user defined parameter, denoted r.

  12. PAM-1 matrices[1-Percent Accepted Mutation]. In PAM-1 matrix it is assumed that the relative mutability r is 1%. That is: 1/100 of the amino acids are mutated. This is modeled by letting n, the total number of pairs of amino acids, be 100f. Using this, we now can compute faa , the number of {a,a} pairs. In this model, p(a,b) is given as the relative frequency of the (unordered) pairs {a,b}:

  13. The log odds-ratio for PAM-1: • The entries in PAM-1 matrix are the logarithms of the odd-ratio: The actual values are 10 times the logarithm, rounded to integer.

  14. Evolutionary distance The choice that r=1% of amino acids change is quite arbitrary. It could fit specific set of proteins whose evolutionary distance is such that indeed 1% of the letters have mutated. This is a unit of evolutionary change, not time because evolution acts differently on distinct sequence types. What is the substitution matrix for k units of evolutionary time ?

  15. A T T C T A C C G G Model of Evolution We make some assumptions: • Each position changes independently of the rest • The probability of mutations is the same in each position • Evolution does not “remember” Time t+ t+2 t+3 t+4 t

  16. Model of Evolution • How do we model such a process? • This process is called a Markov Chain A chain is defined by the transition probability • P(Xt+=b|Xt=a) - the probability that the next state is b given that the current state is a • We often describe these probabilities by a matrix:M[]ab =P(Xt+=b|Xt=a)

  17. Using Conditional independence (No memory) Multi-Step Changes • Based on Pab, we can compute the probabilities of changes over two time periods Thus P[2] = P[]P[] • By induction (HMW exercise): P[n] = P[] n

  18. More on Markov Chains – next week

  19. Comments regarding PAM • Historically researchers use PAM-250. (The only one published in the original paper.) • Original PAM matrices were based on small number of proteins (circa 1978). Later versions use many more examples. • Used to be the most popular scoring rule, but there are some problems with PAM matrices.

  20. Degrees of freedom in PAM definition With r=1/100 the 1-PAM matrix is obtained from the matrix With r=1/50 the basic matrix is different, namely: Thus we have two different ways to estimate the matrix M[4] : Use the PAM-1 matrix to the fourth power: M[4] = M[] 4 Or Use the K=50 matrix to the second power: M[4] = M’[2] 2

  21. Problems in building distance matrices • How do we find pairs of aligned sequences? • How far is the ancestor ? • earlier divergence  low sequence similarity • later divergence  high sequence similarity E.g., M[250] is known not to reflect well long period changes (see p. 43 at Durbin et al). • Does one letter mutate to the other or are they both mutations of a third letter ?

  22. BLOSUM Outline(Durbin et. al, p. 43-44) PAM matrices does not capture the difference between short and long time mutations. BLOck-SUbstitution-Matrices (1992) attempt to model long time mutations. • Idea: use aligned ungapped regions of protein families.These are assumed to have a common ancestor. Similar ideas but better statistics and modeling. • Procedure: • Cluster together sequences in a family whenever more than L% identical residues are shared. • Count number of substitutions across different clusters (in the same family). • Estimate frequencies using the counts. • Practice: Blosum50 (ie L=50) and Blosum62 are wildly used. (See page 43-44 in Durbin et al).

  23. A - T A G - G T T G G G G T G G - - T - A T T A - - A - T A C C A C C C - G C - G - Possible alignment Possible alignment Multiple Sequence Alignment S1=AGGTC S2=GTTCG S3=TGAAC

  24. Multiple Sequence Alignment Aligning more than two sequences. • Definition: Given strings S1,S2,…,Ska multiple (global) alignment map them to strings S’1,S’2,…,S’k that may contain gaps, where: • |S’1|= |S’2|=…= |S’k| • The removal of gaps from S’ileaves Si

  25. x1 x2 x3 x4 Multiple alignments We use a matrix to represent the alignment of k sequences, K=(x1,...,xk). We assume no columns consists solely of gaps. The common scoring functions give a score to each column, and set: score(K)=∑i score(column(i)) The scoring function is symmetric - the order of arguments need not matter: score(I,_,I,V) = score(_,I,I,V). Still, for k=10, a scoring function has > 2k -1 > 1000 entries to specify.

  26. SUM OF PAIRS A common scoring function is SP – sum of scores of the projected pairwise alignments: SPscore(K)=∑i<j score(xi,xj). Note that we need to specify the score(-,-) because a column may have several blanks (as long as not all entries are blanks). In order for this score to be written as∑i score(column(i)), we set score(-,-) = 0. Why ? Because these entries appear in the sum of columns but not in the sum of projected pairwise alignments (lines).

  27. Definition: The sum-of-pairs (SP) value for a multiple global alignment A of k strings is the sum of the values of all projected pairwise alignments induced by A where the pairwise alignment function score(xi,xj) is additive. SUM OF PAIRS

  28. 3 +4 3 + 4 + 5 = 12 3 Example Consider the following alignment: a c - c d b - - c - a d b d a - b c d a d Using the edit distance and for , this alignment has a SP value of

  29. For each cell i =(i1,..,ik), compute an optimal multiple alignment for the k prefix sequences x1(1,..,i1),...,xk(1,..,ik). The adjacent cells are those that differ in their indices by one or zero. Each cell depends on 2k-1 adjacent cells. Multiple Sequence Alignment Given k strings of length n, there is a natural generalization of the dynamic programming algorithm that finds an alignment that maximizes SP-score(K) = ∑i<j score(xi,xj). Instead of a 2-dimensional table, we now have a k-dimensional table to fill.

  30. The idea via K=2 Recall the notation: and the following recurrence for V: Note that the new cell index (i+1,j+1) differs from previous indices by one of 2k-1 non-zero binary vectors (1,1), (1,0), (0,1).

  31. The idea for arbitrary k Order the cells i =(i1,..,ik) by increasing order of the sum ∑ij. Set s(0,..,0)=0, and for i > (0,...,0): Where • b ranges over all non-zero binary k-tuples. • The cell i-b is the non-negative difference of i and b. • column(i,b)is the k-tuple (c1,…,ck) where:cj=xj(ij) if bi=1, and cj= ‘-’ otherwise. • (Reflecting that bj=1 iff the sequence xj was extended by a letter in the corresponding case).

  32. Complexity of the DP approach Number of cells nk. Number of adjacent cells O(2k). Computation of SP score for each column(i,b) is O(k2) Total run time is O(k22knk) which is utterly unacceptable ! Not much hope for a polynomial algorithm because the problem has been shown to be NP complete. Need heuristic to reduce time.

  33. x1 x2 x3 x4 Time saving heuristics: Relevance tests Heuristic: Avoid computing score(i) for irrelevant cells. Let L be a lower bound on the optimal SP score of a multiple alignment of the k sequences (L can be obtained from an arbitrary multiple alignment, computed in any way).

  34. Time saving heuristics: Relevance tests Heuristic: Avoid computing score(i) for irrelevant cells. Relevance test for cell i=(..iu,..iv…): For each two indices 1  u < v  k, compute the optimal score of an alignment of s=xu and t=xv, which goes via cell i. If the sum of these scores over all pairs (u,v) is smaller than L, then i is irrelevant. An alignment via s=xu and t=xv, which goes via cell i is an alignment of s and t which has a prefix which aligns (s[1],…s[iu]) with (t[1],..,t[iv]). How do we find an optimal such alignment?

  35. t s Recall the Linear Space algorithm • V[i,j] = d(s[1..i],t[1..j]) • B[i,j] = d(s[i+1..n],t[j+1..m]) • F[i,j] + B[i,j] = score of best alignment through (i,j) These computations done in linear space. Build such a table for every two sequences s=xu and t=xv, 1  u, v  k. This entry encodes the optimum through (iu,iv).

  36. Time saving heuristics:Relevance test But can we go over all cells determine if they are relevant or not ? No. Start with (0,…,0) and add to the list relevant entries until reaching (n1,…,nk)

  37. Multiple Sequence Alignment – Approximation Algorithm In tutorial time you saw an O(k2n2) multiple alignment algorithm for the SP-score that errs by a factor of at most 2(1-1/k) < 2.

  38. Star Alignments Rather then summing up all pairwise alignments, select a fixed sequence x0 as a center, and set Star-score(K) = ∑j>0score(x0,xj). The algorithm to find optimal alignment: at each step, add another sequence aligned with x0, keeping old gaps and possibly adding new ones.

  39. Tree Alignments • Assume that there is a tree T=(V,E) whose leaves are the sequences. • Associate a sequence in each internal node. • Tree-score(K) = ∑(i,j)Escore(xi,xj). • Finding the optimal assignment of sequences to the internal nodes is NP Hard. We will meet again this problem in the study of Phylogenetic trees

More Related