|
|
Line 1: |
Line 1: |
| '''Bayesian inference in [[Phylogenetics|phylogeny]]''' generates a posterior distribution for a parameter, composed of a [[phylogenetic tree]] and a model of evolution, based on the prior for that parameter and the likelihood of the data, generated by a multiple alignment. The Bayesian approach has become more popular due to advances in computational machinery, especially, [[Markov chain Monte Carlo]] algorithms. [[Bayesian inference]] has a number of applications in [[molecular phylogenetics]], for example, estimation of [[species]] phylogeny and species divergence times.
| | Her title is Felicidad Ahmad. The occupation he's been occupying for years is a messenger. Alabama is exactly where he and his wife live and he has everything that he needs there. To perform badminton is some thing he truly enjoys doing.<br><br>Also visit my website extended car warranty ([http://Discordia.Cwsurf.de/index.php?mod=users&action=view&id=61 mouse click the up coming website]) |
| | |
| ==Basic Bayesian theory==
| |
| Recall that for Bayesian inference:
| |
| | |
| : <math> p(\theta | D) = \frac{p(D|\theta)p(\theta)}{p(D)}\ </math>
| |
| | |
| The denominator <math>p(D)\ </math> is the ''marginal probability of the data'', averaged over all possible parameter values weighted by their prior distribution. Formally,
| |
| | |
| : <math>p(D) = \int_\Theta p(D|\theta)p(\theta)d\theta\ </math>
| |
| | |
| where <math>\Theta\ </math> is the parameter space for <math>\theta\ </math>.
| |
| | |
| In the original [[Metropolis algorithm]], given a current <math>\theta\ </math>-value <math>x\ </math>, and a new <math>\theta\ </math>-value <math>y\ </math>, the new value is accepted with probability:
| |
| | |
| :<math>h(y)/h(x) = \frac{p(D|y)p(y)}{p(D|x)p(x)}\ </math>
| |
| | |
| ==The LOCAL algorithm of Larget and Simon==
| |
| | |
| The LOCAL algorithm begins by selecting an internal branch of the tree at random. The nodes at the ends of this branch are each connected to two other branches. One of each pair is chosen at random. Imagine taking these three selected edges and stringing them like a clothesline from left to right, where the direction (left/right) is also selected at random. The two endpoints of the first branch selected will have a sub-tree hanging like a piece of clothing strung to the line. The algorithm proceeds by multiplying the three selected branches by a common random amount, akin to stretching or shrinking the clothesline. Finally the leftmost of the two hanging sub-trees is disconnected and reattached to the clothesline at a location selected uniformly at random. ''This is the candidate tree''.
| |
| | |
| Suppose we began by selecting the internal branch with length <math>t_8\ </math> (in Figure (a) (to be added)) that separates taxa <math>A\ </math> and <math>B\ </math> from the rest. Suppose also that we have (randomly) selected branches with lengths <math>t_1\ </math> and <math>t_9\ </math> from each side, and that we oriented these branches as shown in Figure(b). Let <math>m = t_1+t_8+t_9\ </math>, be the current length of the clothesline. We select the new length to be <math>m^{\star} = m\exp(\lambda(U_1-0.5))\ </math>, where <math>U_1\ </math> is a uniform random variable on <math>(0,1)\ </math>. Then for the LOCAL algorithm, the acceptance probability can be computed to be:
| |
| | |
| :<math>\frac{h(y)}{h(x)} \times \frac{{m^{\star}}^3}{m^3}\ </math>
| |
| | |
| ===Assessing convergence===
| |
| | |
| Suppose we want to estimate a branch length of a 2-taxon tree under JC, in which <math>n_1</math> sites are unvaried and <math>n_2</math> are variable. Assume exponential prior distribution with rate <math>\lambda\ </math>. The density is <math>p(t) = \lambda e^{-\lambda t}\ </math>. The probabilities of the possible site patterns are:
| |
| | |
| :<math>1/4\left(1/4+3/4e^{-4/3t}\right)\ </math>
| |
| | |
| for unvaried sites, and
| |
| | |
| :<math> 1/4\left(1/4-1/4e^{-4/3t}\right)\ </math>
| |
| | |
| Thus the unnormalized posterior distribution is:
| |
| | |
| : <math> h(t) = \left(1/4\right)^{n_1+n_2}\left(1/4+3/4{e^{-4/3t}}^{n_1}\right)\ </math>
| |
| | |
| or, alternately,
| |
| | |
| : <math> h(t) = \left(1/4-1/4{e^{-4/3t}}^{n_2}\right)(\lambda e^{-\lambda t})\ </math>
| |
| | |
| Update branch length by choosing new value uniformly at random from a window of half-width <math>w\ </math> centered at the current value:
| |
| | |
| :<math> t^\star = |t+U|\ </math>
| |
| | |
| where <math>U\ </math>is uniformly distributed between <math>-w\ </math> and <math>w\ </math>. The acceptance | |
| probability is:
| |
| | |
| :<math> h(t^\star)/h(t)\ </math>
| |
| | |
| Example: <math>n_1 = 70\ </math>, <math>n_2 = 30\ </math>. We will compare results for two values of <math>w\ </math>, <math>w = 0.1\ </math> and <math>w = 0.5\ </math>. In each case, we will begin with an initial length of <math>5\ </math> and update the length <math>2000\ </math> times. (See Figure 3.2 (to be added) for results.)
| |
| | |
| ==Metropolis-coupled MCMC (Geyer)==
| |
| | |
| If the target distribution has multiple peaks, separated by low valleys, the Markov chain may have difficulty in moving from one peak to another. As a result, the chain may get stuck on one peak and the resulting samples will not approximate the posterior density correctly. This is a serious practical concern for phylogeny reconstruction, as multiple local peaks are known to exist in the tree space during heuristic tree search under maximum parsimony (MP), maximum likelihood (ML), and minimum evolution (ME) criteria, and the same can be expected for stochastic tree search using MCMC. Many strategies have been proposed to improve mixing of Markov chains in presence of multiple local peaks in the posterior density. One of the most successful algorithms is the Metropolis-coupled MCMC (or <math>\mathrm{MC}^3\ </math>).
| |
| | |
| In this algorithm, <math>m\ </math> chains are run in parallel, with different stationary distributions <math>\pi_j(.)\ </math>, <math>j = 1, 2, \ldots, m\ </math>, where the first one, <math>\pi_1 = \pi\ </math> is the target density, while <math>\pi_j\ </math>, <math>j = 2, 3, \ldots, m\ </math> are chosen to improve mixing. For example, one can choose incremental heating of the form:
| |
| | |
| : <math> \pi_j(\theta) = \pi(\theta)^{1/[1+\lambda(j-1)]}, \ \ \lambda > 0, </math>
| |
| | |
| so that the first chain is the cold chain with the correct target density, while chains <math>2, 3, \ldots, m</math> are heated chains. Note that raising the density <math>\pi(.)</math> to the power <math>1/T\ </math> with <math>T>1\ </math> has the effect of flattening out the distribution, similar to heating a metal. In such a distribution, it is easier to traverse between peaks (separated by valleys) than in the original distribution. After each iteration, a swap of states between two randomly chosen chains is proposed through a Metropolis-type step. Let <math>\theta^{(j)}\ </math> be the current state in chain <math>j\ </math>, <math>j = 1, 2, \ldots, m\ </math>. A swap between the states of chains <math>i\ </math> and <math>j\ </math> is accepted with probability:
| |
| | |
| :<math> \alpha = \frac{\pi_i(\theta^{(j)})\pi_j(\theta^{(i)})}{\pi_i(\theta^{(i)})\pi_j(\theta^{(j)})}\ </math>
| |
| | |
| At the end of the run, output from only the cold chain is used, while those from the hot chains are discarded. Heuristically, the hot chains will visit the local peaks rather easily, and swapping states between chains will let the cold chain occasionally jump valleys, leading to better mixing. However, if <math>\pi_i(\theta)/\pi_j(\theta)\ </math> is unstable, proposed swaps will seldom be accepted. This is the reason for using several chains which differ only incrementally. (See Figure3.3 (to be added)).
| |
| | |
| An obvious disadvantage of the algorithm is that <math>m\ </math> chains are run and only one chain is used for inference. For this reason, <math>\mathrm{MC}^3\ </math> is ideally suited for implementation on parallel machines, since each chain will in general require the same amount of computation per iteration.
| |
| | |
| {{nofootnotes|date=April 2011}}
| |
| == References ==
| |
| *{{cite book |author=Geyer, C.J. |chapter=Markov chain Monte Carlo maximum likelihood |editor=Keramidas, E.M. |title=Computing Science and Statistics: Proceedings of the 23rd Symposium of the Interface |publisher=Interface Foundation |location=Fairfax Station VA |year=1991 |isbn= |pages=156–163 |oclc=26603816}}
| |
| *{{cite journal |author=Yang Z, Rannala B |title=Bayesian phylogenetic inference using DNA sequences: a Markov Chain Monte Carlo Method |journal=Mol. Biol. Evol. |volume=14 |issue=7 |pages=717–24 |date=July 1997|pmid=9214744 |url=http://mbe.oxfordjournals.org/cgi/pmidlookup?view=long&pmid=9214744}}
| |
| *{{cite journal |author=Larget B, Simon DL |title=Markov chain Monte Carlo algorithms for the Bayesian analysis of phylogenetic trees |journal= |volume=16 |issue=6 |pages=750–9 |date=June 1999 |format=PDF |url=http://mbe.oxfordjournals.org/content/16/6/750.full.pdf+html}}
| |
| *{{cite journal |author=Huelsenbeck JP, Ronquist F |title=MRBAYES: Bayesian inference of phylogenetic trees |journal=Bioinformatics |volume=17 |issue=8 |pages=754–5 |date=August 2001|pmid=11524383 |url=http://bioinformatics.oxfordjournals.org/cgi/pmidlookup?view=long&pmid=11524383}}
| |
| *{{cite journal |author=Ronquist F, Huelsenbeck JP |title=MrBayes 3: Bayesian phylogenetic inference under mixed models |journal=Bioinformatics |volume=19 |issue=12 |pages=1572–4 |date=August 2003|pmid=12912839 |url=http://bioinformatics.oxfordjournals.org/cgi/pmidlookup?view=long&pmid=12912839}}
| |
| | |
| [[Inferencia bayesiana en filogenia|es:Español]]
| |
| | |
| {{Phylogenetics}}
| |
| | |
| [[Category:Bioinformatics]]
| |
| [[Category:Computational phylogenetics]]
| |
| [[Category:Applications of Bayesian inference|Phylogeny]]
| |
Her title is Felicidad Ahmad. The occupation he's been occupying for years is a messenger. Alabama is exactly where he and his wife live and he has everything that he needs there. To perform badminton is some thing he truly enjoys doing.
Also visit my website extended car warranty (mouse click the up coming website)