1 Statistical Mechanics and MultiScale Simulation Methods ChBE 591-009 Prof. C. Heath Turner Lecture 12 • Some materials adapted from Prof. Keith E. Gubbins: http://gubbins.ncsu.edu • Some materials adapted from Prof. David Kofke: http://www.cbe.buffalo.edu/kofke.htm Monte Carlo Integration: Review Taken from Dr. D. A. Kofke’s lectures @ SUNY Buffalo: http://www.eng.buffalo.edu/~kofke/ce530/index.html Stochastic evaluation of integrals • sum integrand evaluated at randomly generated points • most appropriate for high-dimensional integrals error vanishes more quickly (1/n1/2) better suited for complex-shaped domains of integration Monte Carlo simulation • Monte Carlo integration for ensemble averages U Importance Sampling • • • • 1 N! dr N N e U ( r U (r ) Z N N) p(rN) emphasizes sampling in domain where integrand is largest it is easy to generate points according to a simple distribution stat mech p distributions are too complex for direct sampling need an approach to generate random multidimensional points according to a complex probability distribution f • then integral is given by I p p 2 Taken from Dr. D. A. Kofke’s lectures @ SUNY Buffalo: http://www.eng.buffalo.edu/~kofke/ce530/index.html Markov Processes Stochastic process • movement through a series of well-defined states in a way that involves some element of randomness • for our purposes,“states” are microstates in the governing ensemble Markov process • stochastic process that has no memory • selection of next state depends only on current state, and not on prior states • process is fully defined by a set of transition probabilities pij pij = probability of selecting state j next, given that presently in state i. Transition-probability matrix P collects all pij 3 Taken from Dr. D. A. Kofke’s lectures @ SUNY Buffalo: http://www.eng.buffalo.edu/~kofke/ce530/index.html Transition-Probability Matrix Example If in state 1, will stay in state 1 with probability 0.1 • system with three states p 11 p 12 p 13 0.1 0.5 0.4 P p 21 p 22 p 23 0.9 0.1 0.0 p 31 p 32 p 33 0.3 0.3 0.4 If in state 1, will move to state 3 with probability 0.4 Never go to state 3 from state 2 Requirements of transition-probability matrix • all probabilities non-negative, and no greater than unity • sum of each row is unity • probability of staying in present state may be non-zero 4 Taken from Dr. D. A. Kofke’s lectures @ SUNY Buffalo: http://www.eng.buffalo.edu/~kofke/ce530/index.html Distribution of State Occupancies Consider process of repeatedly moving from one state to the next, choosing each subsequent state according to P • 1 2 2 1 3 2 2 3 3 1 2 3 etc. Histogram the occupancy number for each state • n1 = 3 • n2 = 5 • n3 = 4 p1 = 0.33 p2 = 0.42 p3 = 0.25 1 2 3 After very many steps, a limiting distribution emerges Click here for an applet that demonstrates a Markov process and its approach to a limiting distribution 5 Taken from Dr. D. A. Kofke’s lectures @ SUNY Buffalo: http://www.eng.buffalo.edu/~kofke/ce530/index.html The Limiting Distribution 1. Consider the product of P with itself p 11 p 12 p 13 p 11 p 12 p 13 All ways of going from state P 2 p 21 p 22 p 23 p 21 p 22 p 23 1 to state 2 in two steps p p p p p p 31 32 33 31 32 33 p 11p 11 p 12p 21 p 13p 31 p 11p 12 p 12p 22 p 13p 32 etc. p 21p 11 p 22p 21 p 23p 31 p 21p 12 p 22p 22 p 23p 32 etc. p p p p p p 31 11 32 21 33 31 p 31p 12 p 32p 22 p 33p 32 etc. Probability of going from state 3 to state 2 in two steps In general Pn is the n-step transition probability matrix • probabilities of going from state i to j in exactly n steps ( n) ( n) ( n) p 11 p 12 p13 ( n) ( n) ( n) n P p 21 p 22 p 23 defines p ( n ) ij ( n) ( n) ( n) p 31 p 32 p 33 6 Taken from Dr. D. A. Kofke’s lectures @ SUNY Buffalo: http://www.eng.buffalo.edu/~kofke/ce530/index.html The Limiting Distribution 2. Define p i(0) as a unit state vector p1(0) 1 0 0 p 2(0) 0 1 0 p 3(0) 0 0 1 Then p i(n) p i(0)P n is a vector of probabilities for ending at each state after n steps if beginning at state i p1( n) ( n) ( n) ( n) p 11 p12 p13 (0) n ( n) ( n) ( n) ( n) ( n) ( n) p 1 P 1 0 0 p 21 p 22 p 23 p 11 p 12 p 13 ( n) ( n) ( n) p 31 p 32 p 33 The limiting distribution corresponds to n () () ( ) • independent of initial state p1 p 2 p 3 p 7 Taken from Dr. D. A. Kofke’s lectures @ SUNY Buffalo: http://www.eng.buffalo.edu/~kofke/ce530/index.html 8 The Limiting Distribution 3. Stationary property of p p lim p i(0)P n n lim p i(0)P n 1 P n pP p is a left eigenvector of P with unit eigenvalue • such an eigenvector is guaranteed to exist for matrices with rows that each sum to unity Equation for elements of limiting distribution p p i p jp ji j p 1 0.1p 1 0.9p 2 0.3p 3 0.1 0.5 0.4 p 2 0.5p 1 .1p 2 0.3p 3 e.g . P 0.9 0.1 0.0 0.3 0.3 0.4 p 3 0.4p 1 0.0p 2 0.4p 3 p1 p 2 p 3 p1 p 2 p 3 not independent Taken from Dr. D. A. Kofke’s lectures @ SUNY Buffalo: http://www.eng.buffalo.edu/~kofke/ce530/index.html 9 Detailed Balance Eigenvector equation for limiting distribution • p i p jp ji j A sufficient (but not necessary) condition for solution is • p ip ij p jp ji • “detailed balance” or “microscopic reversibility” Thus • p i p jp ji j p ip ij j p i p ij p i j 0.1 P 0.9 0.3 0.5 0.1 0.3 0.4 0.0 0.4 For a given P, it is not always possible to satisfy detailed balance; e.g. for this P p 3p 32 p 2p 23 zero Taken from Dr. D. A. Kofke’s lectures @ SUNY Buffalo: http://www.eng.buffalo.edu/~kofke/ce530/index.html 10 Deriving Transition Probabilities Turn problem around... …given a desired p, what transition probabilities will yield this as a limiting distribution? Construct transition probabilities to satisfy detailed balance Many choices are possible • e.g. p 0.25 0.5 0.25 • try them out 0 1 0 P 0.5 0 0.5 0 1 0 Most efficient 0.97 0.02 0.01 P 0.01 0.98 0.01 0.01 0.02 0.97 Least efficient 0.0 0.5 0.5 0.42 0.33 0.25 P 0.17 0.66 0.17 P 0.25 0.5 0.25 0.5 0.5 0.0 0.25 0.33 0.42 Barker Metropolis Taken from Dr. D. A. Kofke’s lectures @ SUNY Buffalo: http://www.eng.buffalo.edu/~kofke/ce530/index.html 11 Metropolis Algorithm 1. Prescribes transition probabilities to satisfy detailed balance, given desired limiting distribution Recipe: From a state i… • with probability tij, choose a trial state j for the move (note: tij = tji) • if pj > pi, accept j as the new state • otherwise, accept state j with probability pj/pi generate a random number R on (0,1); accept if R < pj/pi • if not accepting j as the new state, take the present state as the next one in the Markov chain p ii 0 Metropolis, Rosenbluth, Rosenbluth, Teller and Teller, J. Chem. Phys., 21 1087 (1953) 12 Metropolis Algorithm 2. Taken from Dr. D. A. Kofke’s lectures @ SUNY Buffalo: http://www.eng.buffalo.edu/~kofke/ce530/index.html What are the transition probabilities for this algorithm? • Without loss of generality, define i as the state of greater probability pj p ij t ij pi p ji t ji p ii 1 p ij p j in general: p t min ij ij p ,1 i pi p j j i Do they obey detailed balance? ? p ip ij p jp ji pj ? p it ij p jt ji pi t ij t ji Yes, as long as the underlying matrix T of the Markov chain is symmetric • this can be violated, but acceptance probabilities must be modified Taken from Dr. D. A. Kofke’s lectures @ SUNY Buffalo: http://www.eng.buffalo.edu/~kofke/ce530/index.html 13 Markov Chains and Importance Sampling 1. Importance sampling specifies the desired limiting distribution We can use a Markov chain to generate quadrature points according to this distribution 1 inside R Example s 0 0.5 r2 0.5 dx 0.5 dy ( x 2 y 2 ) s ( x, y ) 0.5 0.5 0.5 0.5 dx 0.5 dys( x, y) outside R r 2s s V V q = normalization constant V • Method 1: let p1( x, y) s( x, y) / q1 • then r 2 s 2 2 r 2 p1 p 1 s p1 p 1 q1r q1 p1 p1 q1 r q1 p1 r 2 p1 Simply sum r2 with points given by Metropolis sampling 14 Markov Chains and Importance Sampling 2. Taken from Dr. D. A. Kofke’s lectures @ SUNY Buffalo: http://www.eng.buffalo.edu/~kofke/ce530/index.html Example (cont’d) p ( x, y ) r 2 s / q2 • Method 2: let r2s • then r 2 p2 p 2 s p2 p 2 q2 p2 q2 / r 2 p2 q2 q2 1/ r 2 p2 1 r 2 p2 Algorithm and transition probabilities • given a point in the region R • generate a new point in the vicinity of given point xnew = x + r(-1,+1)dx ynew = y + r(-1,+1)dy • accept with probability min(1,p new / p old ) p 1new s new / q1 s new • note p 1old s old / q1 s old Normalization constants cancel! • Method 1: accept all moves that stay in R 2 new r 2 old r • Method 2: if in R, accept with probability Taken from Dr. D. A. Kofke’s lectures @ SUNY Buffalo: http://www.eng.buffalo.edu/~kofke/ce530/index.html Markov Chains and Importance Sampling 3. Subtle but important point • Underlying matrix T is set by the trial-move algorithm (select new point uniformly in vicinity of present point) • It is important that new points are selected in a volume that is independent of the Different-sized trial sampling present position regions • If we reject configurations outside R, without taking the original point as the “new” one, then the underlying matrix becomes asymmetric 15 Taken from Dr. D. A. Kofke’s lectures @ SUNY Buffalo: http://www.eng.buffalo.edu/~kofke/ce530/index.html Evaluating Areas with Metropolis Sampling What if we want the absolute area of the region R, not an average over it? 0.5 0.5 A 0.5 dx 0.5 dys( x, y ) s V • Let p1( x, y) s( x, y) / q1 • then A s q q 1 p 1 p 1 p1 1 • We need to know the normalization constant q1 • but this is exactly the integral that we are trying to solve! Absolute integrals difficult by MC • relates to free-energy evaluation 16 Taken from Dr. D. A. Kofke’s lectures @ SUNY Buffalo: http://www.eng.buffalo.edu/~kofke/ce530/index.html Summary Markov process is a stochastic process with no memory Full specification of process is given by a matrix of transition probabilities P A distribution of states are generated by repeatedly stepping from one state to another according to P A desired limiting distribution can be used to construct transition probabilities using detailed balance • Many different P matrices can be constructed to satisfy detailed balance • Metropolis algorithm is one such choice, widely used in MC simulation Markov Monte Carlo is good for evaluating averages, but not absolute integrals Next up: Monte Carlo simulation 17
© Copyright 2026 Paperzz