An introduction to markov chains
Download
1 / 36

www2.spsu - PowerPoint PPT Presentation


  • 342 Views
  • Updated On :

An Introduction to Markov Chains Homer and Marge repeatedly play a gambling game. Each time they play, the probability that Homer wins is 0.4, and the probability that Homer loses is 0.6 A “Drunkard’s Walk” 0 1 2 3 4 0 1 2 3 4 P(Homer wins) = .4 P(Homer loses) = .6

Related searches for www2.spsu

loader
I am the owner, or an agent authorized to act on behalf of the owner, of the copyrighted work described.
capcha
Download Presentation

PowerPoint Slideshow about 'www2.spsu' - andrew


An Image/Link below is provided (as is) to download presentation

Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.


- - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - -
Presentation Transcript

Slide2 l.jpg

Homer and Marge repeatedly play a gambling game. Each time they play,the probability that Homer wins is 0.4, and the probability that Homer loses is 0.6

A “Drunkard’s Walk”


Slide3 l.jpg

0 1 2 3 4 they play,

0

1

2

3

4

P(Homer wins) = .4 P(Homer loses) = .6

Homer and Marge both start with $2


Slide12 l.jpg

A they play,Markov Chain is a mathematical model for a process which moves step by step through various states.

In a Markov chain, the probability that the process moves from any given state to any other particular state is always the same, regardless of the history of the process.


Slide13 l.jpg

A they play,Markov chain consists of states and transitionprobabilities.

Each transition probability is the probability of moving from one state to another in one step. The transition probabilities are independent of the past, and depend only on the two states involved. The matrix of transition probabilities is called the transition matrix.


Slide14 l.jpg

0 1 2 3 4 they play,

P(Homer wins) = .4 P(Homer loses) = .6

Homer and Marge both start with $2


Slide15 l.jpg

If they play,Pis the transition matrix for a Markov Chain, then the nth power of P gives the probabilities of going from state to state in exactly n steps.


Slide16 l.jpg

If the vector they play,v represents the initial state, then the probabilities of winding up in the various states in exactly n steps are exactly v times the nth power of P .


Slide17 l.jpg

When they both start with $2, the probability that Homer is ruined is 9/13.

If Homer starts with $ x and Marge starts with $ N-x, and P(Homer wins) = p, P(Homer loses) = q, then the probability Homer is ruined is


Slide18 l.jpg

Suppose you be on red in roulette. ruined is 9/13.P(win) = 18/38 = 9/19; P(lose) = 10/19.

Suppose you and the house each have $10

Now suppose you have $ 10 and the house has $20



Slide20 l.jpg

Andrei Markov (1856-1922) ruined is 9/13.

Paul Eherenfest: Diffusion model, early 1900sStatistical interpretation of the second law of thermodynamics: The entropy of a closed system can only increase.Proposed the “Urn Model” to explain diffusion.

Albert Einstein, 1905Realized Brownian motion would provide a magnifying glass into the world of the atom. Brownian motion has been extensively modeled by Markov Chains


Slide22 l.jpg

Particles are separated by a semi-permeable membrane, which they can pass through in either direction.

Suppose that there are N black particles inside the membrane, and N white particles outside the membrane. Each second, one random molecule goes from outside the membrane to inside, and vice versa.

There are N+1 states, given by the number of white molecules inside.

Osmosis


Slide24 l.jpg

0 1 2 3 4 5 they can pass through in either direction.

5 molecules


Slide25 l.jpg

N molecules they can pass through in either direction.


Slide26 l.jpg

N molecules they can pass through in either direction.


Slide27 l.jpg

If this process runs for a while, an interesting question is: How much time, on average, is the process in each state?

A Markov chain with transition matrix P is said to be regular if some power of P has all positive entries for some n. In a regular Markov chain, it is possible to get from any state to any other state in n steps.


Slide28 l.jpg

The Markov chain for our osmosis process is regular. Even starting with all black particles inside, if a white particle entered at every step, then the process would pass from zero white inside through all possible states.


Slide29 l.jpg

For a regular Markov chain, the amount of time the process spends in each state is given by the fixed probability vector, which is the vector a such that Pa = a.

Moreover, for any probability vector w,

No matter what the starting state, if the process runs for a long time, the probability of being in a given state is given by a.


Slide30 l.jpg

I spends in each state is given by the n the long run, the fraction of time the process spends in each state is given by the fixed probability vector.


Slide31 l.jpg

For N particles, the fixed vector is: spends in each state is given by the

1

1 1

1 2 1

1 3 3 1

1 4 6 4 1

1 5 10 10 5 1

1 6 15 20 15 6 1


Slide32 l.jpg

Fixed vectors spends in each state is given by the

N = 4 (1/70, 16/70, 36/70, 16/70, 1/70)


Slide35 l.jpg

Now suppose 500 molecules spends in each state is given by the

The percent of the time that there are between 225 and 275 black molecules inside is 0.999.

The percent of the time that there are either fewer than 100 black or more than 400 black molecules inside is


Slide36 l.jpg

If the universe is 15 billion years old, the average amount of time that a system with 500 molecules will have fewer than 100 black or more than 400 black molecules inside the membrane is


ad