* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download Markov - Mathematics
Computer simulation wikipedia , lookup
Algorithm characterizations wikipedia , lookup
Mathematical economics wikipedia , lookup
Mathematical physics wikipedia , lookup
Particle filter wikipedia , lookup
Brownian motion wikipedia , lookup
Financial economics wikipedia , lookup
Fisher–Yates shuffle wikipedia , lookup
Monte Carlo method wikipedia , lookup
Markov Chains Brian Carrico The Mathematical Markovs  Vladimir Andreyevich Markov (1871-1897)   Andrey Markov’s younger brother  With Andrey, developed the Markov brothers’ inequality Andrey Andreyevich Markov Jr (1903-1979)  Andrey Markov’s son  One of the key founders of the Russian school of constructive mathematics and logic  Also made contributions to differential equations,topology, mathematical logic and the foundations of mathematics  Which brings us to: Andrey Andreyevich Markov Андрей Андреевич Марков June 14, 1856 – July 20, 1922  Born in Ryazan  (roughly 170 miles Southeast of Moscow)  Began Grammar School     in 1866 Started at St Petersburg University in 1874 Defended his Masters Thesis in 1880 Doctoral Thesis in 1885 Excommunicated from the Russian Orthodox Church Precursors to Markov Chains  Bernoulli Series  Brownian Motion  Random Walks Bernoulli Series  Jakob Bernoulli (1654-1705)  Sequence independent random variables X1, X2,X3,... such that  For every i, Xi is either 0 or 1  For every i, P(Xi)=1 is the same  Markov’s first discussions of chains, a 1906 paper, considers only chains with two states  Closely related to Random Walks Brownian Motion  Described as early as 60 BC by Roman poet Lucretius  Formalized and officially discovered by botanist Robert Brown in 1827  The seemingly random movement of particles suspended in a fluid Random Walks  Formalized in 1905 by Karl Pearson  The formalization of a trajectory that consists of taking successive random steps  The results of random walk analysis have been applied to computer science, physics, ecology, economics, and a number of other fields as a fundamental model for random processes in time  Turns out to be a specific Markov chain So what is a Markov Chain?  A random process where all information about the future is contained in the present state  Or less formally: a process where future states depend only on the present state, and are independent of past states  Mathematically: Applications of Markov Chains  Science  Statistics  Economics and Finance  Gambling and games of chance  Baseball  Monte Carlo Science  Physics  Thermodynamic systems generally have timeinvariant dynamics  All relevant information is in the state description  Chemistry  An algorithm based on a Markov chain was used to focus the fragment-based growth of chemicals in silico towards a desired class of compounds such as drugs or natural products Economics and Finance  Markov Chains are used model a variety of different phenomena, including asset prices and market crashes.  Regime-switching model of James D. Hamilton  Markov Switching Multifractal asset pricing model  Dynamic macroeconomics Gambling and Games of Chance  In most card games each hand is independent  Board games like Snakes and Ladders Baseball  Use of Markov chain models in baseball analysis began in 1960  Each at bat can be taken as a Markov chain Monte Carlo  A Markov chain with a large number of steps is used to create the algorithm for the basis of the Monte Carlo simulation Statistics  Many important statistics measure independent trials, which can be represented by Markov chains An Example from Statistics  A thief is in a dungeon with three identical doors. Once the thief chooses a door and passes through it, the door locks behind him. The three doors lead to:  A 6 hour tunnel leading to freedom  A 3 hour tunnel that returns to the dungeon  A 9 hour tunnel that returns to the dungeon  Each door is chosen with equal probability. When he is dropped back into the dungeon by the second and third doors there is a memoryless choice of doors. He isn’t able to mark the doors in any way. What is his expected time of escape?  Note: Example (cont)  We plug the values in for xi and p(xi) to get:  E(X)=6*(1/3)+x2*(1/3)+x3*(1/3)  But what are x2 and x3?  Because the decision is memoryless, the expected time after returning from tunnels 2 or 3 doesn’t change from the initial expected time. So, x2=x3=E(X).  So,  E(X)=6*(1/3)+E(X)*(1/3)+E(X)*(1/3)  Now we’re back in Algebra 1 Sources  Wikipedia  The Life and Work of A.A. Markov. Basharin, Gely P. et al. http://decision.csl.illinois.edu/~meyn/pages /Markov-Work-and-life.pdf  Leemis (2009), Probability