Random walk

A random walk is a object, known as a stochastic or, that describes a path that consists of a succession of  steps on some mathematical space such as the integers. An elementary example of a random walk is the random walk on the number line, $$\mathbb Z$$, which starts at 0 and at each step moves +1 or −1 with equal probability. Other examples include the path traced by a as it travels in a liquid or a gas, the search path of a  animal, the price of a fluctuating  and the financial status of a : all can be approximated by random walk models, even though they may not be truly random in reality. As illustrated by those examples, random walks have applications to and many scientific fields including, , , , ,  as well as. Random walks explain the observed behaviors of many processes in these fields, and thus serve as a fundamental for the recorded. As a more mathematical application, the value of $\pi$ can be approximated by the use of random walk in an agent-based modeling environment. The term random walk was first introduced by in 1905.

Various types of random walks are of interest, which can differ in several ways. The term itself most often refers to a special category of, but many time-dependent processes are referred to as random walks, with a modifier indicating their specific properties. Random walks (Markov or not) can also take place on a variety of spaces: commonly studied ones include, others on the integers or the real line, in the plane or higher-dimensional vector spaces, on or higher-dimensional s, and also on  finite,  or. The time parameter can also be manipulated. In the simplest context the walk is in discrete time, that is a sequence of s (X$t$) = (X$1$, X$2$, ...) indexed by the natural numbers. However, it is also possible to define random walks which take their steps at random times, and in that case, the position X$t$ has to be defined for all times t &isin; [0,+&infin;). Specific cases or limits of random walks include the and  models such as.

Random walks are a fundamental topic in discussions of Markov processes. Their mathematical study has been extensive. Several properties, including dispersal distributions, first-passage or hitting times, encounter rates, recurrence or transience, have been introduced to quantify their behavior.

Lattice random walk
A popular random walk model is that of a random walk on a regular lattice, where at each step the location jumps to another site according to some probability distribution. In a simple random walk, the location can only jump to neighboring sites of the lattice, forming a. In simple symmetric random walk on a locally finite lattice, the probabilities of the location jumping to each one of its immediate neighbors are the same. The best-studied example is of random walk on the d-dimensional integer lattice (sometimes called the hypercubic lattice) $$\mathbb Z^d$$.

If the state space is limited to finite dimensions, the random walk model is called simple bordered symmetric random walk, and the transition probabilities depend on the location of the state because on margin and corner states the movement is limited.

One-dimensional random walk
An elementary example of a random walk is the random walk on the number line, $$\mathbb Z$$, which starts at 0 and at each step moves +1 or −1 with equal probability.

This walk can be illustrated as follows. A marker is placed at zero on the number line, and a fair coin is flipped. If it lands on heads, the marker is moved one unit to the right. If it lands on tails, the marker is moved one unit to the left. After five flips, the marker could now be on 1, −1, 3, −3, 5, or −5. With five flips, three heads and two tails, in any order, it will land on 1. There are 10 ways of landing on 1 (by flipping three heads and two tails), 10 ways of landing on −1 (by flipping three tails and two heads), 5 ways of landing on 3 (by flipping four heads and one tail), 5 ways of landing on −3 (by flipping four tails and one head), 1 way of landing on 5 (by flipping five heads), and 1 way of landing on −5 (by flipping five tails). See the figure below for an illustration of the possible outcomes of 5 flips.

To define this walk formally, take independent random variables $$Z_1, Z_2,\dots$$, where each variable is either 1 or −1, with a 50% probability for either value, and set $$S_0 = 0\,\!$$ and $$S_n =\sum_{j=1}^nZ_j.$$ The $$\{S_n\}\,\!$$ is called the simple random walk on $$\mathbb Z$$. This series (the sum of the sequence of −1s and 1s) gives the distance walked, if each part of the walk is of length one. The $$E(S_n)\,\!$$ of $$S_n\,\!$$ is zero. That is, the mean of all coin flips approaches zero as the number of flips increases. This follows by the finite additivity property of expectation:


 * $$E(S_n)=\sum_{j=1}^n E(Z_j)=0.$$

A similar calculation, using the independence of the random variables and the fact that $$E(Z_n^2)=1$$, shows that:


 * $$E(S_n^2)=\sum_{i=1}^n \sum_{j=1}^n E(Z_j Z_i)=n.$$

This hints that $$E(|S_n|)\,\!$$, the translation distance after n steps, should be  $$\sqrt n$$. In fact,


 * $$\lim_{n\to\infty} \frac{E(|S_n|)}{\sqrt n}= \sqrt{\frac {2}{\pi}}.$$

This result shows that diffusion is ineffective for mixing because of the way the square root behaves for large $$N$$.

How many times will a random walk cross a boundary line if permitted to continue walking forever? A simple random walk on $$\mathbb Z$$ will cross every point an infinite number of times. This result has many names: the level-crossing phenomenon, recurrence or the . The reason for the last name is as follows: a gambler with a finite amount of money will eventually lose when playing a fair game against a bank with an infinite amount of money. The gambler's money will perform a random walk, and it will reach zero at some point, and the game will be over.

If a and b are positive integers, then the expected number of steps until a one-dimensional simple random walk starting at 0 first hits b or −a is ab. The probability that this walk will hit b before −a is $$a/(a+b)$$, which can be derived from the fact that simple random walk is a.

Some of the results mentioned above can be derived from properties of. The number of different walks of n steps where each step is +1 or −1 is 2n. For the simple random walk, each of these walks is equally likely. In order for Sn to be equal to a number k it is necessary and sufficient that the number of +1 in the walk exceeds those of −1 by k. It follows +1 must appear (n + k)/2 times among n steps of a walk, hence the number of walks which satisfy $$S_n=k$$ equals the number of ways of choosing (n + k)/2 elements from an n element set, denoted $$n \choose (n+k)/2$$. For this to have meaning, it is necessary that n + k be an even number, which implies n and k are either both even or both odd. Therefore, the probability that $$S_n=k$$ is equal to $$2^{-n}{n\choose (n+k)/2}$$. By representing entries of Pascal's triangle in terms of s and using, one can obtain good estimates for these probabilities for large values of $$n$$.

If space is confined to $$\mathbb Z$$+ for brevity, the number of ways in which a random walk will land on any given number having five flips can be shown as {0,5,0,4,0,1}.

This relation with Pascal's triangle is demonstrated for small values of n. At zero turns, the only possibility will be to remain at zero. However, at one turn, there is one chance of landing on −1 or one chance of landing on 1. At two turns, a marker at 1 could move to 2 or back to zero. A marker at −1, could move to −2 or back to zero. Therefore, there is one chance of landing on −2, two chances of landing on zero, and one chance of landing on 2.

The and the  describe important aspects of the behavior of simple random walks on $$\mathbb Z$$. In particular, the former entails that as n increases, the probabilities (proportional to the numbers in each row) approach a.

As a direct generalization, one can consider random walks on crystal lattices (infinite-fold abelian covering graphs over finite graphs). Actually it is possible to establish the central limit theorem and large deviation theorem in this setting.

As a Markov chain
A one-dimensional random walk can also be looked at as a whose state space is given by the integers $$i=0,\pm 1,\pm 2,\dots .$$ For some number p satisfying $$\,0 < p < 1$$, the transition probabilities (the probability Pi,j of moving from state i to state j)  are given by
 * $$\,P_{i,i+1}=p=1-P_{i,i-1}.$$

Higher dimensions
In higher dimensions, the set of randomly walked points has interesting geometric properties. In fact, one gets a discrete, that is, a set which exhibits stochastic on large scales. On small scales, one can observe "jaggedness" resulting from the grid on which the walk is performed. Two books of Lawler referenced below are a good source on this topic. The trajectory of a random walk is the collection of points visited, considered as a set with disregard to when the walk arrived at the point. In one dimension, the trajectory is simply all points between the minimum height and the maximum height the walk achieved (both are, on average, on the order of $$ \sqrt{n} $$).

To visualize the two-dimensional case, one can imagine a person walking randomly around a city. The city is effectively infinite and arranged in a square grid of sidewalks. At every intersection, the person randomly chooses one of the four possible routes (including the one originally travelled from). Formally, this is a random walk on the set of all points in the with.

Will the person ever get back to the original starting point of the walk? This is the 2-dimensional equivalent of the level crossing problem discussed above. In 1921 proved that the person  would in a 2-dimensional random walk, but for 3 dimensions or higher, the probability of returning to the origin decreases as the number of dimensions increases. In 3 dimensions, the probability decreases to roughly 34%.

The asymptotic function for a two-dimensional random walk as the number of steps increases is given by a. The probability distribution is a function of the radius from the origin and the step length is constant for each step.


 * $$P(r) = \frac{2r}{N} e^{-r^2/N} $$

Relation to Wiener process
A is a stochastic process with similar behavior to, the physical phenomenon of a minute particle diffusing in a fluid. (Sometimes the is called "Brownian motion", although this is strictly speaking a .)

A Wiener process is the of random walk in dimension 1. This means that if you take a random walk with very small steps, you get an approximation to a Wiener process (and, less accurately, to Brownian motion). To be more precise, if the step size is ε, one needs to take a walk of length L/ε2 to approximate a Wiener length of L. As the step size tends to 0 (and the number of steps increases proportionally), random walk converges to a Wiener process in an appropriate sense. Formally, if B is the space of all paths of length L with the maximum topology, and if M is the space of measure over B with the norm topology, then the convergence is in the space M. Similarly, a Wiener process in several dimensions is the scaling limit of random walk in the same number of dimensions.

A random walk is a discrete (a function with integer dimensions; 1, 2, ...), but a Wiener process trajectory is a true fractal, and there is a connection between the two. For example, take a random walk until it hits a circle of radius r times the step length. The average number of steps it performs is r2. This fact is the discrete version of the fact that a Wiener process walk is a fractal of 2.

In two dimensions, the average number of points the same random walk has on the boundary of its trajectory is r4/3. This corresponds to the fact that the boundary of the trajectory of a Wiener process is a fractal of dimension 4/3, a fact predicted by using simulations but proved only in 2000 by, and.

A Wiener process enjoys many random walk does not. For example, a Wiener process walk is invariant to rotations, but the random walk is not, since the underlying grid is not (random walk is invariant to rotations by 90 degrees, but Wiener processes are invariant to rotations by, for example, 17 degrees too). This means that in many cases, problems on a random walk are easier to solve by translating them to a Wiener process, solving the problem there, and then translating back. On the other hand, some problems are easier to solve with random walks due to its discrete nature.

Random walk and can be, namely manifested on the same probability space in a dependent way that forces them to be quite close. The simplest such coupling is the Skorokhod embedding, but there exist more precise couplings, such as theorem.

The convergence of a random walk toward the Wiener process is controlled by the, and by. For a particle in a known fixed position at t = 0, the central limit theorem tells us that after a large number of  steps in the random walk, the walker's position is distributed according to a  of total :


 * $$\sigma^2 = \frac{t}{\delta t}\,\varepsilon^2,$$

where t is the time elapsed since the start of the random walk, $$\varepsilon$$ is the size of a step of the random walk, and $$\delta t$$ is the time elapsed between two successive steps.

This corresponds to the of the  that controls the Wiener process, which suggests that, after a large number of steps, the random walk converges toward a Wiener process.

In 3D, the variance corresponding to the of the diffusion equation is:


 * $$\sigma^2 = 6\,D\,t.$$

By equalizing this quantity with the variance associated to the position of the random walker, one obtains the equivalent diffusion coefficient to be considered for the asymptotic Wiener process toward which the random walk converges after a large number of steps:


 * $$D = \frac{\varepsilon^2}{6 \delta t}$$ (valid only in 3D).

Remark: the two expressions of the variance above correspond to the distribution associated to the vector $$\vec R$$ that links the two ends of the random walk, in 3D. The variance associated to each component $$R_x$$, $$R_y$$ or $$R_z$$ is only one third of this value (still in 3D).

For 2D:


 * $$D = \frac{\varepsilon^2}{4 \delta t}.$$

For 1D:


 * $$D = \frac{\varepsilon^2}{2 \delta t}.$$

Gaussian random walk
A random walk having a step size that varies according to a is used as a model for real-world time series data such as financial markets. The formula for modeling option prices, for example, uses a Gaussian random walk as an underlying assumption.

Here, the step size is the inverse cumulative normal distribution $$\Phi^{-1}(z,\mu,\sigma)$$ where 0 ≤ z ≤ 1 is a uniformly distributed random number, and μ and σ are the mean and standard deviations of the normal distribution, respectively.

If μ is nonzero, the random walk will vary about a linear trend. If vs is the starting value of the random walk, the expected value after n steps will be vs + nμ.

For the special case where μ is equal to zero, after n steps, the translation distance's probability distribution is given by N(0, nσ2), where N is the notation for the normal distribution, n is the number of steps, and σ is from the inverse cumulative normal distribution as given above.

Proof: The Gaussian random walk can be thought of as the sum of a sequence of independent and identically distributed random variables, Xi from the inverse cumulative normal distribution with mean equal zero and σ of the original inverse cumulative normal distribution:
 * Z = $$\sum_{i=0}^n {X_i}$$,

but we have the distribution for the sum of two independent normally distributed random variables, Z = X + Y, is given by
 * $$\mathcal{N}$$(μX + μY, σ2X + σ2Y).

In our case, μX = μY = 0 and σ2X = σ2Y = σ2 yield
 * $$\mathcal{N}$$(0, 2σ2)

By induction, for n steps we have
 * Z ~ $$\mathcal{N}$$(0, nσ2).

For steps distributed according to any distribution with zero mean and a finite variance (not necessarily just a normal distribution), the translation distance after n steps is
 * $$\sqrt{E|S_n^2|} = \sigma \sqrt{n}.$$

But for the Gaussian random walk, this is just the standard deviation of the translation distance's distribution after n steps. Hence, if μ is equal to zero, and since the root mean square(RMS) translation distance is one standard deviation, there is 68.27% probability that the RMS translation distance after n steps will fall between ± σ$$\sqrt{n}$$. Likewise, there is 50% probability that the translation distance after n steps will fall between ± 0.6745σ$$\sqrt{n}$$.

Anomalous diffusion
In disordered systems such as porous media and fractals $$\sigma^2$$ may not be proportional to $$t$$ but to $$t^{2 / d_w}$$. The exponent $$d_w$$ is called the exponent and can be larger or smaller than 2. may also be expressed as σr2 ~ Dtα where α is the anomaly parameter. Some diffusions in random environment  are even proportional to a power of the logarithm of the time,  see for example Sinai's walk or Brox diffusion.

Number of distinct sites
The number of distinct sites visited by a single random walker $$S(t)$$ has been studied extensively for square and cubic lattices and for fractals. This quantity is useful for the analysis of problems of trapping and kinetic reactions. It is also related to the vibrational density of states, diffusion reactions processes and spread of populations in ecology. The generalization of this problem to the number of distinct sites visited by $$N$$ random walkers, $$S_N(t)$$, has recently been studied for d-dimensional Euclidean lattices. The number of distinct sites visited by N walkers is not simply related to the number of distinct sites visited by each walker.

Information rate
The of a Gaussian random walk with respect to the squared error distance, i.e. its quadratic, is given parametrically by
 * $$ R(D_\theta) = \frac{1}{2} \int_0^1 \max\{0, \log_2\left(S(\varphi)/\theta \right) \} \, d\varphi, $$
 * $$ D_\theta = \int_0^1 \min\{S(\varphi),\theta\} \, d\varphi, $$

where $$S(\varphi) = \left(2 \sin (\pi \varphi/2) \right)^{-2}$$. Therefore, it is impossible to encode $${\{Z_n\}_{n=1}^N}$$ using a of less than $$NR(D_\theta)$$ s and recover it with expected mean squared error less than $$D_\theta$$. On the other hand, for any $$\varepsilon>0$$, there exists an $$N \in \mathbb N$$ large enough and a of no more than $$2^{N R(D_{\theta})}$$ distinct elements such that the expected mean sqaured error in recovering $${\{Z_n\}_{n=1}^N}$$ from this code is at most $$D_\theta - \varepsilon$$.

Applications
As mentioned the range of natural phenomena which have been subject to attempts at description by some flavour of random walks is considerable, in particular in physics and chemistry,, biology and various other fields. The following are some specific applications of random walk: In all these cases, random walk is often substituted for Brownian motion.
 * In, the "" is used to model shares prices and other factors. Empirical studies found some deviations from this theoretical model, especially in short term and long term correlations. See s.
 * In, random walk describes the statistical properties of
 * In, random walks are used as simplified models of physical and diffusion such as the   of  in liquids and gases. See for example . Also in physics, random walks and some of the self interacting walks play a role in.
 * In, random walks are used to describe individual animal movements, to empirically support processes of , and occasionally to model.
 * In, random walk describes an . It is the simplest model to study.
 * In other fields of mathematics, random walk is used to calculate solutions to, to estimate the , and for various constructions in and.
 * In, random walks are used to estimate the size of the . In the World Wide Web conference-2006, Bar-Yossef et al. published their findings and algorithms for the same.
 * In, random walks are used to determine the labels (i.e., "object" or "background") to associate with each pixel. This algorithm is typically referred to as the  segmentation algorithm.
 * In, random walks and reinforced random walks are used to model cascades of neuron firing in the brain.
 * In, ocular drift tends to behave like a random walk. According to some authors,  in general are also well described by a random walk.
 * In, random walks explain accurately the relation between the time needed to make a decision and the probability that a certain decision will be made.
 * Random walks can be used to sample from a state space which is unknown or very large, for example to pick a random page off the internet or, for research of working conditions, a random worker in a given country.
 * When this last approach is used in it is known as  or MCMC for short. Often, sampling from some complicated state space also allows one to get a probabilistic estimate of the space's size. The estimate of the  of a large  of zeros and ones was the first major problem tackled using this approach.


 * Random walks have also been used to massive online graphs such as s.
 * In, a random walk is used to model node movement.
 * engage in a.
 * Random walks are used to model.
 * In physics, random walks underlie the method of.
 * On the web, the Twitter website uses random walks to make suggestions of who to follow
 * and have proven that 7  are enough to mix a pack of cards (see more details under ). This result translates to a statement about random walk on the  which is what they prove, with a crucial use of the group structure via Fourier analysis.

Variants
A number of types of es have been considered that are similar to the pure random walks but where the simple structure is allowed to be more generalized. The pure structure can be characterized by the steps being defined by.

On graphs
A random walk of length k on a possibly infinite G with a root 0 is a stochastic process with random variables $$X_1,X_2,\dots,X_k$$ such that $$X_1=0$$ and $$ {X_{i+1}} $$ is a vertex chosen uniformly at random from the neighbors of $$X_i$$. Then the number $$p_{v,w,k}(G)$$ is the probability that a random walk of length k starting at v ends at w. In particular, if G is a graph with root 0, $$p_{0,0,2k}$$ is the probability that a $$2k$$-step random walk returns to 0.

Building on the analogy from the earlier section on higher dimensions, assume now that our city is no longer a perfect square grid. When our person reaches a certain junction, he picks between the variously available roads with equal probability. Thus, if the junction has seven exits the person will go to each one with probability one-seventh. This is a random walk on a graph. Will our person reach his home? It turns out that under rather mild conditions, the answer is still yes. For example, if the lengths of all the blocks are between a and b (where a and b are any two finite positive numbers), then the person will, almost surely, reach his home. Notice that we do not assume that the graph is, i.e. the city may contain tunnels and bridges. One way to prove this result is using the connection to. Take a map of the city and place a one  on every block. Now measure the "resistance between a point and infinity." In other words, choose some number R and take all the points in the electrical network with distance bigger than R from our point and wire them together. This is now a finite electrical network, and we may measure the resistance from our point to the wired points. Take R to infinity. The limit is called the resistance between a point and infinity. It turns out that the following is true (an elementary proof can be found in the book by Doyle and Snell):

Theorem: ''a graph is transient if and only if the resistance between a point and infinity is finite. It is not important which point is chosen if the graph is connected.''

In other words, in a transient system, one only needs to overcome a finite resistance to get to infinity from any point. In a recurrent system, the resistance from any point to infinity is infinite.

This characterization of is very useful, and specifically it allows us to analyze the case of a city drawn in the plane with the distances bounded.

A random walk on a graph is a very special case of a. Unlike a general Markov chain, random walk on a graph enjoys a property called time symmetry or reversibility. Roughly speaking, this property, also called the principle of, means that the probabilities to traverse a given path in one direction or the other have a very simple connection between them (if the graph is , they are just equal). This property has important consequences.

Starting in the 1980s, much research has gone into connecting properties of the graph to random walks. In addition to the electrical network connection described above, there are important connections to, see more , functional inequalities such as and  inequalities and properties of solutions of. A significant portion of this research was focused on s of. In many cases these discrete results carry over to, or are derived from s and s.

In the context of s, particularly that of the, analytical results to some properties of random walkers have been obtained. These include the distribution of first and last hitting times of the walker, where the first hitting time is given by the first time the walker steps into a previously visited site of the graph, and the last hitting time corresponds the first time the walker cannot perform an additional move without revisiting a previously visited site.

A good reference for random walk on graphs is the online book by Aldous and Fill. For groups see the book of Woess. If the transition kernel $$p(x,y)$$ is itself random (based on an environment $$\omega$$) then the random walk is called a "random walk in random environment". When the law of the random walk includes the randomness of $$\omega$$, the law is called the annealed law; on the other hand, if $$\omega$$ is seen as fixed, the law is called a quenched law. See the book of Hughes, the book of Revesz, or the lecture notes of Zeitouni.

We can think about choosing every possible edge with the same probability as maximizing uncertainty (entropy) locally. We could also do it globally – in maximal entropy random walk (MERW) we want all paths to be equally probable, or in other words: for every two vertexes, each path of given length is equally probable. This random walk has much stronger localization properties.

Self-interacting random walks
There are a number of interesting models of random paths in which each step depends on the past in a complicated manner. All are more complex for solving analytically than the usual random walk; still, the behavior of any model of a random walker is obtainable using computers. Examples include: The self-avoiding walk of length n on Z^d is the random n-step path which starts at the origin, makes transitions only between adjacent sites in Z^d, never revisit a site, and is chosen uniformly among all such paths. In two dimensions, due to self-trapping, a typical self-avoiding walk is very short, while in higher dimension it grows beyond all bounds. This model has often been used in (since the 1960s).
 * The.
 * The.
 * The.
 * The.
 * The.

Long-range correlated walks
Long-range correlated time series are found in many biological, climatological and economic systems.
 * Heartbeat records
 * Non-coding DNA sequences
 * Volatility time series of stocks
 * Temperature records around the globe

Maximal entropy random walk
Random walk chosen to maximize, has much stronger localization properties.

Correlated random walks
Random walks where the direction of movement at one time is with the direction of movement at the next time. It is used to model animal movements.