In the mathematical discipline of graph theory, the expander walk sampling theorem states that sampling vertices in an expander graph by doing a random walk is almost as good as sampling the vertices independently from a uniform distribution. The earliest version of this theorem is due to Ajtai, Komlós & Szemerédi (1987), and the more general version is typically attributed to Gillman (1998).
Let G = ( V , E ) be an expander graph with normalized second-largest eigenvalue λ . Let n denote the number of vertices in G . Let f : V → [ 0 , 1 ] be a function on the vertices of G . Let μ = E [ f ] denote the mean of f , i.e. μ = 1 n ∑ v ∈ V f ( v ) . Then, if we let Y 0 , Y 1 , … , Y k denote the vertices encountered in a k -step random walk on G starting at a random vertex Y 0 , we have the following for all γ > 0 :
Pr [ 1 k ∑ i = 0 k f ( Y i ) − μ > γ ] ≤ e − Ω ( γ 2 ( 1 − λ ) k ) . Here the Ω hides an absolute constant ≥ 1 / 10 . An identical bound holds in the other direction:
Pr [ 1 k ∑ i = 0 k f ( Y i ) − μ < − γ ] ≤ e − Ω ( γ 2 ( 1 − λ ) k ) . This theorem is useful in randomness reduction in the study of derandomization. Sampling from an expander walk is an example of a randomness-efficient sampler. Note that the number of bits used in sampling k independent samples from f is k log n , whereas if we sample from an infinite family of constant-degree expanders this costs only log n + O ( k ) . Such families exist and are efficiently constructible, e.g. the Ramanujan graphs of Lubotzky-Phillips-Sarnak.