Continuous-Time Markov Chains & Poisson Processes
Continuous-time Markov chains model systems that jump between states at random times — from radioactive decay to queuing systems to neural spike trains. The Poisson process is the canonical example, and its properties underlie survival analysis, event modeling, and the connection between discrete probabilistic models and differential equations.
Concepts
A hospital emergency room, a packet router, a radioactive nucleus — all are systems that jump between states at random times, with no memory of how long they have been waiting. The exponential distribution's memorylessness is not a modeling convenience but a mathematical necessity: it is the unique continuous distribution where . Continuous-time Markov chains are the precise framework for such systems, with the generator matrix encoding the rate of every possible transition.
Continuous-Time Markov Chains: Generator and Kolmogorov Equations
A continuous-time Markov chain (CTMC) on state space satisfies the Markov property in continuous time: for all , .
The chain is characterized by the generator matrix (or -matrix): for (transition rate from to ), and (so each row sums to zero). The holding time in state is exponential with rate .
The transition semigroup (matrix exponential) satisfies (Chapman-Kolmogorov). The probability vector evolves as:
Equivalently, .
The matrix exponential is forced by two requirements working together: the Chapman-Kolmogorov equation (from the Markov property) and differentiability at (from the exponential holding-time distribution). Any solution satisfying both is a matrix exponential; the generator is the unique finite object encoding instantaneous transition rates. Without , you would need to specify for every separately — with , one matrix determines all finite-time probabilities.
Stationary distribution: , i.e., for all . Detailed balance: .
Relationship to discrete-time chain: the embedded chain has transitions for . The CTMC visits states in the order of the embedded chain but spends exponential holding times at each.
The Poisson Process
The Poisson process with rate is the canonical CTMC with states and generator , (only upward jumps).
Definition via inter-arrivals: inter-arrival times are iid ; .
Marginal distribution: , i.e., .
Independent increments: for , and is independent of .
Superposition: the merge of independent Poisson processes with rates is Poisson with rate .
Thinning: independently keep each arrival with probability ; the result is Poisson with rate .
Conditioning: given , the arrival times are distributed as the order statistics of iid random variables.
Birth-Death Chains
A birth-death chain has states with transitions only between adjacent states: birth rate from state to , death rate from to , with .
The detailed balance equations reduce to a product formula for the stationary distribution:
M/M/1 queue: arrivals at rate , service at rate , . Stationary distribution: for . Mean queue length: ; mean wait time (Little's law): .
Renewal Theory
A renewal process generalizes Poisson by allowing non-exponential inter-arrival times with mean .
Renewal theorem: as .
Inspection paradox: the renewal interval containing a fixed time has length distribution stochastically greater than . This is why buses always seem late — you are more likely to arrive during a long gap.
Worked Example
Example 1: Waiting Times for Rare Events
In a network, packet errors occur as a Poisson process with rate errors/second.
Mean time to first error: seconds.
Probability of error-free period of length s: .
Expected errors in 1000 seconds: , with (Poisson: mean = variance).
Overdispersion check: if observed variance mean, the Poisson assumption fails (e.g., bursty errors form a clustered process). Variance-to-mean ratio is the Poisson diagnostic.
Example 2: M/M/1 Queue Near Saturation
A web server handles requests: /sec, /sec, .
Mean queue length: requests. Mean response time: seconds.
At : , second — 10× worse for a 10% load increase. Near , the system enters congestion collapse. Real CDN capacity planning provisions to maintain sub-100ms latency.
Example 3: Gillespie Algorithm for Chemical Kinetics
A system has two species and with reactions and . At state , the total rate is .
The Gillespie algorithm: time to next event ; event type with probability . The stationary distribution is binomial: . This is exact stochastic kinetics with no mean-field approximation.
Connections
Where Your Intuition Breaks
The Poisson process is memoryless in the sense that the waiting time for the next arrival is independent of how long you have been waiting. But this does not mean the process is memoryless about the total count: grows steadily with time, and the process is not stationary. More subtly, the inspection paradox undermines the intuition that "a random time lands in an interval of average length": if you sample a uniformly random time in , you are more likely to land inside a long inter-arrival interval than a short one. The average length of the interval containing your random time is — strictly greater than unless all intervals are equal. The memorylessness of the exponential eliminates the inspection paradox for the Poisson process specifically, but it returns for any renewal process with non-exponential inter-arrivals.
The generator is the continuous-time analog of . The Kolmogorov forward equation is the probability analog of a first-order linear ODE. When has eigenvalues with , the transient terms decay as , and the system relaxes to equilibrium at rate — the spectral gap of the generator, analogous to the discrete case. This makes eigendecomposition of the primary tool for analyzing relaxation times.
Poisson processes are maximally unstructured arrival streams. The Poisson process is the unique renewal process with both independent and stationary increments simultaneously. It is the maximum entropy point process for a given mean rate — knowing only the mean intensity fully specifies the distribution. This is why Poisson is the default model for counts of independent events: it uses the least information beyond the mean. Departures from Poisson (overdispersion, clustering, periodic structure) are meaningful signals about structure in the data-generating process.
Little's law is exact but requires stationarity. (average number in system = arrival rate × average time in system) holds for any stable queueing system under very general assumptions — not just M/M/1. But it requires the system to be in steady state. During transients (e.g., a server starting up or experiencing a traffic spike), can fail dramatically. In production observability, computing latency from throughput and queue depth via Little's law is valid only if the system is stationary; measurement during traffic spikes gives misleading estimates.
Enjoying these notes?
Get new lessons delivered to your inbox. No spam.