The Central Binomial Coefficient and Laplace's Method
The central binomial coefficient is the sequence down the "center" of Pascal's triangle, namely
(n2n).
These numbers show up, for example, in a random walk on the integers: if we start off at 0 and make 2n independent increments of +1 or −1, giving each choice even odds, then we will find ourselves back at the origin with probability
2−2n(n2n).
Some basic Markov process reasoning tells us that a random walk like this one can either be recurrent or dissipative: trajectories will either visit each position infinitely many times, or escape to infinity, spending only a finite number of steps at each position. Now, let {X0,X1,…} be our trajectory and C be the number of times that Xn=0. Since
E(C)=n=0∑∞P(Xn=0)=n=0∑∞P(X2n=0)=n=1∑∞2−2n(n2n),
the question of whether the random walk we have described is recurrent comes down to whether the central binomial coefficient grows quickly enough to make this series diverge. Does it? We can also consider the random walk on Zd, moving at each step by an increment in the set {−1,1}d, in which case the expected number of visits to the origin is
E(C)=n=0∑∞2−2nd(n2n)d.
When are these processes recurrent? To answer either question, we need to figure out something about the asymptotic growth of central binomial coefficient.
A basic "statistical intuition" tells us that (n2n) may on the order of 22n/n. To see why, choose 0<ϵ<1 and put
F(n,ϵ)=P(∣Xn∣<ϵ)
so that
F(2n,ϵ)=P(X2n=0)=2−2n(n2n).
Meanwhile, the central limit theorem tells us that F(n,nϵ) converges to P(Z≤ϵ) for large n, where Z is normally distributed with the variance of an increment—in our case, V(Z)=1. Since the density of Z at 0 is 1/2π, we may expect that
F(n,ϵ)≈P(Z≤nϵ)∼2πnϵ.
Unfortunately, it seems quite difficult to get anything rigorous from this idea. In fact, the closer I look at the approximation above, the more I despair: not only does the expression ϵ/2πn depend on ϵ while F(n,ϵ) does not for ϵ<1, but in fact F(n,ϵ)cannot be asymptotic to C/n for any constant C for the simple reason that F(n,ϵ)=0 for odd n.
Although the central limit theorem is useless to make rigorous conclusions here—as far as I can tell—it "miraculously" gives us the right order of growth. In fact, it is true that
(n2n)∼πn22n.
Plugging this approximation into the sums above gives the classic result that the random walk on Zd is recurrent for d≤2 but dissipative for all larger d. Moreover, an asymptotic series can be computed for the relative error, giving (for example) the remarkably accurate approximation
(n2n)=πn22n(1−81n−1+1281n−2+10245n−3+o(n−3))
which gives a maximum relative error of around 10−8. (You may play the "guess the sequence" game with those coefficients at your own risk.)
These estimates can be derived from Stirling's approximation for n!, but in this post I want to focus on a simple idea from which various asymptotic series—including the Stirling series—can be derived independently.
Our asymptotic expansion for the central binomial coefficient will rely on approximating a certain exponential integral. As a warmup, let's consider the problem of approximating a Laplace transform-type integral
I(x)=∫0∞f(t)e−xtdt
for large x. Assuming that f(t) does not grow too quickly for large t, our integral will vanish exponentially over the interval [δ,+∞), so the behavior of I(x) comes down to the behavior of f near 0. If f is also bounded near 0, then I(x) tends to 0 overall since
∫0∞e−xtdx=x1.
Cancelling out this term, xI(x) can be understood as an expectation of f over an exponential distribution with mean 1/x. Indeed, where Y is exponentially distributed with density e−t,
xI(x)=∫0∞f(t)xe−xtdt=∫0∞f(x−1t)e−tdt=E(f(x−1Y)).
When f(t)=tk, this expectation is easy to compute:
E(x−kYk)=x−kE(Yk)=x−kk!.
If f admits a Taylor series to order n with remainder rn(t)=o(tn),
I(x)=x1(k=0∑nE(k!f(k)(0)x−kYk)+E(rn(x−1Y)))=k=0∑nxk+1f(k)(0)+x1E(rn(x−1Y)).
All that remains is to bound the expectation of the remainder. We know that ∣rn(t)∣<ϵ∣t∣n for sufficiently small t<δ, so let's consider the events Y<xδ and Y≥xδ separately—call then S and L. We will decompose the expectation into integrals over these two sets:
E(rn(x−1Y))=ES(rn(x−1Y))+EL(rn(x−1Y)).
For the expectation over L, we only need a bound like ∣rn(t)∣≤CtK for some C and K, since then
∣EL(rn(x−1Y))∣≤∫δx∞∣rn(x−1t)∣e−tdt≤∫δx∞Cx−KtKe−tdt
is guaranteed to vanish exponentially for large x, as we had mentioned above. (We call this kind of error beyond all orders, since it is dominated by x−n for all n.) Meanwhile, to bound the expectation over S we can use the bound on rn:
∣ES(rn(x−1Y))∣≤ϵES(x−nYn)≤ϵxnE(Yn).
We conclude overall that
E(rn(x−1Y))=o(x−n),
and so the whole approximation is
I(x)=k=0∑nxk+1f(k)(0)+o(x−n−1).
The reasoning above can naturally be generalized to any variable Y with quickly decaying tails.
Proposition: Suppose f(t) is of polynomial order for large t and verifiesf(t)=k=0∑naktk+o(tn)for small t. Then, for any variable Y with exponentially bounded tails in the sense that P(∣Y∣≥M)≤Ce−δM for some C>0, δ>0, we also haveE(f(tY))=k=0∑nE(Yk)aktk+o(tn).
This idea comes in comes in handy in the general task of approximating an integral like
I(x)=∫Df(t)e−xϕ(t)dt
for large x. Again, if f does not grow quickly for large t, this integral is determined (up to an exponentially vanishing error) by the integrand near the minimizers of ϕ. For an isolated minimum at t0 with ϕ(t)=(t−t0)k+O((t−t0)k+1), changing variables to ϕ(t)=(s−t0)k in a neighborhood of t0 gives us an integral we can evaluate with the proposition above. This technique is known as Laplace's method. We'll see an example of this now.
Let's return to our random walk on the integers. The characteristic function of our increment is
2e−it+eit=cos(t),
and so cosn(t) will be the characteristic function of Xn. Since Xn is supported on the integers, the characteristic function is a Fourier polynomial and we can extract probability mass functions easily; for example,
P(Xn=0)=2π1∫−ππcosn(θ)dθ.
This integral vanishes exponentially except near the points θ=0 and θ=±π. For odd n these two peaks cancel out—indeed, P(Xn=0)=0 for odd n—and for even n they are equal. We conclude that
P(X2n=0)≈π1∫−ϵϵcos2n(θ)dθ
for ϵ<π, ignoring an error beyond all orders of n.
Since lncos(θ)≈−θ2/2, we can reparameterize by a variable s solving the equation
lncos(θ)=−2s2.
Near θ=s=0, the solutions to this equation are a union of two smooth curves. We can choose one branch by defining
s=θθ2−2lncos(θ)=θ1+O(θ2).
We can now change variables from θ to s and write our integral as a certain expectation over a Gaussian variable:
P(X2n=0)≈π1∫−ϵϵexp(2nlncos(θ))dθ=π1∫−ϵϵdsdθexp(−ns2)ds.=π2n1∫−2nϵ2nϵdsdθ(2ns)exp(−s2/2)ds=πn1E(dsdθ((2n)−1/2Z)).
(In the last expectation, we're defining dθ/ds(s)=0 for ∣s∣≥ϵ as a convenience of notation.) Our leading order approximation for P(X2n=0) is thus
P(X2n=0)=πn1+o(n−3/2),
and in particular we have an asymptotic equivalence
P(X2n=0)∼πn1.
In terms of the central binomial coefficient, this proves
(n2n)=22nP(X2n=0)∼πn22n,
as promised. The full asymptotic expansion will be
(n2n)≈πn22nE(dsdθ((2n)−1/2Z))=πn22n(k=0∑Ndsk+1dθk+1(0)k!(2n)k/2E(Zk)+o(n−N/2)).
Since the odd moments of Z vanish and the even moments are E(Zk)=(k−1)!! for k>0, we can rewrite this as
πn22n(1+k=1∑Nds2k+1dθ2k+1(0)(2k)!!2k1n−k+o(n−N)).
All that remains is to compute the series expansion of θ in terms of s. As far as I know, there is no simple closed-form for this series. Fortunately, if we happen to want coefficients, these computations can be easily done by a computer algebra system. Expanding
s=θθ2−2lncos(θ)=θ+12θ3+1603θ5+40320209θ7+O(θ9)
and inverting the series gives
θ=s−12s3+480s5+2688s7+O(s9)=s−213!s3+415!s5+8157!s7+O(s9).
Altogether, we have
(n2n)=πn22n(1−212!!×21n−1+414!!×41n−2+1586!!×81n−3+o(n−3))=πn22n(1−81n−1+1281n−2+10245n−3+o(n−3)),
as claimed above.