This post documents some facts about bootstrapping that I encountered during my PhD. In particular, this post highlights some asymptotic results relating to
The mean absolute deviation of normal variables
The quantile of absolute deviations from the mean
The distribution of standard deviations
Consider the setting where we sample T observations and create B samples of size T by sampling with replacement. In other words, we have:
The original sample S:={Xi}i∈[T]
μˉ is the sample mean statistic computed over S
The samples from S with replacement: Sb:={Xi(b)}i=1T,b∈[B]:={1,2,...,B}
μˉ(b) is the statistic computed over the T observations associated with batch b∈[B]
The central limit theorem states that if {Xi}i∈[T] are T samples drawn independently and identically distributed from a distribution with mean μ and variance σ2, then Z=limT→∞T(T1∑t∈[T]Xt−μ)/σ follows a standard normal distribution. The sample estimate of the mean for batch b∈[B] is given by μˉ(b):=T1∑t∈[T]Xt(b), therefore, for large T:
T1t∈[T]∑Xt(b)→dN(μ,Tσ)
In our setting, the distribution is the empirical distribution defined by the sample S, implying that μ and σ are given by
μ=ES[X]=T1x∈S∑x
and
σ=ES[(X−μ)2]=T1x∈S∑(x−μ)2
Therefore μˉ(b)=T1∑t∈[T]Xt(b) can be thought of as a sample from N(μ,Tσ)
where β∈R+N and αi∈R+∀i∈[N] denote the penalty parameters. Ho, M., Sun, Z., and Xin, J. show that αi and βi correspond to the sizes of box-uncertainty sets used in the robust counterpart of the mean-variance optimization problem. Specifically, αi and βi correspond to uncertainty in the diagonal of the estimated covariance matrix Σii and expected return μi respectively for asset i=1,2,...,N. Ho, M., Sun, Z., and Xin, J. use a bootstrapping approach to select the values of βi and αi; they proceed as follows: Let {r(t)∈RN}t∈[T] denote T observations of the random variable r used to estimate Σ and μ. The estimation errors are computed by re-sampling T observations with replacement from the original set of observations B times. Each re-sampling yields estimates of the mean and covariance μˉ(b),Σˉ(b)b∈[B]. Let p1 and p2 denote the investor's aversion to estimation risk of the mean and variance, respectively. The values for βi and αi are defined as the quantiles of the bootstrapped deviations
One can express the quantile of absolute deviation of a normal variable in a closed-form equation. The results can be derived as follows: letting Z∼N(0,1) then the quantile of the absolute deviation satisfies P(∣Z∣≤x)=p which is the same as 2Φ(x)−1=p which implies
The case of determining an approximation for the distribution of Σ is more complex. For simplicity, we consider the case where Σ is a scalar denoted by σ. An approach referred to as the delta method can be used to obtain the asymptotic distribution of Σ=T1∑i∈[T]Xi2−(T1∑i∈[T]Xi)2.
Let W(2) and W(1) denote T1∑i∈[T]Xi2 and T1∑tXt respectively. W(1) and W(2) converge to a joint normal distribution by the multivariate central limit theorem with covariance ΞW(j),W(k)=E[W(j)W(k)]−E[W(j)]E[W(k)] and expected value μW i.e T(W−μW)→dN(0,Ξ).
In general, for any differentiable g(W) one can write the following first-order Taylor expansion g(W)≈g(μW)+∇g(μW)(W−μW). One can approximate the covariance of g(W) by V[g(W)]≈∇g(W)⊺Ξ∇g(W), and as such T(g(W)−g(μW))→dN(0,∇g(W)⊺Ξ∇g(W)).
One can set g(W1,W2)=W(2)−(W(1))2 and use the first-order delta method as described above to obtain asymptotic convergence condition. The first-order delta method does not always work. Convergence to normality does not hold for Bernoulli variables with p=0.5 because the variance of g ends up being zero.
To circumvent this, the second-order Taylor approximation must be used. In the case of Bernoulli random variables g can be expressed as a function of a single random parameter:
p^:=T1t∑Xt,
since T1∑tXt=T1∑tXt2. In this case,
g(p^)=T1t∑Xt2−(T1t∑Xt)2=p^(1−p^).
If the mean is p=1/2, then taking the second-order taylor expansion around the mean implies p^(1−p^)=1/4+0∗(p^−0.5)+1/2(−2)(p^−1/2)2. It is then true that p^(1−p^) is asymptotically negative χ2 and is not normal because p^=T1∑tXt is asymptotically normal by the central limit theorem and p^(1−p^) is a constant minus an asymptotically normal variable squared (as shown in the above Taylor series. In the case that the first-order delta method is applicable and the first-order Taylor series introduces uncertainty in g, it follows: