Student Puzzle Editor Anirban DasGupta poses another two problems, and says, “Our statistics problem this time is on universally consistent estimation of the regression coefficients for not necessarily Gaussian errors, and the probability problem is on probabilistic graph theory. The statistics problem is a practically important problem that will make you think, and the probability problem is a fun, unusual problem, but a simple one.” Send us your solution, to either or both, by September 15, 2023.

Puzzle 46.1.
Consider the usual linear model Yi=β0+β1xi,1++βpxi,p+ϵi,i=1,2,n,1p<. We assume that ϵiiidf(z), where f(z)=c(α)e|z|α,<z<,0<α<, and c(α) is the normalizing constant. Provide infinitely many explicit consistent estimators of the vector of regression coefficients, estimators that are consistent under all error densities f stated above.

Puzzle 46.2.
Suppose X,Y,Z are iid Poisson with mean λ>0. Let f(λ)=Pλ(X,Y,Zare the degrees of a nonempty graph on 3 vertices). Find supλ>0f(λ).

Solution to Puzzle 45

Thanks to Bilol Banerjee (ISI Kolkata) and Soham Bonnerjee (University of Chicago), who sent in answers, and a special well done to Bishakh Bhattacharya (ISI Kolkata) whose solution to 45.1 was “admirably detailed.” Anirban DasGupta explains:

Puzzle 45.1.
The key is to show separately that (almost surely) max{X12,X22,,Xn2}2logn and i=1nXi2n, where we say that for two positive sequences an,bn,anbn if anbn1. That i=1nXi2n is just the strong law. To show that max{X12,X22,,Xn2}2logn, let F(x)=2Φ(x)1 denote the CDF of |X1| and qn=F1(11n)=Φ1(112n)2logn. By using the formula for F(x), we show easily that for each c>1,n=1[1F(cqn)]<. This suffices to show that max{|X1|,|X2|,,|Xn|}F1(11n)2logn, and hence, max{X12,X22,,Xn2}2logn. Therefore, max{X12,X22,,Xn2}/i=1nXi22lognn. On the other hand, by the prime number theorem, π(n)nlogn. This finishes the proof of puzzle 45.1.

Puzzle 45.2.
Denoting the t-percentile tα/2,n1 by c, the t-confidence interval at level 1α is Cn=X¯±csn. Hence, μCnμ>X¯+csnμ<X¯csn. Therefore, by symmetry considerations, E[dist(μ,Cn)|μCn]=2×E[μ(X¯+csn)|μ>X¯+csn] \[ = 2\,\times \frac{2}{\alpha }\, \int \frac{s {\sqrt{n}}\,(t-c)\,I_{t > c}\,f_n(s,t)dsdt, \] where t denotes the t-statistic n(X¯μ)s, and fn(s,t) denotes the joint density of (s,t), and this can be calculated by the usual Jacobian method, starting with the joint density of (X¯,s). A (long) calculation then shows that sn(tc)It>cfn(s,t)dsdt =σn[ϕ(zα2)α2zα2]+O(n3/2), In order to get this above expression in terms of zα2, one will need to use the fact that for any given α,c=zα2+O(1n).