2 min read

Another view of the ‘nearly true’ model

Ok, so to recap, we have a large model (such as ‘we know the marginal sampling probabilities’) and a small model (such as the subset of the large model with logitP[Y=1]=xβ).  Under the large model, we would use the estimator β^L, but under the small model there is a more efficient estimator β^S. That is, under the small model
n(β^Sβ0)dN(0,σ2)
and
n(β^Lβ0)dN(0,σ2+ω2)

We’re worried that the small model might be slightly misspecified. One test of model misspecification is based on D=β^Sβ^L.  Under the small model, nDdN(0,τ2) for some τ2. This test isn’t a straw man – for example, DuMouchel and Duncan recommended it in the context of survey regression in a 1983 JASA paper.

If we assume that β^S is (locally, semiparametric) efficient in the small model then τ=ω.  Now suppose the small model is slightly untrue so that nDdN(Δ,ω2) with Δ>0. If, say, Δ=ω, then approximately
β^SN(ω,σ2)
and
β^LN(0,σ2+ω2)
so the two estimators have the same asymptotic mean squared error. Since β^L is asymptotically unbiased it would probably be preferred, but the test based on D has noncentrality parameter 1 and very poor power. If we relied on the test, we would probably end up choosing β^S

So the test based on D is not very useful if we want to protect against small amounts of model misspecification. We should use a better test. 

But sometimes the test based on D is the most powerful test or not far from it. Since we know what β^S and β^L look like as functionals of the distribution, we could try to maliciously arrange for the model misspecification to be in the direction that maximised β^Sβ^L, and D would then be the Neyman-Pearson most powerful test – that’s what UMP tests look like for Gaussian shift alternatives. We can’t quite do that, but in large enough sample sizes we can come as close as we need.