Ok, so to recap, we have a large model (such as ‘we know the marginal sampling probabilities’) and a small model (such as the subset of the large model with ). Under the large model, we would use the estimator , but under the small model there is a more efficient estimator . That is, under the small model
and
We’re worried that the small model might be slightly misspecified. One test of model misspecification is based on . Under the small model, for some . This test isn’t a straw man – for example, DuMouchel and Duncan recommended it in the context of survey regression in a 1983 JASA paper.
If we assume that is (locally, semiparametric) efficient in the small model then . Now suppose the small model is slightly untrue so that with . If, say, , then approximately
and
so the two estimators have the same asymptotic mean squared error. Since is asymptotically unbiased it would probably be preferred, but the test based on has noncentrality parameter 1 and very poor power. If we relied on the test, we would probably end up choosing
So the test based on is not very useful if we want to protect against small amounts of model misspecification. We should use a better test.
But sometimes the test based on is the most powerful test or not far from it. Since we know what and look like as functionals of the distribution, we could try to maliciously arrange for the model misspecification to be in the direction that maximised , and would then be the Neyman-Pearson most powerful test – that’s what UMP tests look like for Gaussian shift alternatives. We can’t quite do that, but in large enough sample sizes we can come as close as we need.