## Inference in components of variance models with low replication.(English)Zbl 1039.62065

Summary: In components of variance models the data are viewed as arising through a sum of two random variables, representing between- and within-group variation, respectively. The former is generally interpreted as a group effect, and the latter as error. It is assumed that these variables are stochastically independent and that the distributions of the group effect and the error do not vary from one instance to another. If each group effect can be replicated a large number of times, then standard methods can be used to estimate the distributions of both the group effect and the error. This cannot be achieved without replication, however. How feasible is distribution estimation if it is not possible to replicate prolifically?
Can the distributions of random effects and errors be estimated consistently from a small number of replications of each of a large number of noisy group effects, for example, in a nonparametric setting? Often extensive replication is practically infeasible, in particular, if inherently small numbers of individuals exhibit any given group effect. Yet it is quite unclear how to conduct inference in this case.
We show that inference is possible, even if the number of replications is as small as 2. Two methods are proposed, both based on Fourier inversion. One, which is substantially more computer intensive than the other, exhibits better performance in numerical experiments.

### MSC:

 62J10 Analysis of variance and covariance (ANOVA) 62E17 Approximations to statistical distributions (nonasymptotic) 62G07 Density estimation 62G08 Nonparametric regression and quantile regression
Full Text:

### References:

 [1] AIRY, G. B. (1861). On the Algebraical and Numerical Theory of Errors of Observations and the Combination of Observations. Macmillan, London. [2] BERAN, R., FEUERVERGER, A. and HALL, P. (1996). On nonparametric estimation of intercept and slope distributions in random coefficient regression. Ann. Statist. 24 2569-2592. · Zbl 0867.62021 [3] BICKEL, P. and RITOV, Y. (1987). Efficient estimation in the errors in variables model. Ann. Statist. 15 513-540. · Zbl 0643.62029 [4] BOWMAN, A. W. (1984). An alternative method of cross-validation for the smoothing of density estimates. Biometrika 71 353-360. JSTOR: [5] BOWMAN, A. W., HALL, P. and PRVAN, T. (1998). Bandwidth selection for the smoothing of distribution functions. Biometrika 84 799-808. JSTOR: · Zbl 0921.62042 [6] CARROLL, R. J. and HALL, P. (1988). Optimal rates of convergence for deconvolving a density. J. Amer. Statist. Assoc. 83 1184-1186. JSTOR: · Zbl 0673.62033 [7] CORNFIELD, J. and TUKEY, J. W. (1956). Average values of mean squares in factorials. Ann. Math. Statist. 27 907-949. · Zbl 0075.29404 [8] COX, D. R. and HALL, P. (2002). Estimation in a simple random effects model with nonnormal distributions. Biometrika 89 831-840. JSTOR: · Zbl 1036.62055 [9] DANIELS, H. E. (1939). The estimation of components of variance. J. Roy. Statist. Soc. Suppl. 6 186-197. · JFM 66.1304.05 [10] EISENHART, C. (1947). The assumptions underlying the analysis of variance. Biometrics 3 1-21. JSTOR: [11] FAN, J. (1991). On the optimal rates of convergence for nonparametric deconvolution problems. Ann. Statist. 19 1257-1272. · Zbl 0729.62033 [12] HECKMAN, M. (1960). Flame photometric determination of calcium in animal feeds. J. Assoc. Official Analy tical Chemists 43 337-340. [13] JOHNSON, N., KOTZ, S. and BALAKRISHNAN, N. (1994). Continuous Univariate Distributions 1, 2nd ed. Wiley, New York. · Zbl 0811.62001 [14] KEMPTHORNE, O. (1975). Fixed and mixed models in the analysis of variance. Biometrics 31 473- 486. JSTOR: · Zbl 0306.62020 [15] KHURI, A. I. and SAHAI, H. (1985). Variance components analysis: A selective literature survey. Internat. Statist. Rev. 53 279-300. JSTOR: · Zbl 0586.62110 [16] MORAN, P. A. P. (1971). Estimating structural and functional relationships. J. Multivariate Anal. 1 232-255. · Zbl 0219.62011 [17] NELDER, J. A. (1977). A reformulation of linear models (with discussion). J. Roy. Statist. Soc. Ser. A 140 48-76. JSTOR: [18] NEy MAN, J. (1951). Existence of consistent estimates of the directional parameter in a linear structural relation between two variables. Ann. Math. Statist. 22 497-512. · Zbl 0043.34902 [19] PLACKETT, R. L. (1960). Models in the analysis of variance (with discussion). J. Roy. Statist. Soc. Ser. B 22 195-217. JSTOR: · Zbl 0109.37802 [20] REIERSØL, O. (1950). Identifiability of a linear relation between variables which are subject to error. Econometrica 18 375-389. · Zbl 0040.22502 [21] RUDEMO, M. (1982). Empirical choice of histograms and kernel density estimators. Scand. J. Statist. 9 65-78. · Zbl 0501.62028 [22] SAHAI, H., KHURI, A. I. and KAPADIA, C. H. (1985). A second bibliography on variance components. Comm. Statist. Theory Methods 14 63-115. · Zbl 0589.62055 [23] SARDA, P. (1993). Smoothing parameter selection for smooth distribution functions. J. Statist. Plann. Inference 35 65-75. · Zbl 0769.62030 [24] SEARLE, S. R., CASELLA, G. and MCCULLOCH, C. E. (1992). Variance Components. Wiley, New York. · Zbl 0850.62007 [25] SPIEGELMAN, C. (1979). On estimating the slope of a straight line when both variables are subject to error. Ann. Statist. 7 201-206. · Zbl 0412.62048 [26] TIPPETT, L. H. C. (1931). The Methods of Statistics. Williams and Norgate, London. · Zbl 0047.37803 [27] WOLFOWITZ, J. (1952). Consistent estimators of the parameters of a linear structural relation. Skand. Aktuarietidskr 35 132-151. · Zbl 0048.36903 [28] YATES, F. (1966). A fresh look at the basic principles of the design and analysis of experiments. Proc. Fifth Berkeley Sy mp. Math. Statist. Probab. 4 777-790. Univ. California Press, Berkeley.
This reference list is based on information provided by the publisher or from digital mathematics libraries. Its items are heuristically matched to zbMATH identifiers and may contain data conversion errors. It attempts to reflect the references listed in the original paper as accurately as possible without claiming the completeness or perfect precision of the matching.