Vajda, I. Minimum divergence principle in statistical estimation. (English) Zbl 0558.62004 Recent results in estimation theory and related topics, Suppl. Issues Stat. Decis. 1, 239-261 (1984). [For the entire collection see Zbl 0542.00015.] Statistical estimators (D-estimators) minimizing f-divergence between theoretical probability \(P_{\theta}\in {\mathcal P}_{\Theta}\) and empirical probability \(P_ n\) on a measurable space (X,\({\mathcal B})\) are introduced. Suitable specifications of a convex function f yield new promising estimators or well-known so far quite diversely motivated ones. Examples are the MLE and other M-estimators, the minimum distance estimators considered in the literature as well as certain ”smooth approximations” to A17-A25 emerging from the Princeton robustness study. The class of D-estimators is homogeneous enough to be studied on the whole. In this sense, the minimum divergence approach is a possible alternative to the loss function approach in a systematic development of statistical estimation theory. The present paper is restricted to non- asymptotic aspects of the theory with emphasis on motivation, examples, existence and measurability of D-estimators. In particular, it is shown that under quite general regularity conditions the D-estimates are continuous functions of sample vectors. Cited in 1 ReviewCited in 3 Documents MSC: 62A01 Foundations and philosophical topics in statistics 62F10 Point estimation Keywords:f-divergence; theoretical probability; empirical probability; M- estimators; minimum distance estimators; minimum divergence approach; non-asymptotic aspects; examples; existence and measurability of D- estimators Citations:Zbl 0542.00015 PDFBibTeX XML