Density estimation; parametric estimation; robust estimation
Résumé :
[en] We consider the problem of estimating the joint distribution of n independent random variables. Given a loss function and a family of candidate probabilities, that we shall call a model, we aim at designing an estimator with values in our model that possesses good estimation properties not only when the distribution of the data belongs to the model but also when it lies close enough to it. The losses we have in mind are the total variation, Hellinger, Wasserstein and L_p-distances to name a few. We show that the risk of our estimator can be bounded by the sum of an approximation term that accounts for the loss between the true distribution and the model and a complexity term that corresponds to the bound we would get if this distribution did belong to the model. Our results hold under mild assumptions on the true distribution of the data and are based on exponential deviation inequalities that are non-asymptotic and involve explicit constants. Interestingly, when the model reduces to two distinct probabilities, our procedure results in a robust test whose errors of first and second kinds only depend on the losses between the true distribution and the two tested probabilities.
Disciplines :
Mathématiques
Auteur, co-auteur :
BARAUD, Yannick ; University of Luxembourg > Faculty of Science, Technology and Communication (FSTC) > Mathematics Research Unit
Co-auteurs externes :
no
Langue du document :
Anglais
Titre :
Tests and estimation strategies associated to some loss functions
Date de publication/diffusion :
2021
Titre du périodique :
Probability Theory and Related Fields
ISSN :
0178-8051
eISSN :
1432-2064
Maison d'édition :
Springer, Allemagne
Volume/Tome :
180
Fascicule/Saison :
3
Pagination :
799-846
Peer reviewed :
Peer reviewed vérifié par ORBi
Projet européen :
H2020 - 811017 - SanDAL - ERA Chair in Mathematical Statistics and Data Science for the University of Luxembourg