Excess risk bounds in robust empirical risk minimization
نویسندگان
چکیده
Abstract This paper investigates robust versions of the general empirical risk minimization algorithm, one core techniques underlying modern statistical methods. Success is based on fact that for a ‘well-behaved’ stochastic process $\left \{ f(X), \ f\in \mathscr F\right \}$ indexed by class functions $f\in F$, averages $\frac{1}{N}\sum _{j=1}^N f(X_j)$ evaluated over sample $X_1,\ldots ,X_N$ i.i.d. copies $X$ provide good approximation to expectations $\mathbb E f(X)$, uniformly large classes F$. However, this might no longer be true if marginal distributions are heavy tailed or contains outliers. We propose version idea replacing proxies and obtain high-confidence bounds excess resulting estimators. In particular, we show estimators can converge $0$ at fast rates with respect size $N$, referring faster than $N^{-1/2}$. discuss implications main results linear logistic regression problems evaluate numerical performance proposed methods simulated real data.
منابع مشابه
Chaining Bounds for Empirical Risk Minimization
This paper extends the standard chaining technique to prove excess risk upper bounds for empirical risk minimization with random design settings even if the magnitude of the noise and the estimates is unbounded. The bound applies to many loss functions besides the squared loss, and scales only with the sub-Gaussian or subexponential parameters without further statistical assumptions such as the...
متن کاملEmpirical Risk Minimization: Abstract Risk Bounds and Rademacher Averages
1. An abstract framework for ERM To study ERM in a general framework, we will adopt a simplified notation often used in the literature. We have a space Z and a class F of functions f : Z→ [0, 1]. Let P(Z) denote the space of all probability distributions on Z. For each sample size n, the training data are in the form of an n-tuple Zn = (Z1, . . . , Zn) of Z-valued random variables drawn accordi...
متن کاملAsymptotics in Empirical Risk Minimization
In this paper, we study a two-category classification problem. We indicate the categories by labels Y = 1 and Y = −1. We observe a covariate, or feature, X ∈ X ⊂ R. Consider a collection {ha} of classifiers indexed by a finite-dimensional parameter a, and the classifier ha∗ that minimizes the prediction error over this class. The parameter a∗ is estimated by the empirical risk minimizer ân over...
متن کاملmixup: BEYOND EMPIRICAL RISK MINIMIZATION
Large deep neural networks are powerful, but exhibit undesirable behaviors such as memorization and sensitivity to adversarial examples. In this work, we propose mixup, a simple learning principle to alleviate these issues. In essence, mixup trains a neural network on convex combinations of pairs of examples and their labels. By doing so, mixup regularizes the neural network to favor simple lin...
متن کاملAggregation via Empirical Risk Minimization
Given a finite set F of estimators, the problem of aggregation is to construct a new estimator whose risk is as close as possible to the risk of the best estimator in F . It was conjectured that empirical minimization performed in the convex hull of F is an optimal aggregation method, but we show that this conjecture is false. Despite that, we prove that empirical minimization in the convex hul...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Information and Inference: A Journal of the IMA
سال: 2021
ISSN: ['2049-8772', '2049-8764']
DOI: https://doi.org/10.1093/imaiai/iaab004