Analysis of the oracle risk in multi-task ridge regression

نویسندگان

  • Matthieu Solnon
  • Theodoros Evgeniou
  • Charles A. Micchelli
  • Sylvain Arlot
چکیده

In this talk1 we will present new results obtained in the multiple kernel ridge regression framework, which we refer to as multi-task regression. Multi-task techniques come into play when experimental limitations makes it impossible to increase the sample size n, which is the classical way to improve the performance of the estimators. However, it is often possible to have access to other closely related regression problems. Then, taking into account the similarity between those p different tasks and treating them in a multi-dimensional setting may lead to some improvement of the procedure, hoping that the heuristic “p tasks with each n observations is equivalent to an n× p sample” is valid. In this setting we use a generalization of the ridge regression procedure proposed by Evgeniou et al. [2005], in which a penalization term containing a matricial parameter expresses the similarity between the tasks. A fully data-driven selection of this matrix was proposed by Solnon et al. [2012], resulting in an oracle inequality, which shows that the selected estimator attains a risk that is close to the best possible risk, the oracle risk. We will show how we obtain an estimation of this multi-task oracle risk. Then, we find several settings in which the single-task oracle risk can also be expressed and compare both risks. This allows us to discriminate several settings, depending on whether the multi-task oracle outperforms the single-task oracle, the opposite, or whether both behave similarly. We also show that our estimation of the multi-task oracle risk is precise enough to be plugged in the latter oracle inequality, thus showing that the estimator of Solnon et al. [2012] has a lower risk than the single-task oracle risk in favorable settings. We finally show simulation results in situations where the oracle risk can no longer be explicitly computed and show that, in those cases, the oracle retains the same virtues and disadvantages as before.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Comparison bewteen multi-task and single-task oracle risks in kernel ridge regression

In this paper we study multi-task kernel ridge regression and try to understand when the multi-task procedure performs better than the single-task one, in terms of averaged quadratic risk. In order to do so, we compare the risks of the estimators with perfect calibration, the oracle risk. We are able to give explicit settings, favorable to the multi-task procedure, where the multi-task oracle p...

متن کامل

Multi-task Regression using Minimal Penalties Multi-task Regression using Minimal Penalties

In this paper we study the kernel multiple ridge regression framework, which we refer to as multi-task regression, using penalization techniques. The theoretical analysis of this problem shows that the key element appearing for an optimal calibration is the covariance matrix of the noise between the different tasks. We present a new algorithm to estimate this covariance matrix, based on the con...

متن کامل

Generalized Ridge Regression Estimator in Semiparametric Regression Models

In the context of ridge regression, the estimation of ridge (shrinkage) parameter plays an important role in analyzing data. Many efforts have been put to develop skills and methods of computing shrinkage estimators for different full-parametric ridge regression approaches, using eigenvalues. However, the estimation of shrinkage parameter is neglected for semiparametric regression models. The m...

متن کامل

Multi-task regression using minimal penalties

In this paper we study the kernel multiple ridge regression framework, which we refer to as multi-task regression, using penalization techniques. The theoretical analysis of this problem shows that the key element appearing for an optimal calibration is the covariance matrix of the noise between the different tasks. We present a new algorithm to estimate this covariance matrix, based on the con...

متن کامل

Diagnostic Measures in Ridge Regression Model with AR(1) Errors under the Stochastic Linear Restrictions

Outliers and influential observations have important effects on the regression analysis. The goal of this paper is to extend the mean-shift model for detecting outliers in case of ridge regression model in the presence of stochastic linear restrictions when the error terms follow by an autoregressive AR(1) process. Furthermore, extensions of measures for diagnosing influential observations are ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2013