Permutation Complexity Bound on Out-Sample Error

نویسنده

  • Malik Magdon-Ismail
چکیده

Assume a standard setting with data D = {(xi, yi)}i=1, where (xi, yi) are sampled iid from the joint distribution p(x, y) on Rd×{±1}. Let H = {h : R 7→ {±1}} be a learning model which produces a hypothesis g ∈ H when given D (we use g for the hypothesis returned by the learning algorithm and h for a generic hypothesis in H). We assume the 0-1 loss, so the in-sample error is ein(h) = 1 2n ∑n i=1(1 − yih(xi)). The out-sample error eout(h) = 1 2 E [(1 − yh(x))]; the expectation is over the joint distribution p(x, y). We wish to bound eout(g). To do so, we will bound |eout(h)− ein(h)| uniformly over H for all distributions p(x, y); however, the bound itself will depend on the data, and hence the distribution. The classic distribution independent bound is the VC-bound (Vapnik and Chervonenkis, 1971); the hope is that by taking into account the data one can get a tighter bound. The data dependent permutation complexity1 for H is defined by:

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Permutation Approach to Validation

We give a permutation approach to validation (estimation of out-sample error). One typical use of validation is model selection. We establish the legitimacy of the proposed permutation complexity by proving a uniform bound on the out-sample error, similar to a VC-style bound. We extensively demonstrate this approach experimentally on synthetic data, standard data sets from the UCI-repository, a...

متن کامل

On lower bounds for integration of multivariate permutation-invariant functions

In this note we study multivariate integration for permutation-invariant functions from a certain Banach space Ed,α of Korobov type in the worst case setting. We present a lower error bound which particularly implies that in dimension d every cubature rule which reduces the initial error necessarily uses at least d+ 1 function values. Since this holds independently of the number of permutation-...

متن کامل

On the tenacity of cycle permutation graph

A special class of cubic graphs are the cycle permutation graphs. A cycle permutation graph Pn(α) is defined by taking two vertex-disjoint cycles on n vertices and adding a matching between the vertices of the two cycles.In this paper we determine a good upper bound for tenacity of cycle permutation graphs.

متن کامل

A Tighter Error Bound for Decision Tree Learning Using PAC Learnability

Error bounds for decision trees are generally based on depth or breadth of the tree. In this paper, we propose a bound for error rate that depends both on the depth and the breadth of a specific decision tree constructed from the training samples. This bound is derived from sample complexity estimate based on PAC learnability. The proposed bound is compared with other traditional error bounds o...

متن کامل

A quantum lower bound for distinguishing random functions from random permutations

The problem of distinguishing between a random function and a random permutation on a domain of size N is important in theoretical cryptography, where the security of many primitives depend on the problem’s hardness. We study the quantum query complexity of this problem, and show that any quantum algorithm that solves this problem with bounded error must make Ω(N1/5/polylogN) queries to the inp...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2010