New Probabilistic Bounds on Eigenvalues and Eigenvectors of Random Kernel Matrices
نویسندگان
چکیده
Kernel methods are successful approaches for different machine learning problems. This success is mainly rooted in using feature maps and kernel matrices. Some methods rely on the eigenvalues/eigenvectors of the kernel matrix, while for other methods the spectral information can be used to estimate the excess risk. An important question remains on how close the sample eigenvalues/eigenvectors are to the population values. In this paper, we improve earlier results on concentration bounds for eigenvalues of general kernel matrices. For distance and inner product kernel functions, e.g. radial basis functions, we provide new concentration bounds, which are characterized by the eigenvalues of the sample covariance matrix. Meanwhile, the obstacles for sharper bounds are accounted for and partially addressed. As a case study, we derive a concentration inequality for sample kernel target-alignment.
منابع مشابه
APPLICATION OF THE RANDOM MATRIX THEORY ON THE CROSS-CORRELATION OF STOCK PRICES
The analysis of cross-correlations is extensively applied for understanding of interconnections in stock markets. Variety of methods are used in order to search stock cross-correlations including the Random Matrix Theory (RMT), the Principal Component Analysis (PCA) and the Hierachical Structures. In this work, we analyze cross-crrelations between price fluctuations of 20 company stocks...
متن کاملOn the eigenvalues of some matrices based on vertex degree
The aim of this paper is to compute some bounds of forgotten index and then we present spectral properties of this index. In continuing, we define a new version of energy namely ISI energy corresponded to the ISI index and then we determine some bounds for it.
متن کاملAsymptotic error bounds for kernel-based Nystrm low-rank approximation matrices
• Many kernel-based learning algorithms have the computational load. • The Nyström low-rank approximation is designed for reducing the computation. • We propose the spectrum decomposition condition with a theoretical justification. • Asymptotic error bounds on eigenvalues and eigenvectors are derived. • Numerical experiments are provided for covariance kernel and Wishart matrix. AMS subject cla...
متن کاملA note on positive deniteness and stability of interval matrices
It is proved that by using bounds of eigenvalues of an interval matrix, someconditions for checking positive deniteness and stability of interval matricescan be presented. These conditions have been proved previously with variousmethods and now we provide some new proofs for them with a unity method.Furthermore we introduce a new necessary and sucient condition for checkingstability of interval...
متن کاملRandom matrix approach to cross correlations in financial data.
We analyze cross correlations between price fluctuations of different stocks using methods of random matrix theory (RMT). Using two large databases, we calculate cross-correlation matrices C of returns constructed from (i) 30-min returns of 1000 US stocks for the 2-yr period 1994-1995, (ii) 30-min returns of 881 US stocks for the 2-yr period 1996-1997, and (iii) 1-day returns of 422 US stocks f...
متن کامل