نتایج جستجو برای: overfitting

تعداد نتایج: 4333  

2016
Michael Opitz Horst Possegger Horst Bischof

Large neural networks trained on small datasets are increasingly prone to overfitting. Traditional machine learning methods can reduce overfitting by employing bagging or boosting to train several diverse models. For large neural networks, however, this is prohibitively expensive. To address this issue, we propose a method to leverage the benefits of ensembles without explicitely training sever...

Journal: :CoRR 2017
Izhar Wallach Abraham Heifets

Undetected overfitting can occur when there are significant redundancies between training and validation data. We describe AVE, a new measure of training-validation redundancy for ligand-based classification problems that accounts for the similarity amongst inactive molecules as well as active. We investigated nine widely-used benchmarks for virtual screening and QSAR, and show that the amount ...

2018
Roger Grosse

When we train a machine learning model, we don’t just want it to learn to model the training data. We want it to generalize to data it hasn’t seen before. Fortunately, there’s a very convenient way to measure an algorithm’s generalization performance: we measure its performance on a held-out test set, consisting of examples it hasn’t seen before. If an algorithm works well on the training set b...

Journal: :CoRR 2017
Akm Ashiquzzaman Abdul Kawsar Tushar Md. Rashedul Islam Jong-Myon Kim

Augmented accuracy in prediction of diabetes will open up new frontiers in health prognostics. Data overfitting is a performance-degrading issue in diabetes prognosis. In this study, a prediction system for the disease of diabetes is presented where the issue of overfitting is minimized by using the dropout method. Deep learning neural network is used where both fully connected layers are follo...

Journal: :J. Classification 2001
Herbert K. H. Lee

Classification rates on out-of-sample predictions can often be improved through the use of model selection when fitting a model on the training data. Using correlated predictors or fitting a model of too high a dimension can lead to overfitting, which in turn leads to poor out-of-sample performance. I will discuss methodology using the Bayesian Information Criterion (BIC) of Schwarz (1978) that...

2018
Dominik Janzing Bernhard Schoelkopf

We consider linear models where d potential causes X1, . . . , Xd are correlated with one target quantity Y and propose a method to infer whether the association is causal or whether it is an artifact caused by overfitting or hidden common causes. We employ the idea that in the former case the vector of regression coefficients has ‘generic’ orientation relative to the covariance matrix ΣXX of X...

2014
Rui LI Yingying LI

In order to solve the overfitting of sample weights and the low detection rate in training process of the traditional AdaBoost algorithm, an improved AdaBoost algorithm based on Haar-like features and LBP features is proposed. This method improves weight updating rule and weights normalization rule of the traditional AdaBoost algorithm. Then combining this method with the AdaBoost algorithm bas...

Journal: :Knowl.-Based Syst. 2002
Max Bramer

The automatic induction of classification rules from examples in the form of a decision tree is an important technique used in data mining. One of the problems encountered is the overfitting of rules to training data. In some cases this can lead to an excessively large number of rules, many of which have very little predictive value for unseen data. This paper is concerned with the reduction of...

2016
Taygun Kekec David M. J. Tax

Word embedding models learn vectorial word representations that can be used in a variety of NLP applications. When training data is scarce, these models risk losing their generalization abilities due to the complexity of the models and the overfitting to finite data. We propose a regularized embedding formulation, called Robust Gram (RG), which penalizes overfitting by suppressing the disparity...

Journal: :CoRR 2015
Jonathan Raiman Szymon Sidor

We present a complimentary objective for training recurrent neural networks (RNN) with gating units that helps with regularization and interpretability of the trained model. Attention-based RNN models have shown success in many difficult sequence to sequence classification problems with long and short term dependencies, however these models are prone to overfitting. In this paper, we describe h...

نمودار تعداد نتایج جستجو در هر سال

با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید