Overfitting and generalization in learning discrete patterns
نویسنده
چکیده
Understanding and preventing overrtting is a very important issue in artiicial neural network design, implementation, and application. Weigend (1994) reports that the presence and absence of overrtting in neural networks depends on how the testing error is measured, and that there is no overrtting in terms of the classiication error (symbolic-level errors). In this paper, we show that, in terms of the classiication error, overrtting does occur for certain representation used to encode the discrete attributes. We design simple Boolean functions with clear rationale, and present experimental results to support our claims. In addition, we report some interesting results on the best generalization ability of networks in terms of their sizes.
منابع مشابه
Adaptive Learning with Binary Neurons
A efficient incremental learning algorithm for classification tasks, called NetLines, well adapted for both binary and real-valued input patterns is presented. It generates small compact feedforward neural networks with one hidden layer of binary units and binary output units. A convergence theorem ensures that solutions with a finite number of hidden units exist for both binary and real-valued...
متن کاملAn Approach to Reducing Overfitting in FCM with Evolutionary Optimization
Fuzzy clustering methods are conveniently employed in constructing a fuzzy model of a system, but they need to tune some parameters. In this research, FCM is chosen for fuzzy clustering. Parameters such as the number of clusters and the value of fuzzifier significantly influence the extent of generalization of the fuzzy model. These two parameters require tuning to reduce the overfitting in the...
متن کاملEfficient Adaptive Learning for Classification Tasks with Binary Units
This article presents a new incremental learning algorithm for classification tasks, called NetLines, which is well adapted for both binary and real-valued input patterns. It generates small, compact feedforward neural networks with one hidden layer of binary units and binary output units. A convergence theorem ensures that solutions with a finite number of hidden units exist for both binary an...
متن کاملCS 269 : Machine Learning Theory Lecture 14 : Generalization Error of Adaboost
In this lecture we will continue our discussion of the Adaboost algorithm and derive a bound on the generalization error. We saw last time that the training error decreases exponentially with respect to the number of rounds T . However, we also want to see the performance of this algorithm on new test data. Today we will show why the Adaboost algorithm generalizes so well and why it avoids over...
متن کاملA Fast Scheme for Feature Subset Selection to Avoid Overfitting in AdaBoost
AdaBoost is a well known, effective technique for increasing the accuracy of learning algorithms. However, it has the potential to overfit the training set because its objective is to minimize error on the training set. We show that with the introduction of a scoring function and the random selection of training data it is possible to create a smaller set of feature vectors. The selection of th...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- Neurocomputing
دوره 8 شماره
صفحات -
تاریخ انتشار 1995