Constructive Neural Networks with Regularization

نویسندگان

  • JANI LAHNAJÄRVI
  • MIKKO LEHTOKANGAS
چکیده

In this paper we present a regularization approach to the training of all the network weights in cascadecorrelation type constructive neural networks. Especially, the case of regularizing the output neuron of the network is presented. In this case, the output weights are trained by employing a regularized objective function containing a penalty term which is proportional to the weight values of the unit being trained. It is shown that the training can still be done with the pseudo-inverse method of linear regression if the output unit employs linear activation function. The degree of regularization and the smoothness of network mapping can be adjusted by changing the value of the regularization parameter. The investigated algorithms were Cascade-Correlation, Modified Cascade-Correlation, Cascade, and Fixed Cascade Error. The algorithms having the regularization of the hidden and output units were compared with the ones having only the regularization of the hidden units and with those having no regularization at all. The simulation results show that the regularization of the output unit is highly beneficial. It leads to better generalization performance and in many cases to lower computational costs when compared to the partially and non-regulated versions of the same algorithms. Key-Words: Constructive neural networks, regularization, generalization, cascade-correlation, classification, regression.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Bayesian Regularization in Constructive Neural Networks

In this paper, we study the incorporation of Bayesian reg-ularization into constructive neural networks. The degree of regulariza-tion is automatically controlled in the Bayesian inference framework and hence does not require manual setting. Simulation shows that regular-ization, with input training using a full Bayesian approach, produces networks with better generalization performance and low...

متن کامل

Transfer functions: hidden possibilities for better neural networks

Sigmoidal or radial transfer functions do not guarantee the best generalization nor fast learning of neural networks. Families of parameterized transfer functions provide flexible decision borders. Networks based on such transfer functions should be small and accurate. Several possibilities of using transfer functions of different types in neural models are discussed, including enhancement of i...

متن کامل

Supervised Models C1.2 Multilayer perceptrons

This section introduces multilayer perceptrons, which are the most commonly used type of neural network. The popular backpropagation training algorithm is studied in detail. The momentum and adaptive step size techniques, which are used for accelerated training, are discussed. Other acceleration techniques are briefly referenced. Several implementation issues are then examined. The issue of gen...

متن کامل

Predictive Abilities of Bayesian Regularization and Levenberg–Marquardt Algorithms in Artificial Neural Networks: A Comparative Empirical Study on Social Data

The objective of this study is to compare the predictive ability of Bayesian regularization with Levenberg–Marquardt Artificial Neural Networks. To examine the best architecture of neural networks, the model was tested with one-, two-, three-, four-, and five-neuron architectures, respectively. MATLAB (2011a) was used for analyzing the Bayesian regularization and Levenberg–Marquardt learning al...

متن کامل

Regularization Parameter Selection for Faulty Neural Networks

Regularization techniques have attracted many researches in the past decades. Most focus on designing the regularization term, and few on the optimal regularization parameter selection, especially for faulty neural networks. As is known that in the real world, the node faults often inevitably take place, which would lead to many faulty network patterns. If employing the conventional method, i.e...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2000