Neural Networks: What Non-linearity to Choose

نویسندگان

  • Vladik Kreinovich
  • Chris Quintana
چکیده

Neural networks are now one of the most successful learning formalisms. Neurons transform inputs x1, ..., xn into an output f(w1x1 + ... + wnxn), where f is a non-linear function and wi are adjustable weights. What f to choose? Usually the logistic function is chosen, but sometimes the use of different functions improves the practical efficiency of the network. We formulate the problem of choosing f as a mathematical optimization problem and solve it under different optimality criteria. As a result, we get a list of functions f that are optimal under these criteria. This list includes both the functions that were empirically proved to be the best for some problems, and some new functions that may be worth trying. 1. FORMULATION OF THE PROBLEM. Neural networks are now one of the most successful learning formalisms (see, e.g., the recent survey in [Hecht-Nielsen 1991]). After the initial success of linear neural models, in which the output y is equal to the linear combination of the input signals xi, i.e. y = w1x1 + w2x2 + ..., it was shown in [Minsky Papert 1968] that if we only have linear neurons, then we end up with only linear functions and this severely limits the number of problems that we can solve using the network. The next step, then, is to consider non-linear neurons, in which the output signal is equal to f(w1x1 +w2x2 + ...), where f(y) is a given non-linear function. A natural question arises: what function f(y) do we choose? Why is this problem important? It is a very important problem because although neural networks help us to design good learning procedures, these procedures are far from being reliable. Sometimes these procedures do not work; sometimes they work but demand too much time, and too big a sample, to learn. Naturally, we might think that this is because the function f that we used was not the best one. Sometimes the use of different functions can improve the practical efficiency of the network (see, e.g., [Wasserman 1989, pp. 15-16]). If a simple guess can really improve the learning performance, then it is natural to suppose that deep mathematical optimization will lead to even better results. Why is this problem difficult? We want to find a function f for which some characteristics J of learning, such as average learning time or average number of errors, is optimal (in these cases

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Novel Radial Basis Function Neural Networks based on Probabilistic Evolutionary and Gaussian Mixture Model for Satellites Optimum Selection

In this study, two novel learning algorithms have been applied on Radial Basis Function Neural Network (RBFNN) to approximate the functions with high non-linear order. The Probabilistic Evolutionary (PE) and Gaussian Mixture Model (GMM) techniques are proposed to significantly minimize the error functions. The main idea is concerning the various strategies to optimize the procedure of Gradient ...

متن کامل

Testing for Stochastic Non- Linearity in the Rational Expectations Permanent Income Hypothesis

The Rational Expectations Permanent Income Hypothesis implies that consumption follows a martingale. However, most empirical tests have rejected the hypothesis. Those empirical tests are based on linear models. If the data generating process is non-linear, conventional tests may not assess some of the randomness properly. As a result, inference based on conventional tests of linear models can b...

متن کامل

Suppliers Selection in Consideration of Risks by a Neural Network

Faced with the dynamic demands of a changing market, companies are facing fierce competition, which forces them to consider more and more new approaches to improve quality, reduce costs, produce on time, control their risks and remain successful in the face of any disruption. It is clear that the choice of appropriate suppliers is one of the key factors in increasing the competitiveness of comp...

متن کامل

On Random Weights for Texture Generation in One Layer Neural Networks

Recent work in the literature has shown experimentally that one can use the lower layers of a trained convolutional neural network (CNN) to model natural textures. More interestingly, it has also been experimentally shown that only one layer with random filters can also model textures although with less variability. In this paper we ask the question as to why one layer CNNs with random filters ...

متن کامل

Chaotic Test and Non-Linearity of Abnormal Stock Returns: Selecting an Optimal Chaos Model in Explaining Abnormal Stock Returns around the Release Date of Annual Financial Statements

For many investors, it is important to predict the future trend of abnormal stock returns. Thus, in this research, the abnormal stock returns of the listed companies in Tehran Stock Exchange were tested since 2008- 2017 using three hypotheses. The first and second hypotheses examined the non-linearity and non-randomness of the abnormal stock returns ′ trend around the release date of annual fin...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1991