An efficient one-layer recurrent neural network for solving a class of nonsmooth optimization problems

Authors

  • Alireza Hosseini School of Mathematics, Statistics and Computer Science, College of Science, University of Tehran/ Faculty member
Abstract:

Constrained optimization problems have a wide range of applications in science, economics, and engineering. In this paper, a neural network model is proposed to solve a class of nonsmooth constrained optimization problems with a nonsmooth convex objective function subject to nonlinear inequality and affine equality constraints. It is a one-layer non-penalty recurrent neural network based on the differential inclusion. Unlike most of the existing neural network models, there is neither a penalty parameter nor a penalty function in its structure. It has less complexity which leads to the easier implementation of the model for solving optimization problems. The equivalence of optimal solutions set of the main optimization problem and the equilibrium points set of the model is proven. Moreover, the global convergence and the stability of the introduced neural network are shown. Some examples including the L1-norm minimization problem are given and solved by the proposed model to illustrate its performance and effectiveness.

Upgrade to premium to download articles

Sign up to access the full text

Already have an account?login

similar resources

An Efficient Neurodynamic Scheme for Solving a Class of Nonconvex Nonlinear Optimization Problems

‎By p-power (or partial p-power) transformation‎, ‎the Lagrangian function in nonconvex optimization problem becomes locally convex‎. ‎In this paper‎, ‎we present a neural network based on an NCP function for solving the nonconvex optimization problem‎. An important feature of this neural network is the one-to-one correspondence between its equilibria and KKT points of the nonconvex optimizatio...

full text

A recurrent neural network for solving a class of generalized convex optimization problems

In this paper, we propose a penalty-based recurrent neural network for solving a class of constrained optimization problems with generalized convex objective functions. The model has a simple structure described by using a differential inclusion. It is also applicable for any nonsmooth optimization problem with affine equality and convex inequality constraints, provided that the objective funct...

full text

An efficient modified neural network for solving nonlinear programming problems with hybrid constraints

This paper presents ‎‎the optimization techniques for solving‎‎ convex programming problems with hybrid constraints‎.‎ According to the saddle point theorem‎, ‎optimization theory‎, ‎convex analysis theory‎, ‎Lyapunov stability theory and LaSalle‎‎invariance principle‎,‎ a neural network model is constructed‎.‎ The equilibrium point of the proposed model is proved to be equivalent to the optima...

full text

A Recurrent Neural Network for Solving Strictly Convex Quadratic Programming Problems

In this paper we present an improved neural network to solve strictly convex quadratic programming(QP) problem. The proposed model is derived based on a piecewise equation correspond to optimality condition of convex (QP) problem and has a lower structure complexity respect to the other existing neural network model for solving such problems. In theoretical aspect, stability and global converge...

full text

A Recurrent Neural Network Model for Solving Linear Semidefinite Programming

In this paper we solve a wide rang of Semidefinite Programming (SDP) Problem by using Recurrent Neural Networks (RNNs). SDP is an important numerical tool for analysis and synthesis in systems and control theory. First we reformulate the problem to a linear programming problem, second we reformulate it to a first order system of ordinary differential equations. Then a recurrent neural network...

full text

A Neural Network Method Based on Mittag-Leffler Function for Solving a Class of Fractional Optimal Control Problems

In this paper, a computational intelligence method is used for the solution of fractional optimal control problems (FOCP)'s with equality and inequality constraints. According to the Ponteryagin minimum principle (PMP) for FOCP with fractional derivative in the Riemann- Liouville sense and by constructing a suitable error function, we define an unconstrained minimization problem. In the optimiz...

full text

My Resources

Save resource for easier access later

Save to my library Already added to my library

{@ msg_add @}


Journal title

volume 6  issue 24

pages  97- 110

publication date 2020-05-21

By following a journal you will be notified via email when a new issue of this journal is published.

Hosted on Doprax cloud platform doprax.com

copyright © 2015-2023