نتایج جستجو برای: sufficient descent condition

تعداد نتایج: 490700  

Journal: :SIAM Journal on Optimization 2011
Yasushi Narushima Hiroshi Yabe John A. Ford

Conjugate gradient methods are widely used for solving large-scale unconstrained optimization problems, because they do not need the storage of matrices. In this paper, we propose a general form of three-term conjugate gradient methods which always generate a sufficient descent direction. We give a sufficient condition for the global convergence of the proposed general method. Moreover, we pres...

Journal: :CoRR 2017
Yunwen Lei Lei Shi Zheng-Chu Guo

In this paper we study the convergence of online gradient descent algorithms in reproducing kernel Hilbert spaces (RKHSs) without regularization. We establish a sufficient condition and a necessary condition for the convergence of excess generalization errors in expectation. A sufficient condition for the almost sure convergence is also given. With high probability, we provide explicit converge...

Journal: :Rairo-operations Research 2022

In this paper, we proposed a new hybrid conjugate gradient algorithm for solving unconstrained optimization problems as convex combination of the Dai-Yuan algorithm, conjugate-descent and Hestenes-Stiefel algorithm. This is globally convergent satisfies sufficient descent condition by using strong Wolfe conditions. The numerical results show that nonlinear efficient robust.

2011
Jinkui Liu Shaoheng Wang

In this paper, an efficient modified nonlinear conjugate gradient method for solving unconstrained optimization problems is proposed. An attractive property of the modified method is that the generated direction in each step is always descending without any line search. The global convergence result of the modified method is established under the general Wolfe line search condition. Numerical r...

2009
Jianguo Zhang Yunhai Xiao Zengxin Wei Joaquim J. Júdice

Two nonlinear conjugate gradient-type methods for solving unconstrained optimization problems are proposed. An attractive property of the methods, is that, without any line search, the generated directions always descend. Under some mild conditions, global convergence results for both methods are established. Preliminary numerical results show that these proposed methods are promising, and comp...

Journal: :Tamkang Journal of Mathematics 2022

In this paper, based on the efficient Conjugate Descent ({\tt CD}) method, two generalized {\tt CD}algorithms are proposed to solve unconstrained optimization problems.These methods three-term conjugate gradient which generateddirections by using parameters and independent of line searchsatisfy in sufficient descent condition. Furthermore, under strong Wolfe search,the global convergence proved...

Farhad Sarani, Hadi Nosratipour

In [1] (Hybrid Conjugate Gradient Algorithm for Unconstrained Optimization J. Optimization. Theory Appl. 141 (2009) 249 - 264), an efficient hybrid conjugate gradient algorithm, the CCOMB algorithm is proposed for solving unconstrained optimization problems. However, the proof of Theorem 2.1 in [1] is incorrect due to an erroneous inequality which used to indicate the descent property for the s...

نمودار تعداد نتایج جستجو در هر سال

با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید