| English | Arabic | Home | Login |

Published Journal Articles

2023

A Combined Conjugate Gradient Quasi-Newton Method with Modification BFGS Formula

2023-03
International Journal of Analysis and Applications (Volume : 21)
Abstract. The conjugate gradient and Quasi-Newton methods have advantages and drawbacks, as although quasi-Newton algorithm has more rapid convergence than conjugate gradient, they require more storage compared to conjugate gradient algorithms. In 1976, Buckley designed a method that combines the CG method with QN updates, which is better than that observed for conjugate gradient algorithms but not as good as the quasi-Newton approach. This type of method is called the pre- conditioned conjugate gradient (PCG) method. In this paper, we introduce two new preconditioned conjugate gradient (PCG) methods that combine conjugate gradient with a new update of quasi- Newton methods. The new quasi-Newton method satisfied the positive define, and the direction of the new preconditioned conjugate gradient is descent direction. In numerical results, it is showing the new preconditioned conjugate gradient method is more effective on several high-dimension test problems than standard preconditioning.

A NEW CONJUGATE GRADIENT WITH GLOBAL CONVERGES FOR NONLINEAR PROBLEMS

2023-01
Journal of University of Duhok (Issue : 2) (Volume : 25)
The conjugate gradient(CG) method is one of the most popular and well-known iterative strategies for solving minimization problems, it has extensive applications in many domains such as machine learning, neural networks, and many other fields, partly because to its simplicity in algebraic formulation and implementation in codes of computer and partially due to their efficiency in solving largescale unconstrained optimization problems.Fletcher/Reeves (C, 1964)expanded the concept to nonlinear problems.In 1964, and this is widely regarded as the first algorithm of nonlinear conjugate gradient.Since then, other conjugate gradient method versions have been proposed. In this paper and in section one, we derivea new conjugate gradient for solving nonlinear minimization problems based on parameter of Perry. In section two we will satisfy some conditions like descent and sufficient descent conditions.In section three ,we will study the global convergence of new suggestion. We present numerical findings in the fourth part to demonstrate the efficacy of the suggestion technique. Finally, we provide a conclusion.
2022

A NEW MODIFIED CONJUGATE GRADIENT FOR NONLINEAR MINIMIZATION PROBLEMS

2022-10
Science Journal of University of Zakho (Issue : 4) (Volume : 10)
The conjugate gradient is a highly effective technique to solve theunconstrained nonlinear minimization problemsand it is one of the most well-known methods. It has a lot of applications. For large-scaleand unconstrained minimization problems, conjugate gradient techniques are widely applied.In this paper, we will suggesta newparameter of conjugate gradient tosolvethenonlinear unconstrained minimization problems,based onthe parameter ofDai and Liao. We will study theproperty of the descent ,the property of the sufficient descent and property of the global convergence of thenew method. We introduce some numerical data to prove the efficacy of the our method.

A Descent Conjugate Gradient Method With Global Converges Properties for Non-Linear Optimization

2022-06
Mathematics and Statistics (Issue : 3) (Volume : 10)
Iterative methods such as the conjugate gradient method are well known methods for solving non-linear unconstrained minimization problems partially because of their capacity to handle large-scale unconstrained optimization problems rapidly, and partly due to their algebraic representation and implementation in computer programs. The conjugate gradient method has wide applications in a lot of fields such as machine learning, neural networks and many other fields. Fletcher and Reeves [1] expanded the approach to nonlinear problems in 1964. It is considered to be the first nonlinear conjugate gradient technique. Since then, lots of new other conjugate gradient methods have been proposed. In this work, we will propose a new coefficient conjugate gradient method to find the minimum of the non-linear unconstrained optimization problems based on parameter of Hestenes Stiefel. Section one in this work contains the derivative of new method. In section two, we will satisfy the descent and sufficient descent conditions. In section three, we will study the property of the global convergence of the new proposed. In the fourth section, we will give some numerical results by using some known test functions and compare the new method with Hestenes S. to demonstrate the effectiveness of the suggestion method. Finally, we will give conclusions.
2021

A Modified Perry's Conjugate Gradient Method Based on Powell's Equation for Solving Large-Scale Unconstrained Optimization

2021-11
Mathematics and Statistics (Issue : 6) (Volume : 9)
It is known that the conjugate gradient method is still a popular method for many researchers who are focused in solving the large-scale unconstrained optimization problems and nonlinear equations because the method avoids the computation and storage of some matrices so the memory’s requirements of the method are very small. In this work, a modified of Perry conjugate gradient method which fulfills a global convergence with standard assumptions is shown and analyzed. The idea of new method is based on Perry method by using the equation which is founded via Powell in 1978. The weak Wolfe–Powell search conditions is used to choose the optimal line search, under the line search and suitable conditions we prove both descent and sufficient descent conditions. In particular, numerical results show that the new conjugate gradient method is more effective and competitive when compared to other of standard conjugate gradient methods including: - CG- Hestenes and Stiefel (H/S) method, CG-Perry method CG- Dai and Yuan (D/Y)method. The comparison is completed under a group of standard test problems with various dimensions from the CUTEst test library and the comparative performances of the methods are evaluated by total the number of iterations and the total number of function evaluations.

A New conjugate gradient method for unconstrained optimization problems with descent property

2021-05
General Letters in Mathematics (GLM) (Issue : 2) (Volume : 9)
In this paper, we propose a new conjugate gradient method for solving nonlinear unconstrained optimization. The new method consists of three parts, the first part of them is the parameter of Hestenes-Stiefel (HS). The proposed method is satisfying the descent condition, sufficient descent condition and conjugacy condition. We give some numerical results to show the efficiency of the suggested method.
2020

A new self-scaling variable metric (DFP) method for unconstrained optimization problems

2020-06
Refaad (Volume : 9)
In this study, a new self-scaling variable metric (VM)-updating method for solving nonlinear unconstrained optimization problems is presented. The general strategy of (New VM-updating) is to propose a new quasi-newton condition used for update the usual DFP Hessian to a number of times in a way to be specified in some iteration with PCG method to improve the performance of the Hessian approximation. We show that it produces a positive definite matrix. Experimental results indicate that the new suggested method was more efficient than the standard DFP method, with respect to the number of functions evaluations (NOF) and number of iterations (NOI).
2019

A new class of three-term conjugate Gradient methods for solving unconstrained minimization problems

2019-12
Refaad
Conjugate gradient (CG) methods which are usually generate descent search directions, are beneficial for large-scale unconstrained optimization models, because of its low memory requirement and simplicity. This paper studies the three-term CG method for unconstrained optimization. The modified a three-term CG method based on the formal 𝒕∗ which is suggested by Kafaki and Ghanbari [11], and using some well-known CG formulas for unconstrained optimization. Our proposed method satisfies both (the descent and the sufficient descent) conditions. Furthermore, if we use the exact line search the new proposed is reduce to the classical CG method. The numerical results show that the suggested method is promising and exhibits a better numerical performance in comparison with the three- term (ZHS-CG) method from an implementation of the suggested method on some normal unconstrained optimization test functions

A New Quasi-Newton (SR1) With PCG Method for Unconstrained Nonlinear Optimization

2019-12
Shareef et al., International Journal of Advanced Trends in Computer Science and Engineering, 8(6), November - December 2019, 3124 - 3128 (Volume : 8)
The quasi-newton equation (QN) plays a key role in contemporary nonlinear optimization. In this paper, we present a new symmetric rank-one (SR1) method by using preconditioning conjugate gradient (PCG) method for solving unconstrained optimization problems. The suggested method has an algorithm in which the usual SR1 Hessian is updated. We show that the new quasi-newton (SR1) method maintains the Quasi- Newton condition and the positive definite property. Numerical experiments are reported which produces by the new algorithm better numerical results than those of the normal (SR1) method by using PCG algorithm based on the number of iterations (NOI) and the number of functions evaluation (NOF).

MODIFIED CONJUGATE GRADIENT METHOD FOR TRAINING NEURAL NETWORKS BASED ON LOGISTIC MAPPING

2019-03
Journal of University of Duhok (Volume : 22)
In this paper, we suggested a modified conjugate gradient method for training neural network which assurance the descent and the sufficient descent conditions. The global convergence of our proposed method has been studied. Finally, the test results present that, in general, the modified method is more superior and efficient when compared to other standard conjugate gradient methods

A NEW CONJUGATE GRADIENT COEFFICIENT FOR UNCONSTRAINED OPTIMIZATION BASED ON DAI-LIAO

2019-03
science journal of university of Zakho (Volume : 7)
Conjugate gradient method plays an enormous role in resolving unconstrained optimization problem, particularly for large scale. In this paper, a new conjugate gradient method for unconstrained optimization based on Dai-Liao (DL) formula by using Barzilai and Borwein step size. Our new method satisfies both descent and sufficient descent conditions. The numerical results show that the proposed algorithm is potentially efficient and performs better than with Polak and Ribiere (PR) algorithm, depending on number of iterations (NOI) and the number of functions evaluation (NOF).
2016

A NEW CONJUGATE GRADIENT FOR UNCONSTRAINED OPTIMIZATION BASED ON STEP SIZE OF BARZILAI AND BORWEIN

2016-04
Journal University of Zakho (Volume : 4)
In this paper, a new formula of 􀢼􀢑 is suggested for conjugate gradient method of solving unconstrained optimization problems based on step size of Barzilai and Borwein. Our new proposed CG-method has descent condition, sufficient descent condition and global convergence properties. Numerical comparisons with a standard conjugate gradient algorithm show that this algorithm very effective depending on the number of iterations and the number of functions evaluation.

Back