phone +7 (3412) 91 60 92

Archive of Issues


Algeria; Oman Guelma; Muscat
Year
2023
Volume
33
Issue
2
Pages
348-364
<<
>>
Section Mathematics
Title A new hybrid conjugate gradient algorithm for unconstrained optimization
Author(-s) Hafaidia I.a, Guebbai H.a, Al-Baali M.b, Ghiat M.a
Affiliations University 8 Mai 1945a, Sultan Qaboos Universityb
Abstract It is well known that conjugate gradient methods are useful for solving large-scale unconstrained nonlinear optimization problems. In this paper, we consider combining the best features of two conjugate gradient methods. In particular, we give a new conjugate gradient method, based on the hybridization of the useful DY (Dai-Yuan), and HZ (Hager-Zhang) methods. The hybrid parameters are chosen such that the proposed method satisfies the conjugacy and sufficient descent conditions. It is shown that the new method maintains the global convergence property of the above two methods. The numerical results are described for a set of standard test problems. It is shown that the performance of the proposed method is better than that of the DY and HZ methods in most cases.
Keywords unconstrained optimization, conjugate gradient methods, conjugacy conditions and sufficient descent conditions
UDC 519.6
MSC 90C30, 90C53
DOI 10.35634/vm230211
Received 2 July 2022
Language English
Citation Hafaidia I., Guebbai H., Al-Baali M., Ghiat M. A new hybrid conjugate gradient algorithm for unconstrained optimization, Vestnik Udmurtskogo Universiteta. Matematika. Mekhanika. Komp'yuternye Nauki, 2023, vol. 33, issue 2, pp. 348-364.
References
  1. Al-Baali M., Narushima Y., Yabe H. A family of three-term conjugate gradient methods with sufficient descent property for unconstrained optimization, Computational Optimization and Applications, 2015, vol. 60, issue 1, pp. 89-110. https://doi.org/10.1007/s10589-014-9662-z
  2. Andrei N. A new three-term conjugate gradient algorithm for unconstrained optimization, Numerical Algorithms, 2015, vol. 68, issue 2, pp. 305-321. https://doi.org/10.1007/s11075-014-9845-9
  3. Djordjević S.S. New hybrid conjugate gradient method as a convex combination of LS and FR methods, Acta Mathematica Scientia, 2019, vol. 39, issue 1, pp. 214-228. https://doi.org/10.1007/s10473-019-0117-6
  4. Liu J.K., Feng Y.M., Zou L.M. Some three-term conjugate gradient methods with the inexact line search condition, Calcolo, 2018, vol. 55, issue 2, article number: 16. https://doi.org/10.1007/s10092-018-0258-3
  5. Chen Yuting, Cao Mingyuan, Yang Yueting. A new accelerated conjugate gradient method for large-scale unconstrained optimization, Journal of Inequalities and Applications, 2019, vol. 2019, issue 1, article number: 300. https://doi.org/10.1186/s13660-019-2238-9
  6. Al-Baali M., Caliciotti A., Fasano G., Roma M. Exploiting damped techniques for nonlinear conjugate gradient methods, Mathematical Methods of Operations Research, 2017, vol. 86, issue 3, pp. 501-522. https://doi.org/10.1007/s00186-017-0593-1
  7. Benrabia N., Laskri Y., Guebbai H., Al-Baali M. Applying the Powell's symmetrical technique to conjugate gradient methods with the generalized conjugacy condition, Numerical Functional Analysis and Optimization, 2016, vol. 37, issue 7, pp. 839-849. https://doi.org/10.1080/01630563.2016.1178142
  8. Hestenes M.R., Stiefel E. Methods of conjugate gradients for solving linear systems, Journal of Research of the National Bureau of Standards, 1952, vol. 49, no. 6, pp. 409-436. https://doi.org/10.6028/jres.049.044
  9. Fletcher R., Reeves C.M. Function minimization by conjugate gradients, The Computer Journal, 1964, vol. 7, issue 2, pp. 149-154. https://doi.org/10.1093/comjnl/7.2.149
  10. Polak E., Ribière G. Note sur la convergence de méthodes de directions conjuguées, Revue française d'informatique et de recherche opérationnelle. Série rouge, 1969, vol. 3, no. 16, pp. 35-43 (in French). https://doi.org/10.1051/m2an/196903r100351
  11. Polyak B.T. The conjugate gradient method in extremal problems, USSR Computational Mathematics and Mathematical Physics, 1969, vol. 9, issue 4, pp. 94-112. https://doi.org/10.1016/0041-5553(69)90035-4
  12. Liu Y., Storey C. Efficient generalized conjugate gradient algorithms, part 1: Theory, Journal of Optimization Theory and Applications, 1991, vol. 69, issue 1, pp. 129-137. https://doi.org/10.1007/BF00940464
  13. Dai Y.H., Yuan Y. A nonlinear conjugate gradient method with a strong global convergence property, SIAM Journal on Optimization, 1999, vol. 10, issue 1, pp. 177-182. https://doi.org/10.1137/S1052623497318992
  14. Hager W.W., Zhang Hongchao. A new conjugate gradient method with guaranteed descent and an efficient line search, SIAM Journal on Optimization, 2005, vol. 16, issue 1, pp. 170-192. https://doi.org/10.1137/030601880
  15. Hager W.W., Zhang Hongchao. Algorithm 851: CG_DESCENT, a conjugate gradient method with guaranteed descent, ACM Transactions on Mathematical Software, 2006, vol. 32, no. 1, pp. 113-137. https://doi.org/10.1145/1132973.1132979
  16. Babaie-Kafaki S., Ghanbari R. A hybridization of the Hestenes-Stiefel and Dai-Yuan conjugate gradient methods based on a least-squares approach, Optimization Methods and Software, 2015, vol. 30, issue 4, pp. 673-681. https://doi.org/10.1080/10556788.2014.966825
  17. Babaie-Kafaki S., Ghanbari R. Two hybrid nonlinear conjugate gradient methods based on a modified secant equation, Optimization, 2014, vol. 63, issue 7, pp. 1027-1042. https://doi.org/10.1080/02331934.2012.693083
  18. Gilbert J.C., Nocedal J. Global convergence properties of conjugate gradient methods for optimization, SIAM Journal on Optimization, 1992, vol. 2, issue 1, pp. 21-42. https://doi.org/10.1137/0802003
  19. Al-Baali M. Descent property and global convergence of the Fletcher-Reeves method with inexact line search, IMA Journal of Numerical Analysis, 1985, vol. 5, issue 1, pp. 121-124. https://doi.org/10.1093/imanum/5.1.121
  20. Fletcher R. Practical methods of optimization, John Wiley and Sons, 2000. https://doi.org/10.1002/9781118723203
  21. Andrei N. Another accelerated conjugate gradient algorithm with guaranteed descent and conjugacy conditions for large-scale unconstrained optimization, ICI Technical Report, 2010.
  22. Andrei N. An unconstrained optimization test functions collection, Advanced Modeling and Optimization, 2008, vol. 10, no. 1, pp. 147-161.
  23. Dolan E.D., Moré J.J. Benchmarking optimization software with performance profiles, Mathematical Programming, 2002, vol. 91, issue 2, pp. 201-213. https://doi.org/10.1007/s101070100263
  24. Zoutendijk G. Nonlinear programming computational methods, Integer and Nonlinear Programming, Amsterdam: North-Holland, 1970, pp. 37-86.
Full text
<< Previous article
Next article >>