Nonlinear Conjugate Gradient Methods for Unconstrained Optimization

For each method, the convergence analysis, the computational performances and the comparisons versus other conjugate gradient methods are given. The theory behind the conjugate gradient algorithms presented as a methodology is developed with a clear, rigorous, and friendly exposition; the reader wil...

Full description

Bibliographic Details
Main Author: Andrei, Neculai
Format: eBook
Language:English
Published: Cham Springer International Publishing 2020, 2020
Edition:1st ed. 2020
Series:Springer Optimization and Its Applications
Subjects:
Online Access:
Collection: Springer eBooks 2005- - Collection details see MPG.ReNa
LEADER 03870nmm a2200325 u 4500
001 EB001898565
003 EBX01000000000000001061564
005 00000000000000.0
007 cr|||||||||||||||||||||
008 200706 ||| eng
020 |a 9783030429508 
100 1 |a Andrei, Neculai 
245 0 0 |a Nonlinear Conjugate Gradient Methods for Unconstrained Optimization  |h Elektronische Ressource  |c by Neculai Andrei 
250 |a 1st ed. 2020 
260 |a Cham  |b Springer International Publishing  |c 2020, 2020 
300 |a XXVIII, 498 p. 93 illus., 90 illus. in color  |b online resource 
505 0 |a 1. Introduction -- 2. Linear Conjugate Gradient Algorithm -- 3. General Convergence Results for Nonlinear Conjugate Gradient Methods -- 4. Standard Conjugate Gradient Methods -- 5. Acceleration of Conjugate Gradient Algorithms -- 6. Hybrid and Parameterized Conjugate Gradient Methods -- 7. Conjugate Gradient Methods as Modifications of the Standard Schemes -- 8. Conjugate Gradient Methods Memoryless BFGS Preconditioned -- 9. Three-Term Conjugate Gradient Methods -- 10. Other Conjugate Gradient Methods -- 11. Discussion and Conclusions -- References -- Author Index -- Subject Index 
653 |a Optimization 
653 |a Mathematical Modeling and Industrial Mathematics 
653 |a Mathematical optimization 
653 |a Mathematical models 
041 0 7 |a eng  |2 ISO 639-2 
989 |b Springer  |a Springer eBooks 2005- 
490 0 |a Springer Optimization and Its Applications 
028 5 0 |a 10.1007/978-3-030-42950-8 
856 4 0 |u https://doi.org/10.1007/978-3-030-42950-8?nosfx=y  |x Verlag  |3 Volltext 
082 0 |a 519.6 
520 |a For each method, the convergence analysis, the computational performances and the comparisons versus other conjugate gradient methods are given. The theory behind the conjugate gradient algorithms presented as a methodology is developed with a clear, rigorous, and friendly exposition; the reader will gain an understanding of their properties and their convergence and will learn to develop and prove the convergence of his/her own methods. Numerous numerical studies are supplied with comparisons and comments on the behavior of conjugate gradient algorithms for solving a collection of 800 unconstrained optimization problems of different structures and complexities with the number of variables in the range [1000,10000]. The book is addressed to all those interested in developing and using new advanced techniques for solving unconstrained optimization complex problems.  
520 |a Two approaches are known for solving large-scale unconstrained optimization problems—the limited-memory quasi-Newton method (truncated Newton method) and the conjugate gradient method. This is the first book to detail conjugate gradient methods, showing their properties and convergence characteristics as well as their performance in solving large-scale unconstrained optimization problems and applications. Comparisons to the limited-memory and truncated Newton methods are also discussed. Topics studied in detail include: linear conjugate gradient methods, standard conjugate gradient methods, acceleration of conjugate gradient methods, hybrid, modifications of the standard scheme, memoryless BFGS preconditioned, and three-term. Other conjugate gradient methods with clustering the eigenvalues or with the minimization of the condition number of the iteration matrix, are also treated.  
520 |a Mathematical programming researchers, theoreticians and practitioners in operations research, practitioners in engineering and industry researchers, as well as graduate students in mathematics,Ph.D. and master students in mathematical programming, will find plenty of information and practical applications for solving large-scale unconstrained optimization problems and applications by conjugate gradient methods