A Derivative-free Two Level Random Search Method for Unconstrained Optimization

It is assumed that the minimizing function is continuous, lowerbounded and its minimum value is known. Chapter 3 proves the convergence of the algorithm. In Chapter 4, the numerical performances of the algorithm are shown for solving 140 unconstrained optimization problems, out of which 16 are real...

Full description

Bibliographic Details
Main Author: Andrei, Neculai
Format: eBook
Language:English
Published: Cham Springer International Publishing 2021, 2021
Edition:1st ed. 2021
Series:SpringerBriefs in Optimization
Subjects:
Online Access:
Collection: Springer eBooks 2005- - Collection details see MPG.ReNa
LEADER 03491nmm a2200337 u 4500
001 EB001960120
003 EBX01000000000000001123022
005 00000000000000.0
007 cr|||||||||||||||||||||
008 210407 ||| eng
020 |a 9783030685171 
100 1 |a Andrei, Neculai 
245 0 0 |a A Derivative-free Two Level Random Search Method for Unconstrained Optimization  |h Elektronische Ressource  |c by Neculai Andrei 
250 |a 1st ed. 2021 
260 |a Cham  |b Springer International Publishing  |c 2021, 2021 
300 |a XI, 118 p. 14 illus., 13 illus. in color  |b online resource 
505 0 |a 1. Introduction -- 2. A Derivative-free Two Level Random Search Method for Unconstrained Optimization -- 3. Convergence of the Algorithm -- 4. Numerical Results -- 5. Conclusions -- Annex A. List of Applications -- Annex B. List of Test Functions -- Annex C. Detailed Results for 30 Large-Scale Problems -- Annex D. Detailed Results for 140 Problems 
653 |a Operations Research, Management Science 
653 |a Operations research 
653 |a Optimization 
653 |a Management science 
653 |a Mathematical optimization 
041 0 7 |a eng  |2 ISO 639-2 
989 |b Springer  |a Springer eBooks 2005- 
490 0 |a SpringerBriefs in Optimization 
028 5 0 |a 10.1007/978-3-030-68517-1 
856 4 0 |u https://doi.org/10.1007/978-3-030-68517-1?nosfx=y  |x Verlag  |3 Volltext 
082 0 |a 519.6 
520 |a It is assumed that the minimizing function is continuous, lowerbounded and its minimum value is known. Chapter 3 proves the convergence of the algorithm. In Chapter 4, the numerical performances of the algorithm are shown for solving 140 unconstrained optimization problems, out of which 16 are real applications. This shows that the optimization process has two phases: the reduction phase and the stalling one. Finally, the performances of the algorithm for solving a number of 30 large-scale unconstrained optimization problems up to 500 variables are presented. These numerical results show that this approach based on the two level random search method for unconstrained optimization is able to solve a large diversity of problems with different structures and complexities.  
520 |a The book is intended for graduate students and researchers in mathematics, computer science, and operational research. The book presents a new derivative-free optimization method/algorithm based on randomly generated trial points in specified domains and where the best ones are selected at each iteration by using a number of rules. This method is different from many other well established methods presented in the literature and proves to be competitive for solving many unconstrained optimization problems with different structures and complexities, with a relative large number of variables. Intensive numerical experiments with 140 unconstrained optimization problems, with up to 500 variables, have shown that this approach is efficient and robust. Structured into 4 chapters, Chapter 1 is introductory. Chapter 2 is dedicated to presenting a two level derivative-free random search method for unconstrained optimization.  
520 |a There are a number of open problems which refer to the following aspects: the selection of the number of trial or the number of the local trial points, the selection of the bounds of the domains where the trial points and thelocal trial points are randomly generated and a criterion for initiating the line search