|
|
|
|
LEADER |
03592nmm a2200445 u 4500 |
001 |
EB002169395 |
003 |
EBX01000000000000001307172 |
005 |
00000000000000.0 |
007 |
cr||||||||||||||||||||| |
008 |
230808 ||| eng |
020 |
|
|
|a 9783031283949
|
100 |
1 |
|
|a Li, Jinna
|
245 |
0 |
0 |
|a Reinforcement Learning
|h Elektronische Ressource
|b Optimal Feedback Control with Industrial Applications
|c by Jinna Li, Frank L. Lewis, Jialu Fan
|
250 |
|
|
|a 1st ed. 2023
|
260 |
|
|
|a Cham
|b Springer International Publishing
|c 2023, 2023
|
300 |
|
|
|a XVI, 310 p. 114 illus., 110 illus. in color
|b online resource
|
505 |
0 |
|
|a 1. Background on Reinforcement Learning and Optimal Control -- 2. H-infinity Control Using Reinforcement Learning -- 3. Robust Tracking Control and Output Regulation -- 4. Interleaved Robust Reinforcement Learning -- 5. Optimal Networked Controller and Observer Design -- 6. Interleaved Q-Learning -- 7. Off-Policy Game Reinforcement Learning -- 8. Game Reinforcement Learning for Process Industries
|
653 |
|
|
|a Complex Systems
|
653 |
|
|
|a Industrial engineering
|
653 |
|
|
|a Engineering mathematics
|
653 |
|
|
|a Computational intelligence
|
653 |
|
|
|a Control and Systems Theory
|
653 |
|
|
|a Computational Intelligence
|
653 |
|
|
|a Industrial and Production Engineering
|
653 |
|
|
|a System theory
|
653 |
|
|
|a Control engineering
|
653 |
|
|
|a Process Engineering
|
653 |
|
|
|a Engineering / Data processing
|
653 |
|
|
|a Mathematical and Computational Engineering Applications
|
653 |
|
|
|a Production engineering
|
700 |
1 |
|
|a Lewis, Frank L.
|e [author]
|
700 |
1 |
|
|a Fan, Jialu
|e [author]
|
041 |
0 |
7 |
|a eng
|2 ISO 639-2
|
989 |
|
|
|b Springer
|a Springer eBooks 2005-
|
490 |
0 |
|
|a Advances in Industrial Control
|
028 |
5 |
0 |
|a 10.1007/978-3-031-28394-9
|
856 |
4 |
0 |
|u https://doi.org/10.1007/978-3-031-28394-9?nosfx=y
|x Verlag
|3 Volltext
|
082 |
0 |
|
|a 6,298,312
|
082 |
0 |
|
|a 3
|
520 |
|
|
|a This book offers a thorough introduction to the basics and scientific and technological innovations involved in the modern study of reinforcement-learning-based feedback control. The authors address a wide variety of systems including work on nonlinear, networked, multi-agent and multi-player systems. A concise description of classical reinforcement learning (RL), the basics of optimal control with dynamic programming and network control architectures, and a brief introduction to typical algorithms build the foundation for the remainder of the book. Extensive research on data-driven robust control for nonlinear systems with unknown dynamics and multi-player systems follows. Data-driven optimal control of networked single- and multi-player systems leads readers into the development of novel RL algorithms with increased learning efficiency. The book concludes with a treatment of how these RL algorithms can achieve optimal synchronization policies for multi-agentsystems with unknown model parameters and how game RL can solve problems of optimal operation in various process industries. Illustrative numerical examples and complex process control applications emphasize the realistic usefulness of the algorithms discussed. The combination of practical algorithms, theoretical analysis and comprehensive examples presented in Reinforcement Learning will interest researchers and practitioners studying or using optimal and adaptive control, machine learning, artificial intelligence, and operations research, whether advancing the theory or applying it in mineral-process, chemical-process, power-supply or other industries
|