|
|
|
|
LEADER |
02255nmm a2200409 u 4500 |
001 |
EB002154901 |
003 |
EBX01000000000000001293027 |
005 |
00000000000000.0 |
007 |
cr||||||||||||||||||||| |
008 |
230404 ||| eng |
020 |
|
|
|a 1484290631
|
050 |
|
4 |
|a QA279.5
|
100 |
1 |
|
|a Liu, Peng
|
245 |
0 |
0 |
|a Bayesian optimization
|b theory and practice using Python
|c Peng Liu
|
260 |
|
|
|a New York, NY
|b Apress
|c 2023
|
300 |
|
|
|a xv, 234 pages
|b illustrations (black and white, and colour)
|
505 |
0 |
|
|a Chapter 1: Bayesian Optimization Overview -- Chapter 2: Gaussian Process -- Chapter 3: Bayesian Decision Theory and Expected Improvement -- Chapter 4 : Gaussian Process Regression with GPyTorch -- Chapter 5: Monte Carlo Acquisition Function with Sobol Sequences and Random Restart -- Chapter 6 : Knowledge Gradient: Nested Optimization versus One-shot Learning -- Chapter 7 : Case Study: Tuning CNN Learning Rate with BoTorch
|
653 |
|
|
|a Optimisation mathématique
|
653 |
|
|
|a Bayesian statistical decision theory / Data processing
|
653 |
|
|
|a Mathematical optimization / http://id.loc.gov/authorities/subjects/sh85082127
|
653 |
|
|
|a Python (Computer program language) / fast
|
653 |
|
|
|a Python (Computer program language) / http://id.loc.gov/authorities/subjects/sh96008834
|
653 |
|
|
|a Mathematical optimization / fast
|
653 |
|
|
|a Théorie de la décision bayésienne / Informatique
|
653 |
|
|
|a Python (Langage de programmation)
|
653 |
|
|
|a Bayesian statistical decision theory / Data processing / fast
|
041 |
0 |
7 |
|a eng
|2 ISO 639-2
|
989 |
|
|
|b OREILLY
|a O'Reilly
|
500 |
|
|
|a Includes index
|
024 |
8 |
|
|a 10.1007/978-1-4842-9063-7
|
776 |
|
|
|z 9781484290620
|
776 |
|
|
|z 1484290623
|
776 |
|
|
|z 9781484290637
|
776 |
|
|
|z 1484290631
|
856 |
4 |
0 |
|u https://learning.oreilly.com/library/view/~/9781484290637/?ar
|x Verlag
|3 Volltext
|
082 |
0 |
|
|a 519.5/42
|
520 |
|
|
|a This book covers the essential theory and implementation of popular Bayesian optimization techniques in an intuitive and well-illustrated manner. The techniques covered in this book will enable you to better tune the hyperparemeters of your machine learning models and learn sample-efficient approaches to global optimization
|