Distributed Machine Learning and Gradient Optimization

This book presents the state of the art in distributed machine learning algorithms that are based on gradient optimization methods. In the big data era, large-scale datasets pose enormous challenges for the existing machine learning systems. As such, implementing machine learning algorithms in a dis...

Full description

Bibliographic Details
Main Authors: Jiang, Jiawei, Cui, Bin (Author), Zhang, Ce (Author)
Format: eBook
Language:English
Published: Singapore Springer Nature Singapore 2022, 2022
Edition:1st ed. 2022
Series:Big Data Management
Subjects:
Online Access:
Collection: Springer eBooks 2005- - Collection details see MPG.ReNa
LEADER 02398nmm a2200349 u 4500
001 EB002011541
003 EBX01000000000000001174440
005 00000000000000.0
007 cr|||||||||||||||||||||
008 220303 ||| eng
020 |a 9789811634208 
100 1 |a Jiang, Jiawei 
245 0 0 |a Distributed Machine Learning and Gradient Optimization  |h Elektronische Ressource  |c by Jiawei Jiang, Bin Cui, Ce Zhang 
250 |a 1st ed. 2022 
260 |a Singapore  |b Springer Nature Singapore  |c 2022, 2022 
300 |a XI, 169 p. 1 illus  |b online resource 
505 0 |a 1 Introduction -- 2 Basics of Distributed Machine Learning -- 3 Distributed Gradient Optimization Algorithms -- 4 Distributed Machine Learning Systems -- 5 Conclusion. 
653 |a Machine learning 
653 |a Machine Learning 
653 |a Database Management 
653 |a Data mining 
653 |a Data Mining and Knowledge Discovery 
653 |a Database management 
700 1 |a Cui, Bin  |e [author] 
700 1 |a Zhang, Ce  |e [author] 
041 0 7 |a eng  |2 ISO 639-2 
989 |b Springer  |a Springer eBooks 2005- 
490 0 |a Big Data Management 
028 5 0 |a 10.1007/978-981-16-3420-8 
856 4 0 |u https://doi.org/10.1007/978-981-16-3420-8?nosfx=y  |x Verlag  |3 Volltext 
082 0 |a 006.31 
520 |a This book presents the state of the art in distributed machine learning algorithms that are based on gradient optimization methods. In the big data era, large-scale datasets pose enormous challenges for the existing machine learning systems. As such, implementing machine learning algorithms in a distributed environment has become a key technology, and recent research has shown gradient-based iterative optimization to be an effective solution. Focusing on methods that can speed up large-scale gradient optimization through both algorithm optimizations and careful system implementations, the book introduces three essential techniques in designing a gradient optimization algorithm to train a distributed machine learning model: parallel strategy, data compression and synchronization protocol. Written in a tutorial style, it covers a range of topics, from fundamental knowledge to a number of carefully designed algorithms and systems of distributed machine learning. It will appeal toa broad audience in the field of machine learning, artificial intelligence, big data and database management