Neural Networks with Model Compression

Deep learning has achieved impressive results in image classification, computer vision and natural language processing. To achieve better performance, deeper and wider networks have been designed, which increase the demand for computational resources. The number of floating-point operations (FLOPs)...

Full description

Bibliographic Details
Main Authors: Zhang, Baochang, Wang, Tiancheng (Author), Xu, Sheng (Author), Doermann, David (Author)
Format: eBook
Language:English
Published: Singapore Springer Nature Singapore 2024, 2024
Edition:1st ed. 2024
Series:Computational Intelligence Methods and Applications
Subjects:
Online Access:
Collection: Springer eBooks 2005- - Collection details see MPG.ReNa
LEADER 03107nmm a2200385 u 4500
001 EB002198970
003 EBX01000000000000001336173
005 00000000000000.0
007 cr|||||||||||||||||||||
008 240301 ||| eng
020 |a 9789819950683 
100 1 |a Zhang, Baochang 
245 0 0 |a Neural Networks with Model Compression  |h Elektronische Ressource  |c by Baochang Zhang, Tiancheng Wang, Sheng Xu, David Doermann 
250 |a 1st ed. 2024 
260 |a Singapore  |b Springer Nature Singapore  |c 2024, 2024 
300 |a IX, 260 p. 101 illus., 67 illus. in color  |b online resource 
505 0 |a Chapter 1. Introduction -- Chapter 2. Binary Neural Networks -- Chapter 3. Binary Neural Architecture Search -- Chapter 4. Quantization of Neural Networks -- Chapter 5. Network Pruning -- Chapter 6. Applications 
653 |a Machine learning 
653 |a Image processing / Digital techniques 
653 |a Machine Learning 
653 |a Computer vision 
653 |a Artificial Intelligence 
653 |a Computer Vision 
653 |a Computer Imaging, Vision, Pattern Recognition and Graphics 
653 |a Artificial intelligence 
700 1 |a Wang, Tiancheng  |e [author] 
700 1 |a Xu, Sheng  |e [author] 
700 1 |a Doermann, David  |e [author] 
041 0 7 |a eng  |2 ISO 639-2 
989 |b Springer  |a Springer eBooks 2005- 
490 0 |a Computational Intelligence Methods and Applications 
028 5 0 |a 10.1007/978-981-99-5068-3 
856 4 0 |u https://doi.org/10.1007/978-981-99-5068-3?nosfx=y  |x Verlag  |3 Volltext 
082 0 |a 006.31 
520 |a Deep learning has achieved impressive results in image classification, computer vision and natural language processing. To achieve better performance, deeper and wider networks have been designed, which increase the demand for computational resources. The number of floating-point operations (FLOPs) has increased dramatically with larger networks, and this has become an obstacle for convolutional neural networks (CNNs) being developed for mobile and embedded devices. In this context, our book will focus on CNN compression and acceleration, which are important for the research community. We will describe numerous methods, including parameter quantization, network pruning, low-rank decomposition and knowledge distillation. More recently, to reduce the burden of handcrafted architecture design, neural architecture search (NAS) has been used to automatically build neural networks by searching over a vast architecture space. Our book will also introduce NAS due to its superiority and state-of-the-art performance in various applications, such as image classification and object detection. We also describe extensive applications of compressed deep models on image classification, speech recognition, object detection and tracking. These topics can help researchers better understand the usefulness and the potential of network compression on practical applications. Moreover, interested readers should have basic knowledge about machine learning and deep learning to better understand the methods described in this book