|
|
|
|
LEADER |
01993nmm a2200277 u 4500 |
001 |
EB002071389 |
003 |
EBX01000000000000001211479 |
005 |
00000000000000.0 |
007 |
cr||||||||||||||||||||| |
008 |
220922 ||| eng |
020 |
|
|
|a 0262351609
|
020 |
|
|
|a 9780262351607
|
050 |
|
4 |
|a QA278
|
100 |
1 |
|
|a Aravkin, Aleksandr
|e editor
|
245 |
0 |
0 |
|a Log-linear models, extensions, and applications
|h Elektronische Ressource
|c edited by Aleksandr Aravkin [and six others]
|
260 |
|
|
|a Cambridge
|b MIT Press
|c 2018
|
300 |
|
|
|a 214 pages
|
653 |
|
|
|a Multivariate analysis
|
653 |
|
|
|a Log-linear models
|
041 |
0 |
7 |
|a eng
|2 ISO 639-2
|
989 |
|
|
|b MITArchiv
|a MIT Press eBook Archive
|
490 |
0 |
|
|a Neural information processing series
|
028 |
5 |
0 |
|a 10.7551/mitpress/10012.001.0001
|
856 |
4 |
0 |
|u https://doi.org/10.7551/mitpress/10012.001.0001?locatt=mode:legacy
|x Verlag
|3 Volltext
|
082 |
0 |
|
|a 519.5/35
|
520 |
|
|
|a Log-linear models play a key role in modern big data and machine learning applications. From simple binary classification models through partition functions, conditional random fields, and neural nets, log-linear structure is closely related to performance in certain applications and influences fitting techniques used to train models. This volume covers recent advances in training models with log-linear structures, cover the underlying geometry, optimization techniques, and multiple applications. The first chapter shows readers the inner workings of machine learning, providing insights into the geometry of log-linear and neural net models. The other chapters range from introductory material to optimization techniques to involved use cases. The book, which grew out of a NIPS workshop, is suitable for graduate students doing research in machine learning, in particular deep learning, variable selection, and applications for speech recognition. The contributors come from academia and industry, allowing readers to view the field from both perspectives
|