Language Models in Plain English

Recent advances in machine learning have lowered the barriers to creating and using ML models. But understanding what these models are doing has only become more difficult. We discuss technological advances with little understanding of how they work and struggle to develop a comfortable intuition fo...

Full description

Bibliographic Details
Main Authors: Eovito, Austin, Danilevsky, Marina (Author)
Format: eBook
Language:English
Published: O'Reilly Media, Inc. 2021
Edition:1st edition
Subjects:
Online Access:
Collection: O'Reilly - Collection details see MPG.ReNa
LEADER 02326nmm a2200289 u 4500
001 EB002003971
003 EBX01000000000000001166872
005 00000000000000.0
007 cr|||||||||||||||||||||
008 211025 ||| eng
050 4 |a Q325.5 
100 1 |a Eovito, Austin 
245 0 0 |a Language Models in Plain English  |h [electronic resource]  |c Eovito, Austin 
250 |a 1st edition 
260 |b O'Reilly Media, Inc.  |c 2021 
300 |a 65 pages 
653 |a Machine learning / http://id.loc.gov/authorities/subjects/sh85079324 
653 |a Machine learning / fast 
653 |a Apprentissage automatique 
700 1 |a Danilevsky, Marina  |e author 
041 0 7 |a eng  |2 ISO 639-2 
989 |b OREILLY  |a O'Reilly 
500 |a Made available through: Safari, an O'Reilly Media Company 
776 |z 9781098109066 
856 4 0 |u https://learning.oreilly.com/library/view/~/9781098109073/?ar  |x Verlag  |3 Volltext 
082 0 |a 006.3/1 
520 |a Recent advances in machine learning have lowered the barriers to creating and using ML models. But understanding what these models are doing has only become more difficult. We discuss technological advances with little understanding of how they work and struggle to develop a comfortable intuition for new functionality. In this report, authors Austin Eovito and Marina Danilevsky from IBM focus on how to think about neural network-based language model architectures. They guide you through various models (neural networks, RNN/LSTM, encoder-decoder, attention/transformers) to convey a sense of their abilities without getting entangled in the complex details. The report uses simple examples of how humans approach language in specific applications to explore and compare how different neural network-based language models work. This report will empower you to better understand how machines understand language. Dive deep into the basic task of a language model to predict the next word, and use it as a lens to understand neural network language models Explore encoder-decoder architecture through abstractive text summarization Use machine translation to understand the attention mechanism and transformer architecture Examine the current state of machine language understanding to discern what these language models are good at and their risks and weaknesses