Information theory meets power laws stochastic processes and language models

"This book introduces mathematical foundations of statistical modeling of natural language. The author attempts to explain a few statistical power laws satisfied by texts in natural language in terms of non-Markovian and non-hidden Markovian discrete stochastic processes with some sort of long-...

Full description

Bibliographic Details
Main Author: Dębowski, Łukasz Jerzy
Format: eBook
Language:English
Published: Hoboken, NJ John Wiley & Sons, Inc. 2021
Subjects:
Online Access:
Collection: O'Reilly - Collection details see MPG.ReNa
LEADER 02746nmm a2200481 u 4500
001 EB002012458
003 EBX01000000000000001175357
005 00000000000000.0
007 cr|||||||||||||||||||||
008 220317 ||| eng
020 |a 9781119625384 
020 |a 1119625378 
020 |a 9781119625360 
020 |a 9781119625377 
020 |a 1119625386 
020 |a 111962536X 
050 4 |a P98 
100 1 |a Dębowski, Łukasz Jerzy 
245 0 0 |a Information theory meets power laws  |b stochastic processes and language models  |c Łukasz Dębowski, Polish Academy of Sciences 
260 |a Hoboken, NJ  |b John Wiley & Sons, Inc.  |c 2021 
300 |a xvi, 368 pages  |b illustrations 
505 0 |a Includes bibliographical references and index 
653 |a Stochastic processes / fast 
653 |a Computational linguistics / fast 
653 |a Processus stochastiques 
653 |a Computational linguistics / http://id.loc.gov/authorities/subjects/sh85077224 
653 |a computational linguistics / aat 
653 |a Linguistique informatique 
653 |a Stochastic processes / http://id.loc.gov/authorities/subjects/sh85128181 
653 |a Stochastic Processes 
041 0 7 |a eng  |2 ISO 639-2 
989 |b OREILLY  |a O'Reilly 
015 |a GBC0D9722 
776 |z 111962536X 
776 |z 1119625378 
776 |z 9781119625278 
776 |z 1119625386 
776 |z 9781119625377 
776 |z 9781119625384 
776 |z 9781119625360 
856 4 0 |u https://learning.oreilly.com/library/view/~/9781119625278/?ar  |x Verlag  |3 Volltext 
082 0 |a 410.1/5195 
520 |a "This book introduces mathematical foundations of statistical modeling of natural language. The author attempts to explain a few statistical power laws satisfied by texts in natural language in terms of non-Markovian and non-hidden Markovian discrete stochastic processes with some sort of long-range dependence. To achieve this, he uses various concepts and technical tools from information theory and probability measures. This book begins with an introduction. The first half of the book is an introduction to probability measures, information theory, ergodic decomposition, and Kolmogorov complexity, which is provided to make the book relatively self-contained. This section also covers less standard concepts and results, such as excess entropy and generalization of conditional mutual information to fields. The second part of the book discusses the results concerning power laws for mutual information and maximal repetition, such as theorems about facts and words. There is also a separate chapter discussing toy examples of stochastic processes, which should inspire future work in statistical language modeling"--