## Concentration Inequalities: A Nonasymptotic Theory of IndependenceConcentration inequalities for functions of independent random variables is an area of probability theory that has witnessed a great revolution in the last few decades, and has applications in a wide variety of areas such as machine learning, statistics, discrete mathematics, and high-dimensional geometry. Roughly speaking, if a function of many independent random variables does not depend too much on any of the variables then it is concentrated in the sense that with high probability, it is close to its expected value. This book offers a host of inequalities to illustrate this rich theory in an accessible way by covering the key developments and applications in the field. The authors describe the interplay between the probabilistic structure (independence) and a variety of tools ranging from functional inequalities to transportation arguments to information theory. Applications to the study of empirical processes, random projections, random matrix theory, and threshold phenomena are also presented. A self-contained introduction to concentration inequalities, it includes a survey of concentration of sums of independent random variables, variance bounds, the entropy method, and the transportation method. Deep connections with isoperimetric problems are revealed whilst special attention is paid to applications to the supremum of empirical processes. Written by leading experts in the field and containing extensive exercise sections this book will be an invaluable resource for researchers and graduate students in mathematics, theoretical computer science, and engineering. |

### What people are saying - Write a review

We haven't found any reviews in the usual places.

### Contents

1 Introduction | 1 |

2 Basic Inequalities | 18 |

3 Bounding the Variance | 52 |

4 Basic Information Inequalities | 83 |

5 Logarithmic Sobolev Inequalities | 117 |

6 The Entropy Method | 168 |

7 Concentration and Isoperimetry | 215 |

8 The Transportation Method | 237 |

11 The Variance of Suprema of Empirical Processes | 312 |

Exponential Inequalities | 341 |

13 The Expected Value of Suprema of Empirical Processes | 362 |

14 934Entropies | 412 |

15 Moment Inequalities | 430 |

451 | |

Author Index | 473 |

477 | |

9 Influences and Threshold Phenomena | 262 |

10 Isoperimetry on the Hypercube and Gaussian Spaces | 290 |

### Other editions - View all

Concentration Inequalities: A Nonasymptotic Theory of Independence Stéphane Boucheron,Gábor Lugosi,Pascal Massart No preview available - 2013 |

### Common terms and phrases

argument assume Bernoulli Bernoulli distributions binary hypercube Bobkov’s bounded differences inequality Cauchy–Schwarz inequality Chapter condition constant convex function Corollary defined denote derive differential inequality duality formula Efron–Stein inequality eigenvalue empirical processes entropy method example Exercise expected value finite follows function f graph Hamming distance Hint hypercube independent random variables inequality of Theorem inequality Theorem integrable isoperimetric theorem Jensen's inequality Ledoux Lemma Let f Let X1 log Ee logarithmic Sobolev inequality matrix modified logarithmic Sobolev moment-generating function monotone set norm obtain Poincaré inequality probability measure proof of Theorem prove random matrix random variables taking random vectors Recall result satisfies Section seT i-1 ſº sub-additivity sub-gamma sub-Gaussian subset sums of independent super XX suprema of empirical supremum symmetric tail bounds tail inequality Talagrand total influence upper bound variance factor vertex Xn are independent Z-EZ