Cos 511: Theoretical Machine Learning 2 Relative Entropy and Chernoff Bounds 2.1 Relative Entropy
نویسنده
چکیده
In the previous lecture, we saw that PAC learning is not enough to model real learning problems. We may not desire or even may not be able to find a hypothesis that is consistent with the training data. So we introduced a more general model in which the data is generated as a pair (x, y) from an unknown distribution D. And we defined the generalization error to be errD(h) = Pr(x,y)∼D[h(x) 6= y] and also the empirical error of the training set x1, ..., xm to be the fraction of mistakes on the training set . ˆ err(h) = 1 m |{i : h(xi) 6= yi|. We also showed that in order to get an appropriate bound for the generalization error, it is enough to show that |err(h) − ˆ err(h)| ≤ . And today, we will introduce some powerful tools to find the desired bounds.
منابع مشابه
Some properties of the parametric relative operator entropy
The notion of entropy was introduced by Clausius in 1850, and some of the main steps towards the consolidation of the concept were taken by Boltzmann and Gibbs. Since then several extensions and reformulations have been developed in various disciplines with motivations and applications in different subjects, such as statistical mechanics, information theory, and dynamical systems. Fujii and Kam...
متن کاملA note on inequalities for Tsallis relative operator entropy
In this short note, we present some inequalities for relative operator entropy which are generalizations of some results obtained by Zou [Operator inequalities associated with Tsallis relative operator entropy, {em Math. Inequal. Appl.} {18} (2015), no. 2, 401--406]. Meanwhile, we also show some new lower and upper bounds for relative operator entropy and Tsallis relative o...
متن کاملEntropy-based Consensus for Distributed Data Clustering
The increasingly larger scale of available data and the more restrictive concerns on their privacy are some of the challenging aspects of data mining today. In this paper, Entropy-based Consensus on Cluster Centers (EC3) is introduced for clustering in distributed systems with a consideration for confidentiality of data; i.e. it is the negotiations among local cluster centers that are used in t...
متن کاملIRWIN AND JOAN JACOBS CENTER FOR COMMUNICATION AND INFORMATION TECHNOLOGIES An Information-Theoretic Perspective of the Poisson Approximation via the Chen- Stein Method
The first part of this work considers the entropy of the sum of (possibly dependent and non-identically distributed) Bernoulli random variables. Upper bounds on the error that follows from an approximation of this entropy by the entropy of a Poisson random variable with the same mean are derived via the Chen-Stein method. The second part of this work derives new lower bounds on the total variat...
متن کاملAn Information-Theoretic Perspective of the Poisson Approximation via the Chen-Stein Method
The first part of this work considers the entropy of the sum of (possibly dependent and non-identically distributed) Bernoulli random variables. Upper bounds on the error that follows from an approximation of this entropy by the entropy of a Poisson random variable with the same mean are derived via the Chen-Stein method. The second part of this work derives new lower bounds on the total variat...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2008