Performance guarantees for regularized maximum entropy density estimation

Publication Type:

Conference Paper

Source:

17th Annual Conference on Learning Theory, COLT, Springer Berlin Heidelberg, Banff, Canada; July 1-4, 2004, p.472-486 (2004)

ISBN:

978-3-540-22282-8

Call Number:

U04DUD01IDUS

URL:

http://www.cs.princeton.edu/~mdudik/DudikPhSc04.pdf

Keywords:

distribution probability, maxent, maximum entropy

Abstract:

The authors consider the problem of estimating an unknown probability distribution from samples using the principle of maximum entropy (maxent). To alleviate overfitting with a very large number of features, they propose applying the maxent principle with relaxed constraints on the expectations of the features. By convex duality, this turns out to be equivalent to finding the Gibbs distribution minimizing a regularized version of the empirical log loss. They prove nonasymptotic bounds showing that, with respect to the true underlying distribution, this relaxed version of maxent produces density estimates that are almost as good as the best possible. These bounds are in terms of the deviation of the feature empirical averages relative to their true expectations, a number that can be bounded using standard uniform-convergence techniques. In particular, this leads to bounds that drop quickly with the number of samples and that depend very moderately on the number or complexity of the features. The authors also derive and prove convergence for both sequential-update and parallel-update algorithms. Finally, they briefly describe experiments on data relevant to the modeling of species’ geographical distributions.

Notes:

ELECTRONIC FILE - Zoology

Book title: Learning theory: proceedings of the 17th Annual Conference on Computational Learning Theory, COLT 2004, Banff, Canada, July 1-4, 2004. 654 pp.
Series: Lecture Notes in Computer Science, Vol. 3120
Subseries: Lecture Notes in Artificial Intelligence