Mixtures of Conditional Maximum Entropy Models

Dmitry Pavlov, Alexandrin Popescul, David M. Pennock, Lyle H. Ungar

Research output: Chapter in Book/Report/Conference proceedingConference contribution

10 Scopus citations

Abstract

Driven by successes in several application areas, maximum entropy modeling has recently gained considerable popularity. We generalize the standard maximum entropy formulation of classification problems to better handle the case where complex data distributions arise from a mixture of simpler underlying (latent) distributions. We develop a theoretical framework for characterizing data as a mixture of maximum entropy models. We formulate a maximum-likelihood interpretation of the mixture model learning, and derive a generalized EM algorithm to solve the corresponding optimization problem. We present empirical results for a number of data sets showing that modeling the data as a mixture of latent maximum entropy models gives significant improvement over the standard, single component, maximum entropy approach. Mixture model, maximum entropy, latent structure, classification.

Original languageEnglish (US)
Title of host publicationProceedings, Twentieth International Conference on Machine Learning
EditorsT. Fawcett, N. Mishra
Pages584-591
Number of pages8
StatePublished - 2003
Externally publishedYes
EventProceedings, Twentieth International Conference on Machine Learning - Washington, DC, United States
Duration: Aug 21 2003Aug 24 2003

Publication series

NameProceedings, Twentieth International Conference on Machine Learning
Volume2

Conference

ConferenceProceedings, Twentieth International Conference on Machine Learning
Country/TerritoryUnited States
CityWashington, DC
Period8/21/038/24/03

All Science Journal Classification (ASJC) codes

  • General Engineering

Fingerprint

Dive into the research topics of 'Mixtures of Conditional Maximum Entropy Models'. Together they form a unique fingerprint.

Cite this