Not Logged In

Generalized Optimal Reverse Prediction

Full Text: white12.pdf PDF

Recently it has been shown that classical supervised and unsupervised training methods can be unified as special cases of so-called “optimal reverse prediction": predicting inputs from target labels while optimizing over both model parameters and missing labels. Although this perspective establishes links between classical training principles, the existing formulation only applies to linear predictors under squared loss, hence is extremely limited. We generalize the formulation of optimal reverse prediction to arbitrary Bregman divergences, and more importantly to nonlinear predictors. This extension is achieved by establishing a new, generalized form of forward-reverse minimization equivalence that holds for arbitrary matching losses. Several benefits follow. First, a new variant of Bregman divergence clustering can be recovered that incorporates a non-linear data reconstruction model. Second, normalized-cut and kernel-based extensions can be formulated coherently. Finally, a new semi-supervised training principle can be recovered for classification problems that demonstrates advantages over the state of the art.

Citation

M. White, D. Schuurmans. "Generalized Optimal Reverse Prediction". Artificial Intelligence and Statistics, (ed: Neil D. Lawrence, Mark A. Girolami), pp 1305-1313, April 2012.

Keywords:  
Category: In Conference
Web Links: PMLR

BibTeX

@incollection{White+Schuurmans:AISTATS12,
  author = {Martha White and Dale Schuurmans},
  title = {Generalized Optimal Reverse Prediction},
  Editor = {Neil D. Lawrence, Mark A. Girolami},
  Pages = {1305-1313},
  booktitle = {Artificial Intelligence and Statistics},
  year = 2012,
}

Last Updated: February 25, 2020
Submitted by Sabina P

University of Alberta Logo AICML Logo