10. J. Lasserre: Principled Hybrids of Generative and Discriminative Models
When labelled training data is plentiful, discriminative techniques are widely used since they give excellent generalization performance. However, for large-scale applications such as object recognition, hand labelling of data is expensive, and there is much interest in semi-supervised techniques based on generative models in which the majority of the training data is unlabelled. Although the generalization performance of generative models can often be improved by 'training them discriminatively', they can then no longer make use of unlabelled data. In an attempt to gain the benefit of both generative and discriminative approaches, heuristic procedures have been proposed which interpolate between these two extremes by taking a convex combination of the generative and discriminative objective functions.
We adopt a new perspective which says that there is only one correct way to
train a model, and that a 'discriminatively trained' generative model is
fundamentally a new model. From this viewpoint, generative and
discriminative models correspond to specific choices for the prior over
parameters. As well as giving a principled interpretation of
'discriminative training', this approach opens door to very general ways of
interpolating between generative and discriminative extremes through
alternative choices of prior.