Previous abstract | Contents | Next abstract

Models, Kernels, and Algorithms for Discrete Data

When working with discrete data such as text and natural language, statistical machine learning offers two fundamental sets of techniques. Model-based approaches have many advantages from a probabilistic point of view, yet purely discriminative methods such as kernel machines, which are generally "model free," are appealing from the perspective of error rates and algorithms. It is attractive to explore methods that combine them.

In this talk we outline a confluence between these two main streams of research. A new family of kernel methods for statistical learning is presented that exploits the geometric structure of statistical models. In particular, based on the heat equation on the Riemannian manifold defined by the Fisher information metric on a statistical family, we propose a family of kernels that provide a natural way of combining generative statistical modeling with non-parametric discriminative learning. As a special case, the kernels give a new approach to designing learning algorithms for discrete data.

In addition to presenting new results, the talk will give an overview of some of the main developments in recent years in these two research areas as they relate to computational language learning.

References


John Lafferty, Models, Kernels, and Algorithms for Discrete Data. Invited talk presented at CoNLL-2002.
Last update: September 08, 2002. erikt@uia.ua.ac.be