Real world data is not random: The variability in the data-sets that arise in computer vision,
signal processing and other areas is often highly constrained and governed by a number of
degrees of freedom that is much smaller than the superficial dimensionality of the data.
Unsupervised learning methods can be used to automatically discover the “true”, underlying
structure in such data-sets and are therefore a central component in many systems that deal
with high-dimensional data.
In this thesis we develop several new approaches to modeling the low-dimensional structure
in data. We introduce a new non-parametric framework for latent variable modelling, that in
contrast to previous methods generalizes learned embeddings beyond the training data and its
latent representatives. We show that the computational complexity for learning and applying
the model is much smaller than that of existing methods, and we illustrate its applicability
on several problems.
We also show how we can introduce supervision signals into latent variable models using
conditioning. Supervision signals make it possible to attach “meaning” to the axes of a latent
representation and to untangle the factors that contribute to the variability in the data. We
develop a model that uses conditional latent variables to extract rich distributed representations
of image transformations, and we describe a new model for learning transformation
features in structured supervised learning problems.
Identifer | oai:union.ndltd.org:TORONTO/oai:tspace.library.utoronto.ca:1807/11118 |
Date | 28 July 2008 |
Creators | Memisevic, Roland |
Contributors | Hinton, Geoffrey |
Source Sets | University of Toronto |
Language | en_ca |
Detected Language | English |
Type | Thesis |
Format | 7564985 bytes, application/pdf |
Page generated in 0.0018 seconds