Return to search

Learning without labels and nonnegative tensor factorization

Supervised learning tasks like building a classifier, estimating the error rate of the
predictors, are typically performed with labeled data. In most cases, obtaining labeled data
is costly as it requires manual labeling. On the other hand, unlabeled data is available in
abundance. In this thesis, we discuss methods to perform supervised learning tasks with
no labeled data. We prove consistency of the proposed methods and demonstrate its applicability
with synthetic and real world experiments. In some cases, small quantities of labeled data maybe easily available and supplemented with large quantities of unlabeled data (semi-supervised learning). We derive the asymptotic efficiency of generative models for semi-supervised learning and quantify the effect of labeled and unlabeled data on the quality of the estimate. Another independent track of the thesis is efficient computational methods for nonnegative tensor factorization (NTF). NTF provides the user with rich modeling capabilities but it comes with an added computational cost. We provide a fast algorithm for performing NTF using a modified active set method called block principle pivoting method and demonstrate its applicability to social network analysis and text
mining.

Identiferoai:union.ndltd.org:GATECH/oai:smartech.gatech.edu:1853/33926
Date08 April 2010
CreatorsBalasubramanian, Krishnakumar
PublisherGeorgia Institute of Technology
Source SetsGeorgia Tech Electronic Thesis and Dissertation Archive
Detected LanguageEnglish
TypeThesis

Page generated in 0.002 seconds