431 |
Machine Learning for Aerial Image LabelingMnih, Volodymyr 09 August 2013 (has links)
Information extracted from aerial photographs has found applications in a wide
range of areas including urban planning, crop and forest management, disaster
relief, and climate modeling. At present, much of the extraction is still
performed by human experts, making the process slow, costly, and error prone.
The goal of this thesis is to develop methods for automatically extracting the
locations of objects such as roads, buildings, and trees directly from aerial
images.
We investigate the use of machine learning methods trained on aligned aerial
images and possibly outdated maps for labeling the pixels of an aerial image
with semantic labels. We show how deep neural networks implemented on modern
GPUs can be used to efficiently learn highly discriminative image features. We
then introduce new loss functions for training neural networks that are
partially robust to incomplete and poorly registered target maps. Finally, we
propose two ways of improving the predictions of our system by introducing
structure into the outputs of the neural networks.
We evaluate our system on the largest and most-challenging road and building
detection datasets considered in the literature and show that it works reliably
under a wide variety of conditions. Furthermore, we are releasing the first
large-scale road and building detection datasets to the public in order to
facilitate future comparisons with other methods.
|
432 |
Learning Accurate Regressors for Predicting Survival Times of Individual Cancer PatientsLin, Hsiu-Chin 06 1900 (has links)
Standard survival analysis focuses on population-based studies. The objective of our work, survival prediction, is different: to find the most accurate model for predicting the survival times for each individual patient. We view this as a regression problem, where we try to map the features for each patient to his/her survival time. This is challenging in medical data due to the presence of irrelevant features, outliers, and missing class labels. Our approach consists of two major steps: (1) apply various grouping methods to segregate patients, and (2) apply different regression to each sub-group we obtained from the first step. We focus our experiments on a data set of 2402 patients (1260 censored). Our final predictor can obtain an average relative absolute error < 0.54. The experimental results verify that we can effectively predict survival times with a combination of statistical and machine learning approaches.
|
433 |
A general framework for reducing variance in agent evaluationWhite, Martha 06 1900 (has links)
In this work, we present a unified, general approach to variance reduction in agent evaluation using machine learning to minimize variance. Evaluating an agent's performance in a stochastic setting is necessary for agent development, scientific evaluation, and competitions. Traditionally, evaluation is done using Monte Carlo estimation (sample averages); the magnitude of the stochasticity in the domain or the high cost of sampling, however, can often prevent the approach from resulting in statistically significant conclusions. Recently, an advantage sum technique based on control variates has been proposed for constructing unbiased, low variance estimates of agent performance. The technique requires an expert to define a value function over states of the system, essentially a guess of the state's unknown value. In this work, we propose learning this value function from past interactions between agents in some target population. Our learned value functions have two key advantages: they can be applied in domains where no expert value function is available and they can result in tuned evaluation for a specific population of agents (e.g., novice versus advanced agents). This work has three main contributions. First, we consolidate previous work in using control variates for variance reduction into one unified, general framework and summarize the connections between this previous work. Second, our framework makes variance reduction practically possible in any sequential decision making task where designing the expert value function is time-consuming, difficult or essentially impossible. We prove the optimality of our approach and extend the theoretical understanding of advantage sum estimators. In addition, we significantly extend the applicability of advantage sum estimators and discuss practical methods for using our framework in real-world scenarios. Finally, we provide low-variance estimators for three poker domains previously without variance reduction and improve strategy selection in the expert-level University of Alberta poker bot.
|
434 |
Learning multi-agent pursuit of a moving targetLu, Jieshan 11 1900 (has links)
In this thesis we consider the task of catching a moving target with multiple pursuers, also known as the “Pursuit Game”, in which coordination among the pursuers is critical. Our testbed is inspired by the pursuit problem in video games, which require fast planning to guarantee fluid frame rates. We apply supervised machine learning methods to automatically derive efficient multi-agent pursuit strategies on rectangular grids. Learning is achieved by computing training data off-line and exploring the game tree on small problems. We also generalize the data to previously unseen and larger problems by learning robust pursuit policies, and run empirical comparisons between several sets of state features using a simple learning architecture. The empirical results show that 1) the application of learning across different maps can help improve game-play performance, especially on non-trivial maps against intelligent targets, and 2) simple heuristic works effectively on simple maps or less intelligent targets.
|
435 |
Data analysis in proteomics novel computational strategies for modeling and interpreting complex mass spectrometry dataSniatynski, Matthew John 11 1900 (has links)
Contemporary proteomics studies require computational approaches to deal with both the complexity of the data generated, and with the volume of data produced. The amalgamation of mass spectrometry -- the analytical tool of choice in proteomics -- with the computational and statistical sciences is still recent, and several avenues of exploratory data analysis and statistical methodology remain relatively unexplored. The current study focuses on three broad analytical domains, and develops novel exploratory approaches and practical tools in each. Data transform approaches are the first explored. These methods re-frame data, allowing for the visualization and exploitation of features and trends that are not immediately evident. An exploratory approach making use of the correlation transform is developed, and is used to identify mass-shift signals in mass spectra. This approach is used to identify and map post-translational modifications on individual peptides, and to identify SILAC modification-containing spectra in a full-scale proteomic analysis. Secondly, matrix decomposition and projection approaches are explored; these use an eigen-decomposition to extract general trends from groups of related spectra. A data visualization approach is demonstrated using these techniques, capable of visualizing trends in large numbers of complex spectra, and a data compression and feature extraction technique is developed suitable for use in spectral modeling. Finally, a general machine learning approach is developed based on conditional random fields (CRFs). These models are capable of dealing with arbitrary sequence modeling tasks, similar to hidden Markov models (HMMs), but are far more robust to interdependent observational features, and do not require limiting independence assumptions to remain tractable. The theory behind this approach is developed, and a simple machine learning fragmentation model is developed to test the hypothesis that reproducible sequence-specific intensity ratios are present within the distribution of fragment ions originating from a common peptide bond breakage. After training, the model shows very good performance associating peptide sequences and fragment ion intensity information, lending strong support to the hypothesis.
|
436 |
Learning and discovery in incremental knowledge acquisitionSuryanto, Hendra, Computer Science & Engineering, Faculty of Engineering, UNSW January 2005 (has links)
Knowledge Based Systems (KBS) have been actively investigated since the early period of AI. There are four common methods of building expert systems: modeling approaches, programming approaches, case-based approaches and machine-learning approaches. One particular technique is Ripple Down Rules (RDR) which may be classified as an incremental case-based approach. Knowledge needs to be acquired from experts in the context of individual cases viewed by them. In the RDR framework, the expert adds a new rule based on the context of an individual case. This task is simple and only affects the expert???s workflow minimally. The rule added fixes an incorrect interpretation made by the KBS but with minimal impact on the KBS's previous correct performance. This provides incremental improvement. Despite these strengths of RDR, there are some limitations including rule redundancy, lack of intermediate features and lack of models. This thesis addresses these RDR limitations by applying automatic learning algorithms to reorganize the knowledge base, to learn intermediate features and possibly to discover domain models. The redundancy problem occurs because rules created in particular contexts which should have more general application. We address this limitation by reorganizing the knowledge base and removing redundant rules. Removal of redundant rules should also reduce the number of future knowledge acquisition sessions. Intermediate features improve modularity, because the expert can deal with features in groups rather than individually. In addition to the manual creation of intermediate features for RDR, we propose the automated discovery of intermediate features to speed up the knowledge acquisition process by generalizing existing rules. Finally, the Ripple Down Rules approach facilitates rapid knowledge acquisition as it can be initialized with a minimal ontology. Despite minimal modeling, we propose that a more developed knowledge model can be extracted from an existing RDR KBS. This may be useful in using RDR KBS for other applications. The most useful of these three developments was the automated discovery of intermediate features. This made a significant difference to the number of knowledge acquisition sessions required.
|
437 |
Behavioural cloning robust goal directed controlIsaac, Andrew Paul, Computer Science & Engineering, Faculty of Engineering, UNSW January 2009 (has links)
Behavioural cloning is a simple and effective technique for automatically and non-intrusively producing comprehensible and implementable models of human control skill. Behavioural cloning applies machine learning techniques to behavioural trace data, in a transparent manner, and has been very successful in a wide range of domains. The limitations of early behavioural cloning work are: that the clones lack goal-structure, are not robust to variation, are sensitive to the nature of the training data and often produce complicated models of the control skill. Recent behavioural cloning work has sought to address these limitations by adopting goal-structured task decompositions and combining control engineering representations with more sophisticated machine learning algorithms. These approaches have had some success but by compromising either transparency or robustness. This thesis addresses these limitations by investigating: new behavioural cloning representations, control structures, data processing techniques, machine learning algorithms, and performance estimation and testing techniques. First a novel hierarchical decomposition of control is developed, where goal settings and the control skill to achieve them are learnt. This decomposition allows feedback control mechanisms to be combined with modular goal-achievement. Data processing limitations are addressed by developing data-driven, correlative and sampling techniques, that also inform the development of the learning algorithm. The behavioural cloning process is developed by performing experiments on simulated aircraft piloting tasks, and then the generality of the process is tested by performing experiments on simulated gantry-crane control tasks. The performance of the behavioural cloning process was compared to existing techniques, and demonstrated a marked improvement over these methods. The system is capable of handling novel goal-settings and task structure, under high noise conditions. The ability to produce successful controllers was greatly improved by using the developed control representation, data processing and learning techniques. The models produced are compact but tend to abstract the originating control behaviour. In conclusion, the control representation and cloning process address current limitations of behavioural cloning, and produce reliable, reusable and readable clones.
|
438 |
Graphical Models: Modeling, Optimization, and Hilbert Space EmbeddingZhang, Xinhua, xinhua.zhang.cs@gmail.com January 2010 (has links)
Over the past two decades graphical models have been widely used as powerful tools for compactly representing distributions. On the other hand, kernel methods have been used extensively to come up with rich representations. This thesis aims to combine graphical models with kernels to produce compact models with rich representational abilities. Graphical models are a powerful underlying formalism in machine learning. Their graph theoretic properties provide both an intuitive modular interface to model the interacting factors, and a data structure facilitating efficient learning and inference. The probabilistic nature ensures the global consistency of the whole framework, and allows convenient interface of models to data. Kernel methods, on the other hand, provide an effective means of representing rich classes of features for general objects, and at the same time allow efficient search for the optimal model. Recently, kernels have been used to characterize distributions by embedding them into high dimensional feature space. Interestingly, graphical models again decompose this characterization and lead to novel and direct ways of comparing distributions based on samples. Among the many uses of graphical models and kernels, this thesis is devoted to the following four areas: Conditional random fields for multi-agent reinforcement learning
Conditional random fields (CRFs) are graphical models for modelling the probability of labels given the observations. They have traditionally been trained with using a set of observation and label pairs. Underlying all CRFs is the assumption that, conditioned on the training data, the label sequences of different training examples are independent and identically distributed (iid ). We extended the use of CRFs to a class of temporal learning algorithms, namely policy gradient reinforcement learning (RL). Now the labels are no longer iid. They are actions that update the environment and affect the next observation. From an RL point of view, CRFs provide a natural way to model joint actions in a decentralized Markov decision process. They define how agents can communicate with each other to choose the optimal joint action. We tested our framework on a synthetic network alignment problem, a distributed sensor network, and a road traffic control system. Using tree sampling by Hamze & de Freitas (2004) for inference, the RL methods employing CRFs clearly outperform those which do not model the proper joint policy.
Bayesian online multi-label classification Gaussian density filtering (GDF) provides fast and effective inference for graphical models (Maybeck, 1982). Based on this natural online learner, we propose a Bayesian online multi-label classification (BOMC) framework which learns a probabilistic model of the linear classifier. The training labels are incorporated to update the posterior of the classifiers via a graphical model similar to TrueSkill (Herbrich et al., 2007), and inference is based on GDF with expectation propagation. Using samples from the posterior, we label the test data by maximizing the expected F-score. Our experiments on Reuters1-v2 dataset show that BOMC delivers significantly higher macro-averaged F-score than the state-of-the-art online maximum margin learners such as LaSVM (Bordes et al., 2005) and passive aggressive online learning (Crammer et al., 2006). The online nature of BOMC also allows us to effciently use a large amount of training data.
Hilbert space embedment of distributions
Graphical models are also an essential tool in kernel measures of independence for non-iid data. Traditional information theory often requires density estimation, which makes it unideal for statistical estimation. Motivated by the fact that distributions often appear in machine learning via expectations, we can characterize the distance between distributions in terms of distances between means, especially means in reproducing kernel Hilbert spaces which are called kernel embedment. Under this framework, the undirected graphical models further allow us to factorize the kernel embedment onto cliques, which yields efficient measures of independence for non-iid data (Zhang et al., 2009). We show the effectiveness of this framework for ICA and sequence segmentation, and a number of further applications and research questions are identified. Optimization in maximum margin models for structured data Maximum margin estimation for structured data, e.g. (Taskar et al., 2004), is an important task in machine learning where graphical models also play a key role. They are special cases of regularized risk minimization, for which bundle methods (BMRM, Teo et al., 2007) and the closely related SVMStruct (Tsochantaridis et al., 2005) are state-of-the-art general purpose solvers. Smola et al. (2007b) proved that BMRM requires O(1/έ) iterations to converge to an έ accurate solution, and we further show that this rate hits the lower bound. By utilizing the structure of the objective function, we devised an algorithm for the structured loss which converges to an έ accurate solution in O(1/√έ) iterations. This algorithm originates from Nesterov's optimal first order methods (Nesterov, 2003, 2005b).
|
439 |
Behavioural cloning robust goal directed controlIsaac, Andrew Paul, Computer Science & Engineering, Faculty of Engineering, UNSW January 2009 (has links)
Behavioural cloning is a simple and effective technique for automatically and non-intrusively producing comprehensible and implementable models of human control skill. Behavioural cloning applies machine learning techniques to behavioural trace data, in a transparent manner, and has been very successful in a wide range of domains. The limitations of early behavioural cloning work are: that the clones lack goal-structure, are not robust to variation, are sensitive to the nature of the training data and often produce complicated models of the control skill. Recent behavioural cloning work has sought to address these limitations by adopting goal-structured task decompositions and combining control engineering representations with more sophisticated machine learning algorithms. These approaches have had some success but by compromising either transparency or robustness. This thesis addresses these limitations by investigating: new behavioural cloning representations, control structures, data processing techniques, machine learning algorithms, and performance estimation and testing techniques. First a novel hierarchical decomposition of control is developed, where goal settings and the control skill to achieve them are learnt. This decomposition allows feedback control mechanisms to be combined with modular goal-achievement. Data processing limitations are addressed by developing data-driven, correlative and sampling techniques, that also inform the development of the learning algorithm. The behavioural cloning process is developed by performing experiments on simulated aircraft piloting tasks, and then the generality of the process is tested by performing experiments on simulated gantry-crane control tasks. The performance of the behavioural cloning process was compared to existing techniques, and demonstrated a marked improvement over these methods. The system is capable of handling novel goal-settings and task structure, under high noise conditions. The ability to produce successful controllers was greatly improved by using the developed control representation, data processing and learning techniques. The models produced are compact but tend to abstract the originating control behaviour. In conclusion, the control representation and cloning process address current limitations of behavioural cloning, and produce reliable, reusable and readable clones.
|
440 |
Building better software the applicability of a professional tool for automating quality assessment and fault detection /Di Stefano, Justin S. January 1900 (has links)
Thesis (M.S.)--West Virginia University, 2008. / Title from document title page. Document formatted into pages; contains vii, 83 p. : ill. (some col.). Vita. Includes abstract. Includes bibliographical references (p. 81-83).
|
Page generated in 0.1394 seconds