181 |
Serial Testing for Detection of Multilocus Genetic InteractionsAl-Khaledi, Zaid T. 01 January 2019 (has links)
A method to detect relationships between disease susceptibility and multilocus genetic interactions is the Multifactor-Dimensionality Reduction (MDR) technique pioneered by Ritchie et al. (2001). Since its introduction, many extensions have been pursued to deal with non-binary outcomes and/or account for multiple interactions simultaneously. Studying the effects of multilocus genetic interactions on continuous traits (blood pressure, weight, etc.) is one case that MDR does not handle. Culverhouse et al. (2004) and Gui et al. (2013) proposed two different methods to analyze such a case. In their research, Gui et al. (2013) introduced the Quantitative Multifactor-Dimensionality Reduction (QMDR) that uses the overall average of response variable to classify individuals into risk groups. The classification mechanism may not be efficient under some circumstances, especially when the overall mean is close to some multilocus means. To address such difficulties, we propose a new algorithm, the Ordered Combinatorial Quantitative Multifactor-Dimensionality Reduction (OQMDR), that uses a series of testings, based on ascending order of multilocus means, to identify best interactions of different orders with risk patterns that minimize the prediction error. Ten-fold cross-validation is used to choose from among the resulting models. Regular permutations testings are used to assess the significance of the selected model. The assessment procedure is also modified by utilizing the Generalized Extreme-Value distribution to enhance the efficiency of the evaluation process. We presented results from a simulation study to illustrate the performance of the algorithm. The proposed algorithm is also applied to a genetic data set associated with Alzheimer's Disease.
|
182 |
Emotion lexicon in the Sepedi, Xitsonga and Tshivenda language groups in South Africa : the impact of culture on emotion / T. NichollsNicholls, Tanja January 2008 (has links)
Thesis (M.A. (Industrial Psychology))--North-West University, Potchefstroom Campus, 2008.
|
183 |
Leveraging attention focus for effective reinforcement learning in complex domainsCobo Rus, Luis Carlos 29 March 2013 (has links)
One of the hardest challenges in the field of machine learning is to build agents, such as robotic assistants in homes and hospitals, that can autonomously learn new tasks that they were not pre-programmed to tackle, without the intervention of an engineer. Reinforcement learning (RL) and learning from demonstration (LfD) are popular approaches for task learning, but they are often ineffective in high-dimensional domains unless provided with either a great deal of problem-specific domain information or a carefully crafted representation of the state and dynamics of the world. Unfortunately, autonomous agents trying to learn new tasks usually do not have access to such domain information nor to an appropriate representation.
We demonstrate that algorithms that focus, at each moment, on the relevant features of the state space can achieve significant speed-ups over previous reinforcement learning algorithms with respect to the number of state features in complex domains. To do so, we introduce and evaluate a family of attention focus algorithms. We show that these algorithms can reduce the dimensionality of complex domains, creating a compact representation of the state space with which satisficing policies can be learned efficiently. Our approach obtains exponential speed-ups with respect to the number of features considered when compared with table-based learning algorithms and polynomial speed-ups when compared with state-of-the-art function approximation RL algorithms such as LSPI or fitted Q-learning.
Our attention focus algorithms are divided in two classes, depending on the source of the focus information they require. Attention focus from human demonstrations infers the features to focus on from a set of demonstrations from human teachers performing the task the agent must learn. We introduce two algorithms within this class. The first one, abstraction from demonstration (AfD), identifies features that can be safely ignored in the whole state space and builds a state-space abstraction where a satisficing policy can be learned efficiently. The second, automatic decomposition and abstraction from demonstration, goes one step further, using the demonstrations to identify a set of subtasks and to find an appropriate abstraction for each subtask found.
The other class of algorithms we present, attention focus with a world model, does not require a set of human demonstrations. Instead, it extracts the attention focus information from an object-based model of the world together with the agent experience in performing the task. Within this class, we introduce object-focused Q-learning (OF-Q), at first with an assumption of object independence that is later removed to support domains where objects interact with each other. Finally, we show that both sources of focus information can be combined for further speed-ups.
|
184 |
Nonlinear Dimensionality Reduction with Side InformationGhodsi Boushehri, Ali January 2006 (has links)
In this thesis, I look at three problems with important applications in data processing. Incorporating side information, provided by the user or derived from data, is a main theme of each of these problems. <br /><br /> This thesis makes a number of contributions. The first is a technique for combining different embedding objectives, which is then exploited to incorporate side information expressed in terms of transformation invariants known to hold in the data. It also introduces two different ways of incorporating transformation invariants in order to make new similarity measures. Two algorithms are proposed which learn metrics based on different types of side information. These learned metrics can then be used in subsequent embedding methods. Finally, it introduces a manifold learning algorithm that is useful when applied to sequential decision problems. In this case we are given action labels in addition to data points. Actions in the manifold learned by this algorithm have meaningful representations in that they are represented as simple transformations.
|
185 |
Nonlinear Dimensionality Reduction with Side InformationGhodsi Boushehri, Ali January 2006 (has links)
In this thesis, I look at three problems with important applications in data processing. Incorporating side information, provided by the user or derived from data, is a main theme of each of these problems. <br /><br /> This thesis makes a number of contributions. The first is a technique for combining different embedding objectives, which is then exploited to incorporate side information expressed in terms of transformation invariants known to hold in the data. It also introduces two different ways of incorporating transformation invariants in order to make new similarity measures. Two algorithms are proposed which learn metrics based on different types of side information. These learned metrics can then be used in subsequent embedding methods. Finally, it introduces a manifold learning algorithm that is useful when applied to sequential decision problems. In this case we are given action labels in addition to data points. Actions in the manifold learned by this algorithm have meaningful representations in that they are represented as simple transformations.
|
186 |
Conjunctive Management of Surface Water and Groundwater ResourcesAbu Rumman, Malek 18 April 2005 (has links)
Surface water and groundwater systems consist of interconnected reservoirs, rivers, and confined and unconfined aquifers. The integrated management of such resources faces several challenges:
High dimensionality refers to the requirement of the large number of variables that need to be considered in the description of surface water and groundwater systems. As the number of these variables increases, the computational requirements quickly saturate the capabilities of the existing management methods.
Uncertainty relates to the imprecise nature of many system inputs and parameters, including reservoir and tributary inflows, precipitation, evaporation, aquifer parameters (e.g., hydraulic conductivity and storage coefficient), and various boundary and initial conditions. Uncertainty complicates very significantly the development and application of efficient management models.
Nonlinearity is intrinsic to some physical processes and also enters through various facility and operational constraints on reservoir storages, releases, and aquifer drawdown and pumping. Nonlinearities compound the previous difficulties.
Multiple objectives pertain to the process of optimizing the use of the integrated surface and groundwater resources to meet various water demands, generate sufficient energy, maintain adequate instream flows, and protect the environment and the ecosystems. Multi-objective decision models and processes continue to challenge professional practice.
This research draws on several disciplines including groundwater flow modeling, hydrology and water resources systems, uncertainty analysis, estimation theory, stochastic optimization of dynamical systems, and policy assessment. A summary of the research contributions made in this work follows:
1.High dimensionality issues related to groundwater aquifers system have been mitigated by the use of transfer functions and their representation by state space approximations.
2.Aquifer response under uncertainty of inputs and aquifer parameters is addressed by a new statistical procedure that is applicable to regions of relatively few measurements and incorporates management reliability considerations.
3.The conjunctive management problem is formulated in a generally applicable way, taking into consideration all relevant uncertainties and system objectives. This problem is solved via an efficient stochastic optimization method that overcomes dimensionality limitations.
4.The methods developed in this Thesis are applied to the Jordanian water resources system, demonstrating their value for operational planning and management.
|
187 |
Estimating the discriminative power of time varying features for EEG BMIMappus, Rudolph Louis, IV 16 November 2009 (has links)
In this work, we present a set of methods aimed at improving the discriminative power of time-varying features of signals that contain noise. These methods use properties of noise signals as well as information theoretic techniques to factor types of noise and support signal inference for electroencephalographic (EEG) based brain-machine interfaces (BMI). EEG data were collected over two studies aimed at addressing Psychophysiological issues involving symmetry and mental rotation processing. The Psychophysiological data gathered in the mental rotation study also tested the feasibility of using dissociations of mental rotation tasks correlated with rotation angle in a BMI. We show the feasibility of mental rotation for BMI by showing comparable bitrates and recognition accuracy to state-of-the-art BMIs. The conclusion is that by using the feature selection methods introduced in this work to dissociate mental rotation tasks, we produce bitrates and recognition rates comparable to current BMIs.
|
188 |
Stochastic modeling and simulation of biochemical reaction kineticsAgarwal, Animesh 21 September 2011 (has links)
Biochemical reactions make up most of the activity in a cell. There is inherent stochasticity in the kinetic behavior of biochemical reactions which in turn governs the fate of various cellular processes. In this work, the precision of a method for dimensionality reduction for stochastic modeling of biochemical reactions is evaluated. Further, a method of stochastic simulation of reaction kinetics is implemented in case of a specific biochemical network involved in maintenance of long-term potentiation (LTP), the basic substrate for learning and memory formation. The dimensionality reduction method diverges significantly from a full stochastic model in prediction the variance of the fluctuations. The application of the stochastic simulation method to LTP modeling was used to find qualitative dependence of stochastic fluctuations on reaction volume and model parameters. / text
|
189 |
Statistical Modeling of Multi-Dimensional Knowledge Diffusion Networks: An ERGM-Based FrameworkJiang, Shan January 2015 (has links)
Knowledge diffusion networks consist of individuals who exchange knowledge and knowledge flows connecting the individuals. By studying knowledge diffusion in a network perspective, it helps us understand how the connections between individuals affect the knowledge diffusion processes. Existing research on knowledge diffusion networks mostly adopts a uni-dimensional perspective, where all the individuals in the networks are assumed to be of the same type. It also assumes that there is only one type of knowledge flow in the network. This dissertation proposes a multi-dimensional perspective of knowledge diffusion networks and examines the patterns of knowledge diffusion with Exponential Random Graph Model (ERGM) based approaches. The objective of this dissertation is to propose a framework that effectively addresses the multi-dimensionality of knowledge diffusion networks, to enable researchers and practitioners to conceptualize the multi-dimensional knowledge diffusion networks in various domains, and to provide implications on how to stimulate and control the knowledge diffusion process. The dissertation consists of three essays, all of which examine the multi-dimensional knowledge diffusion networks in a specific context, but each focuses on a different aspect of knowledge diffusion. Chapter 2 focuses on how structural properties of networks affect various types of knowledge diffusion processes in the domain of commercial technology. The study uses ERGM to simultaneously model multiple types of knowledge flows and examine their interactions. The objective is to understand the impacts of network structures on knowledge diffusion processes. Chapter 3 focuses on examining the impact of individual attributes and the attributes of knowledge on knowledge diffusion in the context of scientific innovation. Based on social capital theory, the study also utilizes ERGM to examine how knowledge transfer and knowledge co-creation can be affected by the attributes of individual researchers and the attributes of scientific knowledge. Chapter 4 considers the dynamic aspect of knowledge diffusion and proposes a novel network model extending ERGM to identify dynamic patterns of knowledge diffusion in social media. In the proposed model, dynamic patterns in social media networks are modeled based on the nodal attributes of individuals and the temporal information of network ties.
|
190 |
Feature extraction via dependence structure optimization / Požymių išskyrimas optimizuojant priklausomumo struktūrąDaniušis, Povilas 01 October 2012 (has links)
In many important real world applications the initial representation of the data is inconvenient,
or even prohibitive for further analysis. For example, in image analysis, text
analysis and computational genetics high-dimensional, massive, structural, incomplete,
and noisy data sets are common. Therefore, feature extraction, or revelation of informative
features from the raw data is one of fundamental machine learning problems.
Efficient feature extraction helps to understand data and the process that generates it,
reduce costs for future measurements and data analysis. The representation of the structured
data as a compact set of informative numeric features allows applying well studied
machine learning techniques instead of developing new ones..
The dissertation focuses on supervised and semi-supervised feature extraction methods,
which optimize the dependence structure of features. The dependence is measured using
the kernel estimator of Hilbert-Schmidt norm of covariance operator (HSIC measure).
Two dependence structures are investigated: in the first case we seek features which
maximize the dependence on the dependent variable, and in the second one, we additionally
minimize the mutual dependence of features. Linear and kernel formulations of
HBFE and HSCA are provided. Using Laplacian regularization framework we construct
semi-supervised variants of HBFE and HSCA.
Suggested algorithms were investigated experimentally using conventional and multilabel
classification data... [to full text] / Daugelis praktiškai reikšmingu sistemu mokymo uždaviniu reikalauja gebeti panaudoti didelio matavimo, strukturizuotus, netiesinius duomenis. Vaizdu, teksto, socialiniu bei verslo ryšiu analize, ivairus bioinformatikos uždaviniai galetu buti tokiu uždaviniu pavyzdžiais. Todel požymiu išskyrimas dažnai yra pirmasis žingsnis, kuriuo pradedama duomenu analize ir nuo kurio priklauso galutinio rezultato sekme. Šio disertacinio darbo tyrimo objektas yra požymiu išskyrimo algoritmai, besiremiantys priklausomumo savoka. Darbe nagrinejamas priklausomumas, nusakytas kovariacinio operatoriaus Hilberto-Šmidto normos (HSIC mato) branduoliniu ivertiniu. Pasiulyti šiuo ivertiniu besiremiantys HBFE ir HSCA algoritmai leidžia dirbti su bet kokios strukturos duomenimis, bei yra formuluojami tikriniu vektoriu terminais (tai leidžia optimizavimui naudoti standartinius paketus), bei taikytini ne tik prižiurimo, bet ir dalinai prižiurimo mokymo imtims. Pastaruoju atveju HBFE ir HSCA modifikacijos remiasi Laplaso reguliarizacija. Eksperimentais su klasifikavimo bei daugiažymio klasifikavimo duomenimis parodyta, jog pasiulyti algoritmai leidžia pagerinti klasifikavimo efektyvuma lyginant su PCA ar LDA.
|
Page generated in 0.0928 seconds