31 |
Subgroup Identification in Clinical TrialsLi, Xiaochen 04 1900 (has links)
Indiana University-Purdue University Indianapolis (IUPUI) / Subgroup analyses assess the heterogeneity of treatment effects in groups of patients defined by patients’ baseline characteristics. Identifying subgroup of patients with differential treatment effect is crucial for tailored therapeutics and personalized medicine. Model-based variable selection methods are well developed and widely applied to select significant treatment-by-covariate interactions for subgroup analyses. Machine learning and data-driven based methods for subgroup identification have also been developed.
In this dissertation, I consider two different types of subgroup identification methods: one is nonparametric machine learning based and the other is model based. In the first part, the problem of subgroup identification was transferred to an optimization problem and a stochastic search technique was implemented to partition the whole population into disjoint subgroups with differential treatment effect. In the second approach, an integrative three-step model-based variable selection method was proposed for subgroup analyses in longitudinal data. Using this three steps variable selection framework, informative features and their interaction with the treatment indicator can be identified for subgroup analysis in longitudinal data. This method can be extended to longitudinal binary or categorical data. Simulation studies and real data examples were used to demonstrate the performance of the proposed methods. / 2022-05-06
|
32 |
Linear Mixed Model Selection by Partial CorrelationAlabiso, Audry 29 April 2020 (has links)
No description available.
|
33 |
Predicting Monthly Precipitation in Ontario using a Multi-Model Ensemble and the XGBoost AlgorithmHadzi-Tosev, Milena January 2020 (has links)
There is a strong interest in the climate community to improve the ability to accurately predict future trends of climate variables. Recently, machine learning methods have proven their ability to contribute to more accurate predictions of historical data on a variety of climate variables. There is also a strong interest in using statistical downscaling to predict local station data from the output of multi-model ensembles. This project looks at using the machine learning algorithm XGBoost and evaluating its ability to accurately predict historical monthly precipitation, with a focus of applying this method to simulate future precipitation trends. / Thesis / Master of Science (MSc)
|
34 |
On Multiplicity Adjustment in Bayesian Variable Selection and An Objective Bayesian Analysis of a Crossover DesignLi, Dandan 23 October 2014 (has links)
No description available.
|
35 |
A Bayesian Approach to Prediction and Variable Selection Using Nonstationary Gaussian ProcessesDavis, Casey Benjamin 28 May 2015 (has links)
No description available.
|
36 |
A Penalized Approach to Mixed Model Selection Via Cross ValidationXiong, Jingwei 05 December 2017 (has links)
No description available.
|
37 |
Two Essays on High-Dimensional Robust Variable Selection and an Application to Corporate Bankruptcy PredictionLi, Shaobo 29 October 2018 (has links)
No description available.
|
38 |
Improving the Accuracy of Variable Selection Using the Whole Solution PathLiu, Yang 23 July 2015 (has links)
No description available.
|
39 |
Multilevel Model Selection: A Regularization Approach Incorporating Heredity ConstraintsStone, Elizabeth Anne January 2013 (has links)
This dissertation focuses on estimation and selection methods for a simple linear model with two levels of variation. This model provides a foundation for extensions to more levels. We propose new regularization criteria for model selection, subset selection, and variable selection in this context. Regularization is a penalized-estimation approach that shrinks the estimate and selects variables for structured data. This dissertation introduces a procedure (HM-ALASSO) that extends regularized multilevel-model estimation and selection to enforce principles of fixed heredity (e.g., including main effects when their interactions are included) and random heredity (e.g., including fixed effects when their random terms are included). The goals in developing this method were to create a procedure that provided reasonable estimates of all parameters, adhered to fixed and random heredity principles, resulted in a parsimonious model, was theoretically justifiable, and was able to be implemented and used in available software. The HM-ALASSO incorporates heredity-constrained selection directly into the estimation process. HM-ALASSO is shown to enjoy the properties of consistency, sparsity, and asymptotic normality. The ability of HM-ALASSO to produce quality estimates of the underlying parameters while adhering to heredity principles is demonstrated using simulated data. The performance of HM-ALASSO is illustrated using a subset of the High School and Beyond (HS&B) data set that includes math-achievement outcomes modeled via student- and school-level predictors. The HM-ALASSO framework is flexible enough that it can be adapted for various rule sets and parameterizations. / Statistics
|
40 |
Spatially Correlated Model Selection (SCOMS)Velasco-Cruz, Ciro 31 May 2012 (has links)
In this dissertation, a variable selection method for spatial data is developed. It is assumed that the spatial process is non-stationary as a whole but is piece-wise stationary. The pieces where the spatial process is stationary are called regions. The variable selection approach accounts for two sources of correlation: (1) the spatial correlation of the data within the regions, and (2) the correlation of adjacent regions.
The variable selection is carried out by including indicator variables that characterize the significance of the regression coefficients. The Ising distribution as prior for the vector of indicator variables, models the dependence of adjacent regions.
We present a case study on brook trout data where the response of interest is the presence/absence of the fish at sites in the eastern United States. We find that the method outperforms the case of the probit regression where the spatial field is assumed stationary and isotropic. Additionally, the method outperformed the case where multiple regions are assumed independent of their neighbors. / Ph. D.
|
Page generated in 0.1883 seconds