121 |
Statistical Models and Analysis of Growth Processes in Biological TissueXia, Jun 15 December 2016 (has links)
The mechanisms that control growth processes in biology tissues have attracted continuous research interest despite their complexity. With the emergence of big data experimental approaches there is an urgent need to develop statistical and computational models to fit the experimental data and that can be used to make predictions to guide future research. In this work we apply statistical methods on growth process of different biological tissues, focusing on development of neuron dendrites and tumor cells.
We first examine the neuron cell growth process, which has implications in neural tissue regenerations, by using a computational model with uniform branching probability and a maximum overall length constraint. One crucial outcome is that we can relate the parameter fits from our model to real data from our experimental collaborators, in order to examine the usefulness of our model under different biological conditions. Our methods can now directly compare branching probabilities of different experimental conditions and provide confidence intervals for these population-level measures. In addition, we have obtained analytical results that show that the underlying probability distribution for this process follows a geometrical progression increase at nearby distances and an approximately geometrical series decrease for far away regions, which can be used to estimate the spatial location of the maximum of the probability distribution. This result is important, since we would expect maximum number of dendrites in this region; this estimate is related to the probability of success for finding a neural target at that distance during a blind search.
We then examined tumor growth processes which have similar evolutional evolution in the sense that they have an initial rapid growth that eventually becomes limited by the resource constraint. For the tumor cells evolution, we found an exponential growth model best describes the experimental data, based on the accuracy and robustness of models. Furthermore, we incorporated this growth rate model into logistic regression models that predict the growth rate of each patient with biomarkers; this formulation can be very useful for clinical trials. Overall, this study aimed to assess the molecular and clinic pathological determinants of breast cancer (BC) growth rate in vivo.
|
122 |
A Study of the Calibration Regression Model with Censored Lifetime Medical CostLu, Min 03 August 2006 (has links)
Medical cost has received increasing interest recently in Biostatistics and public health. Statistical analysis and inference of life time medical cost have been challenging by the fact that the survival times are censored on some study subjects and their subsequent cost are unknown. Huang (2002) proposed the calibration regression model which is a semiparametric regression tool to study the medical cost associated with covariates. In this thesis, an inference procedure is investigated using empirical likelihood ratio method. The unadjusted and adjusted empirical likelihood confidence regions are constructed for the regression parameters. We compare the proposed empirical likelihood methods with normal approximation based method. Simulation results show that the proposed empirical likelihood ratio method outperforms the normal approximation based method in terms of coverage probability. In particular, the adjusted empirical likelihood is the best one which overcomes the under coverage problem.
|
123 |
New Non-Parametric Methods for Income DistributionsLuo, Shan 26 April 2013 (has links)
Low income proportion (LIP), Lorenz curve (LC) and generalized Lorenz curve (GLC) are important indexes in describing the inequality of income distribution. They have been widely used for measuring social stability by governments around the world. The accuracy of estimating those indexes is essential to quantify the economics of a country. Established statistical inferential methods for these indexes are based on an asymptotic normal distribution, which may have poor performance when the real income data is skewed or has outliers. Recent applications of nonparametric methods, though, allow researchers to utilize techniques without giving data the parametric distribution assumption. For example, existing research proposes the plug-in empirical likelihood (EL)-based inferences for LIP, LC and GLC. However, this method becomes computationally intensive and mathematically complex because of the presence of nonlinear constraints in the underlying optimization problem. Meanwhile, the limiting distribution of the log empirical likelihood ratio is a scaled Chi-square distribution. The estimation of the scale constant will affect the overall performance of the plug-in EL method. To improve the efficiency of the existing inferential methods, this dissertation first proposes kernel estimators for LIP, LC and GLC, respectively. Then the cross-validation method is proposed to choose bandwidth for the kernel estimators. These kernel estimators are proved to have asymptotic normality. The smoothed jackknife empirical likelihood (SJEL) for LIP, LC and GLC are defined. Then the log-jackknife empirical likelihood ratio statistics are proved to follow the standard Chi-square distribution. Extensive simulation studies are conducted to evaluate the kernel estimators in terms of Mean Square Error and Asymptotic Relative Efficiency. Next, the SJEL-based confidence intervals and the smoothed bootstrap-based confidence intervals are proposed. The coverage probability and interval length for the proposed confidence intervals are calculated and compared with the normal approximation-based intervals. The proposed kernel estimators are found to be competitive estimators, and the proposed inferential methods are observed to have better finite-sample performance. All inferential methods are illustrated through real examples.
|
124 |
Antiretroviral Regimens in HIV-Infected Adults Receiving Medical Care in the United States: Medical Monitoring Project, 2009Tie, Yunfeng 19 April 2013 (has links)
Effective antiretroviral therapy (ART) is essential for viral suppression (VS) in HIV-infected patients. However, there is a lack of nationally representative data on types of ART regimens used and their impact on VS. This thesis used self-reported interview and abstracted medical record from 2009 Medical Monitoring Project (MMP) to study ART regimen type and related health outcomes. Results showed that 88.6% of HIV-infected adults in care was prescribed ART, and about half took regimens designated as ‘preferred’ according to U.S ART guidelines. Among MMP participants prescribed ART, 62.7% achieved durable VS, 77.8% achieved recent VS, 83.5% were 100% dose-adherent, and 17.1% reported side effects. Multivariate regression analyses revealed that although ART was critical for VS, there were minor differences in health outcomes among the major ART classes in the U.S. ART guidelines or six most-commonly used regimens. This study could be potentially useful for future strategic planning of HIV care.
|
125 |
NONPARAMETRIC INFERENCES FOR THE HAZARD FUNCTION WITH RIGHT TRUNCATIONAkcin, Haci Mustafa 03 May 2013 (has links)
Incompleteness is a major feature of time-to-event data. As one type of incompleteness, truncation refers to the unobservability of the time-to-event variable because it is smaller (or greater) than the truncation variable. A truncated sample always involves left and right truncation.
Left truncation has been studied extensively while right truncation has not received the same level of attention. In one of the earliest studies on right truncation, Lagakos et al. (1988) proposed to transform a right truncated variable to a left truncated variable and then apply existing methods to the transformed variable. The reverse-time hazard function is introduced through transformation. However, this quantity does not have a natural interpretation. There exist gaps in the inferences for the regular forward-time hazard function with right truncated data. This dissertation discusses variance estimation of the cumulative hazard estimator, one-sample log-rank test, and comparison of hazard rate functions among finite independent samples under the context of right truncation.
First, the relation between the reverse- and forward-time cumulative hazard functions is clarified. This relation leads to the nonparametric inference for the cumulative hazard function. Jiang (2010) recently conducted a research on this direction and proposed two variance estimators of the cumulative hazard estimator. Some revision to the variance estimators is suggested in this dissertation and evaluated in a Monte-Carlo study.
Second, this dissertation studies the hypothesis testing for right truncated data. A series of tests is developed with the hazard rate function as the target quantity. A one-sample log-rank test is first discussed, followed by a family of weighted tests for comparison between finite $K$-samples. Particular weight functions lead to log-rank, Gehan, Tarone-Ware tests and these three tests are evaluated in a Monte-Carlo study.
Finally, this dissertation studies the nonparametric inference for the hazard rate function for the right truncated data. The kernel smoothing technique is utilized in estimating the hazard rate function. A Monte-Carlo study investigates the uniform kernel smoothed estimator and its variance estimator. The uniform, Epanechnikov and biweight kernel estimators are implemented in the example of blood transfusion infected AIDS data.
|
126 |
Nonparametric Inferences for the Hazard Function with Right TruncationAkcin, Haci Mustafa 03 May 2013 (has links)
Incompleteness is a major feature of time-to-event data. As one type of incompleteness, truncation refers to the unobservability of the time-to-event variable because it is smaller (or greater) than the truncation variable. A truncated sample always involves left and right truncation.
Left truncation has been studied extensively while right truncation has not received the same level of attention. In one of the earliest studies on right truncation, Lagakos et al. (1988) proposed to transform a right truncated variable to a left truncated variable and then apply existing methods to the transformed variable. The reverse-time hazard function is introduced through transformation. However, this quantity does not have a natural interpretation. There exist gaps in the inferences for the regular forward-time hazard function with right truncated data. This dissertation discusses variance estimation of the cumulative hazard estimator, one-sample log-rank test, and comparison of hazard rate functions among finite independent samples under the context of right truncation.
First, the relation between the reverse- and forward-time cumulative hazard functions is clarified. This relation leads to the nonparametric inference for the cumulative hazard function. Jiang (2010) recently conducted a research on this direction and proposed two variance estimators of the cumulative hazard estimator. Some revision to the variance estimators is suggested in this dissertation and evaluated in a Monte-Carlo study.
Second, this dissertation studies the hypothesis testing for right truncated data. A series of tests is developed with the hazard rate function as the target quantity. A one-sample log-rank test is first discussed, followed by a family of weighted tests for comparison between finite $K$-samples. Particular weight functions lead to log-rank, Gehan, Tarone-Ware tests and these three tests are evaluated in a Monte-Carlo study.
Finally, this dissertation studies the nonparametric inference for the hazard rate function for the right truncated data. The kernel smoothing technique is utilized in estimating the hazard rate function. A Monte-Carlo study investigates the uniform kernel smoothed estimator and its variance estimator. The uniform, Epanechnikov and biweight kernel estimators are implemented in the example of blood transfusion infected AIDS data.
|
127 |
Evaluation of Hedge Funds PerformanceQian, Jing 03 August 2006 (has links)
Hedge funds are private investment funds characterized by unconventional strategies. This thesis employed multi-factor CAPM to evaluate the performance, or manager skill of hedge funds investment segments by using CSFB/Tremont Hedge Fund Indices from January 1994 to September 2005. The performance evaluation is based on the concept of ¡°Jansen¡¯s alpha¡±, which is estimated by applying Generalized Method of Moment. The finding is that hedge funds industry in general displayed the ability to outperform market proxy. Global Macro shows the strongest manager skill, followed by Event Driven, Equity Market Neutral and Long/Short Equity. This thesis also investigates the consistency of hedge funds performance over market environment. It was discovered that the hedge funds industry in general and all the sub-category investment segments except Convertibly Arbitrage, Emerging Market and Fix income Arbitrage displayed the ability to cushion the impact of financial shocks.
|
128 |
SVD and PCA in Image ProcessingRenkjumnong, Wasuta - 16 July 2007 (has links)
The Singular Value Decomposition is one of the most useful matrix factorizations in applied linear algebra, the Principal Component Analysis has been called one of the most valuable results of applied linear algebra. How and why principal component analysis is intimately related to the technique of singular value decomposition is shown. Their properties and applications are described. Assumptions behind this techniques as well as possible extensions to overcome these limitations are considered. This understanding leads to the real world applications, in particular, image processing of neurons. Noise reduction, and edge detection of neuron images are investigated.
|
129 |
Semi-Empirical Likelihood Confidence Intervals for the ROC Curve with Missing DataLiu, Xiaoxia 11 June 2010 (has links)
The receiver operating characteristic (ROC) curve is one of the most commonly used methods to compare the diagnostic performances of two or more laboratory or diagnostic tests. In this thesis, we propose semi-empirical likelihood based confidence intervals for ROC curves of two populations, where one population is parametric while the other one is non-parametric and both populations have missing data. After imputing missing values, we derive the semi-empirical likelihood ratio statistic and the corresponding likelihood equations. It has been shown that the log-semi-empirical likelihood ratio statistic is asymptotically chi-square distributed. The estimating equations are solved simultaneously to obtain the estimated lower and upper bounds of semi-empirical likelihood confidence intervals. Simulation studies are conducted to evaluate the finite sample performance of the proposed empirical likelihood confidence intervals with various sample sizes and different missing rates.
|
130 |
Data Driven Approaches to Testing Homogeneity of Intraclass Correlation CoefficientsWu, Baohua 01 December 2010 (has links)
The test of homogeneity for intraclass correlation coefficients has been one of the active topics in statistical research. Several chi-square tests have been proposed to test the homogeneity of intraclass correlations in the past few decades. The big concern for them is that these methods are seriously biased when sample sizes are not large. In this thesis, data driven approaches are proposed to testing the homogeneity of intraclass correlation coefficients of several populations. Through simulation study, data driven methods have been proved to be less biased and accurate than some commonly used chi-square tests.
|
Page generated in 0.1477 seconds