• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 51
  • 26
  • 20
  • 10
  • 4
  • 3
  • 3
  • 2
  • 1
  • 1
  • Tagged with
  • 144
  • 144
  • 37
  • 37
  • 33
  • 27
  • 18
  • 16
  • 14
  • 14
  • 14
  • 13
  • 13
  • 12
  • 12
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
31

Behind Left and Right. The meaning of left-right orientation in Europe

Weber, Wiebke 21 January 2013 (has links)
The left-right concept is considered to facilitate communication and orientation in the political world. It has a long tradition in European politics due to its capacity to absorb different associations over time. However, this dynamic nature of the concept requires periodical reassessments in order to assure that a common conceptualisation endures. This dissertation focuses on reassign what individual left-right orientation means. Usually, this is measured by asking people to place themselves on a scale labelled ‘left’ and ‘right’ at its endpoints. The first empirical analysis of this dissertation shows that this measure is comparable across groups and countries. Thereafter, the relationship between an individual’s issue preference and left-right orientation is considered. The second empirical analysis shows that this relationship is conditioned by the importance people assign to the respective issues. The final analysis demonstrates that what explains left-right orientation is contingent on individual and contextual factors. This implies that in order to understand left-right orientation, it is not enough to identify what has an impact on a person’s position but also account for all those factors that predict variation between individuals. Given this complexity, my conclusion is that the left-right concept runs the risk of becoming too complicated to serve as an analytical tool to shed light on political attitudes and behaviour. / El concepte esquerra-dreta és considerat com un factor facilitador de la comunicació en el món polític. Té una llarga tradició’ en la política europea degut a la seva capacitat d’absorbir diferents associacions a través del temps. Tanmateix, aquesta natura dinàmica del concepte requereix revisions periòdiques per assegurar que persisteix una conceptualització comuna. La present tesi es centra en resignar el que significa l’orientació esquerra-dreta. Normalment, es mesura tot demanant als enquestats posicionar-se a ells mateixos en una escala que va de l’esquerra a la dreta. El primer anàlisi empíric de la present tesi mostra que aquesta mesura és comparable entre grups i països. Seguidament, es considera la relació entre les preferències temàtiques dels individus i llur orientació esquerra-dreta. El segon anàlisi empíric mostra que aquesta relació està condicionada per la importància que les persones assignen als temes respectius. L’anàlisi final demostra que el que explica l’orientació esquerra-dreta depèn de factors contextuals i individuals. Això implica que per entendre l’orientació esquerra-dreta no és suficient identificar què té un impacte en la posició d’una persona sinó també una explicació per a tots aquells factors que preveuen la variació entre individus. Donada aquesta complexitat, la meva conclusió és que el concepte esquerra-dreta corre el risc de convertir-se en massa complicat per a servir com a eina analítica per a l’estudi de les actituds i el comportament politics.
32

Measurement Error and Misclassification in Interval-Censored Life History Data

White, Bethany Joy Giddings January 2007 (has links)
In practice, data are frequently incomplete in one way or another. It can be a significant challenge to make valid inferences about the parameters of interest in this situation. In this thesis, three problems involving such data are addressed. The first two problems involve interval-censored life history data with mismeasured covariates. Data of this type are incomplete in two ways. First, the exact event times are unknown due to censoring. Second, the true covariate is missing for most, if not all, individuals. This work focuses primarily on the impact of covariate measurement error in progressive multi-state models with data arising from panel (i.e., interval-censored) observation. These types of problems arise frequently in clinical settings (e.g. when disease progression is of interest and patient information is collected during irregularly spaced clinic visits). Two and three state models are considered in this thesis. This work is motivated by a research program on psoriatic arthritis (PsA) where the effects of error-prone covariates on rates of disease progression are of interest and patient information is collected at clinic visits (Gladman et al. 1995; Bond et al. 2006). Information regarding the error distributions were available based on results from a separate study conducted to evaluate the reliability of clinical measurements that are used in PsA treatment and follow-up (Gladman et al. 2004). The asymptotic bias of covariate effects obtained ignoring error in covariates is investigated and shown to be substantial in some settings. In a series of simulation studies, the performance of corrected likelihood methods and methods based on a simulation-extrapolation (SIMEX) algorithm (Cook \& Stefanski 1994) were investigated to address covariate measurement error. The methods implemented were shown to result in much smaller empirical biases and empirical coverage probabilities which were closer to the nominal levels. The third problem considered involves an extreme case of interval censoring known as current status data. Current status data arise when individuals are observed only at a single point in time and it is then determined whether they have experienced the event of interest. To complicate matters, in the problem considered here, an unknown proportion of the population will never experience the event of interest. Again, this type of data is incomplete in two ways. One assessment is made on each individual to determine whether or not an event has occurred. Therefore, the exact event times are unknown for those who will eventually experience the event. In addition, whether or not the individuals will ever experience the event is unknown for those who have not experienced the event by the assessment time. This problem was motivated by a series of orthopedic trials looking at the effect of blood thinners in hip and knee replacement surgeries. These blood thinners can cause a negative serological response in some patients. This response was the outcome of interest and the only available information regarding it was the seroconversion time under current status observation. In this thesis, latent class models with parametric, nonparametric and piecewise constant forms of the seroconversion time distribution are described. They account for the fact that only a proportion of the population will experience the event of interest. Estimators based on an EM algorithm were evaluated via simulation and the orthopedic surgery data were analyzed based on this methodology.
33

Measurement Error and Misclassification in Interval-Censored Life History Data

White, Bethany Joy Giddings January 2007 (has links)
In practice, data are frequently incomplete in one way or another. It can be a significant challenge to make valid inferences about the parameters of interest in this situation. In this thesis, three problems involving such data are addressed. The first two problems involve interval-censored life history data with mismeasured covariates. Data of this type are incomplete in two ways. First, the exact event times are unknown due to censoring. Second, the true covariate is missing for most, if not all, individuals. This work focuses primarily on the impact of covariate measurement error in progressive multi-state models with data arising from panel (i.e., interval-censored) observation. These types of problems arise frequently in clinical settings (e.g. when disease progression is of interest and patient information is collected during irregularly spaced clinic visits). Two and three state models are considered in this thesis. This work is motivated by a research program on psoriatic arthritis (PsA) where the effects of error-prone covariates on rates of disease progression are of interest and patient information is collected at clinic visits (Gladman et al. 1995; Bond et al. 2006). Information regarding the error distributions were available based on results from a separate study conducted to evaluate the reliability of clinical measurements that are used in PsA treatment and follow-up (Gladman et al. 2004). The asymptotic bias of covariate effects obtained ignoring error in covariates is investigated and shown to be substantial in some settings. In a series of simulation studies, the performance of corrected likelihood methods and methods based on a simulation-extrapolation (SIMEX) algorithm (Cook \& Stefanski 1994) were investigated to address covariate measurement error. The methods implemented were shown to result in much smaller empirical biases and empirical coverage probabilities which were closer to the nominal levels. The third problem considered involves an extreme case of interval censoring known as current status data. Current status data arise when individuals are observed only at a single point in time and it is then determined whether they have experienced the event of interest. To complicate matters, in the problem considered here, an unknown proportion of the population will never experience the event of interest. Again, this type of data is incomplete in two ways. One assessment is made on each individual to determine whether or not an event has occurred. Therefore, the exact event times are unknown for those who will eventually experience the event. In addition, whether or not the individuals will ever experience the event is unknown for those who have not experienced the event by the assessment time. This problem was motivated by a series of orthopedic trials looking at the effect of blood thinners in hip and knee replacement surgeries. These blood thinners can cause a negative serological response in some patients. This response was the outcome of interest and the only available information regarding it was the seroconversion time under current status observation. In this thesis, latent class models with parametric, nonparametric and piecewise constant forms of the seroconversion time distribution are described. They account for the fact that only a proportion of the population will experience the event of interest. Estimators based on an EM algorithm were evaluated via simulation and the orthopedic surgery data were analyzed based on this methodology.
34

Analysis of Correlated Data with Measurement Error in Responses or Covariates

Chen, Zhijian January 2010 (has links)
Correlated data frequently arise from epidemiological studies, especially familial and longitudinal studies. Longitudinal design has been used by researchers to investigate the changes of certain characteristics over time at the individual level as well as how potential factors influence the changes. Familial studies are often designed to investigate the dependence of health conditions among family members. Various models have been developed for this type of multivariate data, and a wide variety of estimation techniques have been proposed. However, data collected from observational studies are often far from perfect, as measurement error may arise from different sources such as defective measuring systems, diagnostic tests without gold references, and self-reports. Under such scenarios only rough surrogate variables are measured. Measurement error in covariates in various regression models has been discussed extensively in the literature. It is well known that naive approaches ignoring covariate error often lead to inconsistent estimators for model parameters. In this thesis, we develop inferential procedures for analyzing correlated data with response measurement error. We consider three scenarios: (i) likelihood-based inferences for generalized linear mixed models when the continuous response is subject to nonlinear measurement errors; (ii) estimating equations methods for binary responses with misclassifications; and (iii) estimating equations methods for ordinal responses when the response variable and categorical/ordinal covariates are subject to misclassifications. The first problem arises when the continuous response variable is difficult to measure. When the true response is defined as the long-term average of measurements, a single measurement is considered as an error-contaminated surrogate. We focus on generalized linear mixed models with nonlinear response error and study the induced bias in naive estimates. We propose likelihood-based methods that can yield consistent and efficient estimators for both fixed-effects and variance parameters. Results of simulation studies and analysis of a data set from the Framingham Heart Study are presented. Marginal models have been widely used for correlated binary, categorical, and ordinal data. The regression parameters characterize the marginal mean of a single outcome, without conditioning on other outcomes or unobserved random effects. The generalized estimating equations (GEE) approach, introduced by Liang and Zeger (1986), only models the first two moments of the responses with associations being treated as nuisance characteristics. For some clustered studies especially familial studies, however, the association structure may be of scientific interest. With binary data Prentice (1988) proposed additional estimating equations that allow one to model pairwise correlations. We consider marginal models for correlated binary data with misclassified responses. We develop “corrected” estimating equations approaches that can yield consistent estimators for both mean and association parameters. The idea is related to Nakamura (1990) that is originally developed for correcting bias induced by additive covariate measurement error under generalized linear models. Our approaches can also handle correlated misclassifications rather than a simple misclassification process as considered by Neuhaus (2002) for clustered binary data under generalized linear mixed models. We extend our methods and further develop marginal approaches for analysis of longitudinal ordinal data with misclassification in both responses and categorical covariates. Simulation studies show that our proposed methods perform very well under a variety of scenarios. Results from application of the proposed methods to real data are presented. Measurement error can be coupled with many other features in the data, e.g., complex survey designs, that can complicate inferential procedures. We explore combining survey weights and misclassification in ordinal covariates in logistic regression analyses. We propose an approach that incorporates survey weights into estimating equations to yield design-based unbiased estimators. In the final part of the thesis we outline some directions for future work, such as transition models and semiparametric models for longitudinal data with both incomplete observations and measurement error. Missing data is another common feature in applications. Developing novel statistical techniques for dealing with both missing data and measurement error can be beneficial.
35

Estimation of Stochastic Degradation Models Using Uncertain Inspection Data

Lu, Dongliang January 2012 (has links)
Degradation of components and structures is a major threat to the safety and reliability of large engineering systems, such as the railway networks or the nuclear power plants. Periodic inspection and maintenance are thus required to ensure that the system is in good condition for continued service. A key element for the optimal inspection and maintenance is to accurately model and forecast the degradation progress, such that inspection and preventive maintenance can be scheduled accordingly. In recently years, probabilistic models based on stochastic process have become increasingly popular in degradation modelling, due to their flexibility in modelling both the temporal and sample uncertainties of the degradation. However, because of the often complex structure of stochastic degradation models, accurate estimate of the model parameters can be quite difficult, especially when the inspection data are noisy or incomplete. Not considering the effect of uncertain inspection data is likely to result in biased parameter estimates and therefore erroneous predictions of future degradation. The main objective of the thesis is to develop formal methods for the parameter estimation of stochastic degradation models using uncertain inspection data. Three typical stochastic models are considered. They are the random rate model, the gamma process model and the Poisson process model, among which the random rate model and the gamma process model are used to model the flaw growth, and the Poisson process model is used to model the flaw generation. Likelihood functions of the three stochastic models given noisy or incomplete inspection data are derived, from which maximum likelihood estimates can be obtained. The thesis also investigates Bayesian inference of the stochastic degradation models. The most notable advantage of Bayesian inference over classical point estimates is its ability to incorporate background information in the estimation process, which is especially useful when inspection data are scarce. A major obstacle for accurate parameter inference of stochastic models from uncertain inspection data is the computational difficulties of the likelihood evaluation, as it often involves calculation of high dimensional integrals or large number of convolutions. To overcome the computational difficulties, a number of numerical methods are developed in the thesis. For example, for the gamma process model subject to sizing error, an efficient maximum likelihood method is developed using the Genz's transform and quasi-Monte Carlo simulation. A Markov Chain Monte Carlo simulation with sizing error as auxiliary variables is developed for the Poisson flaw generation model, A sequential Bayesian updating using approximate Bayesian computation and weighted samples is also developed for Bayesian inference of the gamma process subject to sizing error. Examples on the degradation of nuclear power plant components are presented to illustrate the use of the stochastic degradation models using practical uncertain inspection data. It is shown from the examples that the proposed methods are very effective in terms of accuracy and computational efficiency.
36

Accrual Noise Ratio as a Measure of Accrual Reliability

Njoroge, Kenneth January 2009 (has links)
<p>I develop an empirical model that estimates a firm-specific accrual noise ratio (ANR), an operational and statistically grounded measure of accrual reliability, and test the measure's construct validity. The model allows accrual reliability to vary across firms, which is particularly important because many reliability determinants vary in cross-section. Unlike metrics that measure relative perceived reliability, ANR measures accrual reliability independent of the perceptions of investors, creditors or auditors. I find that ANR relates in expected ways with multiple proxies of accounting reliability, that ANR's relation with the proxies of other accounting constructs is consistent with theory, and that ANR's sensitivity to percentage changes of accrual components is consistent with a subjective ordinal ranking of the components' reliability from prior literature.</p> / Dissertation
37

Efficient inference in general semiparametric regression models

Maity, Arnab 15 May 2009 (has links)
Semiparametric regression has become very popular in the field of Statistics over the years. While on one hand more and more sophisticated models are being developed, on the other hand the resulting theory and estimation process has become more and more involved. The main problems that are addressed in this work are related to efficient inferential procedures in general semiparametric regression problems. We first discuss efficient estimation of population-level summaries in general semiparametric regression models. Here our focus is on estimating general population-level quantities that combine the parametric and nonparametric parts of the model (e.g., population mean, probabilities, etc.). We place this problem in a general context, provide a general kernel-based methodology, and derive the asymptotic distributions of estimates of these population-level quantities, showing that in many cases the estimates are semiparametric efficient. Next, motivated from the problem of testing for genetic effects on complex traits in the presence of gene-environment interaction, we consider developing score test in general semiparametric regression problems that involves Tukey style 1 d.f form of interaction between parametrically and non-parametrically modeled covariates. We develop adjusted score statistics which are unbiased and asymptotically efficient and can be performed using standard bandwidth selection methods. In addition, to over come the difficulty of solving functional equations, we give easy interpretations of the target functions, which in turn allow us to develop estimation procedures that can be easily implemented using standard computational methods. Finally, we take up the important problem of estimation in a general semiparametric regression model when covariates are measured with an additive measurement error structure having normally distributed measurement errors. In contrast to methods that require solving integral equation of dimension the size of the covariate measured with error, we propose methodology based on Monte Carlo corrected scores to estimate the model components and investigate the asymptotic behavior of the estimates. For each of the problems, we present simulation studies to observe the performance of the proposed inferential procedures. In addition, we apply our proposed methodology to analyze nontrivial real life data sets and present the results.
38

Deconvolution in Random Effects Models via Normal Mixtures

Litton, Nathaniel A. 2009 August 1900 (has links)
This dissertation describes a minimum distance method for density estimation when the variable of interest is not directly observed. It is assumed that the underlying target density can be well approximated by a mixture of normals. The method compares a density estimate of observable data with a density of the observable data induced from assuming the target density can be written as a mixture of normals. The goal is to choose the parameters in the normal mixture that minimize the distance between the density estimate of the observable data and the induced density from the model. The method is applied to the deconvolution problem to estimate the density of $X_{i}$ when the variable $% Y_{i}=X_{i}+Z_{i}$, $i=1,\ldots ,n$, is observed, and the density of $Z_{i}$ is known. Additionally, it is applied to a location random effects model to estimate the density of $Z_{ij}$ when the observable quantities are $p$ data sets of size $n$ given by $X_{ij}=\alpha _{i}+\gamma Z_{ij},~i=1,\ldots ,p,~j=1,\ldots ,n$, where the densities of $\alpha_{i} $ and $Z_{ij}$ are both unknown. The performance of the minimum distance approach in the measurement error model is compared with the deconvoluting kernel density estimator of Stefanski and Carroll (1990). In the location random effects model, the minimum distance estimator is compared with the explicit characteristic function inversion method from Hall and Yao (2003). In both models, the methods are compared using simulated and real data sets. In the simulations, performance is evaluated using an integrated squared error criterion. Results indicate that the minimum distance methodology is comparable to the deconvoluting kernel density estimator and outperforms the explicit characteristic function inversion method.
39

Bayesian Methods in Nutrition Epidemiology and Regression-based Predictive Models in Healthcare

Zhang, Saijuan 2010 December 1900 (has links)
This dissertation has mainly two parts. In the first part, we propose a bivariate nonlinear multivariate measurement error model to understand the distribution of dietary intake and extend it to a multivariate model to capture dietary patterns in nutrition epidemiology. In the second part, we propose regression-based predictive models to accurately predict surgery duration in healthcare. Understanding the distribution of episodically consumed dietary components is an important problem in public health. Short-term measurements of episodically consumed dietary components are zero-inflated skewed distributions. So-called two-part models have been developed for such data. However, there is much greater public health interest in the usual intake adjusted for caloric intake. Recently a nonlinear mixed effects model has been developed and fit by maximum likelihood using nonlinear mixed effects programs. However, the fitting is slow and unstable. We develop a Monte-Carlo-based fitting method in Chapter II. We demonstrate numerically that our methods lead to increased speed of computation, converge to reasonable solutions, and have the flexibility to be used in either a frequentist or a Bayesian manner. Diet consists of numerous foods, nutrients and other components, each of which have distinctive attributes. Increasingly nutritionists are interested in exploring them collectively to capture overall dietary patterns. We thus extend the bivariate model described in Chapter III to multivariate level. We use survey-weighted MCMC computations to fit the model, with uncertainty estimation coming from balanced repeated replication. The methodology is illustrated through an application of estimating the population distribution of the Healthy Eating Index-2005 (HEI-2005), a multi-component dietary quality index , among children aged 2-8 in the United States. The second part of this dissertation is to accurately predict surgery duration. Prior research has identified the current procedural terminology (CPT) codes as the most important factor when predicting surgical case durations but there has been little reporting of a general predictive methodology using it effectively. In Chapter IV, we propose two regression-based predictive models. However, the naively constructed design matrix is singular. We thus devise a systematic procedure to construct a fullranked design matrix. Using surgical data from a central Texas hospital, we compare the proposed models with a few benchmark methods and demonstrate that our models lead to a remarkable reduction in prediction errors.
40

Modeling Building Height Errors In 3d Urban Environments

Ergin, Ozge 01 December 2007 (has links) (PDF)
A great interest in 3-D modeling in Geographic Information Technologies (GIS) has emerged in recent years, because many GIS related implementations, ranging from urban area design to environmental analysis require 3-D models. Especially the need for 3-D models is quite urgent in urban areas. However, numerous applications in GIS only represent two-dimensional information. The GIS community has been struggling with solving complex problems dealing with 3-D objects using a 2-D approach. This research focused on finding most accurate method which is used for getting height information that is used in 3D modeling of man made structures in urban areas. The first method is estimating height information from floor numbers of the buildings data from municipal database systems. The second method is deriving heights of buildings from Digital Elevation Model (DEM) that is generated from stereo satellite images. The third method is measuring height values of the buildings from 3D view of stereo IKONOS satellite images by operators. The comparisons between these three methods are done with respect to height data collected from field study, and according to these comparisons, the amount of the error is determined. The error is classified according to floor numbers of buildings, so that, the quantified errors can be applied for similar works in future. Lastly, the third method is utilized by the assistance of 10 people who have different experience level about 3D viewing, in order to see the error amount changes according to different operators. Several results are presented with a discussion of evaluation of the methods applied. It is found that, if there is an updated floor number database, obtaining building height is the most accurate way from this database. The second most accurate method is found to be getting height information by using 3D view of stereo IKONOS images through experienced users.

Page generated in 0.0928 seconds