Spelling suggestions: "subject:"1ieasurement error"" "subject:"1ieasurement arror""
31 |
Behind Left and Right. The meaning of left-right orientation in EuropeWeber, Wiebke 21 January 2013 (has links)
The left-right concept is considered to facilitate communication and
orientation in the political world. It has a long tradition in European
politics due to its capacity to absorb different associations over
time. However, this dynamic nature of the concept requires
periodical reassessments in order to assure that a common
conceptualisation endures. This dissertation focuses on reassign
what individual left-right orientation means. Usually, this is
measured by asking people to place themselves on a scale labelled
‘left’ and ‘right’ at its endpoints. The first empirical analysis of this
dissertation shows that this measure is comparable across groups
and countries. Thereafter, the relationship between an individual’s
issue preference and left-right orientation is considered. The second
empirical analysis shows that this relationship is conditioned by the
importance people assign to the respective issues. The final analysis
demonstrates that what explains left-right orientation is contingent
on individual and contextual factors. This implies that in order to
understand left-right orientation, it is not enough to identify what
has an impact on a person’s position but also account for all those
factors that predict variation between individuals. Given this
complexity, my conclusion is that the left-right concept runs the risk
of becoming too complicated to serve as an analytical tool to shed
light on political attitudes and behaviour. / El concepte esquerra-dreta és considerat com un factor facilitador
de la comunicació en el món polític. Té una llarga tradició’ en la
política europea degut a la seva capacitat d’absorbir diferents
associacions a través del temps. Tanmateix, aquesta natura dinàmica
del concepte requereix revisions periòdiques per assegurar que
persisteix una conceptualització comuna. La present tesi es centra
en resignar el que significa l’orientació esquerra-dreta.
Normalment, es mesura tot demanant als enquestats posicionar-se a
ells mateixos en una escala que va de l’esquerra a la dreta. El primer
anàlisi empíric de la present tesi mostra que aquesta mesura és
comparable entre grups i països. Seguidament, es considera la
relació entre les preferències temàtiques dels individus i llur
orientació esquerra-dreta. El segon anàlisi empíric mostra que
aquesta relació està condicionada per la importància que les
persones assignen als temes respectius. L’anàlisi final demostra que
el que explica l’orientació esquerra-dreta depèn de factors
contextuals i individuals. Això implica que per entendre l’orientació
esquerra-dreta no és suficient identificar què té un impacte en la
posició d’una persona sinó també una explicació per a tots aquells
factors que preveuen la variació entre individus. Donada aquesta
complexitat, la meva conclusió és que el concepte esquerra-dreta
corre el risc de convertir-se en massa complicat per a servir com a
eina analítica per a l’estudi de les actituds i el comportament
politics.
|
32 |
Measurement Error and Misclassification in Interval-Censored Life History DataWhite, Bethany Joy Giddings January 2007 (has links)
In practice, data are frequently incomplete in one way or another. It can be a significant challenge to make valid inferences about the parameters of interest in this situation. In this thesis, three
problems involving such data are addressed. The first two problems involve interval-censored life history data with mismeasured
covariates. Data of this type are incomplete in two ways. First, the exact event times are unknown due to censoring. Second, the true covariate is missing for most, if not all, individuals. This work
focuses primarily on the impact of covariate measurement error in progressive multi-state models with data arising from panel (i.e., interval-censored) observation. These types of problems arise frequently in clinical settings (e.g. when disease progression is of interest and patient information is collected during irregularly spaced clinic visits). Two and three state models are considered in this thesis. This work is motivated by a research program on psoriatic arthritis (PsA) where the effects of error-prone covariates on rates of disease progression are of interest and patient information is collected at clinic visits (Gladman et al. 1995; Bond et al. 2006). Information regarding the error distributions were available based on results from a separate study conducted to evaluate the reliability of clinical measurements that are used in PsA treatment and follow-up (Gladman et al. 2004). The asymptotic bias of covariate effects obtained ignoring error in covariates is investigated and shown to be substantial in some settings. In a series of simulation studies, the performance of corrected likelihood methods and methods based on a simulation-extrapolation (SIMEX) algorithm (Cook \& Stefanski 1994) were investigated to address covariate measurement error. The methods implemented were shown to result in much smaller empirical biases and empirical coverage probabilities which were closer to the nominal levels.
The third problem considered involves an extreme case of interval censoring known as current status data. Current status data arise when individuals are observed only at a single point in time and it is then determined whether they have experienced the event of interest. To complicate matters, in the problem considered here, an unknown proportion of the population will never experience the event of interest. Again, this type of data is incomplete in two ways. One assessment is made on each individual to determine whether or not an event has occurred. Therefore, the exact event times are unknown for those who will eventually experience the event. In addition, whether or not the individuals will ever experience the event is unknown for those who have not experienced the event by the assessment time. This problem was motivated by a series of orthopedic trials looking at the effect of blood thinners in hip and knee replacement surgeries. These blood thinners can cause a negative serological response in some patients. This response was the outcome of interest and the only available information regarding it was the seroconversion time under current status observation. In this thesis, latent class models with parametric, nonparametric and piecewise constant forms of the seroconversion time distribution are described. They account for the fact that only a proportion of the population will experience the event of interest. Estimators based on an EM algorithm were evaluated via simulation and the orthopedic surgery data were analyzed based on this methodology.
|
33 |
Measurement Error and Misclassification in Interval-Censored Life History DataWhite, Bethany Joy Giddings January 2007 (has links)
In practice, data are frequently incomplete in one way or another. It can be a significant challenge to make valid inferences about the parameters of interest in this situation. In this thesis, three
problems involving such data are addressed. The first two problems involve interval-censored life history data with mismeasured
covariates. Data of this type are incomplete in two ways. First, the exact event times are unknown due to censoring. Second, the true covariate is missing for most, if not all, individuals. This work
focuses primarily on the impact of covariate measurement error in progressive multi-state models with data arising from panel (i.e., interval-censored) observation. These types of problems arise frequently in clinical settings (e.g. when disease progression is of interest and patient information is collected during irregularly spaced clinic visits). Two and three state models are considered in this thesis. This work is motivated by a research program on psoriatic arthritis (PsA) where the effects of error-prone covariates on rates of disease progression are of interest and patient information is collected at clinic visits (Gladman et al. 1995; Bond et al. 2006). Information regarding the error distributions were available based on results from a separate study conducted to evaluate the reliability of clinical measurements that are used in PsA treatment and follow-up (Gladman et al. 2004). The asymptotic bias of covariate effects obtained ignoring error in covariates is investigated and shown to be substantial in some settings. In a series of simulation studies, the performance of corrected likelihood methods and methods based on a simulation-extrapolation (SIMEX) algorithm (Cook \& Stefanski 1994) were investigated to address covariate measurement error. The methods implemented were shown to result in much smaller empirical biases and empirical coverage probabilities which were closer to the nominal levels.
The third problem considered involves an extreme case of interval censoring known as current status data. Current status data arise when individuals are observed only at a single point in time and it is then determined whether they have experienced the event of interest. To complicate matters, in the problem considered here, an unknown proportion of the population will never experience the event of interest. Again, this type of data is incomplete in two ways. One assessment is made on each individual to determine whether or not an event has occurred. Therefore, the exact event times are unknown for those who will eventually experience the event. In addition, whether or not the individuals will ever experience the event is unknown for those who have not experienced the event by the assessment time. This problem was motivated by a series of orthopedic trials looking at the effect of blood thinners in hip and knee replacement surgeries. These blood thinners can cause a negative serological response in some patients. This response was the outcome of interest and the only available information regarding it was the seroconversion time under current status observation. In this thesis, latent class models with parametric, nonparametric and piecewise constant forms of the seroconversion time distribution are described. They account for the fact that only a proportion of the population will experience the event of interest. Estimators based on an EM algorithm were evaluated via simulation and the orthopedic surgery data were analyzed based on this methodology.
|
34 |
Analysis of Correlated Data with Measurement Error in Responses or CovariatesChen, Zhijian January 2010 (has links)
Correlated data frequently arise from epidemiological studies, especially familial
and longitudinal studies. Longitudinal design has been used by researchers to investigate the changes of certain characteristics over time at the individual level as well as how potential factors influence the changes. Familial studies are often designed to investigate the dependence of health conditions among family members. Various models have been developed for this type of multivariate data, and a wide variety
of estimation techniques have been proposed. However, data collected from observational
studies are often far from perfect, as measurement error may arise from different
sources such as defective measuring systems, diagnostic tests without gold references,
and self-reports. Under such scenarios only rough surrogate variables are measured. Measurement error in covariates in various regression models has been discussed extensively in the literature. It is well known that naive approaches ignoring covariate error often lead to inconsistent estimators for model parameters.
In this thesis, we develop inferential procedures for analyzing correlated data with
response measurement error. We consider three scenarios: (i) likelihood-based inferences for generalized linear mixed models when the continuous response is subject to nonlinear measurement errors; (ii) estimating equations methods for binary responses with misclassifications; and (iii) estimating equations methods for ordinal
responses when the response variable and categorical/ordinal covariates are subject
to misclassifications.
The first problem arises when the continuous response variable is difficult to measure.
When the true response is defined as the long-term average of measurements, a single measurement is considered as an error-contaminated surrogate. We focus on generalized linear mixed models with nonlinear response error and study the induced bias in naive estimates. We propose likelihood-based methods that can yield consistent and efficient estimators for both fixed-effects and variance parameters. Results of simulation studies and analysis of a data set from the Framingham Heart Study
are presented.
Marginal models have been widely used for correlated binary, categorical, and ordinal data. The regression parameters characterize the marginal mean of a single outcome, without conditioning on other outcomes or unobserved random effects. The generalized estimating equations (GEE) approach, introduced by Liang and Zeger (1986), only models the first two moments of the responses with associations being
treated as nuisance characteristics. For some clustered studies especially familial
studies, however, the association structure may be of scientific interest. With binary
data Prentice (1988) proposed additional estimating equations that allow one to
model pairwise correlations. We consider marginal models for correlated binary data
with misclassified responses. We develop “corrected” estimating equations approaches
that can yield consistent estimators for both mean and association parameters. The
idea is related to Nakamura (1990) that is originally developed for correcting bias
induced by additive covariate measurement error under generalized linear models. Our approaches can also handle correlated misclassifications rather than a simple
misclassification process as considered by Neuhaus (2002) for clustered binary data
under generalized linear mixed models. We extend our methods and further develop
marginal approaches for analysis of longitudinal ordinal data with misclassification in both responses and categorical covariates. Simulation studies show that our proposed methods perform very well under a variety of scenarios. Results from application of the proposed methods to real data are presented.
Measurement error can be coupled with many other features in the data, e.g., complex survey designs, that can complicate inferential procedures. We explore combining
survey weights and misclassification in ordinal covariates in logistic regression
analyses. We propose an approach that incorporates survey weights into estimating
equations to yield design-based unbiased estimators.
In the final part of the thesis we outline some directions for future work, such as
transition models and semiparametric models for longitudinal data with both incomplete
observations and measurement error. Missing data is another common feature in applications. Developing novel statistical techniques for dealing with both missing
data and measurement error can be beneficial.
|
35 |
Estimation of Stochastic Degradation Models Using Uncertain Inspection DataLu, Dongliang January 2012 (has links)
Degradation of components and structures is a major threat to the safety and reliability of large engineering systems, such as the railway networks or the nuclear power plants. Periodic inspection and maintenance are thus required to ensure that the system is in good condition for continued service. A key element for the optimal inspection and maintenance is to accurately model and forecast the degradation progress, such that inspection and preventive maintenance can be scheduled accordingly.
In recently years, probabilistic models based on stochastic process have become increasingly popular in degradation modelling, due to their flexibility in modelling both the temporal and sample uncertainties of the degradation. However, because of the often complex structure of stochastic degradation models, accurate estimate of the model parameters can be quite difficult, especially when the inspection data are noisy or incomplete. Not considering the effect of uncertain inspection data is likely to result in biased parameter estimates and therefore erroneous predictions of future degradation.
The main objective of the thesis is to develop formal methods for the parameter estimation of stochastic degradation models using uncertain inspection data. Three typical stochastic models are considered. They are the random rate model, the gamma process model and the Poisson process model, among which the random rate model and the gamma process model are used to model the flaw growth, and the Poisson process model is used to model the flaw generation. Likelihood functions of the three stochastic models given noisy or incomplete inspection data are derived, from which maximum likelihood estimates can be obtained.
The thesis also investigates Bayesian inference of the stochastic degradation models. The most notable advantage of Bayesian inference over classical point estimates is its ability to incorporate background information in the estimation process, which is especially useful when inspection data are scarce.
A major obstacle for accurate parameter inference of stochastic models from uncertain inspection data is the computational difficulties of the likelihood evaluation, as it often involves calculation of high dimensional integrals or large number of convolutions. To overcome the computational difficulties, a number of numerical methods are developed in the thesis. For example, for the gamma process model subject to sizing error, an efficient maximum likelihood method is developed using the Genz's transform and quasi-Monte Carlo simulation. A Markov Chain Monte Carlo simulation with sizing error as auxiliary variables is developed for the Poisson flaw generation model, A sequential Bayesian updating using approximate Bayesian computation and weighted samples is also developed for Bayesian inference of the gamma process subject to sizing error.
Examples on the degradation of nuclear power plant components are presented to illustrate the use of the stochastic degradation models using practical uncertain inspection data. It is shown from the examples that the proposed methods are very effective in terms of accuracy and computational efficiency.
|
36 |
Accrual Noise Ratio as a Measure of Accrual ReliabilityNjoroge, Kenneth January 2009 (has links)
<p>I develop an empirical model that estimates a firm-specific accrual noise ratio (ANR), an operational and statistically grounded measure of accrual reliability, and test the measure's construct validity. The model allows accrual reliability to vary across firms, which is particularly important because many reliability determinants vary in cross-section. Unlike metrics that measure relative perceived reliability, ANR measures accrual reliability independent of the perceptions of investors, creditors or auditors. I find that ANR relates in expected ways with multiple proxies of accounting reliability, that ANR's relation with the proxies of other accounting constructs is consistent with theory, and that ANR's sensitivity to percentage changes of accrual components is consistent with a subjective ordinal ranking of the components' reliability from prior literature.</p> / Dissertation
|
37 |
Efficient inference in general semiparametric regression modelsMaity, Arnab 15 May 2009 (has links)
Semiparametric regression has become very popular in the field of Statistics over the
years. While on one hand more and more sophisticated models are being developed,
on the other hand the resulting theory and estimation process has become more and
more involved. The main problems that are addressed in this work are related to
efficient inferential procedures in general semiparametric regression problems.
We first discuss efficient estimation of population-level summaries in general semiparametric
regression models. Here our focus is on estimating general population-level
quantities that combine the parametric and nonparametric parts of the model (e.g.,
population mean, probabilities, etc.). We place this problem in a general context,
provide a general kernel-based methodology, and derive the asymptotic distributions
of estimates of these population-level quantities, showing that in many cases the estimates
are semiparametric efficient.
Next, motivated from the problem of testing for genetic effects on complex traits in
the presence of gene-environment interaction, we consider developing score test in
general semiparametric regression problems that involves Tukey style 1 d.f form of
interaction between parametrically and non-parametrically modeled covariates. We
develop adjusted score statistics which are unbiased and asymptotically efficient and
can be performed using standard bandwidth selection methods. In addition, to over come the difficulty of solving functional equations, we give easy interpretations of the
target functions, which in turn allow us to develop estimation procedures that can be
easily implemented using standard computational methods.
Finally, we take up the important problem of estimation in a general semiparametric
regression model when covariates are measured with an additive measurement error
structure having normally distributed measurement errors. In contrast to methods
that require solving integral equation of dimension the size of the covariate measured
with error, we propose methodology based on Monte Carlo corrected scores to estimate
the model components and investigate the asymptotic behavior of the estimates.
For each of the problems, we present simulation studies to observe the performance of
the proposed inferential procedures. In addition, we apply our proposed methodology
to analyze nontrivial real life data sets and present the results.
|
38 |
Deconvolution in Random Effects Models via Normal MixturesLitton, Nathaniel A. 2009 August 1900 (has links)
This dissertation describes a minimum distance method for density estimation when the variable of interest is not directly observed. It is assumed that the underlying target density can be well approximated by a mixture of normals. The method compares a density estimate of observable data with a density of the observable data induced from assuming the target density can be written as a mixture of normals. The goal is to choose the parameters in the normal mixture that minimize the distance between the density estimate of the observable data and the induced density from the model. The method is applied to the deconvolution problem to estimate the density of $X_{i}$ when the variable $% Y_{i}=X_{i}+Z_{i}$, $i=1,\ldots ,n$, is observed, and the density of $Z_{i}$ is known. Additionally, it is applied to a location random effects model to estimate the density of $Z_{ij}$ when the observable quantities are $p$ data sets of size $n$ given by $X_{ij}=\alpha _{i}+\gamma Z_{ij},~i=1,\ldots ,p,~j=1,\ldots ,n$, where the densities of $\alpha_{i} $ and $Z_{ij}$ are both unknown.
The performance of the minimum distance approach in the measurement error model is compared with the deconvoluting kernel density estimator of Stefanski and Carroll (1990). In the location random effects model, the minimum distance estimator is compared with the explicit characteristic function inversion method from Hall and Yao (2003). In both models, the methods are compared using simulated and real data sets. In the simulations, performance is evaluated using an integrated squared error criterion. Results indicate that the minimum distance methodology is comparable to the deconvoluting kernel density estimator and outperforms the explicit characteristic function inversion method.
|
39 |
Bayesian Methods in Nutrition Epidemiology and Regression-based Predictive Models in HealthcareZhang, Saijuan 2010 December 1900 (has links)
This dissertation has mainly two parts. In the first part, we propose a bivariate nonlinear multivariate measurement error model to understand the distribution of dietary intake and extend it to a multivariate model to capture dietary patterns in nutrition epidemiology. In the second part, we propose regression-based predictive models to accurately predict surgery duration in healthcare.
Understanding the distribution of episodically consumed dietary components is an important problem in public health. Short-term measurements of episodically consumed dietary components are zero-inflated skewed distributions. So-called two-part models have been developed for such data. However, there is much greater public health interest in the usual intake adjusted for caloric intake. Recently a nonlinear mixed effects model has been developed and fit by maximum likelihood using nonlinear mixed effects programs. However, the fitting is slow and unstable. We develop a Monte-Carlo-based fitting method in Chapter II. We demonstrate numerically that our methods lead to increased speed of computation, converge to reasonable solutions, and have the flexibility to be used in either a frequentist or a Bayesian manner. Diet consists of numerous foods, nutrients and other components, each of which have distinctive attributes. Increasingly nutritionists are interested in exploring them collectively to capture overall dietary patterns. We thus extend the bivariate model described in Chapter III to multivariate level. We use survey-weighted MCMC computations to fit the model, with uncertainty estimation coming from balanced repeated replication. The methodology is illustrated through an application of estimating the population distribution of the Healthy Eating Index-2005 (HEI-2005), a multi-component dietary quality index , among children aged 2-8 in the United States.
The second part of this dissertation is to accurately predict surgery duration. Prior research has identified the current procedural terminology (CPT) codes as the
most important factor when predicting surgical case durations but there has been little reporting of a general predictive methodology using it effectively. In Chapter IV, we propose two regression-based predictive models. However, the naively constructed design matrix is singular. We thus devise a systematic procedure to construct a fullranked design matrix. Using surgical data from a central Texas hospital, we compare the proposed models with a few benchmark methods and demonstrate that our models lead to a remarkable reduction in prediction errors.
|
40 |
Modeling Building Height Errors In 3d Urban EnvironmentsErgin, Ozge 01 December 2007 (has links) (PDF)
A great interest in 3-D modeling in Geographic Information Technologies
(GIS) has emerged in recent years, because many GIS related
implementations, ranging from urban area design to environmental
analysis require 3-D models. Especially the need for 3-D models is quite
urgent in urban areas.
However, numerous applications in GIS only represent two-dimensional
information. The GIS community has been struggling with solving
complex problems dealing with 3-D objects using a 2-D approach. This
research focused on finding most accurate method which is used for
getting height information that is used in 3D modeling of man made
structures in urban areas. The first method is estimating height
information from floor numbers of the buildings data from municipal
database systems. The second method is deriving heights of buildings
from Digital Elevation Model (DEM) that is generated from stereo satellite
images. The third method is measuring height values of the buildings
from 3D view of stereo IKONOS satellite images by operators. The
comparisons between these three methods are done with respect to
height data collected from field study, and according to these
comparisons, the amount of the error is determined. The error is
classified according to floor numbers of buildings, so that, the quantified
errors can be applied for similar works in future. Lastly, the third method
is utilized by the assistance of 10 people who have different experience
level about 3D viewing, in order to see the error amount changes
according to different operators. Several results are presented with a
discussion of evaluation of the methods applied. It is found that, if there
is an updated floor number database, obtaining building height is the
most accurate way from this database. The second most accurate
method is found to be getting height information by using 3D view of
stereo IKONOS images through experienced users.
|
Page generated in 0.134 seconds