51 |
Optimal inspection and maintenance for stochastically deteriorating systemsDagg, Richard Andrew January 1999 (has links)
This thesis concerns the optimisation of maintenance and inspection for stochastically deteriorating systems. The motivation for this thesis is the problem of determining condition based maintenance policies, for systems whose degradation may be modelled by a continuous time stochastic process. Our emphasis is mainly on using the information gained from inspecting the degradation to determine efficient maintenance and inspection policies. The system we shall consider is one in which the degradation is modelled by a Levy process, and in which failure is defined to occur when the degradation reaches a critical level. It is assumed that the system may be inspected or repaired at any time, and that the costs of inspections and repairs may depend on the level of system degradation. Initially we look at determining optimal inspection policies for systems whose degradation may be directly and perfectly observed, before extending this analysis to the case where the degradation is unobservable, and a related covariate process is used to determine maintenance decisions. In both cases it is assumed the replacement policy is fixed and known in advance. Finally we consider the case of joint optimisation of maintenance and inspection, for cases in which the maintenance action has either deterministic or random effect on the degradation level. In all of these cases we use the properties of the Levy process degradation model to form a recursive relationship which allows us to determine integral and functional equations for the maintenance cost of the system. Solutions to these determine optimal periodic and non-periodic inspection and maintenance policies. Throughout the thesis we use the gamma process degradation model as an example. For this model we determine optimal perfect inspection policies for the cases when inspections are periodic and non-periodic. As a special case of a covariate process we consider the optimal imperfect periodic inspection policy. Finally we obtain jointly optimal deterministic-maintenance and periodic-inspection policies.
|
52 |
Methods for demographic inference from single-nucleotide polymorphism dataMair, Colette January 2012 (has links)
The distribution of the current human population is the result of many complex historical and prehistorical demographic events that have shaped variation in the human genome. Genomic dissimilarities between individuals from different geographical regions can potentially unveil something of these processes. The greatest differences lie between, and within, African populations and most research suggests the origin of modern humans lies within Africa. However, differing models have been proposed to model the evolutionary processes leading to humans inhabiting most of the world. This thesis develops a hypothesis test shown to be powerful in distinguishing between two such models. The first ("migration") model assumes the population of interest is divided into subpopulations that exchange migrants at a constant rate arbitrarily far back in the past, whilst the second ("isolation") model assumes that an ancestral population iteratively segregates into subpopulations that evolve independently. Although both models are simplistic, they do capture key aspects of the opposing theories of the history of modern humans. Given single nucleotide polymorphism (SNP) data from two subpopulations, the method described here tests a global null hypothesis that the data are from an isolation model. The test takes a parametric bootstrap approach, iteratively simulating data under the null hypothesis and computing a set of summary statistics shown to be able to distinguish between the two models. Each summary statistic forms the basis of a statistical hypothesis test where the observed value of the statistic is compared to the simulated values. The global null hypothesis is accepted if each individual test is accepted. A correction for multiple comparisons is used to control the type I error rate of this compound test. Extensions to this hypothesis test are given which adapt it to deal with SNP ascertainment and to better handle large genomic data sets. The methods are illustrated on data from the HapMap project using two Kenyan populations and the Japanese and Yoruba populations, after the method has been validated by simulation, where the `true' model is known.
|
53 |
Evaluation of sampling and monitoring designs for water qualityHaggarty, Ruth Alison January 2012 (has links)
Assessing water quality is of crucial importance to both society and the environment. Deterioration in water quality through issues such as eutrophication presents substantial risk to human health, plant and animal life, and can have detrimental effects on the local economy. Long-term data records across multiple sites can be used to investigate water quality and risk factors statistically, however, identification of underlying changes can only be successful if there is a sufficient quantity of data available. As vast amounts of resources are required for the implementation and maintenance of a monitoring network, logistically and financially it is not possible to employ continuous monitoring of all water environments. This raises the question as to the optimal design for long-term monitoring networks which are capable of capturing underlying changes. Two of the main design considerations are clearly where to sample, and how frequently to sample. The principal aim of this thesis is to use statistical analysis to investigate frequently used environmental monitoring networks, developing new methodology where appropriate, so that the design and implementation of future networks can be made as effective and cost efficient as possible. Using data which have been provided by the Scottish Environment Protection Agency, several data from Scottish lakes and rivers and a range of determinands are considered in order to explore water quality monitoring in Scotland. Chapter 1 provides an introduction to environmental monitoring and both existing statistical techniques, and potential challenges which are commonly encountered in the analysis of environmental data are discussed. Following this, Chapter 2 presents a simulation study which has been designed and implemented in order to evaluate the nature and statistical power for commonly used environmental sampling and monitoring designs for surface waters. The aim is to answer questions regarding how many samples to base the chemical classification of standing waters, and how appropriate the currently available data in Scotland are for detecting trends and seasonality. The simulation study was constructed to investigate the ability to detect the different underlying features of the data under several different sampling conditions. After the assessment of how often sampling is required to detect change, the remainder of the thesis will attempt to address some of the questions associated with where the optimal sampling locations are. The European Union Water Framework Directive (WFD) was introduced in 2003 to set compliance standards for all water bodies across Europe, with an aim to prevent deterioration, and ensure all sites reach `good' status by 2015. One of the features of the WFD is that water bodies can be grouped together and the classification of all members of the group is then based on the classification of a single representative site. The potential misclassification of sites means one of the key areas of interest is how well the existing groups used by SEPA for classification capture differences between the sites in terms of several chemical determinands. This will be explored in Chapter 3 where a functional data analysis approach will be taken in order to investigate some of the features of the existing groupings. An investigation of the effect of temporal autocorrelation on our ability to distinguish groups of sites from one another will also be presented here. It is also of interest to explore whether fewer, or indeed more groups would be optimal in order to accurately represent the trends and variability in the water quality parameters. Different statistical approaches for grouping standing waters will be presented in Chapter 4, where the question of how many groups is statistically optimal is also addressed. As in Chapter 3, these approaches for grouping sites will be based on functional data in order to include the temporal dynamics of the variable of interest within any analysis of group structure obtained. Both hierarchical and model based functional clustering are considered here. The idea of functional clustering is also extended to the multivariate setting, thus enabling information from several determinands of interest to be used within formation of groups. This is something which is of particular importance in view of the fact that the WFD classification encompasses a range of different determinands. In addition to the investigation of standing waters, an entirely different type of water quality monitoring network is considered in Chapter 5. While standing waters are assumed to be spatially independent of one another there are several situations where this assumption is not appropriate and where spatial correlation between locations needs to be accounted for. Further developments of the functional clustering methods explored in Chapter 4 are presented here in order to obtain groups of stations that are not only similar in terms of mean levels and temporal patterns of the determinand of interest, but which are also spatially homogenous. The river network data explored in Chapter 5 introduces a set of new challenges when considering functional clustering that go beyond the inclusion of Euclidean distance based spatial correlation. Existing methodology for estimating spatial correlation are combined with functional clustering approaches and developed to be suitable for application on sites which lie along a river network. The final chapter of this thesis provides a summary of the work presented and discussion of limitations and suggestions for future directions.
|
54 |
Latent variable models for mixed manifest variablesMoustaki, Irini January 1996 (has links)
Latent variable models are widely used in social sciences in which interest is centred on entities such as attitudes, beliefs or abilities for which there e)dst no direct measuring instruments. Latent modelling tries to extract these entities, here described as latent (unobserved) variables, from measurements on related manifest (observed) variables. Methodology already exists for fitting a latent variable model to manifest data that is either categorical (latent trait and latent class analysis) or continuous (factor analysis and latent profile analysis). In this thesis a latent trait and a latent class model are presented for analysing the relationships among a set of mixed manifest variables using one or more latent variables. The set of manifest variables contains metric (continuous or discrete) and binary items. The latent dimension is continuous for the latent trait model and discrete for the latent class model. Scoring methods for allocating individuals on the identified latent dimen-sions based on their responses to the mixed manifest variables are discussed. ' Item nonresponse is also discussed in attitude scales with a mixture of binary and metric variables using the latent trait model. The estimation and the scoring methods for the latent trait model have been generalized for conditional distributions of the observed variables given the vector of latent variables other than the normal and the Bernoulli in the exponential family. To illustrate the use of the naixed model four data sets have been analyzed. Two of the data sets contain five memory questions, the first on Thatcher's resignation and the second on the Hillsborough football disaster; these five questions were included in BMRBI's August 1993 face to face omnibus survey. The third and the fourth data sets are from the 1990 and 1991 British Social Attitudes surveys; the questions which have been analyzed are from the sexual attitudes sections and the environment section respectively.
|
55 |
A statistical analysis of low birthweight in GlasgowMurray, Barbara A. January 1999 (has links)
The percentage of singleton livebirths resulting in low birthweight deliveries has remained constant in the last 20 years, with between 6 and 10% of singleton pregnancies resulting in such a delivery. Low birthweight infants have been shown to develop medical problems in infancy and childhood, such as visual impairment, lower IQs and neuromotor problems, and as such it is important to identify those pregnancies that may result in low birthweight infants. This thesis considers factors that may be related to low birthweight, and uses these factors in the construction of a model to predict the probability of a woman delivering a low birthweight infant in order to identify high risk mothers. One factor that may be thought of as being related to low birthweight is deprivation. In this thesis a new deprivation measure is proposed which updates previous work in the area by using the 1991 small area census data to create a continuous deprivation measure, based on postcode area of residence, within the Greater Glasgow Health Board. This new measure of deprivation is included in the model referred to above. As there are many possible risk factors involved in modelling the probability of delivering a low birthweight infant multiple comparisons are involved in the production of the model and it is important to produce a model that incorporates most of the relevant factors and relatively few of the unimportant factors. The first order Bonferroni bound is one method used to correct for multiple comparisons by giving an upper bound on the actual p-value. This thesis considers the second order Bonferroni bound which gives a lower bound on the p-value and, when used in conjunction with the first order bound, gives a better correction method than the first order bound alone. These two bounds are then extended into logistic regression models.
|
56 |
The statistical analysis of exercise test data : a critical reviewMcConnachie, Alex January 2003 (has links)
Exercise tests have played a prominent role in the evaluation of therapies currently used for the management of patients with angina, such as nitrates, b-blockers, and calcium antagonists. Such evaluations have shown dramatic improvements in exercise tolerance, most commonly measured by the time spent exercising until the occurrence of anginal pain or ECG signs of ischaemia, and often amongst patients with severe disease. However, the statistical methods used have generally been based on Normal theory, such as the t-test, or non-parametric equivalents, such as the Wilcoxon rank sum test. Such methods make no allowance for the fact that ischaemic endpoints may not occur in all patients, particularly when patients are under active treatment or in patients with less severe symptoms. In the current situation, where there are several therapeutic options of proven clinical effectiveness, new treatments must be evaluated in opposition, or in addition to existing therapies. Thus it is of particular importance that the statistician responsible for an analysis of exercise test data should use appropriate and efficient techniques, since the benefits of new treatments may be small. Since exercise times may be censored, in that the event of interest need not occur, it has been recognised that methods for the analysis of survival data are the most appropriate for analyses of exercise test data. Using data from the TIBET Study, a large clinical trial of two anti-anginal therapies administered singly or in combination, this thesis examines in detail the appropriateness of the Cox proportional hazards model, the most popular method for survival regression in the medical literature, to this type of data. It then considers alternatives to this model, and addresses the implications of some common features of exercise test data, in particular the presence of interval censoring and the possibility of multiple exercise tests being conducted on the same patient, using data from the TIBET Study and through simulation studies. Finally, using real data examples, two methods that appear to have received little or not attention with respect to exercise test data are explored, namely competing risks and repeated measures analyses.
|
57 |
Optimal cutpoint determination via design theory for regression modelsNguyen, The M. January 2006 (has links)
The main focus of this thesis is how to optimally choose a set or sets of cutpoints (in a categorized survey question) which are offered to respondents. In the case of several sets, a further issue is how to allocate sampled subjects to these sets (design points). Applications include Contingent Valuation (CV) studies (surveys on a population’s willingness to pay for a service or public good) and market research studies which might include for example a question on individual incomes.
|
58 |
Bayesian analysis of finite mixture distributions using the allocation samplerFearnside, Alastair T. January 2007 (has links)
Finite mixture distributions are receiving more and more attention from statisticians in many different fields of research because they are a very flexible class of models. They are typically used for density estimation or to model population heterogeneity. One can think of a finite mixture distribution as grouping the observations into components from which they are assumed to have arisen. In certain settings these groups have a physical interpretation. The interest in these distributions has been boosted recently because of the ever increasing computer power available to researchers to carry out the computationally intensive tasks required in their analysis. In order to fit a finite mixture distribution taking a Bayesian approach a posterior distribution has to be evaluated. When the number of components in the model is assumed known this posterior distribution can be sampled from using methods such as Data Augmentation or Gibbs sampling (Tanner and Wong (1987) and Gelfand and Smith (1990)) and the Metropolis-Hastings algorithm (Hastings (1970)). However, the number of components in the model can also be considered an unknown and an object of inference. Richardson and Green (1997) and Stephens (2000a) both describe Bayesian methods to sample across models with different numbers of components. This enables an estimate of the posterior distribution of the number of components to be evaluated. Richardson and Green (1997) define a reversible jump Markov chain Monte Carlo (RJMCMC) sampler while Stephens (2000a) uses a Markov birth-death process approach sample from the posterior distribution. In this thesis a Markov chain Monte Carlo method, named the allocation sampler. This sampler differs from the RJMCMC method reported in Richardson and Green (1997) because the state space of the sampler is simplified by the assumption that the components' parameters and weights can be analytically integrated out of the model. This in turn has the advantage that only minimal changes are required to the sampler for mixtures of components from other parametric families. This thesis illustrates the allocation sampler's performance on both simulated and real data sets. Chapter 1 provides a background to finite mixture distributions and gives an overview of some inferential techniques that have already been used to analyse these distributions. Chapter 2 sets out the Bayesian model framework that is used throughout this thesis and defines all the required distributional results. Chapter 3 describes the allocation sampler. Chapter 4 tests the performance of the allocation sampler using simulated datasets from a collection of 15 different known mixture distributions. Chapter 5 illustrates the allocation sampler with real datasets from a number of different research fields. Chapter 6 summarises the research in the thesis and provides areas of possible future research.
|
59 |
Vision erosion toward a new model for understanding E-learning in higher edication: the Israeli Open University 1992-2000 as a case study /Golan, Hanan Shay. January 2008 (has links) (PDF)
Thesis (Ph.D.)--The London School of Economics and Political Science University of London 2001.
|
60 |
Theory and applications of delayed censoring models in survival analysisHeydari, Fariborz January 1997 (has links)
The objective of this thesis is to develop new statistical models for the analysis of censored survival data, particularly for the study of recidivism data, such as the reoffence data used in the analysis here. This has been an area of great interest in criminology in recent years. There is a growing literature on survival analysis in criminology, where interest centres on the time from an offender's conviction, or release from prison, to the first reconviction or reimprisonment. In deciding whether to release a prisoner on parole, the Parole Board is provided with a statistical score which estimates the chance that the prisoner will reoffend within the period of time that he or she would otherwise be in prison. This score is based on a survival analysis of data on a sample of releases from long-term prison sentences. To capture most reoffences which occur within 2 years of release, follow-up must continue for at least 3 years to allow for the delay between offence and conviction. We reanalyse the data by using a model which explicitly allows for this delay. We refer to this as 'delayed censoring model'. The new analysis can be applied to data with a substantially shorter length of follow-up. This means that risk scores can be constructed from more up-to-date data and at less cost. It is models of this kind that we shall be concerned with in this thesis, and this is the principal motivation of the work done. The statistical models that this thesis provides bring in a number of new ideas which are undoubtedly useful both at a theoretical level and in applications. Other major divisions of the work include: (i) Assessing the possibility of an association between the delay and reoffence times by studying truncated distributions fitted to these data, by parametric, semi-parametric and nonparametric models. With the nonparametric approach we have developed a 'backward regression model' which is similar to the Cox model. (ii) We have also discussed delayed censoring modification to the Cox model, and developed a more general semi-parametric model for all the data including both observed and censored cases. In this model the delay and reoffence times are allowed to be correlated. We refer to this as the 'generalized weighted hazards model'. (iii) Finally, we have compared the results by applying all these models to the data. Although the parametric models give a good fit to the data, the semi-parametric and nonparametric models give a slightly better fit, as expected.
|
Page generated in 0.0433 seconds