Spelling suggestions: "subject:"confidence intervals"" "subject:"konfidence intervals""
61 |
Exploring Confidence Intervals in the Case of Binomial and Hypergeometric DistributionsMojica, Irene 01 January 2011 (has links)
The objective of this thesis is to examine one of the most fundamental and yet important methodologies used in statistical practice, interval estimation of the probability of success in a binomial distribution. The textbook confidence interval for this problem is known as the Wald interval as it comes from the Wald large sample test for the binomial case. It is generally acknowledged that the actual coverage probability of the standard interval is poor for values of p near 0 or 1. Moreover, recently it has been documented that the coverage properties of the standard interval can be inconsistent even if p is not near the boundaries. For this reason, one would like to study the variety of methods for construction of confidence intervals for unknown probability p in the binomial case. The present thesis accomplishes the task by presenting several methods for constructing confidence intervals for unknown binomial probability p. It is well known that the hypergeometric distribution is related to the binomial distribution. In particular, if the size of the population, N, is large and the number of items of interest k is such that k/N tends to p as N grows, then the hypergeometric distribution can be approximated by the binomial distribution. Therefore, in this case, one can use the confidence intervals constructed for p in the case of the binomial distribution as a basis for construction of the confidence intervals for the unknown value k = pN. The goal of this thesis is to study this approximation and to point out several confidence intervals which are designed specifically for the hypergeometric distribution. In particular, this thesis considers several confidence intervals which are based on estimation of a binomial proportion as well as Bayesian credible sets based on various priors.
|
62 |
Multiple Comparisons under Unequal Variances and Its Application to Dose Response StudiesLi, Hong 28 September 2009 (has links)
No description available.
|
63 |
Quality Control Using Inferential Statistics in Weibull Analyses for Components Fabricated from Monolithic CeramicsParikh, Ankurben H. 04 April 2012 (has links)
No description available.
|
64 |
STATISTICAL INFERENCE ON BINOMIAL PROPORTIONSZHAO, SHUHONG 13 July 2005 (has links)
No description available.
|
65 |
Feasible Generalized Least Squares: theory and applicationsGonzález Coya Sandoval, Emilio 04 June 2024 (has links)
We study the Feasible Generalized Least-Squares (FGLS) estimation of the parameters of a linear regression model in which the errors are allowed to exhibit heteroskedasticity of unknown form and to be serially correlated. The main contribution
is two fold; first we aim to demystify the reasons often advanced to use OLS instead of FGLS by showing that the latter estimate is robust, and more efficient and precise. Second, we devise consistent FGLS procedures, robust to misspecification, which achieves a lower mean squared error (MSE), often close to that of the correctly
specified infeasible GLS.
In the first chapter we restrict our attention to the case with independent heteroskedastic errors. We suggest a Lasso based procedure to estimate the skedastic function of the residuals. This estimate is then used to construct a FGLS estimator. Using extensive Monte Carlo simulations, we show that this Lasso-based FGLS procedure has better finite sample properties than OLS and other linear regression-based FGLS estimates. Moreover, the FGLS-Lasso estimate is robust to misspecification of
both the functional form and the variables characterizing the skedastic function.
The second chapter generalizes our investigation to the case with serially correlated errors. There are three main contributions; first we show that GLS is consistent requiring only pre-determined regressors, whereas OLS requires exogenous regressors to be consistent. The second contribution is to show that GLS is much more robust that OLS; even a misspecified GLS correction can achieve a lower MSE than OLS. The third contribution is to devise a FGLS procedure valid whether or not the regressors are exogenous, which achieves a MSE close to that of the correctly specified infeasible GLS. Extensive Monte Carlo experiments are conducted to assess the performance of our FGLS procedure against OLS in finite samples. FGLS achieves important reductions in MSE and variance relative to OLS.
In the third chapter we consider an empirical application; we re-examine the Uncovered Interest Parity (UIP) hypothesis, which states that the expected rate of return to speculation in the forward foreign exchange market is zero. We extend the FGLS procedure to a setting in which lagged dependent variables are included as regressors. We thus provide a consistent and efficient framework to estimate the parameters of a general k-step-ahead linear forecasting equation. Finally, we apply our FGLS procedures to the analysis of the two main specifications to test the UIP.
|
66 |
The Accuracy of River Bed Sediment SamplesPetrie, John Eric 19 January 1999 (has links)
One of the most important factors that influences a stream's hydraulic and ecological health is the streambed's sediment size distribution. This distribution affects streambed stability, sediment transport rates, and flood levels by defining the roughness of the stream channel. Adverse effects on water quality and wildlife can be expected when excessive fine sediments enter a stream. Many chemicals and toxic materials are transported through streams by binding to fine sediments. Increases in fine sediments also seriously impact the survival of fish species present in the stream. Fine sediments fill tiny spaces between larger particles thereby denying fish embryos the necessary fresh water to survive. Reforestation, constructed wetlands, and slope stabilization are a few management practices typically utilized to reduce the amount of sediment entering a stream. To effectively gauge the success of these techniques, the sediment size distribution of the stream must be monitored.
Gravel bed streams are typically stratified vertically, in terms of particle size, in three layers, with each layer having its own distinct grain size distribution. The top two layers of the stream bed, the pavement and subpavement, are the most significant in determining the characteristics of the stream. These top two layers are only as thick as the largest particle size contained within each layer. This vertical stratification by particle size makes it difficult to characterize the grain size distribution of the surface layer. The traditional bulk or volume sampling procedure removes a specified volume of material from the stream bed. However, if the bed exhibits vertical stratification, the volume sample will mix different populations, resulting in inaccurate sample results. To obtain accurate results for the pavement size distribution, a surface oriented sampling technique must be employed. The most common types of surface oriented sampling are grid and areal sampling. Due to limitations in the sampling techniques, grid samples typically truncate the sample at the finer grain sizes, while areal samples typically truncate the sample at the coarser grain sizes. When combined with an analysis technique, either frequency-by-number or frequency-by-weight, the sample results can be represented in terms of a cumulative grain size distribution. However, the results of different sampling and analysis procedures can lead to biased results, which are not equivalent to traditional volume sampling results. Different conversions, dependent on both the sampling and analysis technique, are employed to remove the bias from surface sample results.
The topic of the present study is to determine the accuracy of sediment samples obtained by the different sampling techniques. Knowing the accuracy of a sample is imperative if the sample results are to be meaningful. Different methods are discussed for placing confidence intervals on grid sample results based on statistical distributions. The binomial distribution and its approximation with the normal distribution have been suggested for these confidence intervals in previous studies. In this study, the use of the multinomial distribution for these confidence intervals is also explored. The multinomial distribution seems to best represent the grid sampling process. Based on analyses of the different distributions, recommendations are made. Additionally, figures are given to estimate the grid sample size necessary to achieve a required accuracy for each distribution. This type of sample size determination figure is extremely useful when preparing for grid sampling in the field.
Accuracy and sample size determination for areal and volume samples present difficulties not encountered with grid sampling. The variability in number of particles contained in the sample coupled with the wide range of particle sizes present make direct statistical analysis impossible. Limited studies have been reported on the necessary volume to sample for gravel deposits. The majority of these studies make recommendations based on empirical results that may not be applicable to different size distributions. Even fewer studies have been published that address the issue of areal sample size. However, using grid sample results as a basis, a technique is presented to estimate the necessary sizes for areal and volume samples. These areal and volume sample sizes are designed to match the accuracy of the original grid sample for a specified grain size percentile of interest. Obtaining grid and areal results with the same accuracy can be useful when considering hybrid samples. A hybrid sample represents a combination of grid and areal sample results that give a final grain size distribution curve that is not truncated. Laboratory experiments were performed on synthetic stream beds to test these theories. The synthetic stream beds were created using both glass beads and natural sediments. Reducing sampling errors and obtaining accurate samples in the field are also briefly discussed. Additionally, recommendations are also made for using the most efficient sampling technique to achieve the required accuracy. / Master of Science
|
67 |
Confidence Intervals and Sample Size Calculations for Studies of Film-reading PerformanceScally, Andy J., Brealey, S. January 2003 (has links)
No / The relaxation of restrictions on the type of professions that can report films has resulted in radiographers and other healthcare professionals becoming increasingly involved in image interpretation in areas such as mammography, ultrasound and plain-film radiography. Little attention, however, has been given to sample size determinations concerning film-reading performance characteristics such as sensitivity, specificity and accuracy. Illustrated with hypothetical examples, this paper begins by considering standard errors and confidence intervals for performance characteristics and then discusses methods for determining sample size for studies of film-reading performance. Used appropriately, these approaches should result in studies that produce estimates of film-reading performance with adequate precision and enable investigators to optimize the sample size in their studies for the question they seek to answer.
|
68 |
A practical introduction to medical statisticsScally, Andy J. 16 October 2013 (has links)
No / Medical statistics is a vast and ever-growing field of academic endeavour, with direct application to developing the robustness of the evidence base in all areas of medicine. Although the complexity of available statistical techniques has continued to increase, fuelled by the rapid data processing capabilities of even desktop/laptop computers, medical practitioners can go a long way towards creating, critically evaluating and assimilating this evidence with an understanding of just a few key statistical concepts. While the concepts of statistics and ethics are not common bedfellows, it should be emphasised that a statistically flawed study is also an unethical study.[1] This review will outline some of these key concepts and explain how to interpret the output of some commonly used statistical analyses. Examples will be confined to two-group tests on independent samples, using both a continuous and a dichotomous/binary outcome measure.
|
69 |
Confidence intervals for estimators of welfare indices under complex samplingKirchoff, Retha 03 1900 (has links)
Thesis (MComm (Statistics and Actuarial Science))--University of Stellenbosch, 2010. / ENGLISH ABSTRACT: The aim of this study is to obtain estimates and confidence intervals for welfare
indices under complex sampling. It begins by looking at sampling in general with
specific focus on complex sampling and weighting. For the estimation of the welfare
indices, two resampling techniques, viz. jackknife and bootstrap, are discussed.
They are used for the estimation of bias and standard error under simple random
sampling and complex sampling. Three con dence intervals are discussed, viz. standard
(asymptotic), percentile and bootstrap-t. An overview of welfare indices and
their estimation is given. The indices are categorized into measures of poverty and
measures of inequality. Two Laeken indices, viz. at-risk-of-poverty and quintile
share ratio, are included in the discussion. The study considers two poverty lines,
namely an absolute poverty line based on percy (ratio of total household income
to household size) and a relative poverty line based on equivalized income (ratio of
total household income to equivalized household size). The data set used as surrogate
population for the study is the Income and Expenditure survey 2005/2006
conducted by Statistics South Africa and details of it are provided and discussed.
An analysis of simulation data from the surrogate population was carried out using
techniques mentioned above and the results were graphed, tabulated and discussed.
Two issues were considered, namely whether the design of the survey should be considered
and whether resampling techniques provide reliable results, especially for
con dence intervals. The results were a mixed bag . Overall, however, it was found
that weighting showed promise in many cases, especially in the improvement of the
coverage probabilities of the con dence intervals. It was also found that the bootstrap
resampling technique was reliable (by looking at standard errors). Further
research options are mentioned as possible solutions towards the mixed results. / AFRIKAANSE OPSOMMING: Die doel van die studie is die verkryging van beramings en vertrouensintervalle vir
welvaartsmaatstawwe onder komplekse steekproefneming. 'n Algemene bespreking
van steekproefneming word gedoen waar daar spesi ek op komplekse steekproefneming
en weging gefokus word. Twee hersteekproefnemingstegnieke, nl. uitsnit
(jackknife)- en skoenlushersteekproefneming, word bespreek as metodes vir die beraming
van die maatstawwe. Hierdie maatstawwe word gebruik vir sydigheidsberaming
asook die beraming van standaardfoute in eenvoudige ewekansige steekproefneming
asook komplekse steekproefneming. Drie vertrouensintervalle word bespreek, nl.
die standaard (asimptotiese), die persentiel en die bootstrap-t vertrouensintervalle.
Daar is ook 'n oorsigtelike bespreking oor welvaartsmaatstawwe en die beraming
daarvan. Hierdie welvaartsmaatstawwe vorm twee kategorieë, nl. maatstawwe van
armoede en maatstawwe van ongelykheid. Ook ingesluit by hierdie bespreking is die
at-risk-of-poverty en quintile share ratio wat deel vorm van die Laekenindekse.
Twee armoedemaatlyne , 'n absolute- en relatiewemaatlyn, word in hierdie studie
gebruik. Die absolute armoedemaatlyn word gebaseer op percy , die verhouding van
die totale huishoudingsinkomste tot die grootte van die huishouding, terwyl die relatiewe
armoedemaatlyn gebasseer word op equivalized income , die verhouding van
die totale huishoudingsinkomste tot die equivalized grootte van die huishouding.
Die datastel wat as surrogaat populasie gedien het in hierdie studie is die Inkomste
en Uitgawe opname van 2005/2006 wat deur Statistiek Suid-Afrika uitgevoer is. Inligting
met betrekking tot hierdie opname word ook gegee. Gesimuleerde data vanuit
die surrogaat populasie is geanaliseer deur middel van die hersteekproefnemingstegnieke
wat genoem is. Die resultate van die simulasie is deur middel van gra eke en
tabelle aangedui en bespreek. Vanuit die simulasie het twee vrae opgeduik, nl. of
die ontwerp van 'n steekproef, dus weging, in ag geneem behoort te word en of die
hersteekproefnemingstegnieke betroubare resultate lewer, veral in die geval van die vertrouensintervalle. Die resultate wat verkry is, het baie gevarieer. Daar is egter
bepaal dat weging in die algemeen belowende resultate opgelewer het vir baie van die
gevalle, maar nie vir almal nie. Dit het veral die dekkingswaarskynlikhede van die
vertrouensintervalle verbeter. Daar is ook bepaal, deur na die standaardfoute van
die skoenlusberamers te kyk, dat die skoenlustegniek betroubare resultate gelewer
het. Verdere navorsingsmoontlikhede is genoem as potensiële verbeteringe op die
gemengde resultate wat verkry is.
|
70 |
Intervalos de confiança para altos quantis oriundos de distribuições de caudas pesadas / Confidence intervals for high quantiles from heavy-tailed distributions.Montoril, Michel Helcias 10 March 2009 (has links)
Este trabalho tem como objetivo calcular intervalos de confiança para altos quantis oriundos de distribuições de caudas pesadas. Para isso, utilizamos os métodos da aproximação pela distribuição normal, razão de verossimilhanças, {\\it data tilting} e gama generalizada. Obtivemos, através de simulações, que os intervalos calculados a partir do método da gama generalizada apresentam probabilidades de cobertura bem próximas do nível de confiança, com amplitudes médias menores do que os outros três métodos, para dados gerados da distribuição Weibull. Todavia, para dados gerados da distribuição Fréchet, o método da razão de verossimilhanças fornece os melhores intervalos. Aplicamos os métodos utilizados neste trabalho a um conjunto de dados reais, referentes aos pagamentos de indenizações, em reais, de seguros de incêndio, de um determinado grupo de seguradoras no Brasil, no ano de 2003 / In this work, confidence intervals for high quantiles from heavy-tailed distributions were computed. More specifically, four methods, namely, normal approximation method, likelihood ratio method, data tilting method and generalised gamma method are used. A simulation study with data generated from Weibull distribution has shown that the generalised gamma method has better coverage probabilities with the smallest average length intervals. However, from data generated from Fréchet distribution, the likelihood ratio method gives the better intervals. Moreover, the methods used in this work are applied on a real data set from 1758 Brazilian fire claims
|
Page generated in 0.0962 seconds