11 |
Statistical inference for inequality measures based on semi-parametric estimatorsKpanzou, Tchilabalo Abozou 12 1900 (has links)
Thesis (PhD)--Stellenbosch University, 2011. / ENGLISH ABSTRACT: Measures of inequality, also used as measures of concentration or diversity, are very popular in economics
and especially in measuring the inequality in income or wealth within a population and between
populations. However, they have applications in many other fields, e.g. in ecology, linguistics, sociology,
demography, epidemiology and information science.
A large number of measures have been proposed to measure inequality. Examples include the Gini
index, the generalized entropy, the Atkinson and the quintile share ratio measures. Inequality measures
are inherently dependent on the tails of the population (underlying distribution) and therefore their
estimators are typically sensitive to data from these tails (nonrobust). For example, income distributions
often exhibit a long tail to the right, leading to the frequent occurrence of large values in samples. Since
the usual estimators are based on the empirical distribution function, they are usually nonrobust to such
large values. Furthermore, heavy-tailed distributions often occur in real life data sets, remedial action
therefore needs to be taken in such cases.
The remedial action can be either a trimming of the extreme data or a modification of the (traditional)
estimator to make it more robust to extreme observations. In this thesis we follow the second option,
modifying the traditional empirical distribution function as estimator to make it more robust. Using results
from extreme value theory, we develop more reliable distribution estimators in a semi-parametric
setting. These new estimators of the distribution then form the basis for more robust estimators of the
measures of inequality. These estimators are developed for the four most popular classes of measures,
viz. Gini, generalized entropy, Atkinson and quintile share ratio. Properties of such estimators
are studied especially via simulation. Using limiting distribution theory and the bootstrap methodology,
approximate confidence intervals were derived. Through the various simulation studies, the proposed
estimators are compared to the standard ones in terms of mean squared error, relative impact of contamination,
confidence interval length and coverage probability. In these studies the semi-parametric
methods show a clear improvement over the standard ones. The theoretical properties of the quintile
share ratio have not been studied much. Consequently, we also derive its influence function as well as
the limiting normal distribution of its nonparametric estimator. These results have not previously been
published.
In order to illustrate the methods developed, we apply them to a number of real life data sets. Using
such data sets, we show how the methods can be used in practice for inference. In order to choose
between the candidate parametric distributions, use is made of a measure of sample representativeness
from the literature. These illustrations show that the proposed methods can be used to reach
satisfactory conclusions in real life problems. / AFRIKAANSE OPSOMMING: Maatstawwe van ongelykheid, wat ook gebruik word as maatstawwe van konsentrasie of diversiteit,
is baie populêr in ekonomie en veral vir die kwantifisering van ongelykheid in inkomste of welvaart
binne ’n populasie en tussen populasies. Hulle het egter ook toepassings in baie ander dissiplines,
byvoorbeeld ekologie, linguistiek, sosiologie, demografie, epidemiologie en inligtingskunde.
Daar bestaan reeds verskeie maatstawwe vir die meet van ongelykheid. Voorbeelde sluit in die Gini
indeks, die veralgemeende entropie maatstaf, die Atkinson maatstaf en die kwintiel aandeel verhouding.
Maatstawwe van ongelykheid is inherent afhanklik van die sterte van die populasie (onderliggende
verdeling) en beramers daarvoor is tipies dus sensitief vir data uit sodanige sterte (nierobuust). Inkomste
verdelings het byvoorbeeld dikwels lang regtersterte, wat kan lei tot die voorkoms van groot
waardes in steekproewe. Die tradisionele beramers is gebaseer op die empiriese verdelingsfunksie, en
hulle is gewoonlik dus nierobuust teenoor sodanige groot waardes nie. Aangesien swaarstert verdelings
dikwels voorkom in werklike data, moet regstellings gemaak word in sulke gevalle.
Hierdie regstellings kan bestaan uit of die afknip van ekstreme data of die aanpassing van tradisionele
beramers om hulle meer robuust te maak teen ekstreme waardes. In hierdie tesis word die
tweede opsie gevolg deurdat die tradisionele empiriese verdelingsfunksie as beramer aangepas word
om dit meer robuust te maak. Deur gebruik te maak van resultate van ekstreemwaardeteorie, word
meer betroubare beramers vir verdelings ontwikkel in ’n semi-parametriese opset. Hierdie nuwe beramers
van die verdeling vorm dan die basis vir meer robuuste beramers van maatstawwe van ongelykheid.
Hierdie beramers word ontwikkel vir die vier mees populêre klasse van maatstawwe, naamlik
Gini, veralgemeende entropie, Atkinson en kwintiel aandeel verhouding. Eienskappe van hierdie
beramers word bestudeer, veral met behulp van simulasie studies. Benaderde vertrouensintervalle
word ontwikkel deur gebruik te maak van limietverdelingsteorie en die skoenlus metodologie. Die
voorgestelde beramers word vergelyk met tradisionele beramers deur middel van verskeie simulasie
studies. Die vergelyking word gedoen in terme van gemiddelde kwadraat fout, relatiewe impak van
kontaminasie, vertrouensinterval lengte en oordekkingswaarskynlikheid. In hierdie studies toon die
semi-parametriese metodes ’n duidelike verbetering teenoor die tradisionele metodes. Die kwintiel
aandeel verhouding se teoretiese eienskappe het nog nie veel aandag in die literatuur geniet nie.
Gevolglik lei ons die invloedfunksie asook die asimptotiese verdeling van die nie-parametriese beramer
daarvoor af.
Ten einde die metodes wat ontwikkel is te illustreer, word dit toegepas op ’n aantal werklike datastelle.
Hierdie toepassings toon hoe die metodes gebruik kan word vir inferensie in die praktyk. ’n Metode
in die literatuur vir steekproefverteenwoordiging word voorgestel en gebruik om ’n keuse tussen die
kandidaat parametriese verdelings te maak. Hierdie voorbeelde toon dat die voorgestelde metodes
met vrug gebruik kan word om bevredigende gevolgtrekkings in die praktyk te maak.
|
12 |
Desenvolvimento e implementação de uma ferramenta computacional aplicada no processo de identificação de sistemas em ambientes Fieldbus foundation / Development and implementation of a computational tool applied to the system identification process in a Fieldbus foundation environmentCunha, Márcio José da 06 October 2004 (has links)
Técnicas experimentais de identificação de sistemas de controle têm despertado interesse na indústria, devido a sua facilidade em se ajustar modelos matemáticos, facilitando a formulação e a resolução de problemas de controle de processos. É proposto o uso de uma técnica experimental de identificação de sistemas, utilizando a estrutura matemática linear ARX. Os parâmetros da estrutura matemática ARX são estimados por meio do algoritmo dos mínimos quadrados recursivo (RLS). A comunicação e a aquisição de dados de redes Fieldbus é feita através do padrão de comunicação OPC. / Experimental system identification techniques are considered interesting by industrial sector due to the simple approach to adjust mathematical models, making it easy the formulation and the solution of process control problems. In this work a computational tool is proposed, the Sintonizador, based on the experimental technic of System Identification, using the linear mathematical structure ARX (Auto-Regressive with eXogenous inputs). The ARX structure parameters are estimated by RLS (Recursive Least Square) algorithm. The data comunication and data aquisition of the fieldbus network has been done through of the OPC comunication standard.
|
13 |
Essays in Efficiency AnalysisDemchuk, Pavlo 16 September 2013 (has links)
Today a standard procedure to analyze the impact of environmental factors on productive efficiency of a decision making unit is to use a two stage approach, where first one estimates the efficiency and then uses regression techniques to explain the variation of efficiency between different units. It is argued that the abovementioned method may produce doubtful results which may distort the truth data represents. In order to introduce economic intuition and to mitigate the problem of omitted variables we introduce the matching procedure which is to be used before the efficiency analysis. We believe that by having comparable decision making units we implicitly control for the environmental factors at the same time cleaning the sample of outliers. The main goal of the first part of the thesis is to compare a procedure including matching prior to efficiency analysis with straightforward two stage procedure without matching as well as an alternative of conditional efficiency frontier. We conduct our study using a Monte Carlo study with different model specifications and despite the reduced sample which may create some complications in the computational stage we strongly agree with a notion of economic meaningfulness of the newly obtained results. We also compare the results obtained by the new method with ones previously produced by Demchuk and Zelenyuk (2009) who compare efficiencies of Ukrainian regions and find some differences between the two approaches.
Second part deals with an empirical study of electricity generating power plants before and after market reform in Texas. We compare private, public and municipal power generators using the method introduced in part one. We find that municipal power plants operate mostly inefficiently, while private and public are very close in their production patterns. The new method allows us to compare decision making units from different groups, which may have different objective schemes and productive incentives. Despite the fact that at a certain point after the reform private generators opted not to provide their data to the regulator we were able to construct tree different data samples comprising two and three groups of generators and analyze their production/efficiency patterns.
In the third chapter we propose a semiparametric approach with shape constrains which is consistent with monotonicity and concavity constraints. Penalized splines are used to maintain the shape constrained via nonlinear transformations of spline basis expansions. The large sample properties, an effective algorithm and method of smoothing parameter selection are presented in the paper. Monte Carlo simulations and empirical examples demonstrate the finite sample performance and the usefulness of the proposed method.
|
14 |
Daugiamačio pasiskirstymo tankio neparametrinis įvertinimas naudojant stebėjimų klasterizavimą / The nonparametric estimation of multivariate distribution density applying clustering proceduresRuzgas, Tomas 14 March 2007 (has links)
The paper is devoted to statistical nonparametric estimation of multivariate distribution density. The influence of data pre-clustering on the estimation accuracy of multimodal density is analysed by means of the Monte-Carlo method.
|
15 |
Daugiamačio pasiskirstymo tankio neparametrinis įvertinimas naudojant stebėjimų klasterizavimą / The nonparametric estimation of multivariate distribution density applying clustering proceduresRuzgas, Tomas 15 March 2007 (has links)
The paper is devoted to statistical nonparametric estimation of multivariate distribution density. The influence of data pre-clustering on the estimation accuracy of multimodal density is analysed by means of the Monte-Carlo method.
|
16 |
Essays in Financial EconometricsDe Lira Salvatierra, Irving January 2015 (has links)
<p>The main goal of this work is to explore the effects of time-varying extreme jump tail dependencies in asset markets. Consequently, a lot of attention has been devoted to understand the extremal tail dependencies between of assets. As pointed by Hansen (2013), the estimation of tail risks dependence is a challenging task and their implications in several sectors of the economy are of great importance. One of the principal challenges is to provide a measure systemic risks that is, in principle, statistically tractable and has an economic meaning. Therefore, there is a need of a standardize dependence measures or at least to provide a methodology that can capture the complexity behind global distress in the economy. These measures should be able to explain not only the dynamics of the most recent financial crisis but also the prior events of distress in the world economy, which is the motivation of this paper. In order to explore the tail dependencies I exploit the information embedded in option prices and intra-daily high frequency data. </p><p>The first chapter, a co-authored work with Andrew Patton, proposes a new class of dynamic copula models for daily asset returns that exploits information from high frequency (intra-daily) data. We augment the generalized autoregressive score (GAS) model of Creal, et al. (2013) with high frequency measures such as realized correlation to obtain a "GRAS" model. We find that the inclusion of realized measures significantly improves the in-sample fit of dynamic copula models across a range of U.S. equity returns. Moreover, we find that out-of-sample density forecasts from our GRAS models are superior to those from simpler models. Finally, we consider a simple portfolio choice problem to illustrate the economic gains from exploiting high frequency data for modeling dynamic dependence.</p><p>In the second chapter using information from option prices I construct two new measures of dependence between assets and industries, the Jump Tail Implied Correlation and the Tail Correlation Risk Premia. The main contribution in this chapter is the construction of a systemic risk factor from daily financial measures using a quantile-regression-based methodology. In this direction, I fill the existing gap between downturns in the financial sector and the real economy. I find that this new index performs well to forecast in-sample and out-of-sample quarterly macroeconomic shocks. In addition, I analyze whether the tail risk of the correlation may be priced. I find that for the S&P500 and its sectors there is an ex ante premium to hedge against systemic risks and changes in the aggregate market correlation. Moreover, I provide evidence that the tails of the implied correlation have remarkable predictive power for future stock market returns.</p> / Dissertation
|
17 |
Statistical analysis and simulation methods related to load-sharing models.Rydén, Patrik January 2000 (has links)
We consider the problem of estimating the reliability of bundles constructed of several fibres, given a particular kind of censored data. The bundles consist of several fibres which have their own independent identically dis-tributed failure stresses (i.e.the forces that destroy the fibres). The force applied to a bundle is distributed between the fibres in the bundle, accord-ing to a load-sharing model. A bundle with these properties is an example of a load-sharing system. Ropes constructed of twisted threads, compos-ite materials constructed of parallel carbon fibres, and suspension cables constructed of steel wires are all examples of load-sharing systems. In par-ticular, we consider bundles where load-sharing is described by either the Equal load-sharing model or the more general Local load-sharing model. In order to estimate the cumulative distribution function of failure stresses of bundles, we need some observed data. This data is obtained either by testing bundles or by testing individual fibres. In this thesis, we develop several theoretical testing methods for both fibres and bundles, and related methods of statistical inference. Non-parametric and parametric estimators of the cumulative distribu-tion functions of failure stresses of fibres and bundles are obtained from different kinds of observed data. It is proved that most of these estimators are consistent, and that some are strongly consistent estimators. We show that resampling, in this case random sampling with replacement from sta-tistically independent portions of data, can be used to assess the accuracy of these estimators. Several numerical examples illustrate the behavior of the obtained estimators. These examples suggest that the obtained estimators usually perform well when the number of observations is moderate.
|
18 |
Nonparametric Markov Random Field Models for Natural Texture ImagesPaget, Rupert Unknown Date (has links)
The underlying aim of this research is to investigate the mathematical descriptions of homogeneous textures in digital images for the purpose of segmentation and recognition. The research covers the problem of testing these mathematical descriptions by using them to generate synthetic realisations of the homogeneous texture for subjective and analytical comparisons with the source texture from which they were derived. The application of this research is in analysing satellite or airborne images of the Earth's surface. In particular, Synthetic Aperture Radar (SAR) images often exhibit regions of homogeneous texture, which if segmented, could facilitate terrain classification. In this thesis we present noncausal, nonparametric, multiscale, Markov random field (MRF) models for recognising and synthesising texture. The models have the ability to capture the characteristics of, and to synthesise, a wide variety of textures, varying from the highly structured to the stochastic. For texture synthesis, we introduce our own novel multiscale approach incorporating a new concept of local annealing. This allows us to use large neighbourhood systems to model complex natural textures with high order statistical characteristics. The new multiscale texture synthesis algorithm also produces synthetic textures with few, if any, phase discontinuities. The power of our modelling technique is evident in that only a small source image is required to synthesise representative examples of the source texture, even when the texture contains long-range characteristics. We also show how the high-dimensional model of the texture may be modelled with lower dimensional statistics without compromising the integrity of the representation. We then show how these models -- which are able to capture most of the unique characteristics of a texture -- can be for the ``open-ended'' problem of recognising textures embedded in a scene containing previously unseen textures. Whilst this technique was developed for the practical application of recognising different terrain types from Synthetic Aperture Radar (SAR) images, it has applications in other image processing tasks requiring texture recognition.
|
19 |
Desenvolvimento e implementação de uma ferramenta computacional aplicada no processo de identificação de sistemas em ambientes Fieldbus foundation / Development and implementation of a computational tool applied to the system identification process in a Fieldbus foundation environmentMárcio José da Cunha 06 October 2004 (has links)
Técnicas experimentais de identificação de sistemas de controle têm despertado interesse na indústria, devido a sua facilidade em se ajustar modelos matemáticos, facilitando a formulação e a resolução de problemas de controle de processos. É proposto o uso de uma técnica experimental de identificação de sistemas, utilizando a estrutura matemática linear ARX. Os parâmetros da estrutura matemática ARX são estimados por meio do algoritmo dos mínimos quadrados recursivo (RLS). A comunicação e a aquisição de dados de redes Fieldbus é feita através do padrão de comunicação OPC. / Experimental system identification techniques are considered interesting by industrial sector due to the simple approach to adjust mathematical models, making it easy the formulation and the solution of process control problems. In this work a computational tool is proposed, the Sintonizador, based on the experimental technic of System Identification, using the linear mathematical structure ARX (Auto-Regressive with eXogenous inputs). The ARX structure parameters are estimated by RLS (Recursive Least Square) algorithm. The data comunication and data aquisition of the fieldbus network has been done through of the OPC comunication standard.
|
20 |
Technical efficiency in noisy multi-output settingsGstach, Dieter January 1998 (has links) (PDF)
This paper surveys four distinct approaches to frontier estimation of multi-output (and simultaneously multi-input) technologies, when nothing but noisy quantity data are available. Parametrized distributions for inefficiency and noise are necessary for identification of inefficiency, when only cross-sectional data are available. In other respects suitable techniques may differ widely, as is shown. A final technique presented rigorously exploits the possibilities from panel-data by dropping parametrization of distributions as well as functional forms. It is illustrated how this last technique can be coupled with the others to provide a state-of-the-art estimation procedure for this setting. (author's abstract) / Series: Department of Economics Working Paper Series
|
Page generated in 0.1501 seconds