1 |
Three essays on econometrics / 計量経済学に関する三つの論文Yi, Kun 23 March 2023 (has links)
京都大学 / 新制・課程博士 / 博士(経済学) / 甲第24375号 / 経博第662号 / 新制||経||302(附属図書館) / 京都大学大学院経済学研究科経済学専攻 / (主査)教授 西山 慶彦, 教授 江上 雅彦, 講師 柳 貴英 / 学位規則第4条第1項該当 / Doctor of Economics / Kyoto University / DFAM
|
2 |
Statistical Methods for Dating Collections of Historical DocumentsTilahun, Gelila 31 August 2011 (has links)
The problem in this thesis was originally motivated by problems presented with documents of Early England Data Set (DEEDS). The central problem with these medieval documents is the lack of methods to assign accurate dates to those documents which bear no date.
With the problems of the DEEDS documents in mind, we present two methods to impute missing features of texts.
In the first method, we suggest a new class of metrics for measuring distances between texts. We then show how to combine the distances between the texts using statistical smoothing. This method can be adapted to settings where the features of the texts are ordered or unordered categoricals (as in the case of, for example, authorship assignment problems).
In the second method, we estimate the probability of occurrences of words in texts using nonparametric regression techniques of local polynomial fitting with kernel weight to generalized linear models. We combine the
estimated probability of occurrences of words of a text to estimate the probability of occurrence of a text as a function of its feature -- the feature in this case being the date in which the text is written. The
application and results of our methods to the DEEDS documents are presented.
|
3 |
Statistical Methods for Dating Collections of Historical DocumentsTilahun, Gelila 31 August 2011 (has links)
The problem in this thesis was originally motivated by problems presented with documents of Early England Data Set (DEEDS). The central problem with these medieval documents is the lack of methods to assign accurate dates to those documents which bear no date.
With the problems of the DEEDS documents in mind, we present two methods to impute missing features of texts.
In the first method, we suggest a new class of metrics for measuring distances between texts. We then show how to combine the distances between the texts using statistical smoothing. This method can be adapted to settings where the features of the texts are ordered or unordered categoricals (as in the case of, for example, authorship assignment problems).
In the second method, we estimate the probability of occurrences of words in texts using nonparametric regression techniques of local polynomial fitting with kernel weight to generalized linear models. We combine the
estimated probability of occurrences of words of a text to estimate the probability of occurrence of a text as a function of its feature -- the feature in this case being the date in which the text is written. The
application and results of our methods to the DEEDS documents are presented.
|
4 |
Estimação de volatilidade em séries financeiras : modelos aditivos semi-paramétricos e GARCHSantos, Douglas Gomes dos January 2008 (has links)
A estimação e previsão da volatilidade de ativos são de suma importância para os mercados financeiros. Temas como risco e incerteza na teoria econômica moderna incentivaram a procura por métodos capazes de modelar uma variância condicional que evolui ao longo do tempo. O objetivo principal desta dissertação é comparar alguns métodos de regressão global e local quanto à extração da volatilidade dos índices Ibovespa e Standard and Poor´s 500. Para isto, são realizadas estimações e previsões com os modelos GARCH paramétricos e com os modelos aditivos semi-paramétricos. Os primeiros, tradicionalmente utilizados na estimação de segundos momentos condicionais, têm sua capacidade sugerida em diversos artigos. Os segundos provêm alta flexibilidade e descrições visualmente informativas das relações entre as variáveis, tais como assimetrias e não linearidades. Sendo assim, testar o desempenho dos últimos frente às estruturas paramétricas consagradas apresenta-se como uma investigação apropriada. A realização das comparações ocorre em períodos selecionados de alta volatilidade no mercado financeiro internacional (crises), sendo a performance dos modelos medida dentro e fora da amostra. Os resultados encontrados sugerem a capacidade dos modelos semi-paramétricos em estimar e prever a volatilidade dos retornos dos índices nos momentos analisados. / Volatility estimation and forecasting are very important matters for the financial markets. Themes like risk and uncertainty in modern economic theory have encouraged the search for methods that allow for the modeling of time varying variances. The main objective of this dissertation is to compare global and local regressions in terms of their capacity to extract the volatility of Ibovespa and Standard and Poor 500 indexes. To achieve this aim, parametric GARCH and semiparametric additive models estimation and forecasting are performed. The first ones, traditionally applied in the estimation of conditional second moments, have their capacity suggested in many papers. The second ones provide high flexibility and visually informative descriptions of the relationships between the variables, like asymmetries and nonlinearities. Therefore, testing the last ones´ performance against the acknowledged parametric structures is an appropriate investigation. Comparisons are made in selected periods of high volatility in the international financial market (crisis), measuring the models´ performance inside and outside sample. The results that were found suggest the capacity of semiparametric models to estimate and forecast the Indexes returns´ volatility at the analyzed moments.
|
5 |
Regression Discontinuity Design with CovariatesKramer, Patrick 07 November 2023 (has links)
This thesis studies regression discontinuity designs with the use of additional covariates for estimation of the average treatment effect. We prove asymptotic normality of the covariate-adjusted estimator under sufficient regularity conditions. In the case of a high-dimensional setting with a large number of covariates depending on the number of observations, we discuss a Lasso-based selection approach as well as alternatives based on calculated correlation thresholds. We present simulation results on those alternative selection strategies.:1. Introduction
2. Preliminaries
3. Regression Discontinuity Designs
4. Setup and Notation
5. Computing the Bias
6. Asymptotic Behavior
7. Asymptotic Normality of the Estimator
8. Including Potentially Many Covariates
9. Simulations
10. Conclusion
|
6 |
Estimação de volatilidade em séries financeiras : modelos aditivos semi-paramétricos e GARCHSantos, Douglas Gomes dos January 2008 (has links)
A estimação e previsão da volatilidade de ativos são de suma importância para os mercados financeiros. Temas como risco e incerteza na teoria econômica moderna incentivaram a procura por métodos capazes de modelar uma variância condicional que evolui ao longo do tempo. O objetivo principal desta dissertação é comparar alguns métodos de regressão global e local quanto à extração da volatilidade dos índices Ibovespa e Standard and Poor´s 500. Para isto, são realizadas estimações e previsões com os modelos GARCH paramétricos e com os modelos aditivos semi-paramétricos. Os primeiros, tradicionalmente utilizados na estimação de segundos momentos condicionais, têm sua capacidade sugerida em diversos artigos. Os segundos provêm alta flexibilidade e descrições visualmente informativas das relações entre as variáveis, tais como assimetrias e não linearidades. Sendo assim, testar o desempenho dos últimos frente às estruturas paramétricas consagradas apresenta-se como uma investigação apropriada. A realização das comparações ocorre em períodos selecionados de alta volatilidade no mercado financeiro internacional (crises), sendo a performance dos modelos medida dentro e fora da amostra. Os resultados encontrados sugerem a capacidade dos modelos semi-paramétricos em estimar e prever a volatilidade dos retornos dos índices nos momentos analisados. / Volatility estimation and forecasting are very important matters for the financial markets. Themes like risk and uncertainty in modern economic theory have encouraged the search for methods that allow for the modeling of time varying variances. The main objective of this dissertation is to compare global and local regressions in terms of their capacity to extract the volatility of Ibovespa and Standard and Poor 500 indexes. To achieve this aim, parametric GARCH and semiparametric additive models estimation and forecasting are performed. The first ones, traditionally applied in the estimation of conditional second moments, have their capacity suggested in many papers. The second ones provide high flexibility and visually informative descriptions of the relationships between the variables, like asymmetries and nonlinearities. Therefore, testing the last ones´ performance against the acknowledged parametric structures is an appropriate investigation. Comparisons are made in selected periods of high volatility in the international financial market (crisis), measuring the models´ performance inside and outside sample. The results that were found suggest the capacity of semiparametric models to estimate and forecast the Indexes returns´ volatility at the analyzed moments.
|
7 |
Estimação de volatilidade em séries financeiras : modelos aditivos semi-paramétricos e GARCHSantos, Douglas Gomes dos January 2008 (has links)
A estimação e previsão da volatilidade de ativos são de suma importância para os mercados financeiros. Temas como risco e incerteza na teoria econômica moderna incentivaram a procura por métodos capazes de modelar uma variância condicional que evolui ao longo do tempo. O objetivo principal desta dissertação é comparar alguns métodos de regressão global e local quanto à extração da volatilidade dos índices Ibovespa e Standard and Poor´s 500. Para isto, são realizadas estimações e previsões com os modelos GARCH paramétricos e com os modelos aditivos semi-paramétricos. Os primeiros, tradicionalmente utilizados na estimação de segundos momentos condicionais, têm sua capacidade sugerida em diversos artigos. Os segundos provêm alta flexibilidade e descrições visualmente informativas das relações entre as variáveis, tais como assimetrias e não linearidades. Sendo assim, testar o desempenho dos últimos frente às estruturas paramétricas consagradas apresenta-se como uma investigação apropriada. A realização das comparações ocorre em períodos selecionados de alta volatilidade no mercado financeiro internacional (crises), sendo a performance dos modelos medida dentro e fora da amostra. Os resultados encontrados sugerem a capacidade dos modelos semi-paramétricos em estimar e prever a volatilidade dos retornos dos índices nos momentos analisados. / Volatility estimation and forecasting are very important matters for the financial markets. Themes like risk and uncertainty in modern economic theory have encouraged the search for methods that allow for the modeling of time varying variances. The main objective of this dissertation is to compare global and local regressions in terms of their capacity to extract the volatility of Ibovespa and Standard and Poor 500 indexes. To achieve this aim, parametric GARCH and semiparametric additive models estimation and forecasting are performed. The first ones, traditionally applied in the estimation of conditional second moments, have their capacity suggested in many papers. The second ones provide high flexibility and visually informative descriptions of the relationships between the variables, like asymmetries and nonlinearities. Therefore, testing the last ones´ performance against the acknowledged parametric structures is an appropriate investigation. Comparisons are made in selected periods of high volatility in the international financial market (crisis), measuring the models´ performance inside and outside sample. The results that were found suggest the capacity of semiparametric models to estimate and forecast the Indexes returns´ volatility at the analyzed moments.
|
8 |
Optimum Savitzky-Golay Filtering for Signal EstimationKrishnan, Sunder Ram January 2013 (has links) (PDF)
Motivated by the classic works of Charles M. Stein, we focus on developing risk-estimation frameworks for denoising problems in both one-and two-dimensions. We assume a standard additive noise model, and formulate the denoising problem as one of estimating the underlying clean signal from noisy measurements by minimizing a risk corresponding to a chosen loss function. Our goal is to incorporate perceptually-motivated loss functions wherever applicable, as in the case of speech enhancement, with the squared error loss being considered for the other scenarios. Since the true risks are observed to depend on the unknown parameter of interest, we circumvent the roadblock by deriving finite-sample un-biased estimators of the corresponding risks based on Stein’s lemma. We establish the link with the multivariate parameter estimation problem addressed by Stein and our denoising problem, and derive estimators of the oracle risks. In all cases, optimum values of the parameters characterizing the denoising algorithm are determined by minimizing the Stein’s unbiased risk estimator (SURE).
The key contribution of this thesis is the development of a risk-estimation approach for choosing the two critical parameters affecting the quality of nonparametric regression, namely, the order and bandwidth/smoothing parameters. This is a classic problem in statistics, and certain algorithms relying on derivation of suitable finite-sample risk estimators for minimization have been reported in the literature (note that all these works consider the mean squared error (MSE) objective). We show that a SURE-based formalism is well-suited to the regression parameter selection problem, and that the optimum solution guarantees near-minimum MSE (MMSE) performance. We develop algorithms for both glob-ally and locally choosing the two parameters, the latter referred to as spatially-adaptive regression. We observe that the parameters are so chosen as to tradeoff the squared bias and variance quantities that constitute the MSE. We also indicate the advantages accruing out of incorporating a regularization term in the cost function in addition to the data error term. In the more general case of kernel regression, which uses a weighted least-squares (LS) optimization, we consider the applications of image restoration from very few random measurements, in addition to denoising of uniformly sampled data. We show that local polynomial regression (LPR) becomes a special case of kernel regression, and extend our results for LPR on uniform data to non-uniformly sampled data also. The denoising algorithms are compared with other standard, performant methods available in the literature both in terms of estimation error and computational complexity.
A major perspective provided in this thesis is that the problem of optimum parameter choice in nonparametric regression can be viewed as the selection of optimum parameters of a linear, shift-invariant filter. This interpretation is provided by deriving motivation out of the hallmark paper of Savitzky and Golay and Schafer’s recent article in IEEE Signal Processing Magazine. It is worth noting that Savitzky and Golay had shown in their original Analytical Chemistry journal article, that LS fitting of a fixed-order polynomial over a neighborhood of fixed size is equivalent to convolution with an impulse response that is fixed and can be pre-computed. They had provided tables of impulse response coefficients for computing the smoothed function and smoothed derivatives for different orders and neighborhood sizes, the resulting filters being referred to as Savitzky-Golay (S-G) filters. Thus, we provide the new perspective that the regression parameter choice is equivalent to optimizing for the filter impulse response length/3dB bandwidth, which are inversely related. We observe that the MMSE solution is such that the S-G filter chosen is of longer impulse response length (equivalently smaller cutoff frequency) at relatively flat portions of the noisy signal so as to smooth noise, and vice versa at locally fast-varying portions of the signal so as to capture the signal patterns. Also, we provide a generalized S-G filtering viewpoint in the case of kernel regression.
Building on the S-G filtering perspective, we turn to the problem of dynamic feature computation in speech recognition. We observe that the methodology employed for computing dynamic features from the trajectories of static features is in fact derivative S-G filtering. With this perspective, we note that the filter coefficients can be pre-computed, and that the whole problem of delta feature computation becomes efficient. Indeed, we observe an advantage by a factor of 104 on making use of S-G filtering over actual LS polynomial fitting and evaluation. Thereafter, we study the properties of first-and second-order derivative S-G filters of certain orders and lengths experimentally. The derivative filters are bandpass due to the combined effects of LPR and derivative computation, which are lowpass and highpass operations, respectively. The first-and second-order S-G derivative filters are also observed to exhibit an approximately constant-Q property. We perform a TIMIT phoneme recognition experiment comparing the recognition accuracies obtained using S-G filters and the conventional approach followed in HTK, where Furui’s regression formula is made use of. The recognition accuracies for both cases are almost identical, with S-G filters of certain bandwidths and orders registering a marginal improvement. The accuracies are also observed to improve with longer filter lengths, for a particular order. In terms of computation latency, we note that S-G filtering achieves delta and delta-delta feature computation in parallel by linear filtering, whereas they need to be obtained sequentially in case of the standard regression formulas used in the literature.
Finally, we turn to the problem of speech enhancement where we are interested in de-noising using perceptually-motivated loss functions such as Itakura-Saito (IS). We propose to perform enhancement in the discrete cosine transform domain using risk-minimization. The cost functions considered are non-quadratic, and derivation of the unbiased estimator of the risk corresponding to the IS distortion is achieved using an approximate Taylor-series analysis under high signal-to-noise ratio assumption. The exposition is general since we focus on an additive noise model with the noise density assumed to fall within the exponential class of density functions, which comprises most of the common densities. The denoising function is assumed to be pointwise linear (modified James-Stein (MJS) estimator), and parallels between Wiener filtering and the optimum MJS estimator are discussed.
|
9 |
Optimum Savitzky-Golay Filtering for Signal EstimationKrishnan, Sunder Ram January 2013 (has links) (PDF)
Motivated by the classic works of Charles M. Stein, we focus on developing risk-estimation frameworks for denoising problems in both one-and two-dimensions. We assume a standard additive noise model, and formulate the denoising problem as one of estimating the underlying clean signal from noisy measurements by minimizing a risk corresponding to a chosen loss function. Our goal is to incorporate perceptually-motivated loss functions wherever applicable, as in the case of speech enhancement, with the squared error loss being considered for the other scenarios. Since the true risks are observed to depend on the unknown parameter of interest, we circumvent the roadblock by deriving finite-sample un-biased estimators of the corresponding risks based on Stein’s lemma. We establish the link with the multivariate parameter estimation problem addressed by Stein and our denoising problem, and derive estimators of the oracle risks. In all cases, optimum values of the parameters characterizing the denoising algorithm are determined by minimizing the Stein’s unbiased risk estimator (SURE).
The key contribution of this thesis is the development of a risk-estimation approach for choosing the two critical parameters affecting the quality of nonparametric regression, namely, the order and bandwidth/smoothing parameters. This is a classic problem in statistics, and certain algorithms relying on derivation of suitable finite-sample risk estimators for minimization have been reported in the literature (note that all these works consider the mean squared error (MSE) objective). We show that a SURE-based formalism is well-suited to the regression parameter selection problem, and that the optimum solution guarantees near-minimum MSE (MMSE) performance. We develop algorithms for both glob-ally and locally choosing the two parameters, the latter referred to as spatially-adaptive regression. We observe that the parameters are so chosen as to tradeoff the squared bias and variance quantities that constitute the MSE. We also indicate the advantages accruing out of incorporating a regularization term in the cost function in addition to the data error term. In the more general case of kernel regression, which uses a weighted least-squares (LS) optimization, we consider the applications of image restoration from very few random measurements, in addition to denoising of uniformly sampled data. We show that local polynomial regression (LPR) becomes a special case of kernel regression, and extend our results for LPR on uniform data to non-uniformly sampled data also. The denoising algorithms are compared with other standard, performant methods available in the literature both in terms of estimation error and computational complexity.
A major perspective provided in this thesis is that the problem of optimum parameter choice in nonparametric regression can be viewed as the selection of optimum parameters of a linear, shift-invariant filter. This interpretation is provided by deriving motivation out of the hallmark paper of Savitzky and Golay and Schafer’s recent article in IEEE Signal Processing Magazine. It is worth noting that Savitzky and Golay had shown in their original Analytical Chemistry journal article, that LS fitting of a fixed-order polynomial over a neighborhood of fixed size is equivalent to convolution with an impulse response that is fixed and can be pre-computed. They had provided tables of impulse response coefficients for computing the smoothed function and smoothed derivatives for different orders and neighborhood sizes, the resulting filters being referred to as Savitzky-Golay (S-G) filters. Thus, we provide the new perspective that the regression parameter choice is equivalent to optimizing for the filter impulse response length/3dB bandwidth, which are inversely related. We observe that the MMSE solution is such that the S-G filter chosen is of longer impulse response length (equivalently smaller cutoff frequency) at relatively flat portions of the noisy signal so as to smooth noise, and vice versa at locally fast-varying portions of the signal so as to capture the signal patterns. Also, we provide a generalized S-G filtering viewpoint in the case of kernel regression.
Building on the S-G filtering perspective, we turn to the problem of dynamic feature computation in speech recognition. We observe that the methodology employed for computing dynamic features from the trajectories of static features is in fact derivative S-G filtering. With this perspective, we note that the filter coefficients can be pre-computed, and that the whole problem of delta feature computation becomes efficient. Indeed, we observe an advantage by a factor of 104 on making use of S-G filtering over actual LS polynomial fitting and evaluation. Thereafter, we study the properties of first-and second-order derivative S-G filters of certain orders and lengths experimentally. The derivative filters are bandpass due to the combined effects of LPR and derivative computation, which are lowpass and highpass operations, respectively. The first-and second-order S-G derivative filters are also observed to exhibit an approximately constant-Q property. We perform a TIMIT phoneme recognition experiment comparing the recognition accuracies obtained using S-G filters and the conventional approach followed in HTK, where Furui’s regression formula is made use of. The recognition accuracies for both cases are almost identical, with S-G filters of certain bandwidths and orders registering a marginal improvement. The accuracies are also observed to improve with longer filter lengths, for a particular order. In terms of computation latency, we note that S-G filtering achieves delta and delta-delta feature computation in parallel by linear filtering, whereas they need to be obtained sequentially in case of the standard regression formulas used in the literature.
Finally, we turn to the problem of speech enhancement where we are interested in de-noising using perceptually-motivated loss functions such as Itakura-Saito (IS). We propose to perform enhancement in the discrete cosine transform domain using risk-minimization. The cost functions considered are non-quadratic, and derivation of the unbiased estimator of the risk corresponding to the IS distortion is achieved using an approximate Taylor-series analysis under high signal-to-noise ratio assumption. The exposition is general since we focus on an additive noise model with the noise density assumed to fall within the exponential class of density functions, which comprises most of the common densities. The denoising function is assumed to be pointwise linear (modified James-Stein (MJS) estimator), and parallels between Wiener filtering and the optimum MJS estimator are discussed.
|
Page generated in 0.0805 seconds