• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 89
  • 24
  • 4
  • 3
  • 3
  • 2
  • 2
  • 2
  • 2
  • 1
  • 1
  • 1
  • Tagged with
  • 150
  • 150
  • 138
  • 45
  • 29
  • 26
  • 26
  • 23
  • 22
  • 20
  • 20
  • 17
  • 16
  • 15
  • 15
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
111

Estimation de l’histoire démographique des populations à partir de génomes entièrement séquencés. / Estimation de l’histoire démographique des populations à partir de génomes entièrement séquencés

Rodriguez Valcarce, Willy 20 June 2016 (has links)
Le développement des nouvelles techniques de séquençage élargit l' horizon de la génétique de populations. Une analyse appropriée des données génétiques peut augmenter notre capacité à reconstruire l'histoire des populations. Cette énorme quantité de données disponibles peut aider les chercheurs en biologie et anthropologie à mieux estimer les changements démographiques subis par une population au cours du temps, mais induit aussi de nouveaux défis. Lorsque les modèles sous-jacents sont trop simplistes il existe unrisque très fort d'être amené à des conclusions erronées sur la population étudiée. Il a été montré que certaines caractéristiques présentes dans l'ADN des individus d'une population structurée se trouvent aussi dans l'ADN de ceux qui proviennent d'une population sans structure dont la taille a changé au cours du temps. Par conséquent il peut s'avérer très difficile de déterminer si les changements de taille inférés à partir des données génétiquesont vraiment eu lieu ou s'il s'agit simplement des effets liés à la structure. D'ailleurs la quasi totalité des méthodes pour inférer les changements de taille d'une population au cours du temps sont basées sur des modèles qui négligent la structure.Dans cette thèse, de nouveaux résultats de génétique de populations sont présentés. Premièrement, nous présentons une méthodologie permettant de faire de la sélection de modèle à partir de l'ADN d'un seul individudiploïde. Cette première étude se limite à un modèle simple de population non structurée avec un changement de taille et à un modèle considérant une population de taille constante mais structurée. Cette nouvelle méthode utilise la distribution des temps de coalescence de deux gènes pour identifier le modèle le plus probable et ouvreainsi la voie pour de nouvelles méthodes de sélection de modèles structurés et non structurés, à partir de données génomiques issues d'un seul individu. Deuxièmement, nous montrons, par une ré-interprétation du taux de coalescence que, pour n'importe quel scénario structuré, et plus généralement n'importe quel modèle, il existe toujours un scénario considérant une population panmictique avec une fonction précise de changements de taille dont la distribution des temps de coalescence de deux gènes est identique a celle du scénario structuré. Cela non seulement explique pourquoi les méthodes d'inférence démographique détectent souvent des changements de taille n'ayant peut-être jamais eu lieu, mais permet aussi de prédire les changements de taille qui seront reconstruits lorsque des méthodes basées sur l'hypothèse de panmixie sont appliquées à des données issues de scénarios plus complexes. Finalement, une nouvelle approche basée sur un processus de Markov est développée et permet de caractériser la distribution du temps de coalescence de deux gènes dans une population structurée soumise à des événements démographiques tel que changement de flux de gènes et changements de taille. Une discussion est menée afin de décrire comment cette méthode donne la possibilité de reconstruire l'histoire démographique à partir de données génomiques tout en considérant la structure. / The rapid development of DNA sequencing technologies is expanding the horizons of population genetic studies. It is expected that genomic data will increase our ability to reconstruct the history of populations.While this increase in genetic information will likely help biologists and anthropologists to reconstruct the demographic history of populations, it also poses big challenges. In some cases, simplicity of the model maylead to erroneous conclusions about the population under study. Recent works have shown that DNA patterns expected in individuals coming from structured populations correspond with those of unstructured populations with changes in size through time. As a consequence it is often difficult to determine whether demographic events such as expansions or contractions (bottlenecks) inferred from genetic data are real or due to the fact that populations are structured in nature. Moreover, almost no inferential method allowing to reconstruct pastdemographic size changes takes into account structure effects. In this thesis, some recent results in population genetics are presented: (i) a model choice procedure is proposed to distinguish one simple scenario of population size change from one of structured population, based on the coalescence times of two genes, showing that for these simple cases, it is possible to distinguish both models using genetic information form one single individual; (ii) by using the notion of instantaneous coalescent rate, it is demonstrated that for any scenario of structured population or any other one, regardless how complex it could be, there always exists a panmitic scenario with a precise function of population size changes havingexactly the same distribution for the coalescence times of two genes. This not only explains why spurious signals of bottlenecks can be found in structured populations but also predicts the demographic history that actual inference methods are likely to reconstruct when applied to non panmitic populations. Finally, (iii) a method based on a Markov process is developed for inferring past demographic events taking the structure into account. This is method uses the distribution of coalescence times of two genes to detect past demographic changes instructured populations from the DNA of one single individual. Some applications of the model to genomic data are discussed.
112

Modelos de regressão lineares mistos sob a classe de distribuições normal-potência / Linear mixed regression models under the power-normal class distributions

Roger Jesus Tovar Falon 27 November 2017 (has links)
Neste trabalho são apresentadas algumas extensões dos modelos potência-alfa assumindo o contexto em que as observações estão censuradas ou limitadas. Inicialmente propomos um novo modelo assimétrico que estende os modelos t-assimétrico (Azzalini e Capitanio, 2003) e t-potência (Zhao e Kim, 2016) e inclui a distribuição t de Student como caso particular. Este novo modelo é capaz de ajustar dados com alto grau de assimetria e curtose, ainda maior do que os modelos t-assimétrico e t-potência. Em seguida estendemos o modelo t-potência às situações em que os dados apresentam censura, com alto grau de assimetria e caudas pesadas. Este modelo generaliza o modelo de regressão linear t de Student para dados censurados por Arellano-Valle et al. (2012). O trabalho também introduz o modelo linear misto normal-potência para dados assimétricos. Aqui a inferência estatística é realizada desde uma perspectiva clássica usando o método de máxima verossimilhança junto com o método de integração numérica de Gauss-Hermite para aproximar as integrais envolvidas na função de verossimilhança. Mais tarde, o modelo linear com interceptos aleatórios para dados duplamente censurados é estudado. Este modelo é desenvolvido sob a suposição de que os erros e os efeitos aleatórios seguem distribuições normal-potência e normal- assimétrica. Para todos os modelos estudados foram realizados estudos de simulação a fim de estudar as suas bondades de ajuste e limitações. Finalmente, ilustram-se todos os métodos propostos com dados reais. / In this work some extensions of the alpha-power models are presented, assuming the context in which the observations are censored or limited. Initially we propose a new asymmetric model that extends the skew-t (Azzalini e Capitanio, 2003) and power-t (Zhao e Kim, 2016) models and includes the Students t-distribution as a particular case. This new model is able to adjust data with a high degree of asymmetry and cursose, even higher than the skew-t and power-t models. Then we extend the power-t model to situations in which the data present censorship, with a high degree of asymmetry and heavy tails. This model generalizes the Students t linear censored regression model t by Arellano-Valle et al. (2012) The work also introduces the power-normal linear mixed model for asymmetric data. Here statistical inference is performed from a classical perspective using the maximum likelihood method together with the Gauss-Hermite numerical integration method to approximate the integrals involved in the likelihood function. Later, the linear model with random intercepts for doubly censored data is studied. This model is developed under the assumption that errors and random effects follow power-normal and skew-normal distributions. For all the models studied, simulation studies were carried out to study their benefits and limitations. Finally, all proposed methods with real data are illustrated.
113

Regressão logística com erro de medida: comparação de métodos de estimação / Logistic regression model with measurement error: a comparison of estimation methods

Agatha Sacramento Rodrigues 27 June 2013 (has links)
Neste trabalho estudamos o modelo de regressão logística com erro de medida nas covariáveis. Abordamos as metodologias de estimação de máxima pseudoverossimilhança pelo algoritmo EM-Monte Carlo, calibração da regressão, SIMEX e naïve (ingênuo), método este que ignora o erro de medida. Comparamos os métodos em relação à estimação, através do viés e da raiz do erro quadrático médio, e em relação à predição de novas observações, através das medidas de desempenho sensibilidade, especificidade, verdadeiro preditivo positivo, verdadeiro preditivo negativo, acurácia e estatística de Kolmogorov-Smirnov. Os estudos de simulação evidenciam o melhor desempenho do método de máxima pseudoverossimilhança na estimação. Para as medidas de desempenho na predição não há diferença entre os métodos de estimação. Por fim, utilizamos nossos resultados em dois conjuntos de dados reais de diferentes áreas: área médica, cujo objetivo está na estimação da razão de chances, e área financeira, cujo intuito é a predição de novas observações. / We study the logistic model when explanatory variables are measured with error. Three estimation methods are presented, namely maximum pseudo-likelihood obtained through a Monte Carlo expectation-maximization type algorithm, regression calibration, SIMEX and naïve, which ignores the measurement error. These methods are compared through simulation. From the estimation point of view, we compare the different methods by evaluating their biases and root mean square errors. The predictive quality of the methods is evaluated based on sensitivity, specificity, positive and negative predictive values, accuracy and the Kolmogorov-Smirnov statistic. The simulation studies show that the best performing method is the maximum pseudo-likelihood method when the objective is to estimate the parameters. There is no difference among the estimation methods for predictive purposes. The results are illustrated in two real data sets from different application areas: medical area, whose goal is the estimation of the odds ratio, and financial area, whose goal is the prediction of new observations.
114

Uma abordagem estatística para o modelo do preço spot da energia elétrica no submercado sudeste/centro-oeste brasileiro / A statistical approach to model the spot price of electric energy: evidende from brazilian southeas/middle-west subsystem.

Ramalho, Guilherme Matiussi 20 March 2014 (has links)
O objetivo deste trabalho e o desenvolvimento de uma ferramenta estatistica que sirva de base para o estudo do preco spot da energia eletrica do subsistema Sudeste/Centro-Oeste do Sistema Interligado Nacional, utilizando a estimacao por regressao linear e teste de razao de verossimilhanca como instrumentos para desenvolvimento e avaliacao dos modelos. Na analise dos resultados estatsticos descritivos dos modelos, diferentemente do que e observado na literatura, a primeira conclusao e a verificacao de que as variaveis sazonais, quando analisadas isoladamente, apresentam resultados pouco aderentes ao preco spot PLD. Apos a analise da componente sazonal e verificada a influencia da energia fornecida e a energia demandada como variaveis de entrada, com o qual conclui-se que especificamente a energia armazenada e producao de energia termeletrica sao as variaveis que mais influenciam os precos spot no subsistema estudado. Entre os modelos testados, o que particularmente ofereceu os melhores resultados foi um modelo misto criado a partir da escolha das melhores variaveis de entrada dos modelos testados preliminarmente, alcancando um coeficiente de determinacao R2 de 0.825, resultado esse que pode ser considerado aderente ao preco spot. No ultimo capitulo e apresentada uma introducao ao modelo de predicao do preco spot, possibilitando dessa forma a analise do comportamento do preco a partir da alteracao das variaveis de entrada. / The objective of this work is the development of a statistical method to study the spot prices of the electrical energy of the Southeast/Middle-West (SE-CO) subsystem of the The Brazilian National Connected System, using the Least Squares Estimation and Likelihood Ratio Test as tools to perform and evaluate the models. Verifying the descriptive statistical results of the models, differently from what is observed in the literature, the first observation is that the seasonal component, when analyzed alone, presented results loosely adherent to the spot price PLD. It is then evaluated the influence of the energy supply and the energy demand as input variables, verifying that specifically the stored water and the thermoelectric power production are the variables that the most influence the spot prices in the studied subsystem. Among the models, the one that offered the best result was a mixed model created from the selection of the best input variables of the preliminarily tested models, achieving a coeficient of determination R2 of 0.825, a result that can be considered adherent to the spot price. At the last part of the work It is presented an introduction to the spot price prediction model, allowing the analysis of the price behavior by the changing of the input variables.
115

none

Chen, Ping-Sen 27 June 2000 (has links)
none
116

Συμβολή στη στατιστική συμπερασματολογία για τις κατανομές γάμα και αντίστροφη κανονική με χρήση της εμπειρικής ροπογεννήτριας συνάρτησης / Contribution to statistical inference for the Gamma distributions and the Inverse Gaussian distributions using the empirical moment generating function

Καλλιώρας, Αθανάσιος Γ. 01 September 2008 (has links)
Το αντικείμενο της παρούσας διατριβής είναι η διερεύνηση μεθόδων στατιστικής συμπερασματολογίας για την προσαρμογή και έλεγχο της κατανομής γάμα και της αντίστροφης κανονικής (inverse Gaussian) κατανομής σε δεδομένα με θετική λοξότητα. Τα πρότυπα αυτά χρησιμοποιούνται ευρέως στην ανάλυση αξιοπιστίας και ελέγχου μακροβιότητας καθώς και σε άλλες εφαρμογές. Αρχικά γίνεται μια περιγραφή εναλλακτικών μεθόδων στατιστικής συμπερασματολογίας για τις διπαραμετρικές και τις τριπαραμετρικές οικογένειες κατανομών γάμα και αντίστροφης κανονικής. Στη συνέχεια διερευνάται η χρήση μεθόδων στατιστικής συμπερασματολογίας για την εκτίμηση των παραμέτρων της διπαραμετρικής γάμα κατανομής με χρήση της εμπειρικής ροπογεννήτριας συνάρτησης. Μέθοδοι εκτιμητικής, όπως είναι η μέθοδος των μικτών ροπών και των γενικευμένων ελαχίστων τετραγώνων, εφαρμόζονται και συγκρίνονται με την μέθοδο της μέγιστης πιθανοφάνειας μέσω πειραμάτων προσομοίωσης Monte Carlo. Επίσης, διερευνώνται έλεγχοι καλής προσαρμογής για τη διπαραμετρική γάμα κατανομή. Οι έλεγχοι αυτοί περιλαμβάνουν τους κλασικούς ελέγχους και έναν έλεγχο που χρησιμοποιεί την εμπειρική ροπογεννήτρια συνάρτηση. Με χρήση πειραμάτων προσομοίωσης Monte Carlo, γίνεται σύγκριση των ελέγχων ως προς το πραγματικό επίπεδο σημαντικότητας και την ισχύ έναντι άλλων λοξών προς τα δεξιά κατανομών. Στη συνέχεια εφαρμόζονται έλεγχοι καλής προσαρμογής γάμα κατανομών σε πραγματικά δεδομένα, τα οποία έχουν αναλυθεί νωρίτερα από άλλους ερευνητές. Για τον έλεγχο της τριπαραμετρικής γάμα κατανομής εφαρμόζεται μόνο ο έλεγχος με χρήση της εμπειρικής ροπογεννήτριας συνάρτησης, αφού δεν είναι γνωστοί κλασικοί έλεγχοι που χρησιμοποιούν την εμπειρική συνάρτηση κατανομής. Τέλος, γίνεται εκτίμηση ποσοστιαίων σημείων της αντίστροφης κανονικής κατανομής. Αρχικά, εκτιμώνται ποσοστιαία σημεία για την τριπαραμετρική κατανομή και στη συνέχεια εφαρμόζονται δύο μέθοδοι υπολογισμού ποσοστιαίων σημείων για την περίπτωση της διπαραμετρικής κατανομής. Η εκτίμηση των ποσοστιαίων σημείων σε κάθε οικογένεια κατανομών χρησιμοποιεί δύο μεθόδους ενδιάμεσης εκτίμησης των παραμέτρων της κατανομής. Οι μέθοδοι συγκρίνονται ως προς το μέσο τετραγωνικό σφάλμα και τη σχετική μεροληψία με τη βοήθεια πειραμάτων προσομοίωσης. / The subject of the present dissertation is the investigation of procedures of statistical inference for fitting and testing the gamma distribution and inverse Gaussian distribution, with data having positive skewness. These distributions are used widely in reliability analysis and lifetime models as well as in other applications. In the beginning, we describe alternative methods of statistical inference for the two and three-parameter families of gamma and inverse Gaussian distributions. Then, we examine methods of statistical inference in order to estimate the parameters of the two-parameter gamma distribution using the empirical moment generating function. Estimation procedures, like the method of mixed moments and the method of generalized least squares, are applied and compared with the method of maximum likelihood through Monte Carlo simulations. Also, we investigate goodness of fit tests for the two-parameter gamma distribution. These tests include the classical tests and a test based on the empirical moment generating function. Using Monte Carlo simulations, we compare the actual level of the tests and the power of the tests against skewed to the right distributions. We apply goodness of fit tests of gamma distributions to real life data, which have been examined earlier by other researchers. For the three-parameter gamma distribution we apply only one test using the empirical moment generating function since there are no classical tests using the empirical distribution function. Finally, we estimate quantiles of the inverse Gaussian distribution. We start estimating quantiles for the three-parameter distribution and then we apply two procedures which estimate quantiles for the two-parameter distribution. The estimates of the quantiles for each family of distributions use two procedures for estimating intermediary the parameters of the distribution. The procedures are compared with respect to the normalized mean square error and the relative bias using simulations.
117

Statistical Modeling for Credit Ratings

Vana, Laura 01 August 2018 (has links) (PDF)
This thesis deals with the development, implementation and application of statistical modeling techniques which can be employed in the analysis of credit ratings. Credit ratings are one of the most widely used measures of credit risk and are relevant for a wide array of financial market participants, from investors, as part of their investment decision process, to regulators and legislators as a means of measuring and limiting risk. The majority of credit ratings is produced by the "Big Three" credit rating agencies Standard & Poors', Moody's and Fitch. Especially in the light of the 2007-2009 financial crisis, these rating agencies have been strongly criticized for failing to assess risk accurately and for the lack of transparency in their rating methodology. However, they continue to maintain a powerful role as financial market participants and have a huge impact on the cost of funding. These points of criticism call for the development of modeling techniques that can 1) facilitate an understanding of the factors that drive the rating agencies' evaluations, 2) generate insights into the rating patterns that these agencies exhibit. This dissertation consists of three research articles. The first one focuses on variable selection and assessment of variable importance in accounting-based models of credit risk. The credit risk measure employed in the study is derived from credit ratings assigned by ratings agencies Standard & Poors' and Moody's. To deal with the lack of theoretical foundation specific to this type of models, state-of-the-art statistical methods are employed. Different models are compared based on a predictive criterion and model uncertainty is accounted for in a Bayesian setting. Parsimonious models are identified after applying the proposed techniques. The second paper proposes the class of multivariate ordinal regression models for the modeling of credit ratings. The model class is motivated by the fact that correlated ordinal data arises naturally in the context of credit ratings. From a methodological point of view, we extend existing model specifications in several directions by allowing, among others, for a flexible covariate dependent correlation structure between the continuous variables underlying the ordinal credit ratings. The estimation of the proposed models is performed using composite likelihood methods. Insights into the heterogeneity among the "Big Three" are gained when applying this model class to the multiple credit ratings dataset. A comprehensive simulation study on the performance of the estimators is provided. The third research paper deals with the implementation and application of the model class introduced in the second article. In order to make the class of multivariate ordinal regression models more accessible, the R package mvord and the complementary paper included in this dissertation have been developed. The mvord package is available on the "Comprehensive R Archive Network" (CRAN) for free download and enhances the available ready-to-use statistical software for the analysis of correlated ordinal data. In the creation of the package a strong emphasis has been put on developing a user-friendly and flexible design. The user-friendly design allows end users to estimate in an easy way sophisticated models from the implemented model class. The end users the package appeals to are practitioners and researchers who deal with correlated ordinal data in various areas of application, ranging from credit risk to medicine or psychology.
118

Modeling based on a reparameterized Birnbaum-Saunders distribution for analysis of survival data / Modelagem baseada na distribuição Birnbaum-Saunders reparametrizada para análise de dados de sobrevivência

Leão, Jeremias da Silva 09 January 2017 (has links)
Submitted by Aelson Maciera (aelsoncm@terra.com.br) on 2017-04-24T18:48:10Z No. of bitstreams: 1 TeseJSL.pdf: 1918523 bytes, checksum: 4d551d58b97032091209f65b7428e992 (MD5) / Approved for entry into archive by Ronildo Prado (ronisp@ufscar.br) on 2017-04-25T18:50:15Z (GMT) No. of bitstreams: 1 TeseJSL.pdf: 1918523 bytes, checksum: 4d551d58b97032091209f65b7428e992 (MD5) / Approved for entry into archive by Ronildo Prado (ronisp@ufscar.br) on 2017-04-25T18:50:23Z (GMT) No. of bitstreams: 1 TeseJSL.pdf: 1918523 bytes, checksum: 4d551d58b97032091209f65b7428e992 (MD5) / Made available in DSpace on 2017-04-25T18:59:25Z (GMT). No. of bitstreams: 1 TeseJSL.pdf: 1918523 bytes, checksum: 4d551d58b97032091209f65b7428e992 (MD5) Previous issue date: 2017-01-09 / Não recebi financiamento / In this thesis we propose models based on a reparameterized Birnbaum-Saunder (BS) distribution introduced by Santos-Neto et al. (2012) and Santos-Neto et al. (2014), to analyze survival data. Initially we introduce the Birnbaum-Saunders frailty model where we analyze the cases (i) with (ii) without covariates. Survival models with frailty are used when further information is nonavailable to explain the occurrence time of a medical event. The random effect is the “frailty”, which is introduced on the baseline hazard rate to control the unobservable heterogeneity of the patients. We use the maximum likelihood method to estimate the model parameters. We evaluate the performance of the estimators under different percentage of censured observations by a Monte Carlo study. Furthermore, we introduce a Birnbaum-Saunders regression frailty model where the maximum likelihood estimation of the model parameters with censored data as well as influence diagnostics for the new regression model are investigated. In the following we propose a cure rate Birnbaum-Saunders frailty model. An important advantage of this proposed model is the possibility to jointly consider the heterogeneity among patients by their frailties and the presence of a cured fraction of them. We consider likelihood-based methods to estimate the model parameters and to derive influence diagnostics for the model. In addition, we introduce a bivariate Birnbaum-Saunders distribution based on a parameterization of the Birnbaum-Saunders which has the mean as one of its parameters. We discuss the maximum likelihood estimation of the model parameters and show that these estimators can be obtained by solving non-linear equations. We then derive a regression model based on the proposed bivariate Birnbaum-Saunders distribution, which permits us to model data in their original scale. A simulation study is carried out to evaluate the performance of the maximum likelihood estimators. Finally, examples with real-data are performed to illustrate all the models proposed here. / Nesta tese propomos modelos baseados na distribuição Birnbaum-Saunders reparametrizada introduzida por Santos-Neto et al. (2012) e Santos-Neto et al. (2014), para análise dados de sobrevivência. Incialmente propomos o modelo de fragilidade Birnbaum-Saunders sem e com covariáveis observáveis. O modelo de fragilidade é caracterizado pela utilização de um efeito aleatório, ou seja, de uma variável aleatória não observável, que representa as informações que não podem ou não foram observadas tais como fatores ambientais ou genéticos, como também, informações que, por algum motivo, não foram consideradas no planejamento do estudo. O efeito aleatório (a “fragilidade”) é introduzido na função de risco de base para controlar a heterogeneidade não observável. Usamos o método de máxima verossimilhança para estimar os parâmetros do modelo. Avaliamos o desempenho dos estimadores sob diferentes percentuais de censura via estudo de simulações de Monte Carlo. Considerando variáveis regressoras, derivamos medidas de diagnóstico de influência. Os métodos de diagnóstico têm sido ferramentas importantes na análise de regressão para detectar anomalias, tais como quebra das pressuposições nos erros, presença de outliers e observações influentes. Em seguida propomos o modelo de fração de cura com fragilidade Birnbaum-Saunders. Os modelos para dados de sobrevivência com proporção de curados (também conhecidos como modelos de taxa de cura ou modelos de sobrevivência com longa duração) têm sido amplamente estudados. Uma vantagem importante do modelo proposto é a possibilidade de considerar conjuntamente a heterogeneidade entre os pacientes por suas fragilidades e a presença de uma fração curada. As estimativas dos parâmetros do modelo foram obtidas via máxima verossimilhança, medidas de influência e diagnóstico foram desenvolvidas para o modelo proposto. Por fim, avaliamos a distribuição bivariada Birnbaum-Saunders baseada na média, como também introduzimos um modelo de regressão para o modelo proposto. Utilizamos os métodos de máxima verossimilhança e método dos momentos modificados, para estimar os parâmetros do modelo. Avaliamos o desempenho dos estimadores via estudo de simulações de Monte Carlo. Aplicações a conjuntos de dados reais ilustram as potencialidades dos modelos abordados.
119

Two essays on Birnbaum-Saunders regression models for censored data / Dois ensaios em modelos de regressão Birnbaum-Saunders para dados censurados

Sousa, Mário Fernando de 06 December 2016 (has links)
Submitted by Luciana Ferreira (lucgeral@gmail.com) on 2017-05-02T15:17:50Z No. of bitstreams: 2 Dissertação - Mário Fernando de Sousa - 2016.pdf: 645506 bytes, checksum: d6fd190570fce6feeb390cfeaf50032f (MD5) license_rdf: 0 bytes, checksum: d41d8cd98f00b204e9800998ecf8427e (MD5) / Approved for entry into archive by Luciana Ferreira (lucgeral@gmail.com) on 2017-05-02T15:18:06Z (GMT) No. of bitstreams: 2 Dissertação - Mário Fernando de Sousa - 2016.pdf: 645506 bytes, checksum: d6fd190570fce6feeb390cfeaf50032f (MD5) license_rdf: 0 bytes, checksum: d41d8cd98f00b204e9800998ecf8427e (MD5) / Made available in DSpace on 2017-05-02T15:18:06Z (GMT). No. of bitstreams: 2 Dissertação - Mário Fernando de Sousa - 2016.pdf: 645506 bytes, checksum: d6fd190570fce6feeb390cfeaf50032f (MD5) license_rdf: 0 bytes, checksum: d41d8cd98f00b204e9800998ecf8427e (MD5) Previous issue date: 2016-12-06 / Coordenação de Aperfeiçoamento de Pessoal de Nível Superior - CAPES / This work aims to fill a gap in the literature on modeling asymmetric and censored data. The main objective is to provide a contribution by developing two models, which will be presented in two papers, respectively. In the first paper, we develop the tobit-Birnbaum-Saunders model, a variation of the standard tobit model. We discuss estimation based on the maximum likelihood method, residuals, diagnostic techniques and an empirical application. In the second paper, we propose the use of a mixture between the Birnbaum-Saunders and Bernoulli distributions. The objective is to generalize the tobit-Birnbaum-Saunders model in order to consider the possibility of partial observations below a cutoff point. For the mixture model, we carry out a Monte Carlo simulation study and an empirical application. The results show that, in both cases, the Birnbaum-Saunders distribution provides the best results. / Este trabalho visa preencher uma lacuna existente na literatura pertinente à modelagem de dados assimétricos e censurados. O objetivo principal é oferecer uma contribuição via o desenvolvimento de dois modelos, os quais serão apresentados em dois artigos. No primeiro artigo é proposto o modelo tobit-Birnbaum-Saunders, ou seja, uma variação do modelo tobit clássico, com estimação baseada no método de máxima verossimilhança, resíduos, técnicas de diagnóstico e uma aplicação a dados reais. No segundo artigo é abordada a utilização de um modelo de mistura entre as distribuições Birnbaum-Saunders e Bernoulli, de modo a generalizar o modelo tobit-Birnbaum-Saunders e considerar a possibilidade de observações parciais abaixo do ponto de corte. Para o modelo de mistura são realizadas simulações de Monte Carlo e uma aplicação a dados reais. Os resultados mostram que, em ambos os casos, a distribuição Birnbaum-Saunders oferece os melhores resultados.
120

Uma abordagem estatística para o modelo do preço spot da energia elétrica no submercado sudeste/centro-oeste brasileiro / A statistical approach to model the spot price of electric energy: evidende from brazilian southeas/middle-west subsystem.

Guilherme Matiussi Ramalho 20 March 2014 (has links)
O objetivo deste trabalho e o desenvolvimento de uma ferramenta estatistica que sirva de base para o estudo do preco spot da energia eletrica do subsistema Sudeste/Centro-Oeste do Sistema Interligado Nacional, utilizando a estimacao por regressao linear e teste de razao de verossimilhanca como instrumentos para desenvolvimento e avaliacao dos modelos. Na analise dos resultados estatsticos descritivos dos modelos, diferentemente do que e observado na literatura, a primeira conclusao e a verificacao de que as variaveis sazonais, quando analisadas isoladamente, apresentam resultados pouco aderentes ao preco spot PLD. Apos a analise da componente sazonal e verificada a influencia da energia fornecida e a energia demandada como variaveis de entrada, com o qual conclui-se que especificamente a energia armazenada e producao de energia termeletrica sao as variaveis que mais influenciam os precos spot no subsistema estudado. Entre os modelos testados, o que particularmente ofereceu os melhores resultados foi um modelo misto criado a partir da escolha das melhores variaveis de entrada dos modelos testados preliminarmente, alcancando um coeficiente de determinacao R2 de 0.825, resultado esse que pode ser considerado aderente ao preco spot. No ultimo capitulo e apresentada uma introducao ao modelo de predicao do preco spot, possibilitando dessa forma a analise do comportamento do preco a partir da alteracao das variaveis de entrada. / The objective of this work is the development of a statistical method to study the spot prices of the electrical energy of the Southeast/Middle-West (SE-CO) subsystem of the The Brazilian National Connected System, using the Least Squares Estimation and Likelihood Ratio Test as tools to perform and evaluate the models. Verifying the descriptive statistical results of the models, differently from what is observed in the literature, the first observation is that the seasonal component, when analyzed alone, presented results loosely adherent to the spot price PLD. It is then evaluated the influence of the energy supply and the energy demand as input variables, verifying that specifically the stored water and the thermoelectric power production are the variables that the most influence the spot prices in the studied subsystem. Among the models, the one that offered the best result was a mixed model created from the selection of the best input variables of the preliminarily tested models, achieving a coeficient of determination R2 of 0.825, a result that can be considered adherent to the spot price. At the last part of the work It is presented an introduction to the spot price prediction model, allowing the analysis of the price behavior by the changing of the input variables.

Page generated in 0.0913 seconds