Spelling suggestions: "subject:"model aselection"" "subject:"model selelection""
1 
Bayesian model selection using exact and approximated posterior probabilities with applications to Star DataPokta, Suriani 15 November 2004 (has links)
This research consists of two parts. The ﬁrst part examines the posterior probability integrals for a family of linear models which arises from the work of Hart, Koen and Lombard (2003). Applying Laplace's method to these integrals is not entirely straightforward. One of the requirements is to analyze the asymptotic behavior of the information matrices as the sample size tends to inﬁnity. This requires a number of analytic tricks, including viewing our covariance matrices as tending to differential operators. The use of differential operators and their Green's functions can provide a convenient and systematic method to asymptotically invert the covariance matrices. Once we have found the asymptotic behavior of the information matrices, we will see that in most cases BIC provides a reasonable approximation to the log of the posterior probability and Laplace's method gives more terms in the expansion and hence provides a slightly better approximation. In other cases, a number of pathologies will arise. We will see that in one case, BIC does not provide an asymptotically consistent estimate of the posterior probability; however, the more general Laplace's method will provide such an estimate. In another case, we will see that a naive application of Laplace's method will give a misleading answer and Laplace's method must be adapted to give the correct answer. The second part uses numerical methods to compute the "exact" posterior probabilities and compare them to the approximations arising from BIC and Laplace's method.

2 
Bayesian model selection using exact and approximated posterior probabilities with applications to Star DataPokta, Suriani 15 November 2004 (has links)
This research consists of two parts. The ﬁrst part examines the posterior probability integrals for a family of linear models which arises from the work of Hart, Koen and Lombard (2003). Applying Laplace's method to these integrals is not entirely straightforward. One of the requirements is to analyze the asymptotic behavior of the information matrices as the sample size tends to inﬁnity. This requires a number of analytic tricks, including viewing our covariance matrices as tending to differential operators. The use of differential operators and their Green's functions can provide a convenient and systematic method to asymptotically invert the covariance matrices. Once we have found the asymptotic behavior of the information matrices, we will see that in most cases BIC provides a reasonable approximation to the log of the posterior probability and Laplace's method gives more terms in the expansion and hence provides a slightly better approximation. In other cases, a number of pathologies will arise. We will see that in one case, BIC does not provide an asymptotically consistent estimate of the posterior probability; however, the more general Laplace's method will provide such an estimate. In another case, we will see that a naive application of Laplace's method will give a misleading answer and Laplace's method must be adapted to give the correct answer. The second part uses numerical methods to compute the "exact" posterior probabilities and compare them to the approximations arising from BIC and Laplace's method.

3 
A Review of Cross Validation and Adaptive Model SelectionSyed, Ali R 27 April 2011 (has links)
We perform a review of model selection procedures, in particular various cross validation procedures and adaptive model selection. We cover important results for these procedures and explore the connections between different procedures and information criteria.

4 
A New Measure For Clustering Model SelectionMcCrosky, Jesse January 2008 (has links)
A new method for determining the number of kmeans clusters in a given data set is presented. The algorithm is developed from a theoretical perspective and then its implementation is examined and compared to existing solutions.

5 
A New Measure For Clustering Model SelectionMcCrosky, Jesse January 2008 (has links)
A new method for determining the number of kmeans clusters in a given data set is presented. The algorithm is developed from a theoretical perspective and then its implementation is examined and compared to existing solutions.

6 
TwoStage SCAD Lasso for Linear Mixed Model SelectionYousef, Mohammed A. 07 August 2019 (has links)
No description available.

7 
Application of a spatially referenced water quality model to predict E. coli flux in two Texas river basins, Deepti 15 May 2009 (has links)
Water quality models are applied to assess the various processes affecting the
concentrations of contaminants in a watershed. SPAtially Referenced Regression On
Watershed attributes (SPARROW) is a nonlinear regression based approach to predict
the fate and transport of contaminants in river basins. In this research SPARROW was
applied to the Guadalupe and San Antonio River Basins of Texas to assess E. coli
contamination. Since SPARROW relies on the measured records of concentrations of
contaminants collected at monitoring stations for the prediction, the effect of the
locations and selections of the monitoring stations was analyzed. The results of
SPARROW application were studied in detail to evaluate the contribution from the
statistically significant sources. For verification of SPARROW application, results were
compared to 303 (d) list of Clean Water Act, 2000. Further, a methodology to maintain
the monitoring records of the highly contaminated areas in the watersheds was explored
with the application of the genetic algorithm. In this study, the importance of the
available scale and details of explanatory variables (sources, landwater delivery and
reservoir/ stream attenuation factors) in predicting the water quality processes were also
analyzed. The effect of uncertainty in the monitored records on SPARROW application
was discussed. The application of SPARROW and genetic algorithm were explored to
design a monitoring network for the study area. The results of this study show that
SPARROW model can be used successfully to predict the pathogen contamination of
rivers. Also, SPARROW can be applied to design the monitoring network for the basins.

8 
CrossValidation for Model Selection in ModelBased ClusteringO'Reilly, Rachel 04 September 2012 (has links)
Clustering is a technique used to partition unlabelled data into meaningful groups. This thesis will focus on the area of clustering called modelbased clustering, where it is assumed that data arise from a finite number of subpopulations, each of which follows a known statistical distribution. The number of groups and shape of each group is unknown in advance, and thus one of the most challenging aspects of clustering is selecting these features.
Crossvalidation is a model selection technique which is often used in regression and classification, because it tends to choose models that predict well, and are not overfit to the data. However, it has rarely been applied in a clustering framework. Herein, crossvalidation is applied to select the number of groups and covariance structure within a family of Gaussian mixture models. Results are presented for both real and simulated data. / Ontario Graduate Scholarship Program

9 
Reduction of Dimensionality in Spatiotemporal ModelsSætrom, Jon January 2010 (has links)
No description available.

10 
Evaluating Automatic Model SelectionPENG, SISI January 2011 (has links)
In this paper, we briefly describe the automatic model selection which is provided by Autometrics in the PcGive program. The modeler only needs to specify the initial model and the significance level at which to reduce the model. Then, the algorithm does the rest. The properties of Autometrics are discussed. We also explain its background concepts and try to see whether the model selected by the Autometrics can perform well. For a given data set, we use Autometrics to find a “new” model, and then compare the “new” model with a previously selected one by another modeler. It is an interesting issue to see whether Autometrics can also find models which fit better to the given data. As an illustration, we choose three examples. It is true that Autometrics is labor saving and always gives us a parsimonious model. It is really an invaluable instrument for social science. But, we still need more examples to strongly support the idea that Autometrics can find a model which fits the data better, just a few examples in this paper is far from enough.

Page generated in 0.1245 seconds