Spelling suggestions: "subject:"inference"" "subject:"lnference""
231 |
Complex Feature Recognition: A Bayesian Approach for Learning to Recognize ObjectsViola, Paul 01 November 1996 (has links)
We have developed a new Bayesian framework for visual object recognition which is based on the insight that images of objects can be modeled as a conjunction of local features. This framework can be used to both derive an object recognition algorithm and an algorithm for learning the features themselves. The overall approach, called complex feature recognition or CFR, is unique for several reasons: it is broadly applicable to a wide range of object types, it makes constructing object models easy, it is capable of identifying either the class or the identity of an object, and it is computationally efficient--requiring time proportional to the size of the image. Instead of a single simple feature such as an edge, CFR uses a large set of complex features that are learned from experience with model objects. The response of a single complex feature contains much more class information than does a single edge. This significantly reduces the number of possible correspondences between the model and the image. In addition, CFR takes advantage of a type of image processing called 'oriented energy'. Oriented energy is used to efficiently pre-process the image to eliminate some of the difficulties associated with changes in lighting and pose.
|
232 |
A Statistical Image-Based Shape Model for Visual Hull Reconstruction and 3D Structure InferenceGrauman, Kristen 22 May 2003 (has links)
We present a statistical image-based shape + structure model for Bayesian visual hull reconstruction and 3D structure inference. The 3D shape of a class of objects is represented by sets of contours from silhouette views simultaneously observed from multiple calibrated cameras. Bayesian reconstructions of new shapes are then estimated using a prior density constructed with a mixture model and probabilistic principal components analysis. We show how the use of a class-specific prior in a visual hull reconstruction can reduce the effect of segmentation errors from the silhouette extraction process. The proposed method is applied to a data set of pedestrian images, and improvements in the approximate 3D models under various noise conditions are shown. We further augment the shape model to incorporate structural features of interest; unknown structural parameters for a novel set of contours are then inferred via the Bayesian reconstruction process. Model matching and parameter inference are done entirely in the image domain and require no explicit 3D construction. Our shape model enables accurate estimation of structure despite segmentation errors or missing views in the input silhouettes, and works even with only a single input view. Using a data set of thousands of pedestrian images generated from a synthetic model, we can accurately infer the 3D locations of 19 joints on the body based on observed silhouette contours from real images.
|
233 |
Models and Methods for Molecular PhylogeneticsCatanzaro, Daniele 28 October 2008 (has links)
Un des buts principaux de la biologie évolutive et de la médecine moléculaire consiste à reconstruire les relations phylogénétiques entre organismes à partir de leurs séquences moléculaires. En littérature, cette question est connue sous le nom d’inférence phylogénétique et a d'importantes applications dans la recherche médicale et pharmaceutique, ainsi que dans l’immunologie, l’épidémiologie, et la dynamique des populations. L’accumulation récente de données de séquences d’ADN dans les bases de données publiques, ainsi que la facilité relative avec laquelle des données nouvelles peuvent être obtenues, rend l’inférence phylogénétique particulièrement difficile (l'inférence phylogénétique est un problème NP-Hard sous tous les critères d’optimalité connus), de telle manière que des nouveaux critères et des algorithmes efficaces doivent être développés. Cette thèse a pour but: (i) d’analyser les limites mathématiques et biologiques des critères utilisés en inférence phylogénétique, (ii) de développer de nouveaux algorithmes efficaces permettant d’analyser de plus grands jeux de données.
|
234 |
Information-Theoretic Variable Selection and Network Inference from Microarray DataMeyer, Patrick E 16 December 2008 (has links)
Statisticians are used to model interactions between variables on the basis of observed
data. In a lot of emerging fields, like bioinformatics, they are confronted with datasets
having thousands of variables, a lot of noise, non-linear dependencies and, only, tens of
samples. The detection of functional relationships, when such uncertainty is contained in
data, constitutes a major challenge.
Our work focuses on variable selection and network inference from datasets having
many variables and few samples (high variable-to-sample ratio), such as microarray data.
Variable selection is the topic of machine learning whose objective is to select, among a
set of input variables, those that lead to the best predictive model. The application of
variable selection methods to gene expression data allows, for example, to improve cancer
diagnosis and prognosis by identifying a new molecular signature of the disease. Network
inference consists in representing the dependencies between the variables of a dataset by
a graph. Hence, when applied to microarray data, network inference can reverse-engineer
the transcriptional regulatory network of cell in view of discovering new drug targets to
cure diseases.
In this work, two original tools are proposed MASSIVE (Matrix of Average Sub-Subset
Information for Variable Elimination) a new method of feature selection and MRNET (Minimum
Redundancy NETwork), a new algorithm of network inference. Both tools rely on
the computation of mutual information, an information-theoretic measure of dependency.
More precisely, MASSIVE and MRNET use approximations of the mutual information
between a subset of variables and a target variable based on combinations of mutual informations
between sub-subsets of variables and the target. The used approximations allow
to estimate a series of low variate densities instead of one large multivariate density. Low
variate densities are well-suited for dealing with high variable-to-sample ratio datasets,
since they are rather cheap in terms of computational cost and they do not require a large
amount of samples in order to be estimated accurately. Numerous experimental results
show the competitiveness of these new approaches. Finally, our thesis has led to a freely
available source code of MASSIVE and an open-source R and Bioconductor package of
network inference.
|
235 |
Essays on Aggregation and Cointegration of Econometric ModelsSilvestrini, Andrea 02 June 2009 (has links)
This dissertation can be broadly divided into two independent parts. The first three chapters analyse issues related to temporal and contemporaneous aggregation of econometric models. The fourth chapter contains an application of Bayesian techniques to investigate whether the post transition fiscal policy of Poland is sustainable in the long run and consistent with an intertemporal budget constraint.
Chapter 1 surveys the econometric methodology of temporal aggregation for a wide range of univariate and multivariate time series models.
A unified overview of temporal aggregation techniques for this broad class of processes is presented in the first part of the chapter and the main results are summarized. In each case, assuming to know the underlying process at the disaggregate frequency, the aim is to find the appropriate model for the aggregated data. Additional topics concerning temporal aggregation of ARIMA-GARCH models (see Drost and Nijman, 1993) are discussed and several examples presented. Systematic sampling schemes are also reviewed.
Multivariate models, which show interesting features under temporal aggregation (Breitung and Swanson, 2002, Marcellino, 1999, Hafner, 2008), are examined in the second part of the chapter. In particular, the focus is on temporal aggregation of VARMA models and on the related concept of spurious instantaneous causality, which is not a time series property invariant to temporal aggregation. On the other hand, as pointed out by Marcellino (1999), other important time series features as cointegration and presence of unit roots are invariant to temporal aggregation and are not induced by it.
Some empirical applications based on macroeconomic and financial data illustrate all the techniques surveyed and the main results.
Chapter 2 is an attempt to monitor fiscal variables in the Euro area, building an early warning signal indicator for assessing the development of public finances in the short-run and exploiting the existence of monthly budgetary statistics from France, taken as "example country".
The application is conducted focusing on the cash State deficit, looking at components from the revenue and expenditure sides. For each component, monthly ARIMA models are estimated and then temporally aggregated to the annual frequency, as the policy makers are interested in yearly predictions.
The short-run forecasting exercises carried out for years 2002, 2003 and 2004 highlight the fact that the one-step-ahead predictions based on the temporally aggregated models generally outperform those delivered by standard monthly ARIMA modeling, as well as the official forecasts made available by the French government, for each of the eleven components and thus for the whole State deficit. More importantly, by the middle of the year, very accurate predictions for the current year are made available.
The proposed method could be extremely useful, providing policy makers with a valuable indicator when assessing the development of public finances in the short-run (one year horizon or even less).
Chapter 3 deals with the issue of forecasting contemporaneous time series aggregates. The performance of "aggregate" and "disaggregate" predictors in forecasting contemporaneously aggregated vector ARMA (VARMA) processes is compared. An aggregate predictor is built by forecasting directly the aggregate process, as it results from contemporaneous aggregation of the data generating vector process. A disaggregate predictor is a predictor obtained from aggregation of univariate forecasts for the individual components of the data generating vector process.
The econometric framework is broadly based on Lütkepohl (1987). The necessary and sufficient condition for the equality of mean squared errors associated with the two competing methods in the bivariate VMA(1) case is provided. It is argued that the condition of equality of predictors as stated in Lütkepohl (1987), although necessary and sufficient for the equality of the predictors, is sufficient (but not necessary) for the equality of mean squared errors.
Furthermore, it is shown that the same forecasting accuracy for the two predictors can be achieved using specific assumptions on the parameters of the VMA(1) structure.
Finally, an empirical application that involves the problem of forecasting the Italian monetary aggregate M1 on the basis of annual time series ranging from 1948 until 1998, prior to the creation of the European Economic and Monetary Union (EMU), is presented to show the relevance of the topic. In the empirical application, the framework is further generalized to deal with heteroskedastic and cross-correlated innovations.
Chapter 4 deals with a cointegration analysis applied to the empirical investigation of fiscal sustainability. The focus is on a particular country: Poland. The choice of Poland is not random. First, the motivation stems from the fact that fiscal sustainability is a central topic for most of the economies of Eastern Europe. Second, this is one of the first countries to start the transition process to a market economy (since 1989), providing a relatively favorable institutional setting within which to study fiscal sustainability (see Green, Holmes and Kowalski, 2001). The emphasis is on the feasibility of a permanent deficit in the long-run, meaning whether a government can continue to operate under its current fiscal policy indefinitely.
The empirical analysis to examine debt stabilization is made up by two steps.
First, a Bayesian methodology is applied to conduct inference about the cointegrating relationship between budget revenues and (inclusive of interest) expenditures and to select the cointegrating rank. This task is complicated by the conceptual difficulty linked to the choice of the prior distributions for the parameters relevant to the economic problem under study (Villani, 2005).
Second, Bayesian inference is applied to the estimation of the normalized cointegrating vector between budget revenues and expenditures. With a single cointegrating equation, some known results concerning the posterior density of the cointegrating vector may be used (see Bauwens, Lubrano and Richard, 1999).
The priors used in the paper leads to straightforward posterior calculations which can be easily performed.
Moreover, the posterior analysis leads to a careful assessment of the magnitude of the cointegrating vector. Finally, it is shown to what extent the likelihood of the data is important in revising the available prior information, relying on numerical integration techniques based on deterministic methods.
|
236 |
Causal assumptions : some responses to Nancy CartwrightKristtorn, Sonje 31 July 2007
The theories of causality put forward by Pearl and the Spirtes-Glymour-Scheines group have entered the mainstream of statistical thinking. These theories show that under ideal conditions, causal relationships can be inferred from purely statistical observational data. Nancy Cartwright advances certain arguments against these causal inference algorithms: the well-known factory example argument against the Causal Markov condition and an argument against faithfulness. We point to the dependence of the first argument on undefined categories external to the technical apparatus of causal inference algorithms. We acknowledge the possible practical implication of her second argument, yet we maintain, with respect to both arguments, that this variety of causal inference, if not universal, is nonetheless eminently useful. Cartwright argues against assumptions that are essential not only to causal inference algorithms but to causal inference generally, even if, as she contends, they are not without exception and that the same is true of other, likewise essential, assumptions. We indicate that causal inference is an iterative process and that causal inference algorithms assist, rather than replace, that process as performed by human beings.
|
237 |
Bayesian Inference for Stochastic Volatility ModelsMen, Zhongxian January 1012 (has links)
Stochastic volatility (SV) models provide a natural framework for a
representation of time series for financial asset returns. As a
result, they have become increasingly popular in the finance
literature, although they have also been applied in other fields
such as signal processing, telecommunications, engineering, biology,
and other areas.
In working with the SV models, an important issue arises as how to
estimate their parameters efficiently and to assess how well they
fit real data. In the literature, commonly used estimation methods
for the SV models include general methods of moments, simulated
maximum likelihood methods, quasi Maximum likelihood method, and
Markov Chain Monte Carlo (MCMC) methods. Among these approaches,
MCMC methods are most flexible in dealing with complicated structure
of the models. However, due to the difficulty in the selection of
the proposal distribution for Metropolis-Hastings methods, in
general they are not easy to implement and in some cases we may also
encounter convergence problems in the implementation stage. In the
light of these concerns, we propose in this thesis new estimation
methods for univariate and multivariate SV models. In the simulation
of latent states of the heavy-tailed SV models, we recommend the
slice sampler algorithm as the main tool to sample the proposal
distribution when the Metropolis-Hastings method is applied. For the
SV models without heavy tails, a simple Metropolis-Hastings method
is developed for simulating the latent states. Since the slice
sampler can adapt to the analytical structure of the underlying
density, it is more efficient. A sample point can be obtained from
the target distribution with a few iterations of the sampler,
whereas in the original Metropolis-Hastings method many sampled
values often need to be discarded.
In the analysis of multivariate time series, multivariate SV models
with more general specifications have been proposed to capture the
correlations between the innovations of the asset returns and those
of the latent volatility processes. Due to some restrictions on the
variance-covariance matrix of the innovation vectors, the estimation
of the multivariate SV (MSV) model is challenging. To tackle this
issue, for a very general setting of a MSV model we propose a
straightforward MCMC method in which a Metropolis-Hastings method is
employed to sample the constrained variance-covariance matrix, where
the proposal distribution is an inverse Wishart distribution. Again,
the log volatilities of the asset returns can then be simulated via
a single-move slice sampler.
Recently, factor SV models have been proposed to extract hidden
market changes. Geweke and Zhou (1996) propose a factor SV model
based on factor analysis to measure pricing errors in the context of
the arbitrage pricing theory by letting the factors follow the
univariate standard normal distribution. Some modification of this
model have been proposed, among others, by Pitt and Shephard (1999a)
and Jacquier et al. (1999). The main feature of the factor SV
models is that the factors follow a univariate SV process, where the
loading matrix is a lower triangular matrix with unit entries on the
main diagonal. Although the factor SV models have been successful in
practice, it has been recognized that the order of the component may
affect the sample likelihood and the selection of the factors.
Therefore, in applications, the component order has to be considered
carefully. For instance, the factor SV model should be fitted to
several permutated data to check whether the ordering affects the
estimation results. In the thesis, a new factor SV model is
proposed. Instead of setting the loading matrix to be lower
triangular, we set it to be column-orthogonal and assume that each
column has unit length. Our method removes the permutation problem,
since when the order is changed then the model does not need to be
refitted. Since a strong assumption is imposed on the loading
matrix, the estimation seems even harder than the previous factor
models. For example, we have to sample columns of the loading matrix
while keeping them to be orthonormal. To tackle this issue, we use
the Metropolis-Hastings method to sample the loading matrix one
column at a time, while the orthonormality between the columns is
maintained using the technique proposed by Hoff (2007). A von
Mises-Fisher distribution is sampled and the generated vector is
accepted through the Metropolis-Hastings algorithm.
Simulation studies and applications to real data are conducted to
examine our inference methods and test the fit of our model.
Empirical evidence illustrates that our slice sampler within MCMC
methods works well in terms of parameter estimation and volatility
forecast. Examples using financial asset return data are provided to
demonstrate that the proposed factor SV model is able to
characterize the hidden market factors that mainly govern the
financial time series. The Kolmogorov-Smirnov tests conducted on
the estimated models indicate that the models do a reasonable job in
terms of describing real data.
|
238 |
On the learnibility of Mildly Context-Sensitive languages using positive data and correction queriesBecerra Bonache, Leonor 06 March 2006 (has links)
Con esta tesis doctoral aproximamos la teoría de la inferencia gramatical y los estudios de adquisición del lenguaje, en pos de un objetivo final: ahondar en la comprensión del modo como los niños adquieren su primera lengua mediante la explotación de la teoría inferencial de gramáticas formales.Nuestras tres principales aportaciones son:1. Introducción de una nueva clase de lenguajes llamada Simple p-dimensional external contextual (SEC). A pesar de que las investigaciones en inferencia gramatical se han centrado en lenguajes regulares o independientes del contexto, en nuestra tesis proponemos centrar esos estudios en clases de lenguajes más relevantes desde un punto de vista lingüístico (familias de lenguajes que ocupan una posición ortogonal en la jerarquía de Chomsky y que son suavemente dependientes del contexto, por ejemplo, SEC).2. Presentación de un nuevo paradigma de aprendizaje basado en preguntas de corrección. Uno de los principales resultados positivos dentro de la teoría del aprendizaje formal es el hecho de que los autómatas finitos deterministas (DFA) se pueden aprender de manera eficiente utilizando preguntas de pertinencia y preguntas de equivalencia. Teniendo en cuenta que en el aprendizaje de primeras lenguas la corrección de errores puede jugar un papel relevante, en nuestra tesis doctoral hemos introducido un nuevo modelo de aprendizaje que reemplaza las preguntas de pertinencia por preguntas de corrección.3. Presentación de resultados basados en las dos previas aportaciones. En primer lugar, demostramos que los SEC se pueden aprender a partir de datos positivos. En segundo lugar, demostramos que los DFA se pueden aprender a partir de correcciones y que el número de preguntas se reduce considerablemente.Los resultados obtenidos con esta tesis doctoral suponen una aportación importante para los estudios en inferencia gramatical (hasta el momento las investigaciones en este ámbito se habían centrado principalmente en los aspectos matemáticos de los modelos). Además, estos resultados se podrían extender a diversos campos de aplicación que gozan de plena actualidad, tales como el aprendizaje automático, la robótica, el procesamiento del lenguaje natural y la bioinformática. / With this dissertation, we bring together the Theory of the Grammatical Inference and Studies of language acquisition, in pursuit of our final goal: to go deeper in the understanding of the process of language acquisition by using the theory of inference of formal grammars. Our main three contributions are:1. Introduction of a new class of languages called Simple p-dimensional external contextual (SEC). Despite the fact that the field of Grammatical Inference has focused its research on learning regular or context-free languages, we propose in our dissertation to focus these studies in classes of languages more relevant from a linguistic point of view (families of languages that occupy an orthogonal position in the Chomsky Hierarchy and are Mildly Context-Sensitive, for example SEC).2. Presentation of a new learning paradigm based on correction queries. One of the main results in the theory of formal learning is that deterministic finite automata (DFA) are efficiently learnable from membership query and equivalence query. Taken into account that in first language acquisition the correction of errors can play an important role, we have introduced in our dissertation a novel learning model by replacing membership queries with correction queries.3. Presentation of results based on the two previous contributions. First, we prove that SEC is learnable from only positive data. Second, we prove that it is possible to learn DFA from corrections and that the number of queries is reduced considerably.The results obtained with this dissertation suppose an important contribution to studies of Grammatical Inference (the current research in Grammatical Inference has focused mainly on the mathematical aspects of the models). Moreover, these results could be extended to studies related directly to machine translation, robotics, natural language processing, and bioinformatics.
|
239 |
Can Induction Strengthen Inference to the Best Explanation?Thomson, Neil A. January 2008 (has links)
In this paper I will argue that the controversial process of inferring to the best explanation (IBE) can be made more coherent if its formulation recognizes and includes a significant inductive component. To do so, I will examine the relationship between Harman’s, Lipton’s, and Fumerton’s positions on IBE, settling ultimately upon a conception that categorically rejects Harman’s account while appropriating potions of both Lipton’s and Fumerton’s accounts. The resulting formulation will be called inductive-IBE, and I will argue that this formulation more accurately describes the inferential practices employed in scientific inquiry. The upshot of my argument, that IBE contains a significant inductive component, will be that any conclusion born from such types of inductive inference must be, at best, likely, and not a necessity. And, although previous accounts of IBE have accepted the defeasibility of IBE, I will argue that inductive-IBE is more descriptive because it tells us why this fallibility exists. That is, although the Liptonian conception of IBE acknowledges that IBE is fallible, my account specifically addresses this characteristic and, thus, is more descriptive and informative in this regard. I will use inductive-IBE to argue, contra van Fraassen, that IBE can be a legitimate form of inference that leads science to true theories and real entities.
|
240 |
Bayesian inference for source determination in the atmospheric environmentKeats, William Andrew January 2009 (has links)
In the event of a hazardous release (chemical, biological, or radiological) in an urban environment, monitoring agencies must have the tools to locate and characterize the source of the emission in order to respond and minimize damage. Given a finite and noisy set of concentration measurements, determining the source location, strength and time of release is an ill-posed inverse problem. We treat this problem using Bayesian inference, a framework under which uncertainties in modelled and measured concentrations can be propagated, in a consistent, rigorous manner, toward a final probabilistic estimate for the source.
The Bayesian methodology operates independently of the chosen dispersion model, meaning it can be applied equally well to problems in urban environments, at regional scales, or at global scales. Both Lagrangian stochastic (particle-tracking) and Eulerian (fixed-grid, finite-volume) dispersion models have been used successfully. Calculations are accomplished efficiently by using adjoint (backward) dispersion models, which reduces the computational effort required from calculating one [forward] plume per possible source configuration to calculating one [backward] plume per detector. Markov chain Monte Carlo (MCMC) is used to efficiently sample from the posterior distribution for the source parameters; both the Metropolis-Hastings and hybrid Hamiltonian algorithms are used.
In this thesis, four applications falling under the rubric of source determination are addressed: dispersion in highly disturbed flow fields characteristic of built-up (urban) environments; dispersion of a nonconservative scalar over flat terrain in a statistically stationary and horizontally homogeneous (turbulent) wind field; optimal placement of an auxiliary detector using a decision-theoretic approach; and source apportionment of particulate matter (PM) using a chemical mass balance (CMB) receptor model. For the first application, the data sets used to validate the proposed methodology include a water-channel simulation of the near-field dispersion of contaminant plumes in a large array of building-like obstacles (Mock Urban Setting Trial) and a full-scale field experiment (Joint Urban 2003) in Oklahoma City. For the second and third applications, the background wind and terrain conditions are based on those encountered during the Project Prairie Grass field experiment; mean concentration and turbulent scalar flux data are synthesized using a Lagrangian stochastic model where necessary. In the fourth and final application, Bayesian source apportionment results are compared to the US Environmental Protection Agency's standard CMB model using a test case involving PM data from Fresno, California. For each of the applications addressed in this thesis, combining Bayesian inference with appropriate computational techniques results in a computationally efficient methodology for performing source determination.
|
Page generated in 0.0586 seconds