Spelling suggestions: "subject:"stochastic analysis"" "subject:"ctochastic analysis""
191 |
Optimal policies in reliability modelling of systems subject to sporadic shocks and continuous healingDEBOLINA CHATTERJEE (14206820) 03 February 2023 (has links)
<p>Recent years have seen a growth in research on system reliability and maintenance. Various studies in the scientific fields of reliability engineering, quality and productivity analyses, risk assessment, software reliability, and probabilistic machine learning are being undertaken in the present era. The dependency of human life on technology has made it more important to maintain such systems and maximize their potential. In this dissertation, some methodologies are presented that maximize certain measures of system reliability, explain the underlying stochastic behavior of certain systems, and prevent the risk of system failure.</p>
<p><br></p>
<p>An overview of the dissertation is provided in Chapter 1, where we briefly discuss some useful definitions and concepts in probability theory and stochastic processes and present some mathematical results required in later chapters. Thereafter, we present the motivation and outline of each subsequent chapter.</p>
<p><br></p>
<p>In Chapter 2, we compute the limiting average availability of a one-unit repairable system subject to repair facilities and spare units. Formulas for finding the limiting average availability of a repairable system exist only for some special cases: (1) either the lifetime or the repair-time is exponential; or (2) there is one spare unit and one repair facility. In contrast, we consider a more general setting involving several spare units and several repair facilities; and we allow arbitrary life- and repair-time distributions. Under periodic monitoring, which essentially discretizes the time variable, we compute the limiting average availability. The discretization approach closely approximates the existing results in the special cases; and demonstrates as anticipated that the limiting average availability increases with additional spare unit and/or repair facility.</p>
<p><br></p>
<p>In Chapter 3, the system experiences two types of sporadic impact: valid shocks that cause damage instantaneously and positive interventions that induce partial healing. Whereas each shock inflicts a fixed magnitude of damage, the accumulated effect of k positive interventions nullifies the damaging effect of one shock. The system is said to be in Stage 1, when it can possibly heal, until the net count of impacts (valid shocks registered minus valid shocks nullified) reaches a threshold $m_1$. The system then enters Stage 2, where no further healing is possible. The system fails when the net count of valid shocks reaches another threshold $m_2 (> m_1)$. The inter-arrival times between successive valid shocks and those between successive positive interventions are independent and follow arbitrary distributions. Thus, we remove the restrictive assumption of an exponential distribution, often found in the literature. We find the distributions of the sojourn time in Stage 1 and the failure time of the system. Finally, we find the optimal values of the choice variables that minimize the expected maintenance cost per unit time for three different maintenance policies.</p>
<p><br></p>
<p>In Chapter 4, the above defined Stage 1 is further subdivided into two parts: In the early part, called Stage 1A, healing happens faster than in the later stage, called Stage 1B. The system stays in Stage 1A until the net count of impacts reaches a predetermined threshold $m_A$; then the system enters Stage 1B and stays there until the net count reaches another predetermined threshold $m_1 (>m_A)$. Subsequently, the system enters Stage 2 where it can no longer heal. The system fails when the net count of valid shocks reaches another predetermined higher threshold $m_2 (> m_1)$. All other assumptions are the same as those in Chapter 3. We calculate the percentage improvement in the lifetime of the system due to the subdivision of Stage 1. Finally, we make optimal choices to minimize the expected maintenance cost per unit time for two maintenance policies.</p>
<p><br></p>
<p>Next, we eliminate the restrictive assumption that all valid shocks and all positive interventions have equal magnitude, and the boundary threshold is a preset constant value. In Chapter 5, we study a system that experiences damaging external shocks of random magnitude at stochastic intervals, continuous degradation, and self-healing. The system fails if cumulative damage exceeds a time-dependent threshold. We develop a preventive maintenance policy to replace the system such that its lifetime is utilized prudently. Further, we consider three variations on the healing pattern: (1) shocks heal for a fixed finite duration $\tau$; (2) a fixed proportion of shocks are non-healable (that is, $\tau=0$); (3) there are two types of shocks---self healable shocks heal for a finite duration, and non-healable shocks. We implement a proposed preventive maintenance policy and compare the optimal replacement times in these new cases with those in the original case, where all shocks heal indefinitely.</p>
<p><br></p>
<p>Finally, in Chapter 6, we present a summary of the dissertation with conclusions and future research potential.</p>
|
192 |
Financial stress in an adaptive system: From empirical validity to theoretical foundationsOet, Mikhail V. 01 June 2016 (has links)
No description available.
|
193 |
MODEL-FREE ALGORITHMS FOR CONSTRAINED REINFORCEMENT LEARNING IN DISCOUNTED AND AVERAGE REWARD SETTINGSQinbo Bai (19804362) 07 October 2024 (has links)
<p dir="ltr">Reinforcement learning (RL), which aims to train an agent to maximize its accumulated reward through time, has attracted much attention in recent years. Mathematically, RL is modeled as a Markov Decision Process, where the agent interacts with the environment step by step. In practice, RL has been applied to autonomous driving, robotics, recommendation systems, and financial management. Although RL has been greatly studied in the literature, most proposed algorithms are model-based, which requires estimating the transition kernel. To this end, we begin to study the sample efficient model-free algorithms under different settings.</p><p dir="ltr">Firstly, we propose a conservative stochastic primal-dual algorithm in the infinite horizon discounted reward setting. The proposed algorithm converts the original problem from policy space to the occupancy measure space, which makes the non-convex problem linear. Then, we advocate the use of a randomized primal-dual approach to achieve O(\eps^-2) sample complexity, which matches the lower bound.</p><p dir="ltr">However, when it comes to the infinite horizon average reward setting, the problem becomes more challenging since the environment interaction never ends and can’t be reset, which makes reward samples not independent anymore. To solve this, we design an epoch-based policy-gradient algorithm. In each epoch, the whole trajectory is divided into multiple sub-trajectories with an interval between each two of them. Such intervals are long enough so that the reward samples are asymptotically independent. By controlling the length of trajectory and intervals, we obtain a good gradient estimator and prove the proposed algorithm achieves O(T^3/4) regret bound.</p>
|
194 |
Algorithmic component and system reliability analysis of truss structuresHashemolhosseini, Sepehr 12 1900 (has links)
Thesis (MScEng)-- Stellenbosch University, 2013. / ENGLISH ABSTRACT: Most of the parameters involved in the design and analysis of structures are of stochastic nature.
This is, therefore, of paramount importance to be able to perform a fully stochastic analysis of
structures both in component and system level to take into account the uncertainties involved
in structural analysis and design. To the contrary, in practice, the (computerised) analysis of
structures is based on a deterministic analysis which fails to address the randomness of design
and analysis parameters. This means that an investigation on the algorithmic methodologies for
a component and system reliability analysis can help pave the way towards the implementation
of fully stochastic analysis of structures in a computer environment. This study is focused
on algorithm development for component and system reliability analysis based on the various
proposed methodologies. Truss structures were selected for this purpose due to their simplicity
as well as their wide use in the industry. Nevertheless, the algorithms developed in this study
can be used for other types of structures such as moment-resisting frames with some simple
modi cations.
For a component level reliability analysis of structures different methods such as First Order
Reliability Methods (FORM) and simulation methods are proposed. However, implementation
of these methods for the statistically indeterminate structures is complex due to the implicit
relation between the response of the structural system and the load effect. As a result, the
algorithm developed for the purpose of component reliability analysis should be based on the
concepts of Stochastic Finite Element Methods (SFEM) where a proper link between the finite
element analysis of the structure and the reliability analysis methodology is ensured. In this
study various algorithms are developed based on the FORM method, Monte Carlo simulation,
and the Response Surface Method (RSM). Using the FORM method, two methodologies are
considered: one is based on the development of a finite element code where required alterations
are made to the FEM code and the other is based on the usage of a commercial FEM package.
Different simulation methods are also implemented: Direct Monte Carlo Simulation (DMCS),
Latin Hypercube Sampling Monte Carlo (LHCSMC), and Updated Latin Hypercube Sampling
Monte Carlo (ULHCSMC). Moreover, RSM is used together with simulation methods. Throughout the thesis, the effciency of these methods was investigated. A Fully Stochastic
Finite Element Method (FSFEM) with alterations to the finite element code seems the fastest
approach since the linking between the FEM package and reliability analysis is avoided. Simulation methods can also be effectively used for the reliability evaluation where ULHCSMC seemed
to be the most efficient method followed by LHCSMC and DMCS. The response surface method
is the least straight forward method for an algorithmic component reliability analysis; however,
it is useful for the system reliability evaluation.
For a system level reliability analysis two methods were considered: the ß-unzipping method
and the branch and bound method. The ß-unzipping method is based on a level-wise system
reliability evaluation where the structure is modelled at different damaged levels according to its
degree of redundancy. In each level, the so-called unzipping intervals are defined for the identification of the critical elements. The branch and bound method is based on the identification
of different failure paths of the structure by the expansion of the structural failure tree. The
evaluation of the damaged states for both of the methods is the same. Furthermore, both of
the methods lead to the development of a parallel-series model for the structural system. The
only difference between the two methods is in the search approach used for the failure sequence
identification.
It was shown that the ß-unzipping method provides a better algorithmic approach for evaluating
the system reliability compared to the branch and bound method. Nevertheless, the branch and
bound method is a more robust method in the identification of structural failure sequences. One
possible way to increase the efficiency of the ß-unzipping method is to define bigger unzipping
intervals in each level which can be possible through a computerised analysis. For such an
analysis four major modules are required: a general intact structure module, a damaged structure
module, a reliability analysis module, and a system reliability module.
In this thesis different computer programs were developed for both system and component
reliability analysis based on the developed algorithms. The computer programs are presented
in the appendices of the thesis. / AFRIKAANSE OPSOMMING: Meeste van die veranderlikes betrokke by die ontwerp en analise van strukture is stogasties in
hul aard. Om die onsekerhede betrokke in ontwerp en analise in ag te neem is dit dus van
groot belang om 'n ten volle stogastiese analise te kan uitvoer op beide komponent asook stelsel
vlak. In teenstelling hiermee is die gerekenariseerde analise van strukture in praktyk gebaseer
op deterministiese analise wat nie suksesvol is om die stogastiese aard van ontwerp veranderlikes
in ag te neem nie. Dit beteken dat die ondersoek na die algoritmiese metodiek vir komponent en
stelsel betroubaarheid analise kan help om die weg te baan na die implementering van ten volle
rekenaarmatige stogastiese analise van strukture. Di e studie se fokus is op die ontwikkeling van
algoritmes vir komponent en stelsel betroubaarheid analise soos gegrond op verskeie voorgestelde
metodes. Vakwerk strukture is gekies vir die doeleinde as gevolg van hulle eenvoud asook hulle
wydverspreide gebruik in industrie. Die algoritmes wat in die studie ontwikkel is kan nietemin
ook vir ander tipes strukture soos moment-vaste raamwerke gebruik word, gegewe eenvoudige
aanpassings.
Vir 'n komponent vlak betroubaarheid analise van strukture word verskeie metodes soos die
"First Order Reliability Methods" (FORM) en simulasie metodes voorgestel. Die implementering
van die metodes vir staties onbepaalbare strukture is ingewikkeld as gevolg van die implisiete
verband tussen die gedrag van die struktuur stelsel en die las effek. As 'n gevolg, moet die algoritme
wat ontwikkel word vir die doel van komponent betroubaarheid analise gebaseer word
op die konsepte van stogastiese eindige element metodes ("SFEM") waar 'n duidelike verband
tussen die eindige element analise van die struktuur en die betroubaarheid analise verseker is. In
hierdie studie word verskeie algoritmes ontwikkel wat gebaseer is op die FORM metode, Monte
Carlo simulasie, en die sogenaamde "Response Surface Method" (RSM). Vir die gebruik van die
FORM metode word twee verdere metodologieë ondersoek: een gebaseer op die ontwikkeling
van 'n eindige element kode waar nodige verandering aan die eindige element kode self gemaak
word en die ander waar 'n kommersiële eindige element pakket gebruik word. Verskillende simulasie
metodes word ook geïmplimenteer naamlik Direkte Monte Carlo Simulasie (DMCS),
"Latin Hypercube Sampling Monte Carlo" (LHCSMC) en sogenaamde "Updated Latin Hypercube
Sampling Monte Carlo" (ULHCSMC). Verder, word RSM tesame met die simulasie
metodes gebruik. In die tesis word die doeltreffendheid van die bostaande metodes deurgaans ondersoek. 'n Ten volle stogastiese eindige element metode ("FSFEM") met verandering aan die eindige element
kode blyk die vinnigste benadering te wees omdat die koppeling tussen die eindige element
metode pakket en die betroubaarheid analise verhoed word. Simulasie metodes kan ook effektief
aangewend word vir die betroubaarheid evaluasie waar ULHCSMC as die mees doeltre end
voorgekom het, gevolg deur LHCSMC en DMCS. The RSM metode is die mees komplekse
metode vir algoritmiese komponent betroubaarheid analise. Die metode is egter nuttig vir
sisteem betroubaarheid analise.
Vir sisteem-vlak betroubaarheid analise is twee metodes oorweeg naamlik die "ß-unzipping"
metode and die "branch-and-bound" metode. Die "ß-unzipping" metode is gebaseer op 'n
sisteem-vlak betroubaarheid ontleding waar die struktuur op verskillende skade vlakke gemodelleer
word soos toepaslik vir die hoeveelheid addisionele las paaie. In elke vlak word die
sogenaamde "unzipping" intervalle gedefinieer vir die identifikasie van die kritiese elemente. Die
"branch-and-bound" metode is gebaseer op die identifikasie van verskillende faling roetes van
die struktuur deur uitbreiding van die falingsboom. The ondersoek van die skade toestande vir
beide metodes is dieselfde. Verder kan beide metodes lei tot die ontwikkeling van 'n parallelserie
model van die strukturele stelsel. Die enigste verskil tussen die twee metodes is in die
soek-benadering vir die uitkenning van falingsmodus volgorde.
Dit word getoon dat die "ß-unzipping" metode 'n beter algoritmiese benadering is vir die ontleding
van sisteem betroubaarheid vergeleke met die "branch-and-bound" metode. Die "branch-and-
bound" metode word nietemin as 'n meer robuuste metode vir die uitkenning van die falings
volgorde beskou. Een moontlike manier om die doeltre endheid van die "ß-unzipping" metode
te verhoog is om groter "unzipping" intervalle te gebruik, wat moontlik is vir rekenaarmatige
analise. Vir so 'n analise word vier hoof modules benodig naamlik 'n algemene heel-struktuur
module, 'n beskadigde-struktuur module, 'n betroubaarheid analise module en 'n sisteem betroubaarheid analise module.
In die tesis word verskillende rekenaar programme ontwikkel vir beide sisteem en komponent
betroubaarheid analise. Die rekenaar programme word in die aanhangsels van die tesis
aangebied.
|
195 |
Numerical analysis of the nonlinear dynamics of a drill-string with uncertainty modeling / Analyse numérique de la dynamique nonlinéaire d'une colonne de forage avec modélisation d'incertitudesRitto, Thiago 07 April 2010 (has links)
On fait l'analyse de la dynamique nonlinéaire d'une colonne de forage avec modélisation d'incertitudes. Une colonne de forage est une structure flexible mince qui tourne et fore des roches en cherchant du pétrole. Un modèle mathématique-mécanique est développeé pour cette structure en incluant de l'interaction fluide-structure, des impact, des nonlinéarités géométriques, et de l'interaction tête de la colonne-roche. Les équations de mouvement sont déduites, puis le système est discretisé par la méthode des éléments finis, et un code numérique est développeé pour les simulations numériques avec le logiciel MATLAB. Les modes normaux de la dynamique à la configuration précontrainte sont utilisés pour construire un modèle reduit pour le système. L'approche probabiliste nonparamétrique, qui est capable de prendre en compte les incertitude des paramètres du système et aussi les incertitude de modèle, est utilisée. Les fonctions densités de probabilité liées aux variables aléatoire sont construites par la méthode du Maximum d'Entropie et la réponse stochastique du système est calculée en utilisant la méthode de Monte Carlo. Une nouvelle approche pour prendre en compte des incertitudes de modèle dans une équation constitutive nonlinéaire (interaction tête de la colonne-roche) est développeé avec l'approche probabiliste nonparamétrique. La méthode du maximum de vraisemblance et la réduction statistique dans le domaine de la fréquence (Analyse de composantes principales) sont utilisés pour identifier le modèle probabiliste de l'interaction tête de la colonne-roche. Finalement, un problème d'optimisation robuste est analysé de façon à trouver les paramètres opérationnels du système qui maximise sa performance et respectent les limites d'intégrités du système, comme fatigue et instabilitée / This thesis analyzes the nonlinear dynamics of a drill-string including uncertainty modeling. A drill-string is a slender flexible structure that rotates and digs into the rock in search of oil. A mathematical-mechanical model is developed for this structure including fluid-structure interaction, impact, geometrical nonlinearities and bit-rock interaction. After the derivation of the equations of motion, the system is discretized by means of the finite element method and a computer code is developed for the numerical computations using the software MATLAB. The normal modes of the dynamical system in the prestressed configuration are used to construct a reduced order model for the system. To take into account uncertainties, the nonparametric probabilistic approach, which is able to take into account both system-parameter and model uncertainties, is used. The probability density functions related to the random variables are constructed using the maximum entropy principle and the stochastic response of the system is calculated using the Monte Carlo method. A novel approach to take into account model uncertainties in a nonlinear constitutive equation (bit-rock interaction model) is developed using the nonparametric probabilistic approach. To identify the probabilistic model of the bit-rock interaction model, the maximum likelihood method together with a statistical reduction in the frequency domain (using the Principal Component Analysis) is applied. Finally, a robust optimization problem is performed to find the operational parameters of the system that maximizes its performance, respecting the integrity limits of the system, such as fatigue and instability
|
196 |
Aspekte unendlichdimensionaler Martingaltheorie und ihre Anwendung in der Theorie der FinanzmärkteSchöckel, Thomas 19 October 2004 (has links)
Wir modellieren einen Finanzmarkt mit unendlich vielen Wertpapieren als stochastischen Prozeß X in stetiger Zeit mit Werten in einem separablen Hilbertraum H. In diesem Rahmen zeigen wir die Äquivalenz von Vollständigkeit des Marktes und der Eindeutigkeit des äquivalenten Martingalmaßes unter der Bedingung, daß X stetige Pfade besitzt. Weiter zeigen wir, daß (unter gewissen technischen Bedingungen) für X die Abwesenheit von asymptotischer Arbitrage der ersten/zweiten Art (im Sinne von Kabanov/Kramkov) äquivalent zur Absolutstetigkeit des Referenzmaßes zu einem eindeutigen, lokal äquivalenten Martingalmaß ist. Hat X stetige Pfade, so ist die Abwesenheit von allgemeiner asymptotischer Arbitrage äquivalent zur Existenz eines äquivalenten lokalen Martingalmaßes. Außerdem geben wir ein Kriterium für die Existenz einer optionalen Zerlegung von X an. Dies wenden wir auf das Problem der Risikominimierung bei vorgegebener Investitionsobergrenze (effizientes Hedgen (Föllmer/Leukert)) an, um dieses im unendlichdimensionalen Kontext zu behandeln. Außerdem stellen wir eine unendlichdimensionale Erweiterung des Heath-Jarrow-Morton-Modells vor und nutzen den Potentialansatz nach Rodgers, um zwei weitere Zinsstrukturmodelle zu konstruieren. Als Beitrag zur allgemeinen stochastischen Analysis in Hilberträumen beweisen wir eine pfadweise Version der Itoformel für stochastische Prozesse mit stetigen Pfaden in einem separablen Hilbertraum. Daraus läßt sich eine pfadweise Version des Satzes über die Vertauschbarkeit von stochastischem und Lebesgue-Integral ableiten. Außerdem zeigen wir eine Version der Clark-Formel für eine Brownsche Bewegung mit Werten in einem Hilbertraum. / We model a financial market with infinitely many assets as a stochastic process X with values in a separable Hilbert space H. In this setting we show the equivalence of market completeness and the uniqueness of the equivalent martingale measure, if X has continuous paths. Another result for our model is, that under some technical conditions, the absence of asymptotic arbitrage of the first/second kind (in the sense of Kabanov/Kramkov) is equivalent to the absolute continuity of the reference measure to a unique, locally equivalent, martingale measure. If X has continuous paths, the absence of general asymptotic arbitrage is equivalent to the existence of an equivalent local martingale measure. Furthermore, we give a sufficient condition for the existence of the optional decomposition of X. We apply this result to the problem of risk minimization with given upper limit for investion (efficient hedging (Föllmer/Leukert)). This allows us to solve this optimization problem in our infinite dimensional context. Another result is an infinite dimensional extension of the Heath-Jarrow-Morton term structure model. Two further term structure models are constructed, using the Markov potential approach developed by Rodgers. As a contribution to the theory of stochastic analysis in Hilbert spaces, we proof a pathwise version of the Ito formula for stochastic processes with continuous paths in a separable Hilbert space. This leads to a pathwise version of the interchangability theorem for stochastic and Lebesgue integrals. We also show a version of the Clark formula for Hilbert space valued Brownian motion.
|
197 |
Os fatores determinantes para a eficiência econômica dos produtores de frango de corte: uma análise estocástica. / Os fatores determinantes para a eficiência econômica dos produtores de frango de corte: uma análise estocástica.Zilli, Julcemar Bruno 16 January 2004 (has links)
A produção de frango de corte tem impressionado pelo dinamismo e pela competência conquistada nas últimas décadas, destacando-se o Brasil como o segundo maior produtor dessa proteína animal. O ganho de produtividade, associado à coordenação da cadeia avícola, colocou o País como um dos mais eficientes produtores. Entretanto, a significativa especialização da atividade tende a excluir do processo produtivo os pequenos avicultores e os produtores menos eficientes. Assim, o estudo buscou medir a eficiência econômica dos produtores de frango de corte das regiões Sul e Centro-Oeste do Brasil, identificando quais fatores influenciam essa medida de desempenho. Para isso, usou-se uma função fronteira de lucro estocástica em um estágio (modelo 2) em que os coeficientes da fronteira e os efeitos da ineficiência são obtidos simultaneamente, assumindo que os termos de erro não são identicamente distribuídos. Concluiu-se que para a região Sul, o preço da mão-de-obra contratada interfere significativamente na lucratividade das unidades produtivas, o que seria um dos fatores associados ao maior uso do trabalho familiar no desenvolvimento das atividades. Além disso, os resultados sugerem a presença de uma melhor utilização das áreas ocupadas com a produção avícola. Os efeitos da ineficiência são sentidos principalmente no baixo nível de educação dos que tomam as decisões e nos índices elevados de conversão alimentar. Já no Centro-Oeste, os coeficientes indicaram que o maior uso de mão-de-obra familiar poderia elevar a lucratividade dos produtores. Pelo fato de ser uma região relativamente nova e possuir condições favoráveis ao investimento em capital e tecnologia, estaria indicando um maior lucro na atividade. Não se identificou ganho de escala por meio do modelo estocástico. Embora não se observa uma tendência contínua associada à escala de eficiência, no Centro-Oeste parece existirem ganhos de eficiência relevante nos estratos de médio e alto escala de produção para os padrões regionais. / The dynamism and ability acquired through the last decade by the broiler production is very impressive. The productivity rate gain associated to the good management of poultry chain in Brazil led the country to be the second biggest producer of this animals protein. However, the significant specialization of this activity tend to exclude the smaller producers and those who are lesser efficient in the productive process. In view of that, this study intends to precise the economic efficiency of the broiler producers in the Southerner and Center-Southerner regions of Brazil. For that, the main factors that influence the economic efficiency were identified. To reach those results, it was considered the stochastic profit function in a stage (model 2) where the coefficients of the frontier and the inefficiency effects are obtained simultaneously, since the terms of error are not identical distributing. It was possible to conclude that in the Southerner region of Brazil, the contracted labor force influence significantly the profitability of the productive farms. Thats one of the factors that explain the great use of the familiar labor force in that region. Moreover, the results pointed a better utilization of the broiler producing areas in Brazil. The effects of the inefficiency are present mainly in the producers and players with low educational level, and also in the high food conversion ratio. In the Center-Westerner of Brazil, the coefficients pointed that the better use of the familiar manpower could raise the probability of the producers. For been a new region that present positive conditions for the investments in capital and technology, it also obtained a larger profit with the activity. It was not identified gains by the stochastic model. None successively trend associated to the efficiency scale were pointed in the Center-Westerner of Brazil. Even so, the region shows efficiency gains in the middle and high scales of production in regional terms.
|
198 |
Contributions to functional inequalities and limit theorems on the configuration space / Inégalités fonctionnelles et théorèmes limites sur l'espace des configurationsHerry, Ronan 03 December 2018 (has links)
Nous présentons des inégalités fonctionnelles pour les processus ponctuels. Nous prouvons une inégalité de Sobolev logarithmique modifiée, une inégalité de Stein et un théorème du moment quatrième sans terme de reste pour une classe de processus ponctuels qui contient les processus binomiaux et les processus de Poisson. Les preuves reposent sur des techniques inspirées de l'approche de Malliavin-Stein et du calcul avec l'opérateur $Gamma$ de Bakry-Émery. Pour mettre en œuvre ces techniques nous développons une analyse stochastique pour les processus ponctuels. Plus généralement, nous mettons au point une théorie d'analyse stochastique sans hypothèse de diffusion. Dans le cadre des processus de Poisson ponctuels, l'inégalité de Stein est généralisée pour étudier la convergence stable vers des limites conditionnellement gaussiennes. Nous appliquons ces résultats pour approcher des processus Gaussiens par des processus de Poisson composés et pour étudier des graphes aléatoires. Nous discutons d'inégalités de transport et de leur conséquence en termes de concentration de la mesure pour les processus binomiaux dont la taille de l'échantillon est aléatoire. Sur un espace métrique mesuré quelconque, nous présentons un développement de la concentration de la mesure qui prend en compte l'agrandissement parallèle d'ensembles disjoints. Cette concentration améliorée donne un contrôle de toutes les valeurs propres du Laplacien métrique. Nous discutons des liens de cette nouvelle notion avec une version de la courbure de Ricci qui fait intervenir le transport à plusieurs marginales / We present functional inequalities and limit theorems for point processes. We prove a modified logarithmic Sobolev inequalities, a Stein inequality and a exact fourth moment theorem for a large class of point processes including mixed binomial processes and Poisson point processes. The proofs of these inequalities are inspired by the Malliavin-Stein approach and the $Gamma$-calculus of Bakry-Emery. The implementation of these techniques requires a development of a stochastic analysis for point processes. As point processes are essentially discrete, we design a theory to study non-diffusive random objects. For Poisson point processes, we extend the Stein inequality to study stable convergence with respect to limits that are conditionally Gaussian. Applications to Poisson approximations of Gaussian processes and random geometry are given. We discuss transport inequalities for mixed binomial processes and their consequences in terms of concentration of measure. On a generic metric measured space, we present a refinement of the notion of concentration of measure that takes into account the parallel enlargement of distinct sets. We link this notion of improved concentration with the eigenvalues of the metric Laplacian and with a version of the Ricci curvature based on multi-marginal optimal transport
|
199 |
Stochastic mapping for chemical plume source localization with application to autonomous hydrothermal vent discoveryJakuba, Michael Vavrousek, 1976- January 2006 (has links)
Thesis (Ph. D.)--Joint Program in Oceanography/Applied Ocean Science and Engineering (Massachusetts Institute of Technology, Dept. of Mechanical Engineering; and the Woods Hole Oceanographic Institution), 2007. / Includes bibliographical references (p. 313-325). / This thesis presents a stochastic mapping framework for autonomous robotic chemical plume source localization in environments with multiple sources. Potential applications for robotic chemical plume source localization include pollution and environmental monitoring, chemical plant safety, search and rescue, anti-terrorism, narcotics control, explosive ordinance removal, and hydrothermal vent prospecting. Turbulent flows make the spatial relationship between the detectable manifestation of a chemical plume source, the plume itself, and the location of its source inherently uncertain. Search domains with multiple sources compound this uncertainty because the number of sources as well as their locations is unknown a priori. Our framework for stochastic mapping is an adaptation of occupancy grid mapping where the binary state of map nodes is redefined to denote either the presence (occupancy) or absence of an active plume source. A key characteristic of the chemical plume source localization problem is that only a few sources are expected in the search domain. The occupancy grid framework allows for both plume detections and non-detections to inform the estimated state of grid nodes in the map, thereby explicitly representing explored but empty portions of the domain as well as probable source locations. / (cont.) However, sparsity in the expected number of occupied grid nodes strongly violates a critical conditional independence assumption required by the standard Bayesian recursive map update rule. While that assumption makes for a computationally attractive algorithm, in our application it results in occupancy grid maps that are grossly inconsistent with the assumption of a small number of occupied cells. To overcome this limitation, several alternative occupancy grid update algorithms are presented, including an exact solution that is computationally tractable for small numbers of detections and an approximate recursive algorithm with improved performance relative to the standard algorithm but equivalent computational cost. Application to hydrothermal plume data collected by the autonomous underwater vehicle ABE during vent prospecting operations in both the Pacific and Atlantic oceans verifies the utility of the approach. The resulting maps enable nested surveys for homing-in on seafloor vent sites to be carried out autonomously. This eliminates inter-dive processing, recharging of batteries, and time spent deploying and recovering the vehicle that would otherwise be necessary with survey design directed by human operators. / by Michael V. Jakuba. / Ph.D.
|
200 |
Data analysis and preliminary model development for an odour detection system based on the behaviour of trained waspsZhou, Zhongkun January 2008 (has links)
Microplitis croceipes, one of the nectar feeding parasitoid wasps, has been found to associatively learn chemical cues through feeding. The experiments on M. croceipes are performed and recorded by a Sony camcorder in the USDA-ARS Biological Control Laboratory in Tifton, GA, USA. The experimental videos have shown that M. croceipes can respond to Coffee odour in this study. Their detection capabilities and the behaviour of M. croceipes with different levels of coffee odours were studied. First, the data that are related to trained M. croceipes behaviour was extracted from the experimental videos and stored in a Microsoft Excel database. The extracted data represent the behaviour of M. croceipes trained to 0.02g and then exposed to 0.001g, 0.005g, 0.01g, 0.02g and 0.04g of coffee. Secondly, indices were developed to uniquely characterise the behaviour of trained M. croceipes under different coffee concentrations. Thirdly, a preliminary model and its parameters were developed to classify the response of trained wasps when exposed to these five different coffee odours. In summary, the success of this thesis demonstrates the usefulness of data analysis for interpreting experimental data, developing indices, as well as understanding the design principles of a simple model based on trained wasps.
|
Page generated in 0.0824 seconds