• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 88
  • 15
  • 15
  • 7
  • 2
  • 1
  • 1
  • 1
  • Tagged with
  • 155
  • 155
  • 70
  • 31
  • 28
  • 24
  • 23
  • 22
  • 17
  • 17
  • 15
  • 15
  • 14
  • 14
  • 14
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
91

TESTING FOR DIFFERENTIALLY EXPRESSED GENES AND KEY BIOLOGICAL CATEGORIES IN DNA MICROARRAY ANALYSIS

SARTOR, MAUREEN A. January 2007 (has links)
No description available.
92

Coupling Physical and Machine Learning Models with High Resolution Information Transfer and  Rapid Update Frameworks for Environmental Applications

Sommerlot, Andrew Richard 13 December 2017 (has links)
Few current modeling tools are designed to predict short-term, high-risk runoff from critical source areas (CSAs) in watersheds which are significant sources of non point source (NPS) pollution. This study couples the Soil and Water Assessment Tool-Variable Source Area (SWAT-VSA) model with the Climate Forecast System Reanalysis (CFSR) model and the Global Forecast System (GFS) model short-term weather forecast, to develop a CSA prediction tool designed to assist producers, landowners, and planners in identifying high-risk areas generating storm runoff and pollution. Short-term predictions for streamflow, runoff probability, and soil moisture levels were estimated in the South Fork of the Shenandoah river watershed in Virginia. In order to allow land managers access to the CSA predictions a free and open source software based web was developed. The forecast system consists of three primary components; (1) the model, which preprocesses the necessary hydrologic forcings, runs the watershed model, and outputs spatially distributed VSA forecasts; (2) a data management structure, which converts high resolution rasters into overlay web map tiles; and (3) the user interface component, a web page that allows the user, to interact with the processed output. The resulting framework satisfied most design requirements with free and open source software and scored better than similar tools in usability metrics. One of the potential problems is that the CSA model, utilizing physically based modeling techniques requires significant computational time to execute and process. Thus, as an alternative, a deep learning (DL) model was developed and trained on the process based model output. The DL model resulted in a 9% increase in predictive power compared to the physically based model and a ten-fold decrease in run time. Additionally, DL interpretation methods applicable beyond this study are described including hidden layer visualization and equation extractions describing a quantifiable amount of variance in hidden layer values. Finally, a large-scale analysis of soil phosphorus (P) levels was conducted in the Chesapeake Bay watershed, a current location of several short-term forecast tools. Based on Bayesian inference methodologies, 31 years of soil P history at the county scale were estimated, with the associated uncertainty for each estimate. These data will assist in the planning and implantation of short term forecast tools with P management goals. The short term modeling and communication tools developed in this work contribute to filling a gap in scientific tools aimed at improving water quality through informing land manager's decisions. / PHD
93

Bayesian Variable Selection with Shrinkage Priors and Generative Adversarial Networks for Fraud Detection

Issoufou Anaroua, Amina 01 January 2024 (has links) (PDF)
This research paper focuses on fraud detection in the financial industry using Generative Adversarial Networks (GANs) in conjunction with Uni and Multi Variate Bayesian Model with Shrinkage Priors (BMSP). The problem addressed is the need for accurate and advanced fraud detection techniques due to the increasing sophistication of fraudulent activities. The methodology involves the implementation of GANs and the application of BMSP for variable selection to generate synthetic fraud samples for fraud detection using the augmented dataset. Experimental results demonstrate the effectiveness of the BMSP GAN approach in detecting fraud with improved performance compared to other methods. The conclusions drawn highlight the potential of GANs and BMSP for enhancing fraud detection capabilities and suggest future research directions for further improvements in the field.
94

A Comprehensive Approach to Posterior Jointness Analysis in Bayesian Model Averaging Applications

Crespo Cuaresma, Jesus, Grün, Bettina, Hofmarcher, Paul, Humer, Stefan, Moser, Mathias 03 1900 (has links) (PDF)
Posterior analysis in Bayesian model averaging (BMA) applications often includes the assessment of measures of jointness (joint inclusion) across covariates. We link the discussion of jointness measures in the econometric literature to the literature on association rules in data mining exercises. We analyze a group of alternative jointness measures that include those proposed in the BMA literature and several others put forward in the field of data mining. The way these measures address the joint exclusion of covariates appears particularly important in terms of the conclusions that can be drawn from them. Using a dataset of economic growth determinants, we assess how the measurement of jointness in BMA can affect inference about the structure of bivariate inclusion patterns across covariates. (authors' abstract) / Series: Department of Economics Working Paper Series
95

Validation des modèles statistiques tenant compte des variables dépendantes du temps en prévention primaire des maladies cérébrovasculaires

Kis, Loredana 07 1900 (has links)
L’intérêt principal de cette recherche porte sur la validation d’une méthode statistique en pharmaco-épidémiologie. Plus précisément, nous allons comparer les résultats d’une étude précédente réalisée avec un devis cas-témoins niché dans la cohorte utilisé pour tenir compte de l’exposition moyenne au traitement : – aux résultats obtenus dans un devis cohorte, en utilisant la variable exposition variant dans le temps, sans faire d’ajustement pour le temps passé depuis l’exposition ; – aux résultats obtenus en utilisant l’exposition cumulative pondérée par le passé récent ; – aux résultats obtenus selon la méthode bayésienne. Les covariables seront estimées par l’approche classique ainsi qu’en utilisant l’approche non paramétrique bayésienne. Pour la deuxième le moyennage bayésien des modèles sera utilisé pour modéliser l’incertitude face au choix des modèles. La technique utilisée dans l’approche bayésienne a été proposée en 1997 mais selon notre connaissance elle n’a pas été utilisée avec une variable dépendante du temps. Afin de modéliser l’effet cumulatif de l’exposition variant dans le temps, dans l’approche classique la fonction assignant les poids selon le passé récent sera estimée en utilisant des splines de régression. Afin de pouvoir comparer les résultats avec une étude précédemment réalisée, une cohorte de personnes ayant un diagnostique d’hypertension sera construite en utilisant les bases des données de la RAMQ et de Med-Echo. Le modèle de Cox incluant deux variables qui varient dans le temps sera utilisé. Les variables qui varient dans le temps considérées dans ce mémoire sont iv la variable dépendante (premier évènement cérébrovasculaire) et une des variables indépendantes, notamment l’exposition / The main interest of this research is the validation of a statistical method in pharmacoepidemiology. Specifically, we will compare the results of a previous study performed with a nested case-control which took into account the average exposure to treatment to : – results obtained in a cohort study, using the time-dependent exposure, with no adjustment for time since exposure ; – results obtained using the cumulative exposure weighted by the recent past ; – results obtained using the Bayesian model averaging. Covariates are estimated by the classical approach and by using a nonparametric Bayesian approach. In the later, the Bayesian model averaging will be used to model the uncertainty in the choice of models. To model the cumulative effect of exposure which varies over time, in the classical approach the function assigning weights according to recency will be estimated using regression splines. In order to compare the results with previous studies, a cohort of people diagnosed with hypertension will be constructed using the databases of the RAMQ and Med-Echo. The Cox model including two variables which vary in time will be used. The time-dependent variables considered in this paper are the dependent variable (first stroke event) and one of the independent variables, namely the exposure.
96

Bayesian and Frequentist Approaches for the Analysis of Multiple Endpoints Data Resulting from Exposure to Multiple Health Stressors.

Nyirabahizi, Epiphanie 08 March 2010 (has links)
In risk analysis, Benchmark dose (BMD)methodology is used to quantify the risk associated with exposure to stressors such as environmental chemicals. It consists of fitting a mathematical model to the exposure data and the BMD is the dose expected to result in a pre-specified response or benchmark response (BMR). Most available exposure data are from single chemical exposure, but living objects are exposed to multiple sources of hazards. Furthermore, in some studies, researchers may observe multiple endpoints on one subject. Statistical approaches to address multiple endpoints problem can be partitioned into a dimension reduction group and a dimension preservative group. Composite scores using desirability function is used, as a dimension reduction method, to evaluate neurotoxicity effects of a mixture of five organophosphate pesticides (OP) at a fixed mixing ratio ray, and five endpoints were observed. Then, a Bayesian hierarchical model approach, as a single unifying dimension preservative method is introduced to evaluate the risk associated with the exposure to mixtures chemicals. At a pre-specied vector of BMR of interest, the method estimates a tolerable area referred to as benchmark dose tolerable area (BMDTA) in multidimensional Euclidean plan. Endpoints defining the BMDTA are determined and model uncertainty and model selection problems are addressed by using the Bayesian Model Averaging (BMA) method.
97

Důchodová elasticita poptávky po vodě: Meta-analýza / Income Elasticity of Water Demand: A Meta-Analysis

Vlach, Tomáš January 2016 (has links)
If policymakers address water scarcity with the demand-oriented approach, the income elasticity of water demand is of pivotal importance. Its estimates, however, differ considerably. We collect 307 estimates of the income elasticity of water demand reported in 62 studies, codify 31 variables describing the estimation design, and employ Bayesian model averaging to address model uncertainty inherent to any meta-analysis. The studies were published between 1972 and 2015, which means that this meta-analysis covers a longer period of time than two previous meta-analyses on this topic combined. Our results suggest that income elasticity estimates for developed countries do not significantly differ from income elasticity estimates for developing countries and that different estimation techniques do not systematically produce different values of the income elasticity of water demand. We find evidence of publication selection bias in the literature on the income elasticity of water demand with the use of both graphical and regression analysis. We correct the estimates for publication selection bias and estimate the true effect beyond bias, which reaches approximately 0.2. 1
98

Ohodnocování a predikce systémového rizika: Systém včasného varovaní navržený pro Českou republiku / Systemic Risks Assessment and Systemic Events Prediction: Early Warning System Design for the Czech Republic

Žigraiová, Diana January 2013 (has links)
This thesis develops an early warning system framework for assessing systemic risks and for predicting systemic events, i.e. periods of extreme financial instability with potential real costs, over the short horizon of six quarters and the long horizon of twelve quarters on the panel of 14 countries both advanced and developing. Firstly, Financial Stress Index is built aggregating indicators from equity, foreign exchange, security and money markets in order to identify starting dates of systemic financial crises for each country in the panel. Secondly, the selection of early warning indicators for assessment and prediction of systemic risks is undertaken in a two- step approach; relevant prediction horizons for each indicator are found by means of a univariate logit model followed by the application of Bayesian model averaging method to identify the most useful indicators. Next, logit models containing useful indicators only are estimated on the panel while their in-sample and out-of-sample performance is assessed by a variety of measures. Finally, having applied the constructed EWS for both horizons to the Czech Republic it was found that even though models for both horizons perform very well in-sample, i.e. both predict 100% of crises, only the long model attains the maximum utility of 0,5 as...
99

Cost-benefit analysis of mitigation of outages caused by squirrels on the overhead electricity distribution systems

Malve, Priyanka January 1900 (has links)
Master of Science / Department of Electrical and Computer Engineering / Anil Pahwa / Unpredictable power outages due to environmental factors such as lighting, wind, trees, and animals, have always been a concern for utilities because they are often unavoidable. This research aims to study squirrel-related outages by modeling past real-life outage data and provide the optimal result which would assist utilities in increasing electric system reliability. This research is a novel approach to benchmark system performance in order to identify areas and durations with higher than expected outages. The model is illustrated with seven years (2005-2011) of animal-related outage data and 14 years of weather data (1998-2011) for four cities in Kansas, used as training data to predict future outages. The past data indicates that the number of outages on any day varies with the seasons and weather conditions on that day. The prediction is based on a Bayesian Model using conditional probability table, which is calculated based on training data. Since future weather conditions are unknown and random, Monte Carlo Simulation is used with the past 14 years of weather data to create different yearly scenarios. These scenarios are then used with the models to predict expected outages. Multiple runs of Monte Carlo analysis provide a probability distribution of expected outages. Further work discusses about cost-to-benefit analysis of implementation of outage mitigation methods. The analysis is performed by considering different combinations of outage reduction and mitigation levels. In this research, eight cases of outage reduction and nine cases of mitigation levels are defined. The probability of benefit is calculated by a statistical approach for every combination. Several optimal strategies are constructed using the probability values and outage history. The outcomes are compared with each other to propose the most beneficial outage mitigation strategy. This research will immensely assist utilities in reducing the outages due to squirrels more effectively with higher benefits and therefore improve reliability of the electricity supply to consumers.
100

A Hierarchical Bayesian Model for the Unmixing Analysis of Compositional Data subject to Unit-sum Constraints

Yu, Shiyong 15 May 2015 (has links)
Modeling of compositional data is emerging as an active area in statistics. It is assumed that compositional data represent the convex linear mixing of definite numbers of independent sources usually referred to as end members. A generic problem in practice is to appropriately separate the end members and quantify their fractions from compositional data subject to nonnegative and unit-sum constraints. A number of methods essentially related to polytope expansion have been proposed. However, these deterministic methods have some potential problems. In this study, a hierarchical Bayesian model was formulated, and the algorithms were coded in MATLABÒ. A test run using both a synthetic and real-word dataset yields scientifically sound and mathematically optimal outputs broadly consistent with other non-Bayesian methods. Also, the sensitivity of this model to the choice of different priors and structure of the covariance matrix of error were discussed.

Page generated in 0.0511 seconds