• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 295
  • 106
  • 91
  • 76
  • 36
  • 21
  • 12
  • 8
  • 7
  • 5
  • 5
  • 5
  • 4
  • 3
  • 2
  • Tagged with
  • 826
  • 826
  • 134
  • 122
  • 119
  • 107
  • 99
  • 95
  • 93
  • 85
  • 77
  • 77
  • 66
  • 62
  • 62
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
1

Sensitivity analysis research of Enterprise accounts receivable

Shih, Tsai- Hsien 21 August 2001 (has links)
none
2

Statistical methods for the analysis of DSMC simulations of hypersonic shocks

Strand, James Stephen 25 June 2012 (has links)
In this work, statistical techniques were employed to study the modeling of a hypersonic shock with the Direct Simulation Monte Carlo (DSMC) method, and to gain insight into how the model interacts with a set of physical parameters. Direct Simulation Monte Carlo (DSMC) is a particle based method which is useful for simulating gas dynamics in rarefied and/or highly non-equilibrium flowfields. A DSMC code was written and optimized for use in this research. The code was developed with shock tube simulations in mind, and it includes a number of improvements which allow for the efficient simulation of 1D, hypersonic shocks. Most importantly, a moving sampling region is used to obtain an accurate steady shock profile from an unsteady, moving shock wave. The code is MPI parallel and an adaptive load balancing scheme ensures that the workload is distributed properly between processors over the course of a simulation. Global, Monte Carlo based sensitivity analyses were performed in order to determine which of the parameters examined in this work most strongly affect the simulation results for two scenarios: a 0D relaxation from an initial high temperature state and a hypersonic shock. The 0D relaxation scenario was included in order to examine whether, with appropriate initial conditions, it can be viewed in some regards as a substitute for the 1D shock in a statistical sensitivity analysis. In both analyses sensitivities were calculated based on both the square of the Pearson correlation coefficient and the mutual information. The quantity of interest (QoI) chosen for these analyses was the NO density profile. This vector QoI was broken into a set of scalar QoIs, each representing the density of NO at a specific point in time (for the relaxation) or a specific streamwise location (for the shock), and sensitivities were calculated for each scalar QoI based on both measures of sensitivity. The sensitivities were then integrated over the set of scalar QoIs to determine an overall sensitivity for each parameter. A weighting function was used in the integration in order to emphasize sensitivities in the region of greatest thermal and chemical non-equilibrium. The six parameters which most strongly affect the NO density profile were found to be the same for both scenarios, which provides justification for the claim that a 0D relaxation can in some situations be used as a substitute model for a hypersonic shock. These six parameters are the pre-exponential constants in the Arrhenius rate equations for the N2 dissociation reaction N2 + N ⇄ 3N, the O2 dissociation reaction O2 + O ⇄ 3O, the NO dissociation reactions NO + N ⇄ 2N + O and NO + O ⇄ N + 2O, and the exchange reactions N2 + O ⇄ NO + N and NO + O ⇄ O2 + N. After identification of the most sensitive parameters, a synthetic data calibration was performed to demonstrate that the statistical inverse problem could be solved for the 0D relaxation scenario. The calibration was performed using the QUESO code, developed at the PECOS center at UT Austin, which employs the Delayed Rejection Adaptive Metropolis (DRAM) algorithm. The six parameters identified by the sensitivity analysis were calibrated successfully with respect to a group of synthetic datasets. / text
3

Quantification of Uncertainties Due to Opacities in a Laser-Driven Radiative-Shock Problem

Hetzler, Adam C 03 October 2013 (has links)
This research presents new physics-based methods to estimate predictive uncertainty stemming from uncertainty in the material opacities in radiative transfer computations of key quantities of interest (QOIs). New methods are needed because it is infeasible to apply standard uncertainty-propagation techniques to the O(105) uncertain opacities in a realistic simulation. The new approach toward uncertainty quantification applies the uncertainty analysis to the physical parameters in the underlying model used to calculate the opacities. This set of uncertain parameters is much smaller (O(102)) than the number of opacities. To further reduce the dimension of the set of parameters to be rigorously explored, we use additional screening applied at two different levels of the calculational hierarchy: first, physics-based screening eliminates the physical parameters that are unimportant from underlying physics models a priori; then, sensitivity analysis in simplified versions of the complex problem of interest screens out parameters that are not important to the QOIs. We employ a Bayesian Multivariate Adaptive Regression Spline (BMARS) emulator for this sensitivity analysis. The high dimension of the input space and large number of samples test the efficacy of these methods on larger problems. Ultimately, we want to perform uncertainty quantification on the large, complex problem with the reduced set of parameters. Results of this research demonstrate that the QOIs for target problems agree at for different parameter screening criteria and varying sample sizes. Since the QOIs agree, we have gained confidence in our results using the multiple screening criteria and sample sizes.
4

Sensitivity Enhanced Model Reduction

Munster, Drayton William 06 June 2013 (has links)
In this study, we numerically explore methods of coupling sensitivity analysis to the reduced model in order to increase the accuracy of a proper orthogonal decomposition (POD) basis across a wider range of parameters. Various techniques based on polynomial interpolation and basis alteration are compared. These techniques are performed on a 1-dimensional reaction-diffusion equation and 2-dimensional incompressible Navier-Stokes equations solved using the finite element method (FEM) as the full scale model. The expanded model formed by expanding the POD basis with the orthonormalized basis sensitivity vectors achieves the best mixture of accuracy and computational efficiency among the methods compared. / Master of Science
5

Investigations on Stabilized Sensitivity Analysis of Chaotic Systems

Taoudi, Lamiae 03 May 2019 (has links)
Many important engineering phenomena such as turbulent flow, fluid-structure interactions, and climate diagnostics are chaotic and sensitivity analysis of such systems is a challenging problem. Computational methods have been proposed to accurately and efficiently estimate the sensitivity analysis of these systems which is of great scientific and engineering interest. In this thesis, a new approach is applied to compute the direct and adjoint sensitivities of time-averaged quantities defined from the chaotic response of the Lorenz system and the double pendulum system. A stabilized time-integrator with adaptive time-step control is used to maintain stability of the sensitivity calculations. A study of convergence of a quantity of interest and its square is presented. Results show that the approach computes accurate sensitivity values with a computational cost that is multiple orders-of-magnitude lower than competing approaches based on least-squares-shadowing approach.
6

A Study of Predicted Energy Savings and Sensitivity Analysis

Yang, Ying 16 December 2013 (has links)
The sensitivity of the important inputs and the savings prediction function reliability for the WinAM 4.3 software is studied in this research. WinAM was developed by the Continuous Commissioning (CC) group in the Energy Systems Laboratory at Texas A&M University. For the sensitivity analysis task, fourteen inputs are studied by adjusting one input at a time within ± 30% compared with its baseline. The Single Duct Variable Air Volume (SDVAV) system with and without the economizer has been applied to the square zone model. Mean Bias Error (MBE) and Influence Coefficient (IC) have been selected as the statistical methods to analyze the outputs that are obtained from WinAM 4.3. For the saving prediction reliability analysis task, eleven Continuous Commissioning projects have been selected. After reviewing each project, seven of the eleven have been chosen. The measured energy consumption data for the seven projects is compared with the simulated energy consumption data that has been obtained from WinAM 4.3. Normalization Mean Bias Error (NMBE) and Coefficient of Variation of the Root Mean Squared Error (CV (RMSE)) statistical methods have been used to analyze the results from real measured data and simulated data. Highly sensitive parameters for each energy resource of the system with the economizer and the system without the economizer have been generated in the sensitivity analysis task. The main result of the savings prediction reliability analysis is that calibration improves the model’s quality. It also improves the predicted energy savings results compared with the results generated from the uncalibrated model.
7

Sensitivity Analysis of the Economic Lot-Sizing Problem

Van Hoesel, Stan, Wagelmans, Albert 11 1900 (has links)
In this paper we study sensitivity analysis of the uncapacitated single level economic lot-sizing problem, which was introduced by Wagner and Whitin about thirty years ago. In particular we are concerned with the computation of the maximal ranges in which the numerical problem parameters may vary individually, such that a solution already obtained remains optimal. Only recently it was discovered that faster algorithms than the Wagner-Whitin algorithm exist to solve the economic lot-sizing problem. Moreover, these algorithms reveal that the problem has more structure than was recognized so far. When performing the sensitivity analysis we exploit these newly obtained insights.
8

Ensaio imunoradiometrico ultra-sensivel de tireotrofina humana (hTsH) obtido mediante a identificacao e minimizacao de ligacoes inespecificas

PERONI, CIBELE N. 09 October 2014 (has links)
Made available in DSpace on 2014-10-09T12:38:00Z (GMT). No. of bitstreams: 0 / Made available in DSpace on 2014-10-09T14:04:41Z (GMT). No. of bitstreams: 1 05581.pdf: 1858679 bytes, checksum: 40e224a27b1e68838662dfa34b14949f (MD5) / Dissertacao (Mestrado) / IPEN/D / Instituto de Pesquisas Energeticas e Nucleares - IPEN/CNEN-SP
9

Ensaio imunoradiometrico ultra-sensivel de tireotrofina humana (hTsH) obtido mediante a identificacao e minimizacao de ligacoes inespecificas

PERONI, CIBELE N. 09 October 2014 (has links)
Made available in DSpace on 2014-10-09T12:38:00Z (GMT). No. of bitstreams: 0 / Made available in DSpace on 2014-10-09T14:04:41Z (GMT). No. of bitstreams: 1 05581.pdf: 1858679 bytes, checksum: 40e224a27b1e68838662dfa34b14949f (MD5) / Dissertacao (Mestrado) / IPEN/D / Instituto de Pesquisas Energeticas e Nucleares - IPEN/CNEN-SP
10

FICST: A Tool for Sensitivity Analysis of SCWR Fuel Isotopic Composition to Nuclear Data

Mostofian, Sara January 2014 (has links)
With an ever-increasing population both in Canada and globally, an improved quality of life will depend on having access to energy. The non-renewable, carbon-based, sources of energy that presently provide a major amount of the world's energy supply are depleting and therefore will be expensive in the future. Nuclear technology is a relatively new technology which can fulfill future energy needs but requires highly specialized skills and knowledge to continue to make it safer, cleaner, more reliable, and more affordable. Thus the nuclear industry puts lots of efforts to develop and improve the next generation of nuclear power plants. The Supercritical Water Reactors (SCWRs) are one of the Generation IV nuclear-reactor systems. The SCWRs, to a large extent, are very similar to light water reactors, but with a simpler design. The main advantage of SCWRs is their higher thermal efficiency. The Canadian SCWR has adopted an innovative fuel concept which is a mixture of plutonium and thorium oxides (Th, Pu) O2. The role of nuclear data in fuel development and reactor-physics analysis is quite significant. With the development of nuclear data files over the years, nuclear cross sections and other parameters are widely available, but their accuracy is still a concern. Also the accuracy of nuclear data is more reliable for uranium-based fuels than for thorium-based fuels. It is not known how the uncertainties in the nuclear data will impact the fuel depletion in a SCWR. Thus a sensitivity analysis tool has been developed to evaluate the impact of uncertainties in the neutron cross-sections of the actinides present in SCWR fuel. This document provides the details on the theory and methodology used to develop this tool (FICST). The objective of this work is to develop a code, not any specific calculation done with it. / Thesis / Master of Applied Science (MASc)

Page generated in 0.1663 seconds