Spelling suggestions: "subject:"[een] STATISTICAL METHODS"" "subject:"[enn] STATISTICAL METHODS""
201 |
STATISTICAL MODELING OF SHIP AIRWAKES INCLUDING THE FEASIBILITY OF APPLYING MACHINE LEARNINGUnknown Date (has links)
Airwakes are shed behind the ship’s superstructure and represent a highly turbulent and rapidly distorting flow field. This flow field severely affects pilot’s workload and such helicopter shipboard operations. It requires both the one-point statistics of autospectrum and the two-point statistics of coherence (normalized cross-spectrum) for a relatively complete description. Recent advances primarily refer to generating databases of flow velocity points through experimental and computational fluid dynamics (CFD) investigations, numerically computing autospectra along with a few cases of cross-spectra and coherences, and developing a framework for extracting interpretive models of autospectra in closed form from a database along with an application of this framework to study the downwash effects. By comparison, relatively little is known about coherences. In fact, even the basic expressions of cross-spectra and coherences for three components of homogeneous isotropic turbulence (HIT) vary from one study to the other, and the related literature is scattered and piecemeal. Accordingly, this dissertation begins with a unified account of all the cross-spectra and coherences of HIT from first principles. Then, it presents a framework for constructing interpretive coherence models of airwake from a database on the basis of perturbation theory. For each velocity component, the coherence is represented by a separate perturbation series in which the basis function or the first term on the right-hand side of the series is represented by the corresponding coherence for HIT. The perturbation series coefficients are evaluated by satisfying the theoretical constraints and fitting a curve in a least squares sense on a set of numerically generated coherence points from a database. Although not tested against a specific database, the framework has a mathematical basis. Moreover, for assumed values of perturbation series constants, coherence results are presented to demonstrate how coherences of airwakes and such flow fields compare to those of HIT. / Includes bibliography. / Dissertation (Ph.D.)--Florida Atlantic University, 2020. / FAU Electronic Theses and Dissertations Collection
|
202 |
Advances in Machine Learning for Compositional DataGordon Rodriguez, Elliott January 2022 (has links)
Compositional data refers to simplex-valued data, or equivalently, nonnegative vectors whose totals are uninformative. This data modality is of relevance across several scientific domains. A classical example of compositional data is the chemical composition of geological samples, e.g., major-oxide concentrations. A more modern example arises from the microbial populations recorded using high-throughput genetic sequencing technologies, e.g., the gut microbiome. This dissertation presents a set of methodological and theoretical contributions that advance the state of the art in the analysis of compositional data.
Our work can be divided along two categories: problems in which compositional data represents the input to a predictive model, and problems in which it represents the output of the model. For the first class of problems, we build on the popular log-ratio framework to develop an efficient learning algorithm for high-dimensional compositional data. Our algorithm runs orders of magnitude faster than competing alternatives, without sacrificing model quality. For the second class of problems, we define a novel exponential family of probability distributions supported on the simplex. This distribution enjoys attractive mathematical properties and provides a performant probability model for simplex-valued outcomes. Taken together, our results constitute a broad contribution to the toolkit of researchers and practitioners studying compositional data.
|
203 |
The estimation of missing values in hydrological records using the EM algorithm and regression methodsMakhuvha, Tondani January 1988 (has links)
Includes bibliography. / The objective of this thesis is to review existing methods for estimating missing values in rainfall records and to propose a number of new procedures. Two classes of methods are considered. The first is based on the theory of variable selection in regression. Here the emphasis is on finding efficient methods to identify the set of control stations which are likely to yield the best regression estimates of the missing values in the target station. The second class of methods is based on the EM algorithm, proposed by Dempster, Laird and Rubin (1977). The emphasis here is to estimate the missing values directly without first making a detailed selection of control stations. All "relevant" stations are included. This method has not previously been applied in the context of estimating missing rainfall values.
|
204 |
An empirical evaluation of the Altman (1968) failure prediction model on South African JSE listed companiesRama, Kavir D. 18 March 2013 (has links)
Credit has become very important in the global economy (Cynamon and Fazzari, 2008).
The Altman (1968) failure prediction model, or derivatives thereof, are often used in the
identification and selection of financially distressed companies as it is recognized as one
of the most reliable in predicting company failure (Eidleman, 1995). Failure of a firm can
cause substantial losses to creditors and shareholders, therefore it is important, to detect
company failure as early as possible. This research report empirically tests the Altman
(1968) failure prediction model on 227 South African JSE listed companies using data
from the 2008 financial year to calculate the Z-score within the model, and measuring
success or failure of firms in the 2009 and 2010 years. The results indicate that the
Altman (1968) model is a viable tool in predicting company failure for firms with positive
Z-scores, and where Z-scores do not fall into the range of uncertainty as specified. The
results also suggest that the model is not reliable when the Z–scores are negative or
when they are in the range of uncertainty (between 2.99 and 1.81). If one is able to
predict firm failure in advance, it should be possible for management to take steps to
avert such an occurrence (Deakin, 1972; Keasey and Watson, 1991; Platt and Platt,
2002).
|
205 |
Is the way forward to step back? A meta-research analysis of misalignment between goals, methods, and conclusions in epidemiologic studies.Kezios, Katrina Lynn January 2021 (has links)
Recent discussion in the epidemiologic methods and teaching literatures centers around the importance of clearly stating study goals, disentangling the goal of causation from prediction (or description), and clarifying the statistical tools that can address each goal. This discussion illuminates different ways in which mismatches can occur between study goals, methods, and interpretations, which this dissertation synthesizes into the concept of “misalignment”; misalignment occurs when the study methods and/or interpretations are inappropriate for (i.e., do not match) the study’s goal. While misalignments can occur and may cause problems, their pervasiveness and consequences have not been examined in the epidemiologic literature. Thus, the overall purpose of this dissertation was to document and examine the effects of misalignment problems seen in epidemiologic practice.
First, a review was conducted to document misalignment in a random sample of epidemiologic studies and explore how the framing of study goals contributes to its occurrence. Among the reviewed articles, full alignment between study goals, methods, and interpretations was infrequently observed, although “clearly causal” studies (those that framed causal goals using causal language) were more often fully aligned (5/13, 38%) than “seemingly causal” ones (those that framed causal goals using associational language; 3/71, 4%).
Next, two simulation studies were performed to examine the potential consequences of different types of misalignment problems seen in epidemiologic practice. They are based on the observation that, often, studies that are causally motivated perform analyses that appear disconnected from, or “misaligned” with, their causal goal.
A primary aim of the first simulation study was to examine goal--methods misalignment in terms of inappropriate variable selection for exposure effect estimation (a causal goal). The main difference between predictive and causal models is the conceptualization and treatment of “covariates”. Therefore, exposure coefficients were compared from regression models built using different variable selection approaches that were either aligned (appropriate for causation) or misaligned (appropriate for prediction) with the causal goal of the simulated analysis. The regression models were characterized by different combinations of variable pools and inclusion criteria to select variables from the pools into the models. Overall, for valid exposure effect estimation in a causal analysis, the creation of the variable pool mattered more than the specific inclusion criteria, and the most important criterion when creating the variable pool was to exclude mediators.
The second simulation study concretized the misalignment problem by examining the consequences of goal--method misalignment in the application of the structured life course approach, a statistical method for distinguishing among different causal life course models of disease (e.g., critical period, accumulation of risk). Although exchangeability must be satisfied for valid results using this approach, in its empirical applications, confounding is often ignored. These applications are misaligned because they use methods for description (crude associations) for a causal goal (identifying causal processes). Simulations were used to mimic this misaligned approach and examined its consequences. On average, when life course data was generated under a “no confounding” scenario - an unlikely real-world scenario - the structured life course approach was quite accurate in identifying the life course model that generated the data. However, in the presence of confounding, the wrong underlying life course model was often identified. Five life course confounding structures were examined; as the complexity of examined confounding scenarios increased, particularly when this confounding was strong, incorrect model selection using the structured life course approach was common.
The misalignment problem is recognized but underappreciated in the epidemiologic literature. This dissertation contributes to the literature by documenting, simulating, and concretizing problems of misalignment in epidemiologic practice.
|
206 |
Marginal modelling of capture-recapture dataTurner, Elizabeth L. January 2007 (has links)
No description available.
|
207 |
Sieve bootstrap unit root testsRichard, Patrick. January 2007 (has links)
No description available.
|
208 |
Asymmetric heavy-tailed distributions : theory and applications to finance and risk managementZhu, Dongming, 1963- January 2007 (has links)
No description available.
|
209 |
Macrovariables in mathematical models of ecosystemsLavallée, Paul January 1976 (has links)
No description available.
|
210 |
Statistical evaluation of water quality measurementsBujatzeck, Baldur January 1998 (has links)
No description available.
|
Page generated in 0.0616 seconds