• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 542
  • 93
  • 78
  • 58
  • 25
  • 25
  • 25
  • 25
  • 25
  • 24
  • 22
  • 15
  • 9
  • 4
  • 3
  • Tagged with
  • 917
  • 917
  • 203
  • 161
  • 133
  • 123
  • 95
  • 92
  • 89
  • 74
  • 71
  • 69
  • 66
  • 65
  • 63
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
201

STATISTICAL MODELING OF SHIP AIRWAKES INCLUDING THE FEASIBILITY OF APPLYING MACHINE LEARNING

Unknown Date (has links)
Airwakes are shed behind the ship’s superstructure and represent a highly turbulent and rapidly distorting flow field. This flow field severely affects pilot’s workload and such helicopter shipboard operations. It requires both the one-point statistics of autospectrum and the two-point statistics of coherence (normalized cross-spectrum) for a relatively complete description. Recent advances primarily refer to generating databases of flow velocity points through experimental and computational fluid dynamics (CFD) investigations, numerically computing autospectra along with a few cases of cross-spectra and coherences, and developing a framework for extracting interpretive models of autospectra in closed form from a database along with an application of this framework to study the downwash effects. By comparison, relatively little is known about coherences. In fact, even the basic expressions of cross-spectra and coherences for three components of homogeneous isotropic turbulence (HIT) vary from one study to the other, and the related literature is scattered and piecemeal. Accordingly, this dissertation begins with a unified account of all the cross-spectra and coherences of HIT from first principles. Then, it presents a framework for constructing interpretive coherence models of airwake from a database on the basis of perturbation theory. For each velocity component, the coherence is represented by a separate perturbation series in which the basis function or the first term on the right-hand side of the series is represented by the corresponding coherence for HIT. The perturbation series coefficients are evaluated by satisfying the theoretical constraints and fitting a curve in a least squares sense on a set of numerically generated coherence points from a database. Although not tested against a specific database, the framework has a mathematical basis. Moreover, for assumed values of perturbation series constants, coherence results are presented to demonstrate how coherences of airwakes and such flow fields compare to those of HIT. / Includes bibliography. / Dissertation (Ph.D.)--Florida Atlantic University, 2020. / FAU Electronic Theses and Dissertations Collection
202

Advances in Machine Learning for Compositional Data

Gordon Rodriguez, Elliott January 2022 (has links)
Compositional data refers to simplex-valued data, or equivalently, nonnegative vectors whose totals are uninformative. This data modality is of relevance across several scientific domains. A classical example of compositional data is the chemical composition of geological samples, e.g., major-oxide concentrations. A more modern example arises from the microbial populations recorded using high-throughput genetic sequencing technologies, e.g., the gut microbiome. This dissertation presents a set of methodological and theoretical contributions that advance the state of the art in the analysis of compositional data. Our work can be divided along two categories: problems in which compositional data represents the input to a predictive model, and problems in which it represents the output of the model. For the first class of problems, we build on the popular log-ratio framework to develop an efficient learning algorithm for high-dimensional compositional data. Our algorithm runs orders of magnitude faster than competing alternatives, without sacrificing model quality. For the second class of problems, we define a novel exponential family of probability distributions supported on the simplex. This distribution enjoys attractive mathematical properties and provides a performant probability model for simplex-valued outcomes. Taken together, our results constitute a broad contribution to the toolkit of researchers and practitioners studying compositional data.
203

The estimation of missing values in hydrological records using the EM algorithm and regression methods

Makhuvha, Tondani January 1988 (has links)
Includes bibliography. / The objective of this thesis is to review existing methods for estimating missing values in rainfall records and to propose a number of new procedures. Two classes of methods are considered. The first is based on the theory of variable selection in regression. Here the emphasis is on finding efficient methods to identify the set of control stations which are likely to yield the best regression estimates of the missing values in the target station. The second class of methods is based on the EM algorithm, proposed by Dempster, Laird and Rubin (1977). The emphasis here is to estimate the missing values directly without first making a detailed selection of control stations. All "relevant" stations are included. This method has not previously been applied in the context of estimating missing rainfall values.
204

An empirical evaluation of the Altman (1968) failure prediction model on South African JSE listed companies

Rama, Kavir D. 18 March 2013 (has links)
Credit has become very important in the global economy (Cynamon and Fazzari, 2008). The Altman (1968) failure prediction model, or derivatives thereof, are often used in the identification and selection of financially distressed companies as it is recognized as one of the most reliable in predicting company failure (Eidleman, 1995). Failure of a firm can cause substantial losses to creditors and shareholders, therefore it is important, to detect company failure as early as possible. This research report empirically tests the Altman (1968) failure prediction model on 227 South African JSE listed companies using data from the 2008 financial year to calculate the Z-score within the model, and measuring success or failure of firms in the 2009 and 2010 years. The results indicate that the Altman (1968) model is a viable tool in predicting company failure for firms with positive Z-scores, and where Z-scores do not fall into the range of uncertainty as specified. The results also suggest that the model is not reliable when the Z–scores are negative or when they are in the range of uncertainty (between 2.99 and 1.81). If one is able to predict firm failure in advance, it should be possible for management to take steps to avert such an occurrence (Deakin, 1972; Keasey and Watson, 1991; Platt and Platt, 2002).
205

Is the way forward to step back? A meta-research analysis of misalignment between goals, methods, and conclusions in epidemiologic studies.

Kezios, Katrina Lynn January 2021 (has links)
Recent discussion in the epidemiologic methods and teaching literatures centers around the importance of clearly stating study goals, disentangling the goal of causation from prediction (or description), and clarifying the statistical tools that can address each goal. This discussion illuminates different ways in which mismatches can occur between study goals, methods, and interpretations, which this dissertation synthesizes into the concept of “misalignment”; misalignment occurs when the study methods and/or interpretations are inappropriate for (i.e., do not match) the study’s goal. While misalignments can occur and may cause problems, their pervasiveness and consequences have not been examined in the epidemiologic literature. Thus, the overall purpose of this dissertation was to document and examine the effects of misalignment problems seen in epidemiologic practice. First, a review was conducted to document misalignment in a random sample of epidemiologic studies and explore how the framing of study goals contributes to its occurrence. Among the reviewed articles, full alignment between study goals, methods, and interpretations was infrequently observed, although “clearly causal” studies (those that framed causal goals using causal language) were more often fully aligned (5/13, 38%) than “seemingly causal” ones (those that framed causal goals using associational language; 3/71, 4%). Next, two simulation studies were performed to examine the potential consequences of different types of misalignment problems seen in epidemiologic practice. They are based on the observation that, often, studies that are causally motivated perform analyses that appear disconnected from, or “misaligned” with, their causal goal. A primary aim of the first simulation study was to examine goal--methods misalignment in terms of inappropriate variable selection for exposure effect estimation (a causal goal). The main difference between predictive and causal models is the conceptualization and treatment of “covariates”. Therefore, exposure coefficients were compared from regression models built using different variable selection approaches that were either aligned (appropriate for causation) or misaligned (appropriate for prediction) with the causal goal of the simulated analysis. The regression models were characterized by different combinations of variable pools and inclusion criteria to select variables from the pools into the models. Overall, for valid exposure effect estimation in a causal analysis, the creation of the variable pool mattered more than the specific inclusion criteria, and the most important criterion when creating the variable pool was to exclude mediators. The second simulation study concretized the misalignment problem by examining the consequences of goal--method misalignment in the application of the structured life course approach, a statistical method for distinguishing among different causal life course models of disease (e.g., critical period, accumulation of risk). Although exchangeability must be satisfied for valid results using this approach, in its empirical applications, confounding is often ignored. These applications are misaligned because they use methods for description (crude associations) for a causal goal (identifying causal processes). Simulations were used to mimic this misaligned approach and examined its consequences. On average, when life course data was generated under a “no confounding” scenario - an unlikely real-world scenario - the structured life course approach was quite accurate in identifying the life course model that generated the data. However, in the presence of confounding, the wrong underlying life course model was often identified. Five life course confounding structures were examined; as the complexity of examined confounding scenarios increased, particularly when this confounding was strong, incorrect model selection using the structured life course approach was common. The misalignment problem is recognized but underappreciated in the epidemiologic literature. This dissertation contributes to the literature by documenting, simulating, and concretizing problems of misalignment in epidemiologic practice.
206

Marginal modelling of capture-recapture data

Turner, Elizabeth L. January 2007 (has links)
No description available.
207

Sieve bootstrap unit root tests

Richard, Patrick. January 2007 (has links)
No description available.
208

Asymmetric heavy-tailed distributions : theory and applications to finance and risk management

Zhu, Dongming, 1963- January 2007 (has links)
No description available.
209

Macrovariables in mathematical models of ecosystems

Lavallée, Paul January 1976 (has links)
No description available.
210

Statistical evaluation of water quality measurements

Bujatzeck, Baldur January 1998 (has links)
No description available.

Page generated in 0.0616 seconds