• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 3032
  • 1002
  • 369
  • 345
  • 273
  • 182
  • 174
  • 160
  • 82
  • 54
  • 30
  • 29
  • 23
  • 23
  • 21
  • Tagged with
  • 6654
  • 2249
  • 1129
  • 917
  • 855
  • 791
  • 743
  • 741
  • 647
  • 544
  • 509
  • 488
  • 445
  • 425
  • 399
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
751

An evaluation of environmental concerns and private land conservation programs

Adhikari, Ram Kumar 01 May 2020 (has links)
Bottomland hardwood forests and pine forests in the southern United States provide valuable ecosystem services such as timber, recreation, wildlife habitat, carbon sequestration, floodwater storage, and sediment and nutrient retention. However, these forest ecosystems are threatened because of intensive forest management, forest land conversion, and urbanization. As private landownership dominates in this region, landowner participation is crucial for success of conservation programs facilitating ecosystem services. This research focused on three different aspects of private land conservation programs. First, it estimated the impact of environmental contextual factors, private land attributes and sociodemographic characteristics on landowner concern about environmental issues. Second, it determined the influence of private land attributes, environmental concerns, frequency of contacts with federal agencies and socioeconomic characteristics on landowner satisfaction with available conservation programs. Third, it estimated the monetary compensation required by landowners to implement conservation practices focused on increasing provision of ecosystem services. Data were collected using a mail survey and from online sources. Data were analyzed using seemingly unrelated regression and logistic regression models. Results indicated that private land attributes, particularly size of agricultural land owned, and landownership goals such as providing ecosystem services and profitability, had a greater magnitude of positive association with landowner concerns about environmental issues than other factors. Similarly, size of agricultural land owned, landownership goals such as profitability and personal recreation, concerns about wildlife habitat losses and frequent contacts with federal agencies were positively related to landowner satisfaction with conservation programs. Landowner willingness to participate in a conservation program was positively related to payment amount, concerns about wildlife habitat losses, frequency of contact with the Natural Resources Conservation Service (NRCS), and education level. Similarly, provision of clean water as landownership objective, concerns about hurricanes and tornadoes, and landowner age were negatively associated with landowner interests to participate in conservation programs. This research also quantified landowner median willingness to accept (WTA) compensation which was $229.98 ha-1 yr-1 for participation in a conservation program related to bottomland hardwood or pine forests. The findings help identify likely participants or landowners for conservation programs facilitating ecosystem services and determine actual conservation costs at a regional level.
752

Using Machine Learning to Predict Form Processing Times : Applied to Swedish pay-as-you-earn tax returns / Maskininlärning som verktyg för att förutspå formulärhandläggningstider : Tillämpat på svenska arbetsgivardeklarationer

Al-Kadhimi, Staffan January 2022 (has links)
Forms are used in many situations. For example, they tend to be ubiquitous in communications between individuals and government agencies. Something which could potentially boost transparency and efficiency is accurate estimates of how long it will take for the receiver to process a given completed form. Unfortunately, such estimates are often not available. This thesis examines the problem of using machine learning to predict form processing times, applied to the context of Swedish pay-as-you-earn tax returns. More specifically, it compares a naive baseline model to several random forest models, some based on the more common batch learning principle, and others on online learning which is typically seen as more suitable for working with data streams and changing conditions. Despite the theoretical advantages of online learning, none of the models using that approach were able to consistently outperform the naive baseline model. Conversely, the two primarily evaluated batch learning models were successful in doing so, although the improvement over the baseline was small. / Formulär används i många sammanhang. De är exempelvis mycket vanligt förekommande i kommunikation mellan privatpersoner och myndigheter. Något som potentiellt skulle kunna innebära ökad transparens och effektivitet är träffsäkra uppskattningar av hur lång tid det tar för mottagaren att handlägga ett givet formulär. Dessvärre är sådana uppskattningar ofta inte tillgängliga. Detta examensarbete undersöker hur maskininlärning kan användas för att förutspå formulärhandläggningstider, tillämpat i kontexten svenska arbetsgivardeklarationer. Mer specifikt jämförs en enkel naiv modell mot flera random forest-modeller, vissa baserade på den vanligare batchinlärningsprincipen, och andra på onlineinlärning som brukar ses som mer passande för dataströmmar och föränderliga förhållanden. Trots de teoretiska fördelarna med onlineinlärning lyckades inte någon av modellerna som använde sig av den tekniken konsekvent ge bättre resultat än den naiva grundmodellen. Däremot visade sig de två primärt utvärderade batchinlärningsmodellerna vara framgångsrika i det avseendet, även om skillnaden mot den naiva modellen var liten.
753

Characterizing the performance of a biological analogue to a digital inverter

Ghosh, Susmit Kumar 05 October 2010 (has links)
No description available.
754

Comparing Variable Selection Algorithms On Logistic Regression – A Simulation

SINGH, KEVIN January 2021 (has links)
When we try to understand why some schools perform worse than others, if Covid-19 has struck harder on some demographics or whether income correlates with increased happiness, we may turn to regression to better understand how these variables are correlated. To capture the true relationship between variables we may use variable selection methods in order to ensure that the variables which have an actual effect have been included in the model. Choosing the right model for variable selection is vital. Without it there is a risk of including variables which have little to do with the dependent variable or excluding variables that are important. Failing to capture the true effects would paint a picture disconnected from reality and it would also give a false impression of what reality really looks like. To mitigate this risk a simulation study has been conducted to find out what variable selection algorithms to apply in order to make more accurate inference. The different algorithms being tested are stepwise regression, backward elimination and lasso regression. Lasso performed worst when applied to a small sample but performed best when applied to larger samples. Backward elimination and stepwise regression had very similar results.
755

Case Study of Discharge Modeling for Nissan River in Halmstad Municipality / Fallstudie av vattenflödesmodellering förvattendraget Nissan i Halmstads kommun

Vega Ezpeleta, Federico January 2022 (has links)
Changes in precipitation patterns, temperature, and other climatic variables have been shown to modify thehydrological cycle and hydrological systems, potentially resulting in a shift in river runoff behavior and an increasedrisk of floods. There have been several instances of devastating floods throughout Europe’s history, which haveresulted in devastation and enormous economic losses. As a result of the effects of climate change, floods areoccurring more frequently in Sweden as well as across Europe. Research on the subject of flood prediction has beengoing on for decades, where particularly data-driven models have advanced in recent years. This study examinedtwo different machine learning (data-driven) models for forecasting river discharge in the Nissan River: Linearregression and Random Forrest regression (RFR), with the use of ECMWF Reanalysis v5 ( ERA5 ) data and historicaldischarge data. The Linear regression model yielded a r2 score of 0.45 and could not be considered an acceptablemodel. The RFR model had a r2 score of 0.71. This implies, given ERA5 reanalysis data, that one might generatea moderately performing machine learning model for Nissan river. An additional investigation was carried out,to see if the trained model could be used with EC-EARTH CMIP6 future projection. The findings resulting fromapplying the EC-EARTH CMIP6 future data on the trained RFR indicated too many uncertainties, necessitatingmore investigation before any conclusions can be drawn.
756

Improved Methods for Interrupted Time Series Analysis Useful When Outcomes are Aggregated: Accounting for heterogeneity across patients and healthcare settings

Ewusie, Joycelyne E January 2019 (has links)
This is a sandwich thesis / In an interrupted time series (ITS) design, data are collected at multiple time points before and after the implementation of an intervention or program to investigate the effect of the intervention on an outcome of interest. ITS design is often implemented in healthcare settings and is considered the strongest quasi-experimental design in terms of internal and external validity as well as its ability to establish causal relationships. There are several statistical methods that can be used to analyze data from ITS studies. Nevertheless, limitations exist in practical applications, where researchers inappropriately apply the methods, and frequently ignore the assumptions and factors that may influence the optimality of the statistical analysis. Moreover, there is little to no guidance available regarding the application of the various methods, and a standardized framework for analysis of ITS studies does not exist. As such, there is a need to identify and compare existing ITS methods in terms of their strengths and limitations. Their methodological challenges also need to be investigated to inform and direct future research. In light of this, this PhD thesis addresses two main objectives: 1) to conduct a scoping review of the methods that have been employed in the analysis of ITS studies, and 2) to develop improved methods that address a major limitation of the statistical methods frequently used in ITS data analysis. These objectives are addressed in three projects. For the first project, a scoping review of the methods that have been used in analyzing ITS data was conducted, with the focus on ITS applications in health research. The review was based on the Arksey and O’Malley framework and the Joanna Briggs Handbook for scoping reviews. A total of 1389 studies were included in our scoping review. The articles were grouped into methods papers and applications papers based on the focus of the article. For the methods papers, we narratively described the identified methods and discussed their strengths and limitations. The application papers were summarized using frequencies and percentages. We identified some limitations of current methods and provided some recommendations useful in health research. In the second project, we developed and presented an improved method for ITS analysis when the data at each time point are aggregated across several participants, which is the most common case in ITS studies in healthcare settings. We considered the segmented linear regression approach, which our scoping review identified as the most frequently used method in ITS studies. When data are aggregated, heterogeneity is introduced due to variability in the patient population within sites (e.g. healthcare facilities) and this is ignored in the segmented linear regression method. Moreover, statistical uncertainty (imprecision) is introduced in the data because of the sample size (number of participants from whom data are aggregated). Ignoring this variability and uncertainty will likely lead to invalid estimates and loss of statistical power, which in turn leads to erroneous conclusions. Our proposed method incorporates patient variability and sample size as weights in a weighted segmented regression model. We performed extensive simulations and assessed the performance of our method using established performance criteria, such as bias, mean squared error, level and statistical power. We also compared our method with the segmented linear regression approach. The results indicated that the weighted segmented regression was uniformly more precise, less biased and more powerful than the segmented linear regression method. In the third project, we extended the weighted method to multisite ITS studies, where data are aggregated at two levels: across several participants within sites as well as across multiple sites. The extended method incorporates the two levels of heterogeneity using weights, where the weights are defined using patient variability, sample size, number of sites as well as site-to-site variability. This extended weighted regression model, which follows the weighted least squares approach is employed to estimate parameters and perform significance testing. We conducted extensive empirical evaluations using various scenarios generated from a multi-site ITS study and compared the performance of our method with that of the segmented linear regression model as well as a pooled analysis method previously developed for multisite studies. We observed that for most scenarios considered, our method produced estimates with narrower 95% confidence intervals and smaller p-values, indicating that our method is more precise and is associated with more statistical power. In some scenarios, where we considered low levels of heterogeneity, our method and the previously proposed method showed comparable results. In conclusion, this PhD thesis facilitates future ITS research by laying the groundwork for developing standard guidelines for the design and analysis of ITS studies. The proposed improved method for ITS analysis, which is the weighted segmented regression, contributes to the advancement of ITS research and will enable researchers to optimize their analysis, leading to more precise and powerful results. / Thesis / Doctor of Philosophy (PhD)
757

Automatic age and gender classification using supervised appearance model

Bukar, Ali M., Ugail, Hassan, Connah, David 01 August 2016 (has links)
Yes / Age and gender classification are two important problems that recently gained popularity in the research community, due to their wide range of applications. Research has shown that both age and gender information are encoded in the face shape and texture, hence the active appearance model (AAM), a statistical model that captures shape and texture variations, has been one of the most widely used feature extraction techniques for the aforementioned problems. However, AAM suffers from some drawbacks, especially when used for classification. This is primarily because principal component analysis (PCA), which is at the core of the model, works in an unsupervised manner, i.e., PCA dimensionality reduction does not take into account how the predictor variables relate to the response (class labels). Rather, it explores only the underlying structure of the predictor variables, thus, it is no surprise if PCA discards valuable parts of the data that represent discriminatory features. Toward this end, we propose a supervised appearance model (sAM) that improves on AAM by replacing PCA with partial least-squares regression. This feature extraction technique is then used for the problems of age and gender classification. Our experiments show that sAM has better predictive power than the conventional AAM.
758

Advances in Applied Econometrics: Binary Discrete Choice Models, Artificial Neural Networks, and Asymmetries in the FAST Multistage Demand System

Bergtold, Jason Scott 27 April 2004 (has links)
The dissertation examines advancements in the methods and techniques used in the field of econometrics. These advancements include: (i) a re-examination of the underlying statistical foundations of statistical models with binary dependent variables. (ii) using feed-forward backpropagation artificial neural networks for modeling dichotomous choice processes, and (iii) the estimation of unconditional demand elasticities using the flexible multistage demand system with asymmetric partitions and fixed effects across time. The first paper re-examines the underlying statistical foundations of statistical models with binary dependent variables using the probabilistic reduction approach. This re-examination leads to the development of the Bernoulli Regression Model, a family of statistical models arising from conditional Bernoulli distributions. The paper provides guidelines for specifying and estimating a Bernoulli Regression Model, as well as, methods for generating and simulating conditional binary choice processes. Finally, the Multinomial Regression Model is presented as a direct extension. The second paper empirically compares the out-of-sample predictive capabilities of artificial neural networks to binary logit and probit models. To facilitate this comparison, the statistical foundations of dichotomous choice models and feed-forward backpropagation artificial neural networks (FFBANNs) are re-evaluated. Using contingent valuation survey data, the paper shows that FFBANNs provide an alternative to the binary logit and probit models with linear index functions. Direct comparisons between the models showed that the FFBANNs performed marginally better than the logit and probit models for a number of within-sample and out-of-sample performance measures, but in the majority of cases these differences were not statistically significant. In addition, guidelines for modeling contingent valuation survey data and techniques for estimating median WTP measures using FFBANNs are examined. The third paper estimates a set of unconditional price and expenditure elasticities for 49 different processed food categories using scanner data and the flexible and symmetric translog (FAST) multistage demand system. Due to the use of panel data and the presence of heterogeneity across time, temporal fixed effects were incorporated into the model. Overall, estimated price elasticities are larger, in absolute terms, than previous estimates. The use of disaggregated product groupings, scanner data, and the estimation of unconditional elasticities likely accounts for these differences. / Ph. D.
759

Functional Data Models for Raman Spectral Data and Degradation Analysis

Do, Quyen Ngoc 16 August 2022 (has links)
Functional data analysis (FDA) studies data in the form of measurements over a domain as whole entities. Our first focus is on the post-hoc analysis with pairwise and contrast comparisons of the popular functional ANOVA model comparing groups of functional data. Existing contrast tests assume independent functional observations within group. In reality, this assumption may not be satisfactory since functional data are often collected continually overtime on a subject. In this work, we introduce a new linear contrast test that accounts for time dependency among functional group members. For a significant contrast test, it can be beneficial to identify the region of significant difference. In the second part, we propose a non-parametric regression procedure to obtain a locally sparse estimate of functional contrast. Our work is motivated by a biomedical study using Raman spectroscopy to monitor hemodialysis treatment near real-time. With contrast test and sparse estimation, practitioners can monitor the progress of the hemodialysis within session and identify important chemicals for dialysis adequacy monitoring. In the third part, we propose a functional data model for degradation analysis of functional data. Motivated by degradation analysis application of rechargeable Li-ion batteries, we combine state-of-the-art functional linear models to produce fully functional prediction for curves on heterogenous domains. Simulation studies and data analysis demonstrate the advantage of the proposed method in predicting degradation measure than existing method using aggregation method. / Doctor of Philosophy / Functional data analysis (FDA) studies complex data structure in the form of curves and shapes. Our work is motivated by two applications concerning data from Raman spectroscopy and battery degradation study. Raman spectra of a liquid sample are curves with measurements over a domain of wavelengths that can identify chemical composition and whose values signify the constituent concentrations in the sample. We first propose a statistical procedure to test the significance of a functional contrast formed by spectra collected at beginning and at later time points during a dialysis session. Then a follow-up procedure is developed to produce a sparse representation of the contrast functional contrast with clearly identified zero and nonzero regions. The use of this method on contrast formed by Raman spectra of used dialysate collected at different time points during hemodialysis sessions can be adapted for evaluating the treatment efficacy in real time. In a third project, we apply state-of-the-art methodologies from FDA to a degradation study of rechargeable Li-ion batteries. Our proposed methods produce fully functional prediction of voltage discharge curves allowing flexibility in monitoring battery health.
760

Nonparametric metamodeling for simulation optimization

Keys, Anthony C. 07 June 2006 (has links)
Optimization of simulation model performance requires finding the values of the model's controllable inputs that optimize a chosen model response. Responses are usually stochastic in nature, and the cost of simulation model runs is high. The literature suggests the use of metamodels to synthesize the response surface using sample data. In particular, nonparametric regression is proposed as a useful tool in the global optimization of a response surface. As the general simulation optimization problem is very difficult and requires expertise from a number of fields, there is a growing consensus in the literature that a knowledge-based approach to solving simulation optimization problems is required. This dissertation examines the relative performance of the principal nonparametric techniques, spline and kernel smoothing, and subsequently addresses the issues involved in implementing the techniques in a knowledge-based simulation optimization system. The dissertation consists of two parts. In the first part, a full factorial experiment is carried out to compare the performance of kernel and spline smoothing on a number of measures when modeling a varied set of surfaces using a range of small sample sizes. In the second part, nonparametric metamodeling techniques are placed in a taxonomy of stochastic search procedures for simulation optimization and a method for their implementation in a knowledge-based system is presented. A sequential design procedure is developed that allows spline smoothing to be used as a search technique. Throughout the dissertation, a two-input, single-response model is considered. Results from the experiment show that spline smoothing is superior to constant-bandwidth kernel smoothing in fitting the response. Kernel smoothing is shown to be more accurate in placing optima in X-space for sample sizes up to 36. Inventory model examples are used to illustrate the results. The taxonomy implies that search procedures can be chosen initially using the parameters of the problem. A process that allows for selection of a search technique and its subsequent evaluation for further use or for substitution of another search technique is given. The success of a sequential design method for spline smooths in finding a global optimum is demonstrated using a bimodal response surface. / Ph. D.

Page generated in 0.0734 seconds