• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 91
  • 23
  • 10
  • 4
  • 2
  • 2
  • 2
  • 1
  • Tagged with
  • 148
  • 148
  • 40
  • 38
  • 35
  • 32
  • 22
  • 19
  • 19
  • 18
  • 18
  • 17
  • 17
  • 14
  • 14
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
61

Individual Differences in Reading Proficiency: Investigating Influencing Factors and How They Interact

Nisbet, Kelly January 2021 (has links)
This thesis investigates individual differences and their impact on reading proficiency using different measures of proficiency, a variety of data collection and statistical methods, and different populations. The goal was to examine the impact that individual differences in certain reading-related skills and cognitive abilities have on reading proficiency and how these differences interact. Through three key studies that make up this thesis, several important discoveries and contributions were made to the field. Chapter 2 introduces an easy-to-use application for measuring cloze probability. ‘ClozApp’, was created and made publicly available, along with a user manual and sample code for programming. Chapter 3 contributed through the development of a novel statistical method used to analyze variance between populations with different linguistic backgrounds. This method was used to demonstrate how an individual’s linguistic background (i.e., whether they were first- or second-language speakers of English) impacted how individual differences in reading skills influence their reading fluency, as indicated through their eye-movements. This statistical prediction method is open source and was made widely available for use along with sample data and code. In Chapter 4, a new connection was found between two important cognitive factors that are well-known in the reading literature: statistical learning and motivation. Using mediation analyses, this project discovered an interaction between these factors that further highlights the ways they impact reading proficiency. This thesis demonstrates a comprehensive approach to investigating individual differences in reading proficiency in the following ways: (i) both reading fluency and comprehension were investigated as measures of reading proficiency, (ii) data collection included a variety of reading-related skills, cognitive abilities, and group differences, and (iii) unique statistical analysis methods were utilized to investigate both individual and group differences. This thesis highlights important new discoveries and makes significant lasting contributions to the field of reading research. / Thesis / Doctor of Philosophy (PhD) / This thesis investigates how individual differences influence reading proficiency. Specifically, it asks how the ways in which people differ on certain reading-related skills and cognitive abilities can determine how well they read. Using different measures of proficiency, a variety of data collection and statistical methods, and looking across different populations, the goal of this thesis was to examine the ways in which people differ in these skills and abilities, how these differences interact, and the resulting impact on reading proficiency. This thesis resulted in three significant contributions to the field. First, it made available a new application for collecting data on an important variable in reading research – cloze probability. In addition, it culminated in the development of a novel statistical method that demonstrates how an individual’s linguistic background can influence their reading fluency. Finally, a new connection was found between two important cognitive factors that interact to influence reading comprehension.
62

CONTENT TRADING AND PRIVACY-AWARE PRICING FOR EFFICIENT SPECTRUM UTILIZATION

Alotaibi, Faisal F. January 2019 (has links)
No description available.
63

An Efficient Implementation of a Robust Clustering Algorithm

Blostein, Martin January 2016 (has links)
Clustering and classification are fundamental problems in statistical and machine learning, with a broad range of applications. A common approach is the Gaussian mixture model, which assumes that each cluster or class arises from a distinct Gaussian distribution. This thesis studies a robust, high-dimensional extension of the Gaussian mixture model that automatically detects outliers and noise, and a computationally efficient implementation thereof. The contaminated Gaussian distribution is a robust elliptic distribution that allows for automatic detection of ``bad points'', and is used to make robust the usual factor analysis model. In turn, the mixtures of contaminated Gaussian factor analyzers (MCGFA) algorithm allows high-dimesional, robust clustering, classification and detection of bad points. A family of MCGFA models is created through the introduction of different constraints on the covariance structure. A new, efficient implementation of the algorithm is presented, along with an account of its development. The fast implementation permits thorough testing of the MCGFA algorithm, and its performance is compared to two natural competitors: parsimonious Gaussian mixture models (PGMM) and mixtures of modified t factor analyzers (MMtFA). The algorithms are tested systematically on simulated and real data. / Thesis / Master of Science (MSc)
64

Statistical learning and predictive modeling in data mining

Li, Bin 13 September 2006 (has links)
No description available.
65

Role of Majorization in Learning the Kernel within a Gaussian Process Regression Framework

Kapat, Prasenjit 21 October 2011 (has links)
No description available.
66

Foundations of Vocabulary: Does Statistical Segmentation of Events Contribute to Word Learning?

Levine, Dani Fara January 2017 (has links)
This dissertation evaluates the untested assumption that the individuation of events into units matters for word learning, particularly the learning of terms which map onto relational event units (Gentner & Boroditsky, 2001; Maguire et al., 2006). We predicted that 3-year-old children’s statistical action segmentation abilities would relate to their verb comprehension and to their overall vocabulary knowledge (Research Question 1). We also hypothesized that statistical action segmentation would facilitate children’s learning of novel verbs (Research Question 2). Largely confirming our first prediction, children who were better able to statistically segment novel action sequences into reliable units had more sophisticated overall vocabularies and were quicker to select the correct referents of overall vocabulary items and verb vocabulary items; nevertheless, they did not have larger verb vocabularies. Unexpectedly, statistical action segmentation did not facilitate children’s learning of verbs for statistically consistent action units. However, children showed greater learning of verbs labeling statistical action part-units than verbs labeling statistical action non-units, providing some evidence for our second prediction. In sum, this dissertation takes an important step towards understanding how event segmentation may contribute to vocabulary acquisition. / Psychology
67

APPLICATIONS OF STATISTICAL LEARNING ALGORITHMS IN ELECTRON SPECTROSCOPY / TOWARDS CALIBRATION-INVARIANT SPECTROSCOPY USING DEEP LEARNING

Chatzidakis, Michael 06 1900 (has links)
Building on the recent advances in computer vision with convolutional neural networks, we have built SpectralNet, a spectroscopy-optimized convolutional neural network architecture capable of classifying spectra despite large temporal (i.e. translational, chemical, calibration) shifts. Present methods of measuring the local chemical environment of atoms at the nano-scale involve manual feature extraction and dimensionality reduction of the original signal such as: using the peak onset, the ratio of peaks, or the full-width half maximum of peaks. Convolutional neural networks like SpectralNet are able to automatically find parts of the spectra (i.e. features) of the spectra which maximally discriminate between the classes without requiring manual feature extraction. The advantage of such a process is to remove bias and qualitative interpretation in spectroscopy analysis which occurs during manual feature extraction. Because of this automated feature extraction process, this method of spectroscopy analysis is also immune to instrument calibration differences since it performs classification based on the shape of the spectra. Convolutional neural networks are an ideal statistical classifier for spectroscopy data (i.e. time-series data) due to its shared weighting scheme in neural network weights which is ideal for identifying local correlations between adjacent dimensions of the time-series data. Over 2000 electron energy loss spectra were collected using a scanning transmission electron microscope of three oxidation states of Mn. SpectralNet was trained to learn the differences between them. We prove generalizability by training SpectralNet on electron energy loss spectroscopy data from one instrument, and test it on a variety of reference spectra found in the literature with perfect accuracy. We also test SpectralNet against a wide variety of high noise samples which a trained human spectroscopist would find incomprehensible. We also compare other neural network architectures used in the literature and determine that SpectralNet, a dense-layer free neural network, is immune to calibration differences whereas other styles of network are not. / Thesis / Master of Applied Science (MASc) / Spectroscopy is the study of the interaction between photons or electrons and a material to determine what that material is made of. One advanced way to make accurate measurements down to the atomic scale is to use high energy electrons in a transmission electron microscope. Using this instrument, a special type of photograph can be taken of the material (a spectrograph or spectrum) which is detailed enough to identify which kinds of atoms are in the material. The spectrographs are very complicated to interpret and the human eye struggles to find patterns in noisy and low resolution data. Depending on which instrument that the spectrographs are taken on, the resulting spectrograph will also change which adds extra difficulty. In this study, advanced algorithms are used to identify which types of atoms can be identified in the noisy signal from the spectrograph regardless of which instrument is used. These algorithms (convolutional neural networks) are also used in self-driving cars for a similar task of identifying objects whereas in this study we use it for identifying atoms.
68

Inteligência estatística na tomada de decisão médica: um estudo de caso em pacientes traumatizados / Statistical intelligence in medical decision making: a case study in traumatized patients

Garcia, Marcelo 22 November 2018 (has links)
O principal objetivo do estudo foi utilizar informações de ocorrência do Traumatismo Crânio Encefálico (TCE) que possam inferir/gerar descobertas associadas ao risco de gravidade do paciente, bem como auxiliar na tomada de decisão médica ao definir o melhor prognóstico, indicando quais as possíveis medidas que podem ser escolhidas para a gravidade na lesão sofrida pela vítima. Inicialmente, foram analisadas as estatísticas descritivas dos dados dos pacientes de TCE de um hospital do interior de São Paulo. Participaram desse estudo 50 pacientes. Os resultados mostraram que a maior frequência do trauma é por acidentes de trânsito (62%), seguidos de acidentes por queda (24%). Traumas em pacientes do sexo masculino (88%) são muito mais frequentes do que em pacientes do sexo feminino. Para modelagem, transformou-se a variável resposta \"Abbreviated Injury Scale (AIS)\" em dicotômica, considerando 0 (zero) aos pacientes fora de risco e 1 (um) aos que apresentaram algum tipo de risco. Em seguida, técnicas de aprendizado estatístico foram utilizadas de modo a comparar o desempenho dos classificadores Regressão Logística sendo um caso do Generalized Linear Model (GLM), Random Forest (RF), Support Vector Machine (SVM) e redes probabilísticas Naïve Bayes (NB). O modelo com melhor desempenho (RF) combinou os índices Accuracy (ACC) , Area Under ROC Curve (AUC) , Sensitivity (SEN), Specificity (SPE) e Matthews Correlation Coefficient (MCC), que apresentaram os resultados mais favoráveis no quesito de apoio no auxílio da tomada de decisão médica, possibilitando escolher o estudo clínico mais adequado das vítimas traumatizadas ao considerar o risco de vida do indivíduo. Conforme o modelo selecionado foi possível gerar um ranking para estimar a probabilidade de risco de vida do paciente. Em seguida foi realizado uma comparação de desempenho entre o modelo RF (novo classificador) e os índices Revisited Trauma Score (RTS), Injury Severity Score (ISS) , Índice de Barthel (IB) referente à classificação de risco dos pacientes. / The main objective of this study was to consider the information related to the occurrence of traumatic brain injury (TBI) that can infer new results associated with the patients risk of severity as well as assisting in the medical decision in order to find the best prognosis; this can lead to indicate possible measures that can be chosen for severity in the injury suffered by the victim. Initially, we have presented descriptive statistics from the patients with TBI from a hospital located in the heartland of São Paulo. Fifty patients were recruited for this study. Descriptive analyzes showed that the highest frequency of trauma is due to traffic accidents (62 %) followed by crashes per accident (24 %). The causes related to trauma occur much more often in male patients (88 %) than in female patients. To order model, the response variable Abbreviated Injury Scale (AIS) was considered as dichotomous, where 0 (zero) was to out-of-risk patients and 1 (one) to those who presented some type of risk. Further, statistical learning techniques were used in order to compare the performance of the Logistic Regression as a Generalized Linear Model (GLM), Random Forest (RF), Support Vector Machine (SVM) and Naive Bayes (NB) model. The best performing (RF) model combined the Accuracy (ACC) , Area Under ROC Curve (AUC) , Sensitivity (SEN), Specificity (SPE) e Matthews Correlation Coefficient (MCC), which presented the most favorable results in terms of support in medical decision, making it possible to choose the most appropriate clinical study of traumatized victims based on the individual life risk. According to the selected model it was possible to generate a rank to estimate the probability of life risk of the patient. Then a performance comparison was performed between the RF model (proposed classifier) and the Revisited Trauma Score (RTS), Injury Severity Score (ISS), Barthel index (IB) referring to the risk classification of patients.
69

Evolutionary algorithms in statistical learning : Automating the optimization procedure / Evolutionära algoritmer i statistisk inlärning : Automatisering av optimeringsprocessen

Sjöblom, Niklas January 2019 (has links)
Scania has been working with statistics for a long time but has invested in becoming a data driven company more recently and uses data science in almost all business functions. The algorithms developed by the data scientists need to be optimized to be fully utilized and traditionally this is a manual and time consuming process. What this thesis investigates is if and how well evolutionary algorithms can be used to automate the optimization process. The evaluation was done by implementing and analyzing four variations of genetic algorithms with different levels of complexity and tuning parameters. The algorithm subject to optimization was XGBoost, a gradient boosted tree model, applied to data that had previously been modelled in a competition. The results show that evolutionary algorithms are applicable in finding good models but also emphasizes the importance of proper data preparation. / Scania har länge jobbat med statistik men har på senare år investerat i att bli ett mer datadrivet företag och använder nu data science i nästan alla avdelningar på företaget. De algoritmer som utvecklas av data scientists måste optimeras för att kunna utnyttjas till fullo och detta är traditionellt sett en manuell och tidskrävade process. Detta examensarbete utreder om och hur väl evolutionära algoritmer kan användas för att automatisera optimeringsprocessen. Utvärderingen gjordes genom att implementera och analysera fyra varianter avgenetiska algoritmer med olika grader av komplexitet och trimningsparameterar. Algoritmen som var målet för optimering var XGBoost, som är en gradient boosted trädbaserad modell. Denna applicerades på data som tidigare hade modellerats i entävling. Resultatet visar att evolutionära algoritmer är applicerbara i att hitta bra modellermen påvisar även hur fundamentalt det är att arbeta med databearbetning innan modellering.
70

Fraud or Not?

Åkerblom, Thea, Thor, Tobias January 2019 (has links)
This paper uses statistical learning to examine and compare three different statistical methods with the aim to predict credit card fraud. The methods compared are Logistic Regression, K-Nearest Neighbour and Random Forest. They are applied and estimated on a data set consisting of nearly 300,000 credit card transactions to determine their performance using classification of fraud as the outcome variable. The three models all have different properties and advantages. The K-NN model preformed the best in this paper but has some disadvantages, since it does not explain the data but rather predict the outcome accurately. Random Forest explains the variables but performs less precise. The Logistic Regression model seems to be unfit for this specific data set.

Page generated in 0.0415 seconds