• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 57
  • 6
  • 3
  • 2
  • 2
  • 2
  • 1
  • 1
  • 1
  • Tagged with
  • 84
  • 84
  • 37
  • 35
  • 20
  • 20
  • 18
  • 18
  • 18
  • 16
  • 16
  • 14
  • 14
  • 13
  • 13
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
11

Disease Mapping with log Gaussian Cox Processes

Li, Ye 16 August 2013 (has links)
One of the main classes of spatial epidemiological studies is disease mapping, where the main aim is to describe the overall disease distribution on a map, for example, to highlight areas of elevated or lowered mortality or morbidity risk, or to identify important social or environmental risk factors adjusting for the spatial distribution of the disease. This thesis focused and proposed solutions to two most commonly seen obstacles in disease mapping applications, the changing census boundaries due to long study period and data aggregation for patients' confidentiality. In disease mapping, when target diseases have low prevalence, the study usually covers a long time period to accumulate sufficient cases. However, during this period, numerous irregular changes in the census regions on which population is reported may occur, which complicates inferences. A new model was developed for the case when the exact location of the cases is available, consisting of a continuous random spatial surface and fixed effects for time and ages of individuals. The process is modelled on a fine grid, approximating the underlying continuous risk surface with Gaussian Markov Random Field and Bayesian inference is performed using integrated nested Laplace approximations. The model was applied to clinical data on the location of residence at the time of diagnosis of new Lupus cases in Toronto, Canada, for the 40 years to 2007, with the aim of finding areas of abnormally high risk. Predicted risk surfaces and posterior exceedance probabilities are produced for Lupus and, for comparison, Psoriatic Arthritis data from the same clinic. Simulation studies are also carried out to better understand the performance of the proposed model as well as to compare with existing methods. When the exact locations of the cases are not known, inference is complicated by the uncertainty of case locations due to data aggregation on census regions for confidentiality. Conventional modelling relies on the census boundaries that are unrelated to the biological process being modelled, and may result in stronger spatial dependence in less populated regions which dominate the map. A new model was developed consisting of a continuous random spatial surface with aggregated responses and fixed covariate effects on census region levels. The continuous spatial surface was approximated by Markov random field, greatly reduces the computational complexity. The process was modelled on a lattice of fine grid cells and Bayesian inference was performed using Markov Chain Monte Carlo with data augmentation. Simulation studies were carried out to assess performance of the proposed model and to compare with the conventional Besag-York-Molli\'e model as well as model assuming exact locations are known. Receiver operating characteristic curves and Mean Integrated Squared Errors were used as measures of performance. For the application, surveillance data on the locations of residence at the time of diagnosis of syphilis cases in North Carolina, for the 9 years to 2007 are modelled with the aim of finding areas of abnormally high risk. Predicted risk surfaces and posterior exceedance probabilities are also produced, identifying Lumberton as a ``syphilis hotspot".
12

Disease Mapping with log Gaussian Cox Processes

Li, Ye 16 August 2013 (has links)
One of the main classes of spatial epidemiological studies is disease mapping, where the main aim is to describe the overall disease distribution on a map, for example, to highlight areas of elevated or lowered mortality or morbidity risk, or to identify important social or environmental risk factors adjusting for the spatial distribution of the disease. This thesis focused and proposed solutions to two most commonly seen obstacles in disease mapping applications, the changing census boundaries due to long study period and data aggregation for patients' confidentiality. In disease mapping, when target diseases have low prevalence, the study usually covers a long time period to accumulate sufficient cases. However, during this period, numerous irregular changes in the census regions on which population is reported may occur, which complicates inferences. A new model was developed for the case when the exact location of the cases is available, consisting of a continuous random spatial surface and fixed effects for time and ages of individuals. The process is modelled on a fine grid, approximating the underlying continuous risk surface with Gaussian Markov Random Field and Bayesian inference is performed using integrated nested Laplace approximations. The model was applied to clinical data on the location of residence at the time of diagnosis of new Lupus cases in Toronto, Canada, for the 40 years to 2007, with the aim of finding areas of abnormally high risk. Predicted risk surfaces and posterior exceedance probabilities are produced for Lupus and, for comparison, Psoriatic Arthritis data from the same clinic. Simulation studies are also carried out to better understand the performance of the proposed model as well as to compare with existing methods. When the exact locations of the cases are not known, inference is complicated by the uncertainty of case locations due to data aggregation on census regions for confidentiality. Conventional modelling relies on the census boundaries that are unrelated to the biological process being modelled, and may result in stronger spatial dependence in less populated regions which dominate the map. A new model was developed consisting of a continuous random spatial surface with aggregated responses and fixed covariate effects on census region levels. The continuous spatial surface was approximated by Markov random field, greatly reduces the computational complexity. The process was modelled on a lattice of fine grid cells and Bayesian inference was performed using Markov Chain Monte Carlo with data augmentation. Simulation studies were carried out to assess performance of the proposed model and to compare with the conventional Besag-York-Molli\'e model as well as model assuming exact locations are known. Receiver operating characteristic curves and Mean Integrated Squared Errors were used as measures of performance. For the application, surveillance data on the locations of residence at the time of diagnosis of syphilis cases in North Carolina, for the 9 years to 2007 are modelled with the aim of finding areas of abnormally high risk. Predicted risk surfaces and posterior exceedance probabilities are also produced, identifying Lumberton as a ``syphilis hotspot".
13

Entwicklung eines Monte-Carlo-Verfahrens zum selbständigen Lernen von Gauß-Mischverteilungen

Lauer, Martin. Unknown Date (has links) (PDF)
Universiẗat, Diss., 2004--Osnabrück.
14

Human pose augmentation for facilitating Violence Detection in videos: a combination of the deep learning methods DensePose and VioNetHuman pose augmentation for facilitating Violence Detection in videos: a combination of the deep learning methods DensePose and VioNet

Calzavara, Ivan January 2020 (has links)
In recent years, deep learning, a critical technology in computer vision, has achieved remarkable milestones in many fields, such as image classification and object detection. In particular, it has also been introduced to address the problem of violence detection, which is a big challenge considering the complexity to establish an exact definition for the phenomenon of violence. Thanks to the ever increasing development of new technologies for surveillance, we have nowadays access to an enormous database of videos that can be analyzed to find any abnormal behavior. However, by dealing with such huge amount of data it is unrealistic to manually examine all of them. Deep learning techniques, instead, can automatically study, learn and perform classification operations. In the context of violence detection, with the extraction of visual harmful patterns, it is possible to design various descriptors to represent features that can identify them. In this research we tackle the task of generating new augmented datasets in order to try to simplify the identification step performed by a violence detection technique in the field of Deep Learning. The novelty of this work is to introduce the usage of DensePose model to enrich the images in a dataset by highlighting (i.e. by identifying and segmenting) all the human beings present in them. With this approach we gained knowledge of how this algorithm performs on videos with a violent context and how the violent detection network benefit from this procedure. Performances have been evaluated from the point of view of segmentation accuracy and efficiency of the violence detection network, as well from the computational point of view. Results shows how the context of the scene is the major indicator that brings the DensePose model to correct segment human beings and how the context of violence does not seem to be the most suitable field for the application of this model since the common overlap of bodies (distinctive aspect of violence) acts as disadvantage for the segmentation. For this reason, the violence detection network does not exploit its full potential. Finally, we understood how such augmented datasets can boost up the training speed by reducing the time needed for the weights-update phase, making this procedure a helpful adds-on for implementations in different contexts where the identification of human beings still plays the major role.
15

A Study on Generative Adversarial Networks Exacerbating Social Data Bias

January 2020 (has links)
abstract: Generative Adversarial Networks are designed, in theory, to replicate the distribution of the data they are trained on. With real-world limitations, such as finite network capacity and training set size, they inevitably suffer a yet unavoidable technical failure: mode collapse. GAN-generated data is not nearly as diverse as the real-world data the network is trained on; this work shows that this effect is especially drastic when the training data is highly non-uniform. Specifically, GANs learn to exacerbate the social biases which exist in the training set along sensitive axes such as gender and race. In an age where many datasets are curated from web and social media data (which are almost never balanced), this has dangerous implications for downstream tasks using GAN-generated synthetic data, such as data augmentation for classification. This thesis presents an empirical demonstration of this phenomenon and illustrates its real-world ramifications. It starts by showing that when asked to sample images from an illustrative dataset of engineering faculty headshots from 47 U.S. universities, unfortunately skewed toward white males, a DCGAN’s generator “imagines” faces with light skin colors and masculine features. In addition, this work verifies that the generated distribution diverges more from the real-world distribution when the training data is non-uniform than when it is uniform. This work also shows that a conditional variant of GAN is not immune to exacerbating sensitive social biases. Finally, this work contributes a preliminary case study on Snapchat’s explosively popular GAN-enabled “My Twin” selfie lens, which consistently lightens the skin tone for women of color in an attempt to make faces more feminine. The results and discussion of the study are meant to caution machine learning practitioners who may unsuspectingly increase the biases in their applications. / Dissertation/Thesis / Masters Thesis Computer Science 2020
16

Machine Learning on Acoustic Signals Applied to High-Speed Bridge Deck Defect Detection

Chou, Yao 06 December 2019 (has links)
Machine learning techniques are being applied to many data-intensive problems because they can accurately provide classification of complex data using appropriate training. Often, the performance of machine learning can exceed the performance of traditional techniques because machine learning can take advantage of higher dimensionality than traditional algorithms. In this work, acoustic data sets taken using a rapid scanning technique on concrete bridge decks provided an opportunity to both apply machine learning algorithms to improve detection performance and also to investigate the ways that training of neural networks can be aided by data augmentation approaches. Early detection and repair can enhance safety and performance as well as reduce long-term maintenance costs of concrete bridges. In order to inspect for non-visible internal cracking (called delaminations) of concrete bridges, a rapid inspection method is needed. A six-channel acoustic impact-echo sounding apparatus is used to generate large acoustic data sets on concrete bridge decks at high speeds. A machine learning data processing architecture is described to accurately detect and map delaminations based on the acoustic responses. The machine learning approach achieves accurate results at speeds between 25 and 45 km/h across a bridge deck and successfully demonstrates the use of neural networks to analyze this type of acoustic data. In order to obtain excellent performance, model training generally requires large data sets. However, in many potentially interesting cases, such as bridge deck defect detection, acquiring enough data for training can be difficult. Data augmentation can be used to increase the effective size of the training data set. Acoustic signal data augmentation is demonstrated in conjunction with a machine learning model for acoustic defect detection on bridge decks. Four different augmentation methods are applied to data using two different augmentation strategies. This work demonstrates that a "goldilocks" data augmentation approach can be used to increase machine learning performance when only a limited data set is available. The major technical contributions of this work include application of machine learning to acoustic data sets relevant to bridge deck inspection, solving an important problem in the field of nondestructive evaluation, and a more generalized approach to data augmentation of limited acoustic data sets to expand the classes of acoustic problems that machine learning can successfully address.
17

Simulating Artificial Recombination for a Deep Convolutional Autoencoder

Levin, Fredrik January 2021 (has links)
Population structure is an important field of study due to its importance in finding underlying genetics of various diseases.This is why this thesis has looked at a newly presented deep convolutional autoencoder that has been showing promising results when compared to the state-of-the-art method for quantifying genetic similarities within population structure. The main focus was to introduce data augmentation in the form of artificial diploid recombination to this autoencoder in an attempt to increase performance and robustness of the network structure.  The training data for the network consist of arrays containing information about single-nucleotide polymorphisms present in an individual. Each instance of augmented data was simulated by randomising cuts based on the distance between the polymorphisms, and then creating a new array by alternating between the arrays of two randomised original data instances. Several networks were then trained using this data augmentation. The performance of the trained networks was compared to networks trained on only original data using several metrics. Both groups of networks had similar performance for most metrics. The main difference was that networks trained on only original data had a low genotype concordance on simulated data. This indicates an underlying risk using the original networks, which can be overcome by introducing the artificial recombination.
18

A Data Augmentation Methodology for Class-imbalanced Image Processing in Prognostic and Health Management

Yang, Shaojie January 2020 (has links)
No description available.
19

Cycling Safety Data Augmentation in the Urban Environment

Costa, Miguel, Roque, Carlos, Marques, Manuel, Moura, Filipe 02 January 2023 (has links)
Cities plan to revitalize sustainable transportation, with a key emphasis on cycling. However, cities need to provide safe environments for cyclists through better infrastru.cture design. education programs, or other interventions to increase cycling nwnbers, as safety concerns greatly discourage people from cycling. Thus, cities' strategies aim to protect and improve the safety ofthose who cycle. Here, cycling research con1ributes to understanding cycling and what factors related to the individua4 the bicycle, and the surrounding environm.ent, in.fluence the risk. cyclists face. Objective cycling safety goals are to i) decrease the outcome severity of accidents involving cyclists and ii) decrease the overall nwnber of accidents. lt is often based on accident records or police reports, yet most incidents are often not reported. Nevertheless, accident statistics are vital because they allow for factors such as demographic data and built environment tobe analyzed to understand cyclists' risk. ofbeing involved or injured in an accident. There is a worldwide need for more data about cycling accidents, their context, and the built environm.ent's influence. Hence complete datasets are required. We mak:e use of CYCLANDS - a collection of 30 datasets comprising 1.58M cycling accident records - to explore how other data and analysis can complement accident records. Thus, a subset of CYCLANDS was augmented to analyze circulation spaces around accident locations. We hope this takes a step in that direction, fostering the mix of authoritative and volunteered data and providing a more complete data set.
20

Synthesis of Pediatric Brain Tumor Image With Mass Effect / Syntes av pediatrisk hjärntumörbild med masseffekt

Zhou, Yu January 2022 (has links)
During the last few years, deep learning-based techniques have made much progress in the medical image processing field, such as segmentation and registration. The main characteristic of these methods is the large demand of medical images to do model training. However, the acquisition of these data is often difficult, due to the high expense and ethical issues. As a consequence, the lack of data may lead to poor performance and overfitting. To tackle this problem, we propose a data augmentation algorithm in this paper to inpaint the tumor on healthy pediatric brain MRI images to simulate pathological images. Since the growth of tumor may cause deformation and edema of the surrounding tissues which is called the 'mass effect', a probabilistic UNet is applied to mimic this deformation field. Then, instead of directly adding the tumor to the image, the GAN-based method is applied to transfer the mask to the image and make it more plausible, both visually and anatomically. Meanwhile, the annotations of the different brain tissues are also obtained by employing the deformation field to the original labels. Finally, the synthesized image together with the real dataset is trained to do the tumor segmentation task, and the results indicate a statistical improvement in accuracy.

Page generated in 0.0348 seconds