• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 2
  • Tagged with
  • 2
  • 2
  • 2
  • 2
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
1

Narrow Pretraining of Deep Neural Networks : Exploring Autoencoder Pretraining for Anomaly Detection on Limited Datasets in Non-Natural Image Domains

Eriksson, Matilda, Johansson, Astrid January 2022 (has links)
Anomaly detection is the process of detecting samples in a dataset that are atypical or abnormal. Anomaly detection can for example be of great use in an industrial setting, where faults in the manufactured products need to be detected at an early stage. In this setting, the available image data might be from different non-natural domains, such as the depth domain. However, the amount of data available is often limited in these domains. This thesis aims to investigate if a convolutional neural network (CNN) can be trained to perform anomaly detection well on limited datasets in non-natural image domains. The attempted approach is to train the CNN as an autoencoder, in which the CNN is the encoder network. The encoder is then extracted and used as a feature extractor for the anomaly detection task, which is performed using Semantic Pyramid Anomaly Detection (SPADE). The results are then evaluated and analyzed. Two autoencoder models were used in this approach. As the encoder network, one of the models uses a MobileNetV3-Small network that had been pretrained on ImageNet, while the other uses a more basic network, which is a few layers deep and initialized with random weights. Both these networks were trained as regular convolutional autoencoders, as well as variational autoencoders. The results were compared to a MobileNetV3-Small network that had been pretrained on ImageNet, but had not been trained as an autoencoder. The models were tested on six different datasets, all of which contained images from the depth and intensity domains. Three of these datasets additionally contained images from the scatter domain, and for these datasets, the combination of all three domains was tested as well. The main focus was however on the performance in the depth domain. The results show that there is generally an improvement when training the more complex autoencoder on the depth domain. Furthermore, the basic network generally obtains an equivalent result to the more complex network, suggesting that complexity is not necessarily an advantage for this approach. Looking at the different domains, there is no apparent pattern to which domain yields the best performance. This rather seems to depend on the dataset. Lastly, it was found that training the networks as variational autoencoders did generally not improve the performance in the depth domain compared to the regular autoencoders. In summary, an improved anomaly detection was obtained in the depth domain, but for optimal anomaly detection with regard to domain and network, one must look at the individual datasets. / <p>Examensarbetet är utfört vid Institutionen för teknik och naturvetenskap (ITN) vid Tekniska fakulteten, Linköpings universitet</p>
2

Siamese Network with Dynamic Contrastive Loss for Semantic Segmentation of Agricultural Lands

Pendotagaya, Srinivas 07 1900 (has links)
This research delves into the application of semantic segmentation in precision agriculture, specifically targeting the automated identification and classification of various irrigation system types within agricultural landscapes using high-resolution aerial imagery. With irrigated agriculture occupying a substantial portion of US land and constituting a major freshwater user, the study's background highlights the critical need for precise water-use estimates in the face of evolving environmental challenges, the study utilizes advanced computer vision for optimal system identification. The outcomes contribute to effective water management, sustainable resource utilization, and informed decision-making for farmers and policymakers, with broader implications for environmental monitoring and land-use planning. In this geospatial evaluation research, we tackle the challenge of intraclass variability and a limited dataset. The research problem centers around optimizing the accuracy in geospatial analyses, particularly when confronted with intricate intraclass variations and constraints posed by a limited dataset. Introducing a novel approach termed "dynamic contrastive learning," this research refines the existing contrastive learning framework. Tailored modifications aim to improve the model's accuracy in classifying and segmenting geographic features accurately. Various deep learning models, including EfficientNetV2L, EfficientNetB7, ConvNeXtXLarge, ResNet-50, and ResNet-101, serve as backbones to assess their performance in the geospatial context. The data used for evaluation consists of high-resolution aerial imagery from the National Agriculture Imagery Program (NAIP) captured in 2015. It includes four bands (red, green, blue, and near-infrared) with a 1-meter ground sampling distance. The dataset covers diverse landscapes in Lonoke County, USA, and is annotated for various irrigation system types. The dataset encompasses diverse geographic features, including urban, agricultural, and natural landscapes, providing a representative and challenging scenario for model assessment. The experimental results underscore the efficacy of the modified contrastive learning approach in mitigating intraclass variability and improving performance metrics. The proposed method achieves an average accuracy of 96.7%, a BER of 0.05, and an mIoU of 88.4%, surpassing the capabilities of existing contrastive learning methods. This research contributes a valuable solution to the specific challenges posed by intraclass variability and limited datasets in the realm of geospatial feature classification. Furthermore, the investigation extends to prominent deep learning architectures such as Segformer, Swin Transformer, Convexnext, and Convolution Vision Transformer, shedding light on their impact on geospatial image analysis. ConvNeXtXLarge emerges as a robust backbone, demonstrating remarkable accuracy (96.02%), minimal BER (0.06), and a high MIOU (85.99%).

Page generated in 0.0626 seconds