51 |
Investigation on how presentation attack detection can be used to increase security for face recognition as biometric identification : Improvements on traditional locking systemÖberg, Fredrik January 2021 (has links)
Biometric identification has already been applied to society today, as today’s mobile phones use fingerprints and other methods like iris and the face itself. With growth for technologies like computer vision, the Internet of Things, Artificial Intelligence, The use of face recognition as a biometric identification on ordinary doors has become increasingly common. This thesis studies is looking into the possibility of replacing regular door locks with face recognition or supplement the locks to increase security by using a pre-trained state-of-the-art face recognition method based on a convolution neural network. A subsequent investigation concluded that a networks based face recognition are is highly vulnerable to attacks in the form of presentation attacks. This study investigates protection mechanisms against these forms of attack by developing a presentation attack detection and analyzing its performance. The obtained results from the proof of concept showed that local binary patterns histograms as a presentation attack detection could help the state of art face recognition to avoid attacks up to 88\% of the attacks the convolution neural network approved without the presentation attack detection. However, to replace traditional locks, more work must be done to detect more attacks in form of both higher percentage of attacks blocked by the system and the types of attack that can be done. Nevertheless, as a supplement face recognition represents a promising technology to supplement traditional door locks, enchaining their security by complementing the authorization with biometric authentication. So the main contributions is that by using simple older methods LBPH can help modern state of the art face regognition to detect presentation attacks according to the results of the tests. This study also worked to adapt this PAD to be suitable for low end edge devices to be able to adapt in an environment where modern solutions are used, which LBPH have.
|
52 |
Image Steganography Using Deep Learning TechniquesAnthony Rene Guzman (12468519) 27 April 2022 (has links)
<p>Digital image steganography is the process of embedding information withina cover image in a secure, imperceptible, and recoverable way.The three main methods of digital image steganography are spatial, transform, and neural network methods. Spatial methods modify the pixel valuesof an image to embed information, while transform methods embed hidden information within the frequency of the image.Neural network-based methods use neural networks to perform the hiding process, which is the focus of the proposed methodology.</p>
<p>This research explores the use of deep convolutional neural networks (CNNs) in digital image steganography. This work extends an existing implementation that used a two-dimensional CNN to perform the preparation, hiding, and extraction phases of the steganography process. The methodology proposed in this research, however, introduced changes into the structure of the CNN and used a gain function based on several image similarity metrics to maximize the imperceptibility between a cover and steganographic image.</p>
<p>The performance of the proposed method was measuredusing some frequently utilized image metrics such as structured similarity index measurement (SSIM), mean square error (MSE), and peak signal to noise ratio (PSNR). The results showed that the steganographic images produced by the proposed methodology areimperceptible to the human eye, while still providing good recoverability. Comparingthe results of the proposed methodologyto theresults of theoriginalmethodologyrevealed that our proposed network greatly improved over the base methodology in terms of SSIM andcompareswell to existing steganography methods.</p>
|
53 |
Deep neural networks for food waste analysis and classification : Subtraction-based methods for the case of data scarcityBrunell, David January 2022 (has links)
Machine learning generally requires large amounts of data, however data is often limited. On the whole the amount of data needed grows with the complexity of the problem to be solved. Utilising transfer learning, data augmentation and problem reduction, acceptable performance can be achieved with limited data for a multitude of tasks. The goal of this master project is to develop an artificial neural network-based model for food waste analysis, an area in which large quantities of data is not yet readily available. Given two images an algorithm is expected to identify what has changed in the image, ignore the uncharged areas even though they might contain objects which can be classified and finally classify the change. The approach chosen in this project was to attempt to reduce the problem the machine learning algorithm has to solve by subtracting the images before they are handled by the neural network. In theory this should resolve both object localisation and filtering of uninteresting objects, which only leaves classification to the neural network. Such a procedure significantly simplifies the task to be resolved by the neural network, which results in reduced need for training data as well as keeping the process of gathering data relatively simple and fast. Several models were assessed and theories of adaptation of the neural network to this particular task were evaluated. Test accuracy of at best 78.9% was achieved with a limited dataset of about 1000 images with 10 different classes. This performance was accomplished by a siamese neural network based on VGG19 utilising triplet loss and training data using subtraction as a basis for ground truth mask creation, which was multiplied with the image containing the changed object. / Maskininlärning kräver generellt mycket data, men stora mängder data står inte alltid till förfogande. Generellt ökar behovet av data med problemets komplexitet. Med hjälp av överföringsinlärning, dataaugumentation och problemreduktion kan dock acceptabel prestanda erhållas på begränsad datamängd för flera uppgifter. Målet med denna masteruppsats är att ta fram en modell baserad på artificiella neurala nätverk för matavfallsanalys, ett område inom vilket stora mängder data ännu inte finns tillgängligt. Givet två bilder väntas en algoritm identifiera vad som ändrats i bilden, ignorera de oförändrade områdena även om dessa innehåller objekt som kan klassificeras och slutligen klassificera ändringen. Tillvägagångssättet som valdes var att försöka reducera problemet som maskininlärningsalgoritmen, i detta fall ett artificiellt neuralt nätverk, behöver hantera genom att subtrahera bilderna innan de hanterades av det neurala nätverket. I teorin bör detta ta hand om både objektslokaliseringen och filtreringen av ointressanta objekt, vilket endast lämnar klassificeringen till det neurala nätverket. Ett sådant tillvägagångssätt förenklar problemet som det neurala nätverket behöver lösa avsevärt och resulterar i minskat behov av träningsdata, samtidigt som datainsamling hålls relativt snabbt och simpelt. Flera olika modeller utvärderades och teorier om specialanpassningar av neurala nätverk för denna uppgift evaluerades. En testnoggrannhet på som bäst 78.9% uppnåddes med begränsad datamängd om ca 1000 bilder med 10 klasser. Denna prestation erhölls med ett siamesiskt neuralt nätverk baserat på VGG19 med tripletförlust och träningsdata som använde subtraktion av bilder som grund för framställning av grundsanningsmasker (eng. Ground truth masks) multiplicerade med bilden innehållande förändringen.
|
54 |
Detection of pulmonary tuberculosis using deep learning convolutional neural networksNorval, Michael John 11 1900 (has links)
If Pulmonary Tuberculosis (PTB) is detected early in a patient, the greater the chances of treating
and curing the disease. Early detection of PTB could result in an overall lower mortality rate.
Detection of PTB is achieved in many ways, for instance, by using tests like the sputum culture
test. The problem is that conducting tests like these can be a lengthy process and takes up precious
time. The best and quickest PTB detection method is viewing the chest X-Ray image (CXR) of
the patient. To make an accurate diagnosis requires a qualified professional Radiologist. Neural
Networks have been around for several years but is only now making ground-breaking
advancements in speech and image processing because of the increased processing power at our
disposal. Artificial intelligence, especially Deep Learning Convolutional Neural Networks
(DLCNN), has the potential to diagnose and detect the disease immediately. If DLCNN can be
used in conjunction with the professional medical institutions, crucial time and effort can be saved.
This project aims to determine and investigate proper methods to identify and detect Pulmonary
Tuberculosis in the patient chest X-Ray images using DLCNN. Detection accuracy and success
form a crucial part of the research. Simulations on an input dataset of infected and healthy patients
are carried out. My research consists of firstly evaluating the colour depth and image resolution of
the input images. The best resolution to use is found to be 64x64. Subsequently, a colour depth of
8 bit is found to be optimal for CXR images. Secondly, building upon the optimal resolution and
colour depth, various image pre-processing techniques are evaluated. In further simulations, the
pre-processed images with the best outcome are used. Thirdly the techniques evaluated are transfer
learning, hyperparameter adjustment and data augmentation. Of these, the best results are obtained
from data augmentation. Fourthly, a proposed hybrid approach. The hybrid method is a mixture
of CAD and DLCNN using only the lung ROI images as training data. Finally, a combination of
the proposed hybrid method, coupled with augmented data and specific hyperparameter
adjustment, is evaluated. Overall, the best result is obtained from the proposed hybrid method
combined with synthetic augmented data and specific hyperparameter adjustment. / Electrical and Mining Engineering
|
55 |
Leveraging Graph Convolutional Networks for Point Cloud UpsamplingQian, Guocheng 16 November 2020 (has links)
Due to hardware limitations, 3D sensors like LiDAR often produce sparse and
noisy point clouds. Point cloud upsampling is the task of converting such point
clouds into dense and clean ones. This thesis tackles the problem of point cloud upsampling
using deep neural networks. The effectiveness of a point cloud upsampling
neural network heavily relies on the upsampling module and the feature extractor used
therein. In this thesis, I propose a novel point upsampling module, called NodeShuffle.
NodeShuffle leverages Graph Convolutional Networks (GCNs) to better encode
local point information from point neighborhoods. NodeShuffle is versatile and can
be incorporated into any point cloud upsampling pipeline. Extensive experiments
show how NodeShuffle consistently improves the performance of previous upsampling
methods. I also propose a new GCN-based multi-scale feature extractor, called Inception
DenseGCN. By aggregating features at multiple scales, Inception DenseGCN
learns a hierarchical feature representation and enables further performance gains. I
combine Inception DenseGCN with NodeShuffle into the proposed point cloud upsampling
network called PU-GCN. PU-GCN sets new state-of-art performance with
much fewer parameters and more efficient inference.
|
56 |
Mass Classification of Digital Mammograms Using Convolutional Neural NetworksFranklin, Elijah 04 May 2018 (has links)
This thesis explores the current deep learning (DL) approaches to computer aided diagnosis (CAD) of digital mammographic images and presents two novel designs for overcoming current obstacles endemic to the field, using convolutional neural networks (CNNs). The first method employed utilizes Bayesian statistics to perform decision level fusion from multiple images of an individual. The second method utilizes a new data pre-processing scheme to artificially expand the limited available training data and reduce model overitting.
|
57 |
Violin Artist Identification by Analyzing Raga-vistaram AudioRamlal, Nandakishor January 2023 (has links)
With the inception of music streaming and media content delivery platforms, there has been a tremendous increase in the music available on the internet and the metadata associated with it. In this study, we address the problem of violin artist identification, which tries to classify the performing artist based on the learned features. Even though numerous previous works studied the problem in detail and developed features and deep learning models that can be used, an interesting fact was that most studies focused on artist identification in western popular music and less on Indian classical music. For the same reason, there was no standardized dataset for this purpose. Hence, we curated a new dataset consisting of audio recordings from 6 renowned South Indian Carnatic violin artists. In this study, we explore the use of log-Mel-spectrogram feature and the embeddings generated by a pre-learned VGGish network on a Convolutional Neural Network and Convolutional Recurrent Neural Network Model. From the experiments, we observe that the Convolutional Recurrent Neural Network model trained using the log-Mel-spectrogram feature gave the optimal performance with a classification accuracy of 71.70%. / Med starten av plattformar för musikströmning och leverans av mediainnehåll har det skett en enorm ökning av musiken tillgänglig på internet och den metadata som är associerad med den. I denna studie tar vi upp problemet med fiolkonstnärsidentifikation, som försöker klassificera den utövande konstnären utifrån de inlärda dragen. Även om många tidigare verk studerade problemet i detalj och utvecklade funktioner och modeller för djupinlärning som kan användas, var ett intressant faktum att de flesta studier fokuserade på artistidentifiering i västerländsk populärmusik och mindre på indisk klassisk musik. Av samma anledning fanns det ingen standardiserad datauppsättning för detta ändamål. Därför kurerade vi en ny datauppsättning bestående av ljudinspelningar från 6 kända sydindiska karnatiska violinkonstnärer. I den här studien utforskar vi användningen av log-Melspektrogramfunktionen och inbäddningarna som genereras av ett förinlärt VGGishnätverk på ett Convolutional Neural Network och Convolutional Recurrent Neural Network Model. Från experimenten observerar vi att modellen Convolutional Recurrent Neural Network tränad med hjälp av log-Mel-spektrogramfunktionen gav optimal prestanda med en klassificeringsnoggrannhet på 71,70%.
|
58 |
Homography Estimation using Deep Learning for Registering All-22 Football Video Frames / Homografiuppskattning med deep learning för registrering av bildrutor från video av amerikansk fotbollFristedt, Hampus January 2017 (has links)
Homography estimation is a fundamental task in many computer vision applications, but many techniques for estimation rely on complicated feature extraction pipelines. We extend research in direct homography estimation (i.e. without explicit feature extraction) by implementing a convolutional network capable of estimating homographies. Previous work in deep learning based homography estimation calculates homographies between pairs of images, whereas our network takes single image input and registers it to a reference view where no image data is available. The application of the work is registering frames from American football video to a top-down view of the field. Our model manages to register frames in a test set with an average corner error equivalent to less than 2 yards. / Homografiuppskattning är ett förkrav för många problem inom datorseende, men många tekniker för att uppskatta homografier bygger på komplicerade processer för att extrahera särdrag mellan bilderna. Vi bygger på tidigare forskning inom direkt homografiuppskattning (alltså, utan att explicit extrahera särdrag) genom att implementera ett Convolutional Neural Network (CNN) kapabelt av att direkt uppskatta homografier. Arbetet tillämpas för att registrera bilder från video av amerikansk fotball till en referensvy av fotbollsplanen. Vår modell registrerar bildramer från ett testset till referensvyn med ett snittfel i bildens hörn ekvivalent med knappt 2 yards.
|
59 |
Self-supervised monocular image depth learning and confidence estimationChen, L., Tang, W., Wan, Tao Ruan, John, N.W. 17 June 2020 (has links)
No / We present a novel self-supervised framework for monocular image depth learning and confidence estimation. Our framework reduces the amount of ground truth annotation data required for training Convolutional Neural Networks (CNNs), which is often a challenging problem for the fast deployment of CNNs in many computer vision tasks. Our DepthNet adopts a novel fully differential patch-based cost function through the Zero-Mean Normalized Cross Correlation (ZNCC) to take multi-scale patches as matching and learning strategies. This approach greatly increases the accuracy and robustness of the depth learning. Whilst the proposed patch-based cost function naturally provides a 0-to-1 confidence, it is then used to self-supervise the training of a parallel network for confidence map learning and estimation by exploiting the fact that ZNCC is a normalized measure of similarity which can be approximated as the confidence of the depth estimation. Therefore, the proposed corresponding confidence map learning and estimation operate in a self-supervised manner and is a parallel network to the DepthNet. Evaluation on the KITTI depth prediction evaluation dataset and Make3D dataset show that our method outperforms the state-of-the-art results.
|
60 |
Wildfire Risk Assessment Using Convolutional Neural Networks and Modis Climate DataNesbit, Sean F 01 June 2022 (has links) (PDF)
Wildfires burn millions of acres of land each year leading to the destruction of homes and wildland ecosystems while costing governments billions in funding. As climate change intensifies drought volatility across the Western United States, wildfires are likely to become increasingly severe. Wildfire risk assessment and hazard maps are currently employed by fire services, but can often be outdated. This paper introduces an image-based dataset using climate and wildfire data from NASA’s Moderate Resolution Imaging Spectroradiometer (MODIS). The dataset consists of 32 climate and topographical layers captured across 0.1 deg by 0.1 deg tiled regions in California and Nevada between 2015 and 2020, associated with whether the region later saw a wildfire incident. We trained a convolutional neural network (CNN) with the generated dataset to predict whether a region will see a wildfire incident given the climate data of that region. Convolutional neural networks are able to find spatial patterns in their multi-dimensional inputs, providing an additional layer of inference when compared to logistic regression (LR) or artificial neural network (ANN) models. To further understand feature importance, we performed an ablation study, concluding that vegetation products, fire history, water content, and evapotranspiration products resulted in increases in model performance, while land information products did not. While the novel convolutional neural network model did not show a large improvement over previous models, it retained the highest holistic measures such as area under the curve and average precision, indicating it is still a strong competitor to existing models. This introduction of the convolutional neural network approach expands the wealth of knowledge for the prediction of wildfire incidents and proves the usefulness of the novel, image-based dataset.
|
Page generated in 0.1501 seconds