• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 441
  • 53
  • Tagged with
  • 494
  • 489
  • 485
  • 417
  • 414
  • 412
  • 409
  • 407
  • 407
  • 166
  • 103
  • 103
  • 98
  • 89
  • 82
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
21

Evaluation of Optical Flow for Estimation of Liquid Glass Flow Velocity

Rudin, Malin January 2021 (has links)
In the glass wool industry, the molten glass flow is monitored for regulation purposes. Given the progress in the computer vision field, the current monitoring solution might be replaced by a camera based solution. The aim of this thesis is to investigate the possibility of using optical flow techniques for estimation of the molten glass flow displacement. Three glass melt flow datasets were recorded, as well as two additional melt flow datasets, using a NIR camera. The block matching techniques Full Search (FS) and Adaptive Rood Pattern Search (ARPS), as well as the local feature methods ORB and A-KAZE were considered. These four techniques were compared to RAFT, the state-of-the-art approach for optical flow estimation, using available pre-trained models, as well as an approach of using the tracking method ECO for the optical flow estimation. The methods have been evaluated using the metrics MAE, MSE, and SSIM to compare the warped flow to the target image. In addition, ground truth for 50 frames from each dataset was manually annotated as to use the optical flow metric End-Point Error. To investigate the computational complexity the average computational time per frame was calculated. The investigation found that RAFT does not perform well on the given data, due to the large displacements of the flows. For simulated displacements of up to about 100 pixels at full resolution, the performance is satisfactory, with results comparable to the traditional methods. Using ECO for optical flow estimation encounters similar problems as RAFT, where the large displacement proved challenging for the tracker. Simulating smaller motions of up to 60 pixels resulted in good performance, though computation time of the used implementation is much too high for a real-time implementation. The four traditional block matching and local feature approaches examined in this thesis outperform the state-of-the-art approaches. FS, ARPS, A-KAZE, and ORB all have similar performance on the glass flow datasets, whereas the block matching approaches fail on the alternative melt flow data as the template extraction approach is inadequate. The two local feature approaches, though working reasonably well on all datasets given full resolution, struggle to identify features on down-sampled data. This might be mitigated by fine-tuning the settings of the methods. Generally, ORB mostly outperforms A-KAZE with respect to the evaluation metrics, and is considerably faster.
22

Image-based fashion recommender systems : Considering Deep learning role in computer vision development

shirkhani, shaghayegh January 2021 (has links)
Fashion is perceived as a meaningful way of self-expressing that people use for different purposes. It seems to be an integral part of every person in modern societies, from everyday life to exceptional events and occasions. Fashionable products are highly demanded, and consequently, fashion is perceived as a desirable and profitable industry. Although this massive demand for fashion products provides an excellent opportunity for companies to invest in fashion-related sectors, it also faces different challenges in answering their customer needs. Fashion recommender systems have been introduced to address these needs. This thesis aims to provide deeper insight into the fashion recommender system domain by conducting a comprehensive literature review on more than 100 papers in this field focusing on image-based fashion recommender systems considering computer vision advancements. Justifying fashion domain-specific characteristics, the subtle notions of this domain and their relevancy have been conceptualized. Four main tasks in image-based fashion recommender systems have been recognized, including cloth-item retrievals, Complementary item recommendation, Outfit recommendation, and Capsule wardrobes. An evolvement trajectory of image-based fashion recommender systems concerning computer vision advancements has been illustrated consists of three main eras and the most recent developments. Finally, a comparison between traditional computer vision techniques and deep learning-based has been made. Although the main objective of this literature review was to perform a comprehensive, integrated overview of researches in this field, there is still a need for conducting further studies considering image-based fashion recommender systems from a more practical perspective.
23

Classification of black plastic granulates using computer vision / Classification of black plastic granulates using computer vision

Persson, Anton, Dymne, Niklas January 2021 (has links)
Pollution and climate change are some of the biggest challenges facing humanity. Moreover, for a sustainable future, recycling is needed. Plas- tic is a big part of the recycled material today, but there are problems that the recycling world is facing. The modern-day recycling facilities can handle plastics of all colours except black plastics. For this reason, most recycling companies have resorted to methods unaffected by colour, like the method used at Stena Nordic Recycling Central. The unawareness of the individual plastics causes the problem that Stena Nordic Recycling Central has to wait until an entire bag of plastic granulates has been run through the production line and sorted to test its purity using a chemistry method. Finding out if the electrostats divider settings are correct using this testing method is costly and causes many re-runs. If the divider set- ting is valid in an earlier state, it will save both time and the number of re-runs needed.This thesis aims to create a system that can classify different types of plas- tics by using image analysis. This thesis will explore two techniques to solve this problem. The two computer vision techniques will be the RGB method see 3.3.2 and machine learning see 3.3.4 using transfer learning with an AlexNet. The aim is the accuracy of at least 95% when classifying the plastics granulates.The Convolutional neural network used in this thesis is an AlexNet. The choice of method to further explore is decided in the method part of this thesis. The results of the computer vision method and RGB method were difficult to determine more about in section 4.2. It was not clear if one plastic was blacker than the other. This uncertainty and the fact that a Convolutional neural network takes more features than just RGB into a count, discussed in section 3.3, makes the computer vision method, Con- volutional neural network, a method to further explore in this thesis. The results gathered from the Convolutional neural network’s training was 95% accuracy in classifying the plastic granulates. A separate test is also needed to make sure the accuracy is close to the network accuracy. The result from the stand-alone test was 86.6% accurate, where the plastic- type Polystyrene had a subpar result of 73.3% and 100% accuracy when classifying Acrylonitrile butadiene styrene. The results from the Convo- lutional neural network show that black plastics could be classified using machine learning and could be an excellent solution for classifying and recycling black plastics if further research on the field is conducted.
24

Point Cloud Data Augmentation for Safe 3D Object Detection using Geometric Techniques

Kapoor, Shrayash January 2021 (has links)
Background: Autonomous navigation has become increasingly popular. This surge in popularity caused a lot of interest in sensor technologies, driving the cost of sensor technology down. This has resulted in increasing developments in deep learning for computer vision. There is, however, not a lot of available, adaptable research for directly performing data augmentation on point cloud data independent of the training process. This thesis focuses on the impact of point cloud augmentation techniques on 3D object detection quality. Objectives: The objectives of this thesis are to evaluate the efficiency of geometric data augmentation techniques for point cloud data. The identified techniques are then implemented on a 3D object detector, and the results obtained are then compared based on selected metrics. Methods: This thesis uses two literature reviews to find the appropriate point cloud techniques to implement for data augmentation and a 3D object detector to implement data augmentation. Subsequently, an experiment is performed to quantitatively discern how much improvement augmentation offers in the detection quality. Metrics used to compare the algorithms include precision, recall, average precision, mean average precision, memory usage and training time. Results: The literature review results indicate flipping, scaling, translation and rotation to be ideal candidates for performing geometric data augmentation and ComplexYOLO to be a capable detector for 3D object detection. Experimental results indicate that at the expense of some training time, the developed library "Aug3D" can boost the detection quality and results of the ComplexYOLO algorithm. Conclusions: After analysis of results, it was found that the implementation of geometric data augmentations (namely flipping, translation, scaling and rotation) yielded an increase of over 50% in the mean average precision for the performance of the ComplexYOLO 3D detection model on the Car and Pedestrian classes.
25

Hyperspectral Image Registration and Construction From Irregularly Sampled Data

Freij, Hannes January 2021 (has links)
Hyperspectral imaging based on the use of an exponentially variable filter gives the possibility to construct a lightweight hyperspectral sensor. The exponentially variable filter captures the whole spectral range in each image where each column captures a different wavelength. To gather the full spectrum for any given point in the image requires the fusion of several gathered images with movement in between captures. The construction of a hyperspectral cube requires registration of the gathered images. With a lightweight sensor comes the possibility to mount the hyperspectral sensor on an unmanned aerial vehicle to collect aerial footage. This thesis presents a registration algorithm capable of constructing a complete hyperspectral cube of almost any chosen area in the captured region. The thesis presents the result of a construction method using a multi-frame super-resolution algorithm trying to increase the spectral resolution and a spline interpolation method interpolating missing spectral data. The result of an algorithm trying to suggest the optimal spectral and spatial resolution before constructing the hyperspectral cube is also presented. Lastly, the result of an algorithm providing information about the quality of the constructed hyperspectral cube is also presented.
26

Domain Adaptation of Unreal Images for Image Classification / Domänöversättning av syntetiska bilder för bildklassificiering

Thornström, Johan January 2019 (has links)
Deep learning has been intensively researched in computer vision tasks like im-age classification. Collecting and labeling images that these neural networks aretrained on is labor-intensive, which is why alternative methods of collecting im-ages are of interest. Virtual environments allow rendering images and automaticlabeling,  which could speed up the process of generating training data and re-duce costs.This  thesis  studies  the  problem  of  transfer  learning  in  image  classificationwhen the classifier has been trained on rendered images using a game engine andtested on real images. The goal is to render images using a game engine to createa classifier that can separate images depicting people wearing civilian clothingor camouflage.  The thesis also studies how domain adaptation techniques usinggenerative  adversarial  networks  could  be  used  to  improve  the  performance  ofthe classifier.  Experiments show that it is possible to generate images that canbe used for training a classifier capable of separating the two classes.  However,the experiments with domain adaptation were unsuccessful.  It is instead recom-mended to improve the quality of the rendered images in terms of features usedin the target domain to achieve better results.
27

Vehicle Detection, at a Distance : Done Efficiently via Fusion of Short- and Long-Range Images / Fordonsdetektion, på avstånd

Luusua, Emil January 2020 (has links)
Object detection is a classical computer vision task, encountered in many practical applications such as robotics and autonomous driving. The latter involves serious consequences of failure and a multitude of challenging demands, including high computational efficiency and detection accuracy. Distant objects are notably difficult to detect accurately due to their small scale in the image, consisting of only a few pixels. This is especially problematic in autonomous driving, as objects should be detected at the earliest possible stage to facilitate handling of hazardous situations. Previous work has addressed small objects via use of feature pyramids and super-resolution techniques, but the efficiency of such methods is limited as computational cost increases with image resolution. Therefore, a trade-off must be made between accuracy and cost. Opportunely though, a common characteristic of driving scenarios is the predominance of distant objects in the centre of the image. Thus, the full-frame image can be downsampled to reduce computational cost, and a crop can be extracted from the image centre to preserve resolution for distant vehicles. In this way, short- and long-range images are generated. This thesis investigates the fusion of such images in a convolutional neural network, particularly the fusion level, fusion operation, and spatial alignment. A novel framework — DetSLR — is proposed for the task and examined via the aforementioned aspects. Through adoption of the framework for the well-established SSD detector and MobileNetV2 feature extractor, it is shown that the framework significantly improves upon the original detector without incurring additional cost. The fusion level is shown to have great impact on the performance of the framework, favouring high-level fusion, while only insignificant differences exist between investigated fusion operations. Finally, spatial alignment of features is demonstrated to be a crucial component of the framework.
28

Obstacle avoidance for platforms in three-dimensional environments / Kollisionsundvikande metoder för plattformar i tredimensionella miljöer

Ekström, Johan January 2016 (has links)
The field of obstacle avoidance is a well-researched area. Despite this, research on obstacle avoidance in three dimensions is surprisingly sparse. For platforms which are able to navigate three-dimensional space, such as multirotor UAVs, such methods will become more common. In this thesis, an obstacle avoidance method, intended for a three-dimensional environment, is presented. First the method reduces the dimensionality of the three-dimensional world into two dimensions by projecting obstacle observations onto a two-dimensional spherical depth map, retaining information on direction and distance to obstacles. Next, the method accounts for the dimensions of the platform by applying a post-processing on the depth map. Finally, knowing the motion model, a look-ahead verification step is taken, using information from the depth map, to ensure that the platform does not collide with any obstacles by not allowing control inputs which leads to collisions. If there are multiple control input candidates after verification that lead to velocity vectors close to a desired velocity vector, a heuristic cost function is used to select one single control input, where the similarity in direction and magnitude of the resulting and desired velocity vector is valued. Evaluation of the method reveals that platforms are able to maintain distances to obstacles. However, more work is suggested in order to improve the reliability of the method and to perform a real world evaluation. / Fältet inom kollisionsundvikande är ett välforskat område. Trots detta så är forskning inom kollisionsundvikande metoder i tre dimensioner förvånansvärt magert. För plattformar som kan navigera det tredimensionella rummet, såsom multirotor-baserade drönare kommer sådana metoder att bli mer vanliga. I denna tes presenteras en kollisionsundvikande metod, menad för det tredimensionella rummet. Först reduceras dimensionaliteten av det tredimensionella rummet genom att projicera hinderobservationer på ett tvådimensionellt sfärisk ark i form av en djupkarta som bibehåller information om riktning och avstånd till hinder. Därefter beaktas plattformens dimensioner genom att tillämpa ett efterbehandlingssteg på djupkartan. Till sist, med kunskap om rörelsemodellen, ett verifieringssteg där information från djupkartan används för att försäkra sig om att plattformen inte kolliderar med några hinder genom att inte tillåta kontrollinmatningar som leder till kollisioner. Om det finns flera kontrollinmatningskandidater efter verifikationssteget som leder till hastighetsvektorer nära en önskad hastighetsvektor så används en heuristisk kostnadsfunktion, där likheten i riktning och magnitud av den resulterande vektorn och önskade hastighetsvektorn värderas, för att välja en av dem. Utvärdering av metoden visar att plattformar kan bibehålla avstånd till hinder. Dock föreslås ytterligare arbete för att förbättra tillförlitligheten av metoden samt att utvärdera metoden i den verkliga världen.
29

Feature-Feature Matching For Object Retrieval in Point Clouds

Staniaszek, Michal January 2015 (has links)
In this project, we implement a system for retrieving instances of objects from point clouds using feature based matching techniques. The target dataset of point clouds consists of approximately 80 full scans of office rooms over a period of one month. The raw clouds are reprocessed to remove regions which are unlikely to contain objects. Using locations determined by one of several possible interest point selection methods, one of a number of descriptors is extracted from the processed clouds. Descriptors from a target cloud are compared to those from a query object using a nearest neighbour approach. The nearest neighbours of each descriptor in the query cloud are used to vote for the position of the object in a 3D grid overlaid on the room cloud. We apply clustering in the voting space and rank the clusters according to the number of votes they contain. The centroid of each of the clusters is used to extract a region from the target cloud which, in the ideal case, corresponds to the query object. We perform an experimental evaluation of the system using various parameter settings in order to investigate factors affecting the usability of the system, and the efficacy of the system in retrieving correct objects. In the best case, we retrieve approximately 50% of the matching objects in the dataset. In the worst case, we retrieve only 10%. We find that the best approach is to use a uniform sampling over the room clouds, and to use a descriptor which factors in both colour and shape information to describe points.
30

Calibration in deep-learning eye tracking / Kalibrering i djupinlärd ögonspårning

Lindén, Erik January 2021 (has links)
Personal variations severely limit the performance of appearance-based gaze tracking. Adapting to these variations using standard neural network model adaptation methods is difficult. The problems range from overfitting, due to small amounts of training data, to underfitting, due to restrictive model architectures. In this thesis, these problems are tackled by introducing the SPatial Adaptive GaZe Estimator (\spaze{}). By modeling personal variations as a low-dimensional latent parameter space, \spaze{} provides just enough adaptability to capture the range of personal variations without being prone to overfitting. Calibrating \spaze{} for a new person reduces to solving a small optimization problem. \spaze{} achieves an error of \ang{2.70} with \num{9} calibration samples on MPIIGaze, improving on the state-of-the-art by \SI{14}{\percent}. In the introductory chapters the history, methods and applications of eye tracking are reviewed, with focus on video-based eye tracking and the use of personal calibration in these methods. Emphasis is placed on methods using neural networks and the strengths and weaknesses of how these methods implement personal calibration. / <p>QC 20210528</p>

Page generated in 0.0697 seconds