• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 446
  • 42
  • 12
  • 9
  • 6
  • 2
  • 1
  • 1
  • 1
  • 1
  • Tagged with
  • 546
  • 546
  • 458
  • 449
  • 446
  • 445
  • 443
  • 439
  • 439
  • 151
  • 97
  • 83
  • 81
  • 73
  • 70
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
211

A feature based face tracker using extended Kalman filtering

Ingemars, Nils January 2007 (has links)
A face tracker is exactly what it sounds like. It tracks a face in a video sequence. Depending on the complexity of the tracker, it could track the face as a rigid object or as a complete deformable face model with face expressions. This report is based on the work of a real time feature based face tracker. Feature based means that you track certain features in the face, like points with special characteristics. It might be a mouth or eye corner, but theoretically it could be any point. For this tracker, the latter is of interest. Its task is to extract global parameters, i.e. rotation and translation, as well as dynamic facial parameters (expressions) for each frame. It tracks feature points using motion between frames and a textured face model (Candide). It then uses an extended Kalman filter to estimate the parameters from the tracked feature points.
212

Evaluation of tone mapping operators for use in real time environments

Hellsten, Jonas January 2007 (has links)
As real time visualizations become more realistic it also becomes more important to simulate the perceptual effects of the human visual system. Such effects include the response to varying illumination, glare and differences between photopic and scotopic vision. This thesis evaluates several different tone mapping methods to allow a greater dynamic range to be used in real time visualisations. Several tone mapping methods have been implemented in the Avalanche Game Engine and evaluated using a small test group. To increase immersion in the visualization several filters aimed to simulate perceptual effects has also been implemented. The primary goal of these filters is to simulate scotopic vision. The tests showed that two tone mapping methods would be suitable for the environment used in the tests. The S-curve tone mapping method gave the best result while the Mean Value method gave good results while being the simplest to implement and the cheapest. The test subjects agreed that the simulation of scotopic vision enhanced the immersion in a visualization. The primary difficulties in this work has been lack of dynamic range in the input images and the challenges in coding real time graphics using a graphics processing unit.
213

Facial Features Tracking using Active Appearance Models

Fanelli, Gabriele January 2006 (has links)
This thesis aims at building a system capable of automatically extracting and parameterizing the position of a face and its features in images acquired from a low-end monocular camera. Such a challenging task is justified by the importance and variety of its possible applications, ranging from face and expression recognition to animation of virtual characters using video depicting real actors. The implementation includes the construction of Active Appearance Models of the human face from training images. The existing face model Candide-3 is used as a starting point, making the translation of the tracking parameters to standard MPEG-4 Facial Animation Parameters easy. The Inverse Compositional Algorithm is employed to adapt the models to new images, working on a subspace where the appearance is "projected out" and thus focusing only on shape. The algorithm is tested on a generic model, aiming at tracking different people’s faces, and on a specific model, considering one person only. In the former case, the need for improvements in the robustness of the system is highlighted. By contrast, the latter case gives good results regarding both quality and speed, with real time performance being a feasible goal for future developments.
214

Implementation and Validation of Independent Vector Analysis

Claesson, Kenji January 2010 (has links)
This Master’s Thesis was part of the project called Multimodalanalysis at the Depart-ment of Biomedical Engineering and Informatics at the Ume˚ University Hospital inUme˚ Sweden. The aim of the project is to develop multivariate measurement anda,analysis methods of the skeletal muscle physiology. One of the methods used to scanthe muscle is functional ultrasound. In a study performed by the project group datawas aquired, where test subjects were instructed to follow a certain exercise scheme,which was measured. Since there currently is no superior method to analyze the result-ing data (in form of ultrasound video sequences) several methods are being looked at.One considered method is called Independent Vector Analysis (IVA). IVA is a statisticalmethod to find independent components in a mix of components. This Master’s Thesisis about segmenting and analyzing the ultrasound images with help of IVA, to validateif it is a suitable method for this kind of tasks.First the algorithm was tested on generated mixed data to find out how well itperformed. The results were very accurate, considering that the method only usesapproximations. Some expected variation from the true value occured though.When the algorithm was considered performing to satisfactory, it was tested on thedata gathered by the study and the result can very well reflect an approximation of truesolution, since the resulting segmented signals seem to move in a possible way. But themethod has weak sides (which have been tried to be minimized) and all error analysishas been done by human eye, which definitly is a week point. But for the time being itis more important to analyze trends in the signals, rather than analyze exact numbers.So as long as the signals behave in a realistic way the result can not be said to becompletley wrong. So the overall results of the method were deemed adequate for the application at hand. / Multimodalanalys
215

Structure from Forward Motion / 3D-struktur från framåtrörelse

Svensson, Fredrik January 2010 (has links)
This master thesis investigates the difficulties of constructing a depth map using one low resolution grayscale camera mounted in the front of a car. The goal is to produce a depth map in real-time to assist other algorithms in the safety system of a car. This has been shown to be difficult using the evaluated combination of camera position and choice of algorithms. The main problem is to estimate an accurate optical flow. Another problem is to handle moving objects. The conclusion is that the implementations, mainly triangulation of corresponding points tracked using a Lucas Kanade tracker, provide information of too poor quality to be useful for the safety system of a car. / I detta examensarbete undersöks svårigheterna kring att skapa en djupbild från att endast använda en lågupplöst gråskalekamera monterad framtill i en bil. Målet är att producera en djupbild i realtid som kan nyttjas i andra delar av bilens säkerhetssystem. Detta har visat sig vara svårt att lösa med den undersökta kombinationen av kameraplacering och val av algoritmer. Det huvudsakliga problemet är att räkna ut ett noggrant optiskt flöde. Andra problem härrör från objekt som rör på sig. Slutsatsen är att implementationerna, mestadels triangulering av korresponderande punktpar som följts med hjälp av en Lucas Kanade-följare, ger resultat av för dålig kvalitet för att vara till nytta för bilens säkerhetssystem.
216

Liver Tumor Segmentation Using Level Sets and Region Growing

Thomasson, Viola January 2011 (has links)
Medical imaging is an important tool for diagnosis and treatment planning today. However as the demand for efficiency increases at the same time as the data volumes grow immensely, the need for computer assisted analysis, such as image segmentation, to help and guide the practitioner increases. Medical image segmentation could be used for various different tasks, the localization and delineation of pathologies such as cancer tumors is just one example. Numerous problems with noise and image artifacts in the generated images make the segmentation a difficult task, and the developer is forced to choose between speed and performance. In clinical practise, however, this is impossible as both speed and performance are crucial. One solution to this problem might be to involve the user more in the segmentation, using interactivite algorithms where the user might influence the segmentation for an improved result. This thesis has concentrated on finding a fast and interactive segmentation method for liver tumor segmentation. Various different methods were explored, and a few were chosen for implementation and further development. Two methods appeared to be the most promising, Bayesian Region Growing (BRG) and Level Set. An interactive Level Set algorithm emerged as the best alternative for the interactivity of the algorithm, and could be used in combination with both BRG and Level Set. A new data term based on a probability model instead of image edges was also explored for the Level Set-method, and proved to be more promising than the original one. The probability based Level Set and the BRG method both provided good quality results, but the fastest of the two was the BRG-method, which could segment a tumor present in 25 CT image slices in less than 10 seconds when implemented in Matlab and mex-C++ code on an ACPI x64-based PC with two 2.4 GHz Intel(R) Core(TM) 2CPU and 8 GB RAM memory. The interactive Level Set could be succesfully used as an interactive addition to the automatic method, but its usefulness was somewhat reduced by its slow processing time ( 1.5 s/slice) and the relative complexity of the needed user interactions.
217

An early fire detection system through registration and analysis of waste station IR-images / Tidig brandetektion vid avfallsbunkrar via registrering och analys av IR-bilder

Söderström, Rikard January 2011 (has links)
In this thesis, an investigation was performed to find ways of differencing between firesand vehicles at waste stations in hope of removing vehicles as a source of error duringearly fire detection. The existing system makes use of a heat camera, which rotates in 48different angles (also known as zones) in a fixed position. If the heat is above a certainvalue within a zone the system sounds the fire alarm.The rotation of the camera results in an unwanted displacement between two successiveframes within the same zone. By use of image registration, this displacement wasremoved. After the registration of an image, segmentation was performed where coldobjects are eliminated as an error source. Lastly, an analysis was performed upon thewarm objects.At the end, it was proven that the image registration had been a successful improvementof the existing system. It was also shown that vehicles can, to some extent, beeliminated as an error source. / I denna uppsats görs en undersökning av sätt att urskilja mellan bränder och fordon vid avfallsbunkrar, i hopp om att ta bortfordon som felkälla under tidig branddetektion. Dagens system använder sig av en värmekamera som roterar i 48 vinklar(även kallade zoner) från en fix position och larmar då det blir för varmt i någon zon.Roteringen av kameran medför en icke önskvärd förskjutning mellan två efterföljande bilder inom samma zon. Processenbildregistrering används för att eliminera denna förskjutning. Efter registreringen utförs en segmentering där kalla objekt tasbort som felkälla. När detta är utfört görs en analys av de varma objekten med en mängd mätningar.I slutet bevisas att registreringen har fungerat mycket väl, likaså att det går till viss del att eliminera fordon som felkällaunder tidig brandetektion.
218

Topics in Content Based Image Retrieval : Fonts and Color Emotions

Solli, Martin January 2009 (has links)
Two novel contributions to Content Based Image Retrieval are presented and discussed. The first is a search engine for font recognition. The intended usage is the search in very large font databases. The input to the search engine is an image of a text line, and the output is the name of the font used when printing the text. After pre-processing and segmentation of the input image, a local approach is used, where features are calculated for individual characters. The method is based on eigenimages calculated from edge filtered character images, which enables compact feature vectors that can be computed rapidly. A system for visualizing the entire font database is also proposed. Applying geometry preserving linear- and non-linear manifold learning methods, the structure of the high-dimensional feature space is mapped to a two-dimensional representation, which can be reorganized into a grid-based display. The performance of the search engine and the visualization tool is illustrated with a large database containing more than 2700 fonts. The second contribution is the inclusion of color-based emotion-related properties in image retrieval. The color emotion metric used is derived from psychophysical experiments and uses three scales: activity, weight and heat. It was originally designed for single-color combinations and later extended to include pairs of colors. A modified approach for statistical analysis of color emotions in images, involving transformations of ordinary RGB-histograms, is used for image classification and retrieval. The methods are very fast in feature extraction, and descriptor vectors are very short. This is essential in our application where the intended use is the search in huge image databases containing millions or billions of images. The proposed method is evaluated in psychophysical experiments, using both category scaling and interval scaling. The results show that people in general perceive color emotions for multi-colored images in similar ways, and that observer judgments correlate with derived values. Both the font search engine and the emotion based retrieval system are implemented in publicly available search engines. User statistics gathered during a period of 20 respectively 14 months are presented and discussed.
219

Colorimetric and Multispectral Image Acquisition

Nyström, Daniel January 2006 (has links)
The trichromatic principle of representing color has for a long time been dominating in color imaging. The reason is the trichromatic nature of human color vision, but as the characteristics of typical color imaging devices are different from those of human eyes, there is a need to go beyond the trichromatic approach. The interest for multi-channel imaging, i.e. increasing the number of color channels, has made it an active research topic with a substantial potential of application. To achieve consistent color imaging, one needs to map the imaging-device data to the device-independent colorimetric representations CIEXYZ or CIELAB, the key concept of color management. As the color coordinates depend not only on the reflective spectrum of the object but also on the spectral properties of the illuminant, the colorimetric representation suffers from metamerism, i.e. objects of the same color under a specific illumination may appear different when they are illuminated by other light sources. Furthermore, when the sensitivities of the imaging device differ from the CIE color matching functions, two spectra that appear different for human observers may result in identical device response. On contrary, in multispectral imaging, color is represented by the object’s physical characteristics namely the spectrum which is illuminant independent. With multispectral imaging, different spectra are readily distinguishable, no matter they are metameric or not. The spectrum can then be transformed to any color space and be rendered under any illumination. The focus of the thesis is high quality image-acquisition in colorimetric and multispectral formats. The image acquisition system used is an experimental system with great flexibility in illumination and image acquisition setup. Besides the conventional trichromatic RGB filters, the system also provides the possibility of acquiring multi-channel images, using 7 narrowband filters. A thorough calibration and characterization of all the components involved in the image acquisition system is carried out. The spectral sensitivity of the CCD camera, which can not be derived by direct measurements, is estimated using least squares regression, optimizing the camera response to measured spectral reflectance of carefully selected color samples. To derive mappings to colorimetric and multispectral representations, two conceptually different approaches are used. In the model-based approach, the physical model describing the image acquisition process is inverted, to reconstruct spectral reflectance from the recorded device response. In the empirical approach, the characteristics of the individual components are ignored, and the functions are derived by relating the device response for a set of test colors to the corresponding colorimetric and spectral measurements, using linear and polynomial least squares regression. The results indicate that for trichromatic imaging, accurate colorimetric mappings can be derived by the empirical approach, using polynomial regression to CIEXYZ and CIELAB. Because of the media-dependency, the characterization functions should be derived for each combination of media and colorants. However, accurate spectral data reconstruction requires for multi-channel imaging, using the model-based approach. Moreover, the model-based approach is general, since it is based on the spectral characteristics of the image acquisition system, rather than the characteristics of a set of color samples. / Report code: LiU-TEK-LIC- 2006:70
220

Multiple Session 3D Reconstruction using RGB-D Cameras / 3D-rekonstruktion med RGB-D kamera över multipla sessioner

Widebäck West, Nikolaus January 2014 (has links)
In this thesis we study the problem of multi-session dense rgb-d slam for 3D reconstruc- tion. Multi-session reconstruction can allow users to capture parts of an object that could not easily be captured in one session, due for instance to poor accessibility or user mistakes. We first present a thorough overview of single-session dense rgb-d slam and describe the multi-session problem as a loosening of the incremental camera movement and static scene assumptions commonly held in the single-session case. We then implement and evaluate sev- eral variations on a system for doing two-session reconstruction as an extension to a single- session dense rgb-d slam system. The extension from one to several sessions is divided into registering separate sessions into a single reference frame, re-optimizing the camera trajectories, and fusing together the data to generate a final 3D model. Registration is done by matching reconstructed models from the separate sessions using one of two adaptations on a 3D object detection pipeline. The registration pipelines are evaluated with many different sub-steps on a challenging dataset and it is found that robust registration can be achieved using the proposed methods on scenes without degenerate shape symmetry. In particular we find that using plane matches between two sessions as constraints for as much as possible of the registration pipeline improves results. Several different strategies for re-optimizing camera trajectories using data from both ses- sions are implemented and evaluated. The re-optimization strategies are based on re- tracking the camera poses from all sessions together, and then optionally optimizing over the full problem as represented on a pose-graph. The camera tracking is done by incrementally building and tracking against a tsdf volume, from which a final 3D mesh model is extracted. The whole system is qualitatively evaluated against a realistic dataset for multi-session re- construction. It is concluded that the overall approach is successful in reconstructing objects from several sessions, but that other fine grained registration methods would be required in order to achieve multi-session reconstructions that are indistinguishable from singe-session results in terms of reconstruction quality.

Page generated in 0.0749 seconds