• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 9
  • 5
  • 4
  • 1
  • 1
  • Tagged with
  • 23
  • 23
  • 8
  • 7
  • 6
  • 6
  • 5
  • 5
  • 4
  • 4
  • 3
  • 3
  • 3
  • 3
  • 3
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
11

Principal Point Determination for Camera Calibration

Alturki, Abdulrahman S. 24 August 2017 (has links)
No description available.
12

3D camera with built-in velocity measurement / 3D-kamera med inbyggd hastighetsmätning

Josefsson, Mattias January 2011 (has links)
In today's industry 3D cameras are often used to inspect products. The camera produces both a 3D model and an intensity image by capturing a series of profiles of the object using laser triangulation. In many of these setups a physical encoder is attached to, for example, the conveyor belt that the product is travelling on. The encoder is used to get an accurate reading of the speed that the product has when it passes through the laser. Without this, the output image from the camera can be distorted due to a variation in velocity. In this master thesis a method for integrating the functionality of this physical encoder into the software of the camera is proposed. The object is scanned together with a pattern, with the help of this pattern the object can be restored to its original proportions. / I dagens industri används ofta 3D-kameror för att inspektera produkter. Kameran producerar en 3D-modell samt en intensitetsbild genom att sätta ihop en serie av profilbilder av objektet som erhålls genom lasertriangulering. I många av dessa uppställningar används en fysisk encoder som återspeglar hastigheten på till exempel transportbandet som produkten ligger på. Utan den här encodern kan bilden som kameran fångar bli förvrängd på grund av hastighetsvariationer. I det här examensarbetet presenteras en metod för att integrera funktionaliteten av encodern in i kamerans mjukvara. För att göra detta krävs att ett mönster placeras längs med objektet som ska bli skannat. Mönstret återfinns i bilden fångad av kameran och med hjälp av detta mönster kan hastigheten bestämmas och objektets korrekta proportioner återställas.
13

Towards an efficient, unsupervised and automatic face detection system for unconstrained environments

Chen, Lihui January 2006 (has links)
Nowadays, there is growing interest in face detection applications for unconstrained environments. The increasing need for public security and national security motivated our research on the automatic face detection system. For public security surveillance applications, the face detection system must be able to cope with unconstrained environments, which includes cluttered background and complicated illuminations. Supervised approaches give very good results on constrained environments, but when it comes to unconstrained environments, even obtaining all the training samples needed is sometimes impractical. The limitation of supervised approaches impels us to turn to unsupervised approaches. In this thesis, we present an efficient and unsupervised face detection system, which is feature and configuration based. It combines geometric feature detection and local appearance feature extraction to increase stability and performance of the detection process. It also contains a novel adaptive lighting compensation approach to normalize the complicated illumination in real life environments. We aim to develop a system that has as few assumptions as possible from the very beginning, is robust and exploits accuracy/complexity trade-offs as much as possible. Although our attempt is ambitious for such an ill posed problem-we manage to tackle it in the end with very few assumptions.
14

A comparison of image processing algorithms for edge detection, corner detection and thinning

Parekh, Siddharth Avinash January 2004 (has links)
Image processing plays a key role in vision systems. Its function is to extract and enhance pertinent information from raw data. In robotics, processing of real-time data is constrained by limited resources. Thus, it is important to understand and analyse image processing algorithms for accuracy, speed, and quality. The theme of this thesis is an implementation and comparative study of algorithms related to various image processing techniques like edge detection, corner detection and thinning. A re-interpretation of a standard technique, non-maxima suppression for corner detectors was attempted. In addition, a thinning filter, Hall-Guo, was modified to achieve better results. Generally, real time data is corrupted with noise. This thesis also incorporates few smoothing filters that help in noise reduction. Apart from comparing and analysing algorithms for these techniques, an attempt was made to implement correlation-based optic flow
15

Camera Based Navigation : Matching between Sensor reference and Video image

Olgemar, Markus January 2008 (has links)
an Internal Navigational System and a Global Navigational Satellite System (GNSS). In navigational warfare the GNSS can be jammed, therefore are a third navigational system is needed. The system that has been tried in this thesis is camera based navigation. Through a video camera and a sensor reference the position is determined. This thesis will process the matching between the sensor reference and the video image. Two methods have been implemented: normalized cross correlation and position determination through a homography. Normalized cross correlation creates a correlation matrix. The other method uses point correspondences between the images to determine a homography between the images. And through the homography obtain a position. The more point correspondences the better the position determination will be. The results have been quite good. The methods have got the right position when the Euler angles of the UAV have been known. Normalized cross correlation has been the best method of the tested methods.
16

Statický model scény / Static model of scene

Sikora, Jan January 2013 (has links)
This thesis deal with various methods of background detection and with it related motion detection in a scene. It's progressing from simplest methods to more comlex. For every one are reviewed the possibilities of using and her drawbacks. In introduction are described various types of scenes according to background and foreground type e.g . according to movement objects speed or presence of movement in background. Is proposed several common or specific improvements for obtaining better background even by using simple method. Next part of work solve real situation of shaking camera. There are tested two basic methods for optical stabilization. The first is registration of images by template matching. Alternative method used interest points (corners). Both methods are closely examinate and is sought best way to match following pictures. Except shaking of camera this work deal with rotating camera and in theory solve detection background from cameras placed on ridden car. Part of work is creation database of different types scenes
17

3D monitor pomocí detekce pozice hlavy / 3D Monitor Based on Head Pose Detection

Zivčák, Jan January 2011 (has links)
With the development of posibilities of image processing, stereoscopy, prices of web cameras and power of computers an opportunity to multiply an experience with working with 3D programs showed. From the picture from webcamera an estimation of a pose of user's head can be made. According to this pose a view on 3D scene can be changed. Then, when user moves his head, he will have a feeling as if monitor was a window through which one can see the scene behind. With the system which is the result of this project it will be possible to easily and cheaply add this kind of behaviour to any 3D application.
18

Τρισδιάστατη ανακατασκευή χώρου από ένα μικρό αριθμό φωτογραφιών

Φλώρου, Ραφαέλλα, Χατούπης, Σταύρος 26 April 2012 (has links)
Η παρούσα διπλωματική εργασία αναπτύχθηκε στα πλαίσια των προπτυχιακών σπουδών του τμήματος Ηλεκτρολόγων Μηχανικών και Τεχνολογίας Υπολογιστών του Πανεπιστημίου Πατρών. Θέμα της είναι η τρισδιάστατη ανακατασκευή του χώρου από τουλάχιστον δύο φωτογραφίες του και αποτελεί μέρος του τομέα της Υπολογιστικής Όρασης. Συγκεκριμένα αναλύεται διεξοδικά η περίπτωση της στερεοσκοπικής όρασης, στην οποία η κάμερα μεταξύ δύο διαδοχικών λήψεων της ίδιας σκηνής, έχει μηδενική σχετική περιστροφή ως προς την αρχική της θέση και μικρή μετατόπιση, περίπου 5 εκατοστά. Με τον τρόπο αυτό, προσπαθούμε να προσομοιώσουμε τη λειτουργία της ανθρώπινης όρασης καθώς πολλές εφαρμογές της Τεχνητής Νοημοσύνης το κρίνουν απαραίτητο. Είναι λογικό ότι ο κάθε άνθρωπος θεωρεί τη στερεοσκοπική όραση αυτονόητη γιατί κινείται στον τρισδιάστατο κόσμο. Όταν αυτός όμως καταγράφεται από μία κάμερα, αυτόματα περνάει στο δισδιάστατο επίπεδο. Και πάλι είναι δυνατόν να εξάγουμε πληροφορίες βάθους από μία μόνο εικόνα, όμως γίνεται καθαρά εμπειρικά και βασίζεται στη σύγκριση διάφορων υφών, σχημάτων και μεγεθών. Ο ηλεκτρονικός υπολογιστής αναγνωρίζει την εικόνα σαν ένα οποιοδήποτε αρχείο. Δεν μπορεί να εξάγει κανένα συμπέρασμα για το τι απεικονίζει στον πραγματικό κόσμο. Χρειάζεται το συνδυασμό τουλάχιστον δύο εικόνων της ίδιας σκηνής από διαφορετικές θέσεις για να μπορέσει να αναγνωρίσει για παράδειγμα το βάθος της σκηνής που απεικονίζεται. Αυτή τη διαδικασία περιγράφει αναλυτικά η εργασία. Στο πρώτο κεφάλαιο εισάγουμε την έννοια και τη χρησιμότητα της στερεοσκοπικής όρασης. Στο δεύτερο κεφάλαιο παρουσιάζονται οι βασικές αρχές της προβολικής γεωμετρίας. Στο τρίτο κεφάλαιο αναφερόμαστε στη μοντελοποίηση της κάμερας και τις παραμέτρους που τη χαρακτηρίζουν. Στο τέταρτο κεφάλαιο αναλύεται η διαδικασία της βαθμονόμησης της κάμερας. Στο πέμπτο κεφάλαιο εξηγείται η διαδικασία αντιστοίχησης των σημείων ενδιαφέροντος στις δύο εικόνες. Στο έκτο κεφάλαιο αναλύονται οι βασικές αρχές της επιπολικής γεωμετρίας. Στο έβδομο κεφάλαιο παρουσιάζεται η πειραματική διαδικασία για την εύρεση του βάθους της σκηνής. Στο όγδοο κεφάλαιο παρουσιάζεται συνοπτικά η τρισδιάστατη ανακατασκευή του χώρου και παρουσιάζονται τα αντίστοιχα πειραματικά αποτελέσματα. Στο ένατο κεφάλαιο διατυπώνουμε τα συμπεράσματα της όλης διαδικασίας. Τόσο το θεωρητικό όσο και το πειραματικό μέρος αυτής της εργασίας καλύπτουν σε ένα μεγάλο ποσοστό τα βασικά στάδια ανακατασκευής του τρισδιάστατου χώρου. Τα αποτελέσματα της πειραματικής διαδικασίας αποδεικνύουν ότι οι υπάρχουσες μέθοδοι λειτουργούν ικανοποιητικά αλλά υπάρχουν πολλά περιθώρια βελτίωσης στο θέμα της Υπολογιστικής Όρασης. Στο σημείο αυτό να ευχαριστήσουμε τον επιβλέποντα καθηγητή μας κ. Δερματά για τη συνεργασία του και την κατανόησή του. / The current thesis has been written as part of the undergraduate studies for the department of Electrical and Computer Engineering of Patras University. Its objective is the three-dimensional (3D) reconstruction from two, at least, photographs, which is part of computer vision. More specifically, this thesis analyzes in detail the case of stereo vision when the camera, among two successive shots of the same image, has zero relative rotation compared to its initial position and an average translation of about 5 cm. In this way, it attempts to simulate human vision since this is essential for many Artificial Intelligence applications. Humans take stereo vision for granted since they live in a three-dimensional world. However, this world becomes two-dimensional when recorded by a camera. We can still get information about the image depth but this is empirically done based on comparing various heights, shapes and sizes. Images are identified by the computer as any other file. Computers cannot draw conclusions about what is depicted in the real world. They need to combine at least two images of the same scene and of different positions to identify the image’s depth. This process is described in the current thesis. The first chapter describes stereo vision and why it is so useful. The second chapter provides the basic principles of projective geometry, the mathematical background for passing from the two-dimensional level to the three-dimensional. The third chapter refers to camera modeling and its parameters (instrisic and extrinsic). Chapter four analyzes the camera calibration process. Chapter five explains the matching process of points of interest in both pictures. The sixth chapter provides the basic principles of epipolar geometry. The seventh chapter shows the experimental procedure that we followed in order to estimate the depth of the scene. Chapter eight shows how the 3D reconstruction is finally done. Chapter nine talks about our conclusions and how the results could improve. Both theoretical and experimental parts of this project cover the key points of 3d reconstruction. The results of the experiments show that the existing methods are satisfying but could improve more. We want to thank our supervisor professor Mr. Dermatas for his collaboration and his understanding.
19

Técnicas de clusterização baseadas em características de cor para a consulta em bancos de dados de imagens / Techniques of cluster-based features for classification of color images

Weber, Juliano Gomes 29 July 2009 (has links)
The current technologies for acquisition, storage and transmission of digital data, generate large amounts of data. This quantitative increase is directly proportional to the expansion of multimedia databases, where the bases are part of images. Factors contributing to this expansion is the generation of data access and multimedia, which are frequently used by the population through the media today. Thus, we find a clear need exists for automated systems, capable of dealing with the storage and retrieval of data in a time acceptable to the current standards. To this end, systems are designed for content retrieval of images, where the content is described through its low-level visual features such as shape, texture and color. To have such a system is considered ideal, it must be efficient and effective. The effectiveness will result from the way the information was obtained as a low level of images, considering different conditions of focus, lighting and occlusion. The efficiency is a consequence of the results obtained using the organization of information extracted. The methods of grouping are in one of the useful techniques to reduce the computational complexity of these systems, reducing the computational complexity of the methods implemented, but without losing the representation of information extracted. This work proposes a method for retrieval of images based on content, using appropriate techniques of clustering, a technique for detecting edges and a method to normalize the images in the aspect of enlightenment, to get through it the image descriptors that are robust and can be applied efficiently in a retrieval system for images by content - CBIR (Content Based Image Retrieval). / As tecnologias atuais de aquisição, armazenamento e transmissão de dados digitais geram grandes quantidades de dados. Esse aumento quantitativo é diretamente proporcional à ampliação das bases de dados multimídia, onde se inserem as bases de imagens. Fatores relevantes que contribuem para esta ampliação são o acesso e a geração de dados multimídia, os quais são freqüentemente utilizados pela população através dos meios de comunicação atuais. Desta forma, percebe-se claramente a necessidade existente por sistemas automatizados, capazes de lidar com o armazenamento e a recuperação destes dados em um tempo aceitável para os padrões atuais. Para este fim, são desenvolvidos sistemas de recuperação de imagens por conteúdo, onde este conteúdo é descrito através de suas características visuais de baixo nível, como forma, textura e cor. Para que um sistema deste tipo seja considerado ideal, ele deve ser eficiente e eficaz. A eficácia será resultado da maneira de como foram obtidas as informações de baixo nível das imagens, considerando diferentes condições de foco, oclusão e iluminação. A eficiência é conseqüência dos resultados obtidos utilizando-se a organização das informações extraídas. Os métodos de agrupamento constituem em uma das técnicas úteis para diminuir a complexidade computacional destes sistemas, uma vez que agrupa informações com características semelhantes, sob determinado critério, porém sem perder a representatividade das informações extraídas. Este trabalho propõe um método para recuperação de imagens baseada em conteúdo, que utiliza apropriadamente as técnicas de agrupamento, uma técnica de detecção de cantos e um método para normalizar as imagens no aspecto da iluminação, visando através disso obter descritores da imagem que sejam robustos e possam ser aplicados eficientemente em um sistema de recuperação de imagens por conteúdo - CBIR(Content Based Image Retrieval).
20

Scale Selection Properties of Generalized Scale-Space Interest Point Detectors

Lindeberg, Tony January 2013 (has links)
Scale-invariant interest points have found several highly successful applications in computer vision, in particular for image-based matching and recognition. This paper presents a theoretical analysis of the scale selection properties of a generalized framework for detecting interest points from scale-space features presented in Lindeberg (Int. J. Comput. Vis. 2010, under revision) and comprising: an enriched set of differential interest operators at a fixed scale including the Laplacian operator, the determinant of the Hessian, the new Hessian feature strength measures I and II and the rescaled level curve curvature operator, as well as an enriched set of scale selection mechanisms including scale selection based on local extrema over scale, complementary post-smoothing after the computation of non-linear differential invariants and scale selection based on weighted averaging of scale values along feature trajectories over scale. A theoretical analysis of the sensitivity to affine image deformations is presented, and it is shown that the scale estimates obtained from the determinant of the Hessian operator are affine covariant for an anisotropic Gaussian blob model. Among the other purely second-order operators, the Hessian feature strength measure I has the lowest sensitivity to non-uniform scaling transformations, followed by the Laplacian operator and the Hessian feature strength measure II. The predictions from this theoretical analysis agree with experimental results of the repeatability properties of the different interest point detectors under affine and perspective transformations of real image data. A number of less complete results are derived for the level curve curvature operator. / <p>QC 20121003</p> / Image descriptors and scale-space theory for spatial and spatio-temporal recognition

Page generated in 0.0882 seconds