• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 137
  • 63
  • 21
  • 14
  • 14
  • 6
  • 4
  • 3
  • 2
  • 1
  • 1
  • 1
  • 1
  • Tagged with
  • 304
  • 304
  • 100
  • 87
  • 51
  • 42
  • 38
  • 38
  • 34
  • 31
  • 30
  • 27
  • 27
  • 27
  • 25
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
61

Towards Intelligent Telerobotics: Visualization and Control of Remote Robot

Fu, Bo 01 January 2015 (has links)
Human-machine cooperative or co-robotics has been recognized as the next generation of robotics. In contrast to current systems that use limited-reasoning strategies or address problems in narrow contexts, new co-robot systems will be characterized by their flexibility, resourcefulness, varied modeling or reasoning approaches, and use of real-world data in real time, demonstrating a level of intelligence and adaptability seen in humans and animals. The research I focused is in the two sub-field of co-robotics: teleoperation and telepresence. We firstly explore the ways of teleoperation using mixed reality techniques. I proposed a new type of display: hybrid-reality display (HRD) system, which utilizes commodity projection device to project captured video frame onto 3D replica of the actual target surface. It provides a direct alignment between the frame of reference for the human subject and that of the displayed image. The advantage of this approach lies in the fact that no wearing device needed for the users, providing minimal intrusiveness and accommodating users eyes during focusing. The field-of-view is also significantly increased. From a user-centered design standpoint, the HRD is motivated by teleoperation accidents, incidents, and user research in military reconnaissance etc. Teleoperation in these environments is compromised by the Keyhole Effect, which results from the limited field of view of reference. The technique contribution of the proposed HRD system is the multi-system calibration which mainly involves motion sensor, projector, cameras and robotic arm. Due to the purpose of the system, the accuracy of calibration should also be restricted within millimeter level. The followed up research of HRD is focused on high accuracy 3D reconstruction of the replica via commodity devices for better alignment of video frame. Conventional 3D scanner lacks either depth resolution or be very expensive. We proposed a structured light scanning based 3D sensing system with accuracy within 1 millimeter while robust to global illumination and surface reflection. Extensive user study prove the performance of our proposed algorithm. In order to compensate the unsynchronization between the local station and remote station due to latency introduced during data sensing and communication, 1-step-ahead predictive control algorithm is presented. The latency between human control and robot movement can be formulated as a linear equation group with a smooth coefficient ranging from 0 to 1. This predictive control algorithm can be further formulated by optimizing a cost function. We then explore the aspect of telepresence. Many hardware designs have been developed to allow a camera to be placed optically directly behind the screen. The purpose of such setups is to enable two-way video teleconferencing that maintains eye-contact. However, the image from the see-through camera usually exhibits a number of imaging artifacts such as low signal to noise ratio, incorrect color balance, and lost of details. Thus we develop a novel image enhancement framework that utilizes an auxiliary color+depth camera that is mounted on the side of the screen. By fusing the information from both cameras, we are able to significantly improve the quality of the see-through image. Experimental results have demonstrated that our fusion method compares favorably against traditional image enhancement/warping methods that uses only a single image.
62

Top-Down Bayesian Modeling and Inference for Indoor Scenes

Del Pero, Luca January 2013 (has links)
People can understand the content of an image without effort. We can easily identify the objects in it, and figure out where they are in the 3D world. Automating these abilities is critical for many applications, like robotics, autonomous driving and surveillance. Unfortunately, despite recent advancements, fully automated vision systems for image understanding do not exist. In this work, we present progress restricted to the domain of images of indoor scenes, such as bedrooms and kitchens. These environments typically have the "Manhattan" property that most surfaces are parallel to three principal ones. Further, the 3D geometry of a room and the objects within it can be approximated with simple geometric primitives, such as 3D blocks. Our goal is to reconstruct the 3D geometry of an indoor environment while also understanding its semantic meaning, by identifying the objects in the scene, such as beds and couches. We separately model the 3D geometry, the camera, and an image likelihood, to provide a generative statistical model for image data. Our representation captures the rich structure of an indoor scene, by explicitly modeling the contextual relationships among its elements, such as the typical size of objects and their arrangement in the room, and simple physical constraints, such as 3D objects do not intersect. This ensures that the predicted image interpretation will be globally coherent geometrically and semantically, which allows tackling the ambiguities caused by projecting a 3D scene onto an image, such as occlusions and foreshortening. We fit this model to images using MCMC sampling. Our inference method combines bottom-up evidence from the data and top-down knowledge from the 3D world, in order to explore the vast output space efficiently. Comprehensive evaluation confirms our intuition that global inference of the entire scene is more effective than estimating its individual elements independently. Further, our experiments show that our approach is competitive and often exceeds the results of state-of-the-art methods.
63

3D geometrijos atstatymas panaudojant Kinect jutiklį / 3D geometry reconstruction using Kinect sensor

Udovenko, Nikita 23 July 2012 (has links)
Šiame darbe yra tiriamos atrankiojo aplinkos 3D geometrijos atstatymo galimybės panaudojant Kinect jutiklio kombinuotą vaizdo-gylio kamerą: pateikiamas matematinis atstatymo modelis, jo parametrizavimui reikalingi koeficientai, apibūdinama tikėtinų paklaidų apimtis, siūloma aktualių scenos duomenų išskyrimo iš scenos procedūra, tiriamas gaunamo modelio triukšmas ir jo pašalinimo galimybės ir metodai. Atstatyta geometrija yra pateikiama metrinėje matų sistemoje, ir kiekvienas 3D scenos taškas papildomai saugo savo spalvinę informaciją. Praktinėje dalyje pateikiama sukurta taikomoji programa yra įgyvendinta naudojant C++ ir OpenCV matematines programavimo bibliotekas. Ji atlieka 3D geometrijos atstatymą pagal pateiktą teorinį modelį, išskiria aktualius scenos duomenis, pašalina triukšmą ir gali išsaugoti gautus duomenis į 3D modeliavimo programoms suprantamą PLY formato bylą. Darbą sudaro: įvadas, 3 skyriai, išvados ir literatūros sąrašas. Darbo apimtis – 61 p. teksto be priedų, 43 paveikslai, 4 lentelės, 22 bibliografiniai šaltiniai. / The purpose of this thesis is to investigate the possibilities of selective 3D geometry reconstruction using Kinect combined image-depth camera: a mathematical reconstruction model is provided, as well as coefficients to parametrize it and estimates on expected precision; a procedure on filtering out the background from depth image is proposed, depth image noise and possibilities for its removal are studied. Resulting reconstructed geometry is provided using metric system of measurement, and each 3D point also retains it's color data. Resulting application is implemented in C++ programming language and uses OpenCV programming library. It implements 3D geometry reconstruction as described in theory section, removes background from depth image, as well as noise, and is able to save the resulting 3D geometry to a 3D modeling applications readable file format. Structure: introduction, 3 chapters, conclusions, references. Thesis consists of – 61 p. of text, 43 figures, 4 tables, 22 bibliographical entries.
64

MONOCULAR POSE ESTIMATION AND SHAPE RECONSTRUCTION OF QUASI-ARTICULATED OBJECTS WITH CONSUMER DEPTH CAMERA

Ye, Mao 01 January 2014 (has links)
Quasi-articulated objects, such as human beings, are among the most commonly seen objects in our daily lives. Extensive research have been dedicated to 3D shape reconstruction and motion analysis for this type of objects for decades. A major motivation is their wide applications, such as in entertainment, surveillance and health care. Most of existing studies relied on one or more regular video cameras. In recent years, commodity depth sensors have become more and more widely available. The geometric measurements delivered by the depth sensors provide significantly valuable information for these tasks. In this dissertation, we propose three algorithms for monocular pose estimation and shape reconstruction of quasi-articulated objects using a single commodity depth sensor. These three algorithms achieve shape reconstruction with increasing levels of granularity and personalization. We then further develop a method for highly detailed shape reconstruction based on our pose estimation techniques. Our first algorithm takes advantage of a motion database acquired with an active marker-based motion capture system. This method combines pose detection through nearest neighbor search with pose refinement via non-rigid point cloud registration. It is capable of accommodating different body sizes and achieves more than twice higher accuracy compared to a previous state of the art on a publicly available dataset. The above algorithm performs frame by frame estimation and therefore is less prone to tracking failure. Nonetheless, it does not guarantee temporal consistent of the both the skeletal structure and the shape and could be problematic for some applications. To address this problem, we develop a real-time model-based approach for quasi-articulated pose and 3D shape estimation based on Iterative Closest Point (ICP) principal with several novel constraints that are critical for monocular scenario. In this algorithm, we further propose a novel method for automatic body size estimation that enables its capability to accommodate different subjects. Due to the local search nature, the ICP-based method could be trapped to local minima in the case of some complex and fast motions. To address this issue, we explore the potential of using statistical model for soft point correspondences association. Towards this end, we propose a unified framework based on Gaussian Mixture Model for joint pose and shape estimation of quasi-articulated objects. This method achieves state-of-the-art performance on various publicly available datasets. Based on our pose estimation techniques, we then develop a novel framework that achieves highly detailed shape reconstruction by only requiring the user to move naturally in front of a single depth sensor. Our experiments demonstrate reconstructed shapes with rich geometric details for various subjects with different apparels. Last but not the least, we explore the applicability of our method on two real-world applications. First of all, we combine our ICP-base method with cloth simulation techniques for Virtual Try-on. Our system delivers the first promising 3D-based virtual clothing system. Secondly, we explore the possibility to extend our pose estimation algorithms to assist physical therapist to identify their patients’ movement dysfunctions that are related to injuries. Our preliminary experiments have demonstrated promising results by comparison with the gold standard active marker-based commercial system. Throughout the dissertation, we develop various state-of-the-art algorithms for pose estimation and shape reconstruction of quasi-articulated objects by leveraging the geometric information from depth sensors. We also demonstrate their great potentials for different real-world applications.
65

Motion Correction Structured Light using Pattern Interleaving Technique

Cavaturu, Raja Kalyan Ram 01 January 2008 (has links)
Phase Measuring Profilometry (PMP) is the most robust scanning technique for static 3D data acquisition. To make this technique robust to the target objects which are in motion during the scan interval a novel algorithm called ‘Pattern Interleaving’ is used to get a high density single scan image and making Phase Measuring Profilometry insensitive to ‘z’ motion and prevent motion banding which is predominant in 3D reconstruction when the object is in motion during the scan time
66

Robust Self-Calibration and Fundamental Matrix Estimation in 3D Computer Vision

Rastgar, Houman 30 September 2013 (has links)
The recent advances in the field of computer vision have brought many of the laboratory algorithms into the realm of industry. However, one problem that still remains open in the field of 3D vision is the problem of noise. The challenging problem of 3D structure recovery from images is highly sensitive to the presence of input data that are contaminated by errors that do not conform to ideal assumptions. Tackling the problem of extreme data, or outliers has led to many robust methods in the field that are able to handle moderate levels of outliers and still provide accurate outputs. However, this problem remains open, especially for higher noise levels and so it has been the goal of this thesis to address the issue of robustness with respect to two central problems in 3D computer vision. The two problems are highly related and they have been presented together within a Structure from Motion (SfM) context. The first, is the problem of robustly estimating the fundamental matrix from images whose correspondences contain high outlier levels. Even though this area has been extensively studied, two algorithms have been proposed that significantly speed up the computation of the fundamental matrix and achieve accurate results in scenarios containing more than 50% outliers. The presented algorithms rely on ideas from the field of robust statistics in order to develop guided sampling techniques that rely on information inferred from residual analysis. The second, problem addressed in this thesis is the robust estimation of camera intrinsic parameters from fundamental matrices, or self-calibration. Self-calibration algorithms are notoriously unreliable for general cases and it is shown that the existing methods are highly sensitive to noise. In spite of this, robustness in self-calibration has received little attention in the literature. Through experimental results, it is shown that it is essential for a real-world self-calibration algorithm to be robust. In order to introduce robustness to the existing methods, three robust algorithms have been proposed that utilize existing constraints for self-calibration from the fundamental matrix. However, the resulting algorithms are less affected by noise than existing algorithms based on these constraints. This is an important milestone since self-calibration offers many possibilities by providing estimates of camera parameters without requiring access to the image acquisition device. The proposed algorithms rely on perturbation theory, guided sampling methods and a robust root finding method for systems of higher order polynomials. By adding robustness to self-calibration it is hoped that this idea is one step closer to being a practical method of camera calibration rather than merely a theoretical possibility.
67

Monocular Obstacle Detection for Moving Vehicles

Lalonde, Jeffrey R. 18 January 2012 (has links)
This thesis presents a 3D reconstruction approach to the detection of static obstacles from a single rear view parking camera. Corner features are tracked to estimate the vehicle’s motion and to perform multiview triangulation in order to reconstruct the scene. We model the camera motion as planar motion and use the knowledge of the camera pose to efficiently solve motion parameters. Based on the observed motion, we selected snapshots from which the scene is reconstructed. These snapshots guarantee a sufficient baseline between the images and result in more robust scene modeling. Multiview triangulation of a feature is performed only if the feature obeys the epipolar constraint. Triangulated features are semantically labelled according to their 3D location. Obstacle features are spatially clustered to reduce false detections. Finally, the distance to the nearest obstacle cluster is reported to the driver.
68

Shape Estimation under General Reflectance and Transparency

Morris, Nigel Jed Wesley 31 August 2011 (has links)
In recent years there has been significant progress in increasing the scope, accuracy and flexibility of 3D photography methods. However there are still significant open problems where complex optical properties of mirroring or transparent objects cause many assumptions of traditional algorithms to break down. In this work we present three approaches that attempt to deal with some of these challenges using a few camera views and simple illumination. First, we consider the problem of reconstructing the 3D position and surface normal of points on a time-varying refractive surface. We show that two viewpoints are sufficient to solve this problem in the general case, even if the refractive index is unknown. We introduce a novel ``stereo matching'' criterion called refractive disparity, appropriate for refractive scenes, and develop an optimization-based algorithm for individually reconstructing the position and normal of each point projecting to a pixel in the input views. Second, we present a new method for reconstructing the exterior surface of a complex transparent scene with inhomogeneous interior. We capture images from each viewpoint while moving a proximal light source to a 2D or 3D set of positions giving a 2D (or 3D) dataset per pixel, called the scatter-trace. The key is that while light transport within a transparent scene's interior can be exceedingly complex, a pixel's scatter trace has a highly-constrained geometry that reveals the direct surface reflection, and leads to a simple ``Scatter-trace stereo'' algorithm for computing the exterior surface geometry. Finally, we develop a reconstruction system for scenes with reflectance properties ranging from diffuse to specular. We capture images of the scene as it is illuminated by a planar, spatially non-uniform light source. Then we show that if the source is translated to a parallel position away from the scene, a particular scene point integrates a magnified region of light from the plane. We observe this magnification at each pixel and show how it relates to the source-relative depth of the surface. Next we show how calibration relating the camera and source planes allows for robustness to specular objects and recovery of 3D surface points.
69

Τρισδιάστατη αναπαράσταση χώρου από βιντεοσκοπήσεις κινούμενου αυτοκινήτου / 3d site reconstruction from moving car video capture

Μανιανής, Διονύσιος 04 October 2011 (has links)
Σκοπός της διπλωματικής εργασίας είναι να παρουσιαστούν μερικές από τις πιο αξιοσημείωτες τεχνικές για τρισδιάστατη αναπαράσταση χώρου από βιντεοσκοπήσεις κινούμενου αυτοκινήτου. Αρχικά, το πρώτο κομμάτι είναι προσανατολισμένο στην περιγραφή των εφαρμογών που ακολουθεί η τρισδιάστατη αναπαράσταση χώρου από βιντεοσκοπήσεις και επίσης γίνεται εισαγωγή στις αρχικές έννοιες και στα βασικά μαθηματικά που χρησιμοποιούνται. Έπειτα περιγράφεται μια μέθοδος τρισδιάστατης αναπαράστασης από βιντεοσκόπηση με την χρήση απλών καμερών. Ο αλγόριθμος που περιγράφεται εδώ χρησιμοποιεί τις εικόνες από την απλή κάμερα και απαιτεί επιπλέον ότι κάθε δείκτης στο αντικείμενο συνδέεται με τουλάχιστον έναν άλλο δείκτη από σωστή απόσταση. Στο τρίτο κομμάτι της εργασίας περιγράφεται μια ακόμη μέθοδος η οποία στηρίζεται σε γραμμικά κινούμενα αντικείμενα. / The goal of this bachelor thesis is to present some of the most remarkable techniques of the three-dimensional reconstruction site from video animation moving car. At first, the first piece is oriented at the description of the application used by the three-dimensional reconstruction site of video and it is also an introduction to the initial concept and to the basic mathematics that are used. After that, a method of 3D video reconstruction using simple cameras is being described. The algorithm being described here, uses the images from the simple camera and requires that in addition that every object indication links with at least one other indicator from a known distance. At the third part one more method is described that is founded in linear moving objects.
70

An application of photogrammetry in the petrochemical industry

Singels, Wynand 03 1900 (has links)
Thesis (MScEng (Mathematical Sciences. Applied Mathematics))--Stellenbosch University, 2008. / When building or improving a petrochemical plant, drawings are used extensively in the design process. However, existing petrochemical plants seldom match their drawings, or the drawings are lost, forcing the need to generate a 3D model of the structure of the plant. In this thesis photogrammetry is investigated as a method of generating a digital 3D model of an existing plant. Camera modeling, target extraction and 3D reconstruction are discussed in detail, and a real-world system is investigated.

Page generated in 0.0453 seconds