• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 151
  • 40
  • 15
  • 11
  • 10
  • 5
  • 4
  • 4
  • 4
  • 3
  • 2
  • 1
  • 1
  • 1
  • 1
  • Tagged with
  • 277
  • 277
  • 107
  • 65
  • 58
  • 54
  • 50
  • 44
  • 42
  • 40
  • 39
  • 38
  • 37
  • 35
  • 35
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
221

Optimierter Einsatz eines 3D-Laserscanners zur Point-Cloud-basierten Kartierung und Lokalisierung im In- und Outdoorbereich / Optimized use of a 3D laser scanner for point-cloud-based mapping and localization in indoor and outdoor areas

Schubert, Stefan 05 March 2015 (has links) (PDF)
Die Kartierung und Lokalisierung eines mobilen Roboters in seiner Umgebung ist eine wichtige Voraussetzung für dessen Autonomie. In dieser Arbeit wird der Einsatz eines 3D-Laserscanners zur Erfüllung dieser Aufgaben untersucht. Durch die optimierte Anordnung eines rotierenden 2D-Laserscanners werden hochauflösende Bereiche vorgegeben. Zudem wird mit Hilfe von ICP die Kartierung und Lokalisierung im Stillstand durchgeführt. Bei der Betrachtung zur Verbesserung der Bewegungsschätzung wird auch eine Möglichkeit zur Lokalisierung während der Bewegung mit 3D-Scans vorgestellt. Die vorgestellten Algorithmen werden durch Experimente mit realer Hardware evaluiert.
222

Point Cloud Registration in Augmented Reality using the Microsoft HoloLens

Kjellén, Kevin January 2018 (has links)
When a Time-of-Flight (ToF) depth camera is used to monitor a region of interest, it has to be mounted correctly and have information regarding its position. Manual configuration currently require managing captured 3D ToF data in a 2D environment, which limits the user and might give rise to errors due to misinterpretation of the data. This thesis investigates if a real time 3D reconstruction mesh from a Microsoft HoloLens can be used as a target for point cloud registration using the ToF data, thus configuring the camera autonomously. Three registration algorithms, Fast Global Registration (FGR), Joint Registration Multiple Point Clouds (JR-MPC) and Prerejective RANSAC, were evaluated for this purpose. It was concluded that despite using different sensors it is possible to perform accurate registration. Also, it was shown that the registration can be done accurately within a reasonable time, compared with the inherent time to perform 3D reconstruction on the Hololens. All algorithms could solve the problem, but it was concluded that FGR provided the most satisfying results, though requiring several constraints on the data.
223

Detekce a sledování polohy hlavy v obraze / Head Pose Estimation and Tracking

Pospíšil, Aleš January 2011 (has links)
Diplomová práce je zaměřena na problematiku detekce a sledování polohy hlavy v obraze jako jednu s možností jak zlepšit možnosti interakce mezi počítačem a člověkem. Hlavním přínosem diplomové práce je využití inovativních hardwarových a softwarových technologií jakými jsou Microsoft Kinect, Point Cloud Library a CImg Library. Na úvod je představeno shrnutí předchozích prací na podobné téma. Následuje charakteristika a popis databáze, která byla vytvořena pro účely diplomové práce. Vyvinutý systém pro detekci a sledování polohy hlavy je založený na akvizici 3D obrazových dat a registračním algoritmu Iterative Closest Point. V závěru diplomové práce je nabídnuto hodnocení vzniklého systému a jsou navrženy možnosti jeho budoucího zlepšení.
224

Deep Learning Semantic Segmentation of 3D Point Cloud Data from a Photon Counting LiDAR / Djupinlärning för semantisk segmentering av 3D punktmoln från en fotonräknande LiDAR

Süsskind, Caspian January 2022 (has links)
Deep learning has shown to be successful on the task of semantic segmentation of three-dimensional (3D) point clouds, which has many interesting use cases in areas such as autonomous driving and defense applications. A common type of sensor used for collecting 3D point cloud data is Light Detection and Ranging (LiDAR) sensors. In this thesis, a time-correlated single-photon counting (TCSPC) LiDAR is used, which produces very accurate measurements over long distances up to several kilometers. The dataset collected by the TCSPC LiDAR used in the thesis contains two classes, person and other, and it comes with several challenges due to it being limited in terms of size and variation, as well as being extremely class imbalanced. The thesis aims to identify, analyze, and evaluate state-of-the-art deep learning models for semantic segmentation of point clouds produced by the TCSPC sensor. This is achieved by investigating different loss functions, data variations, and data augmentation techniques for a selected state-of-the-art deep learning architecture. The results showed that loss functions tailored for extremely imbalanced datasets performed the best with regard to the metric mean intersection over union (mIoU). Furthermore, an improvement in mIoU could be observed when some combinations of data augmentation techniques were employed. In general, the performance of the models varied heavily, with some achieving promising results and others achieving much worse results.
225

Deep Learning for Semantic Segmentation of 3D Point Clouds from an Airborne LiDAR / Semantisk segmentering av 3D punktmoln från en luftburen LiDAR med djupinlärning

Serra, Sabina January 2020 (has links)
Light Detection and Ranging (LiDAR) sensors have many different application areas, from revealing archaeological structures to aiding navigation of vehicles. However, it is challenging to interpret and fully use the vast amount of unstructured data that LiDARs collect. Automatic classification of LiDAR data would ease the utilization, whether it is for examining structures or aiding vehicles. In recent years, there have been many advances in deep learning for semantic segmentation of automotive LiDAR data, but there is less research on aerial LiDAR data. This thesis investigates the current state-of-the-art deep learning architectures, and how well they perform on LiDAR data acquired by an Unmanned Aerial Vehicle (UAV). It also investigates different training techniques for class imbalanced and limited datasets, which are common challenges for semantic segmentation networks. Lastly, this thesis investigates if pre-training can improve the performance of the models. The LiDAR scans were first projected to range images and then a fully convolutional semantic segmentation network was used. Three different training techniques were evaluated: weighted sampling, data augmentation, and grouping of classes. No improvement was observed by the weighted sampling, neither did grouping of classes have a substantial effect on the performance. Pre-training on the large public dataset SemanticKITTI resulted in a small performance improvement, but the data augmentation seemed to have the largest positive impact. The mIoU of the best model, which was trained with data augmentation, was 63.7% and it performed very well on the classes Ground, Vegetation, and Vehicle. The other classes in the UAV dataset, Person and Structure, had very little data and were challenging for most models to classify correctly. In general, the models trained on UAV data performed similarly as the state-of-the-art models trained on automotive data.
226

Optimierter Einsatz eines 3D-Laserscanners zur Point-Cloud-basierten Kartierung und Lokalisierung im In- und Outdoorbereich

Schubert, Stefan 30 September 2014 (has links)
Die Kartierung und Lokalisierung eines mobilen Roboters in seiner Umgebung ist eine wichtige Voraussetzung für dessen Autonomie. In dieser Arbeit wird der Einsatz eines 3D-Laserscanners zur Erfüllung dieser Aufgaben untersucht. Durch die optimierte Anordnung eines rotierenden 2D-Laserscanners werden hochauflösende Bereiche vorgegeben. Zudem wird mit Hilfe von ICP die Kartierung und Lokalisierung im Stillstand durchgeführt. Bei der Betrachtung zur Verbesserung der Bewegungsschätzung wird auch eine Möglichkeit zur Lokalisierung während der Bewegung mit 3D-Scans vorgestellt. Die vorgestellten Algorithmen werden durch Experimente mit realer Hardware evaluiert.
227

Application development of 3D LiDAR sensor for display computers

Ekstrand, Oskar January 2023 (has links)
A highly accurate sensor for measuring distances, used for creating high-resolution 3D maps of the environment, utilize “Light Detection And Ranging” (LiDAR) technology. This degree project aims to investigate the implementation of 3D LiDAR sensors into off-highway vehicle display computers, called CCpilots. This involves a study of available low-cost 3D LiDAR sensors on the market and development of an application for visualizing real time data graphically, with room for optimization algorithms. The selected LiDAR sensor is “Livox Mid-360”, a hybrid-solid technology and a field of view of 360° horizontally and 59° vertically. The LiDAR application was developed using Livox SDK2 combined with a C++ back-end, in order to visualize data using Qt QML as the Graphical User Interface design tool. A filter was utilized from the Point Cloud Library (PCL), called a voxel grid filter, for optimization purpose. Real time 3D LiDAR sensor data was graphically visualized on the display computer CCpilot X900. The voxel grid filter had a few visual advantages, although it consumed more processor power compared to when no filter was used. Whether a filter was used or not, all points generated by the LiDAR sensor could be processed and visualized by the developed application without any latency.
228

Modélisation 3D automatique d'environnements : une approche éparse à partir d'images prises par une caméra catadioptrique / Automatic 3d modeling of environments : a sparse approach from images taken by a catadioptric camera

Yu, Shuda 03 June 2013 (has links)
La modélisation 3d automatique d'un environnement à partir d'images est un sujet toujours d'actualité en vision par ordinateur. Ce problème se résout en général en trois temps : déplacer une caméra dans la scène pour prendre la séquence d'images, reconstruire la géométrie, et utiliser une méthode de stéréo dense pour obtenir une surface de la scène. La seconde étape met en correspondances des points d'intérêts dans les images puis estime simultanément les poses de la caméra et un nuage épars de points 3d de la scène correspondant aux points d'intérêts. La troisième étape utilise l'information sur l'ensemble des pixels pour reconstruire une surface de la scène, par exemple en estimant un nuage de points dense.Ici nous proposons de traiter le problème en calculant directement une surface à partir du nuage épars de points et de son information de visibilité fournis par l'estimation de la géométrie. Les avantages sont des faibles complexités en temps et en espace, ce qui est utile par exemple pour obtenir des modèles compacts de grands environnements comme une ville. Pour cela, nous présentons une méthode de reconstruction de surface du type sculpture dans une triangulation de Delaunay 3d des points reconstruits. L'information de visibilité est utilisée pour classer les tétraèdres en espace vide ou matière. Puis une surface est extraite de sorte à séparer au mieux ces tétraèdres à l'aide d'une méthode gloutonne et d'une minorité de points de Steiner. On impose sur la surface la contrainte de 2-variété pour permettre des traitements ultérieurs classiques tels que lissage, raffinement par optimisation de photo-consistance ... Cette méthode a ensuite été étendue au cas incrémental : à chaque nouvelle image clef sélectionnée dans une vidéo, de nouveaux points 3d et une nouvelle pose sont estimés, puis la surface est mise à jour. La complexité en temps est étudiée dans les deux cas (incrémental ou non). Dans les expériences, nous utilisons une caméra catadioptrique bas coût et obtenons des modèles 3d texturés pour des environnements complets incluant bâtiments, sol, végétation ... Un inconvénient de nos méthodes est que la reconstruction des éléments fins de la scène n'est pas correcte, par exemple les branches des arbres et les pylônes électriques. / The automatic 3d modeling of an environment using images is still an active topic in Computer Vision. Standard methods have three steps : moving a camera in the environment to take an image sequence, reconstructing the geometry of the environment, and applying a dense stereo method to obtain a surface model of the environment. In the second step, interest points are detected and matched in images, then camera poses and a sparse cloud of 3d points corresponding to the interest points are simultaneously estimated. In the third step, all pixels of images are used to reconstruct a surface of the environment, e.g. by estimating a dense cloud of 3d points. Here we propose to generate a surface directly from the sparse point cloud and its visibility information provided by the geometry reconstruction step. The advantages are low time and space complexities ; this is useful e.g. for obtaining compact models of large and complete environments like a city. To do so, a surface reconstruction method by sculpting 3d Delaunay triangulation of the reconstructed points is proposed.The visibility information is used to classify the tetrahedra in free-space and matter. Then a surface is extracted thanks to a greedy method and a minority of Steiner points. The 2-manifold constraint is enforced on the surface to allow standard surface post-processing such as denoising, refinement by photo-consistency optimization ... This method is also extended to the incremental case : each time a new key-frame is selected in the input video, new 3d points and camera pose are estimated, then the reconstructed surface is updated.We study the time complexity in both cases (incremental or not). In experiments, a low-cost catadioptric camera is used to generate textured 3d models for complete environments including buildings, ground, vegetation ... A drawback of our methods is that thin scene components cannot be correctly reconstructed, e.g. tree branches and electric posts.
229

Hantering och modellering av laserskanningsdata i FME : Automatisering av modellering av tunnlar / : Automation of modelling of tunnels

Lindqvist, Linus, Pantesjö, Jesper January 2019 (has links)
Bygg- och anläggningsbranschens implementering av BIM har resulterat i ett ökat behov att digitaliserat relationsunderlag. Äldre relationshandlingar, som mestadels utgörs av pappersritningar, saknar digitala motsvarigheter vilket gör att insamlingar av ny information, från pappersritningar, kan bli aktuell. Terrester laserskanning (TLS) är en teknik som tillämpas för insamling av data i punktmolnsform och är en allt mer förekommande insamlingsmetod vid införskaffning av relationsunderlag. Modellering från tredimensionella punktmolnsdata är ofta komplicerad och på så vis införstått med manuellt arbete för att producera ett godtyckligt resultat. Syftet med examensarbetet var att undersöka möjligheten att skapa en CAD-modell av en tunnels ytskikt från ett punktmoln med hjälp av programvaran FME. Studieområdet är ett mindre tunnelsegment och den insamlade datamängden utgörs av tidigare framarbetat punktmoln. Punktmolnet är obearbetat och innehåller brus i form av avvikande punkter samt installations- och konstruktionsobjekt. Tidigare producerat relationsunderlag, i form av CAD-modell, tilldelades också för att möjliggöra en jämförelse mot de modeller som skapats i arbetet. FME tillhandahåller ett flertal verktyg för bearbetning av punktmoln och arbetet har omfattats av tester där de olika verktygen utvärderats. Det huvudsakliga fokuset har legat på verktyget PointCloudSurfaceBuilder, vars funktion är att rekonstruera punktmoln till en mesh. En metod för filtrering av punktmolnet utformades och utreddes också under arbetet. Flertalet försök utfördes för att testa vad som fungerade bäst och ett antal modeller av varierande kvalitet kunde skapas. Metoden Poisson i verktyget PointCloudSurfaceBuilder visade bäst resultat då den skapar en “vattentät” modell som följer punktmolnets rumsliga förhållande bättre än det tilldelade relationsunderlaget. För metoden Poisson var Maximum Depth den parameter som hade störst inverkan på resultatets kvalitet. För varje höjning med 1 i parametern Maximum Depth så ökade upplösningen kvadratiskt i varje dimension för x, y och z. De totala värdena för tidsåtgång, filstorlek och antal trianglar ökade även potentiellt med upplösningen. Värden över 9 blir svåra, om inte omöjliga, att hantera i CAD-miljöer på grund av för detaljerade data i förhållande studieområdets storlek. Därav rekommenderas 7 och 8 som parametervärden vid modellering i miljöer likartade med tunnelsegmentet. / The building and construction industries implementation of BIM has resulted in an increased need to digitalise as-built basis. Older as-built documents, which is mostly made of paper plans, are missing their digital counterparts, which makes it that collection of new information, from the paper plans, can be vital. Terrestrial laser scanning (TLS) is a technique that is applied for collection of data in the form of data point clouds and is a more frequent collection method for obtaining supplies of as-built. Modelling from three-dimensional point cloud data is usually a complicated matter and therefore connected with manual labour to produce an arbitrary result. The purpose with the bachelor thesis was to research the possibility to create a CAD-model of the layer of a tunnel from a point cloud with the use of a software called FME. The study area is a smaller tunnel segment and the collected data set is based from an earlier created point cloud. The point cloud is unprocessed and contains noise from deviant points and object of installations and construction. The earlier produced as-built, in form of a CAD-model, was applied as well to enable a comparison parallel to the newly created models in this thesis. FME contains several tools for handling point clouds and the work have included several tests where the different tools have been evaluated. The primary focus of the work has been to evaluate the possibilities of the tool PointCloudSurfaceBuilder, which function is to reconstruct point clouds to a mesh. A method was also created and examined to clean the point cloud from noise. Several tests were executed to see what kind of method works the best and models of different qualities were rendered. The construction method Poisson in the transformer PointCloudSurfaceBuilder produced the best results whereas it creates a “water tight” model that follows the point clouds spatial conditions in a better way than the as-built model. In the method of Poisson there is a parameter called Maximum Depth which showed the greatest impact for the quality of the result. For every increase of 1 in the parameter Maximum Depth was the resolution increased by a factor of two in every direction of x, y and z. The total values for amount of time, file size and number of triangles increased as well in a way parallel to the potential increase of the resolution. It is hard, if not impossible, to handle the models in CAD-environments above the value 9. That is because of too high detail in the data in relation to the size of the study area. Therefore, are the recommended values of the parameter 7 and 8 in case of modelling of similar environments in tunnel complexes.
230

A three-dimensional representation method for noisy point clouds based on growing self-organizing maps accelerated on GPUs

Orts-Escolano, Sergio 21 January 2014 (has links)
The research described in this thesis was motivated by the need of a robust model capable of representing 3D data obtained with 3D sensors, which are inherently noisy. In addition, time constraints have to be considered as these sensors are capable of providing a 3D data stream in real time. This thesis proposed the use of Self-Organizing Maps (SOMs) as a 3D representation model. In particular, we proposed the use of the Growing Neural Gas (GNG) network, which has been successfully used for clustering, pattern recognition and topology representation of multi-dimensional data. Until now, Self-Organizing Maps have been primarily computed offline and their application in 3D data has mainly focused on free noise models, without considering time constraints. It is proposed a hardware implementation leveraging the computing power of modern GPUs, which takes advantage of a new paradigm coined as General-Purpose Computing on Graphics Processing Units (GPGPU). The proposed methods were applied to different problem and applications in the area of computer vision such as the recognition and localization of objects, visual surveillance or 3D reconstruction.

Page generated in 0.0884 seconds