• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 31
  • 6
  • 6
  • 2
  • 2
  • 1
  • 1
  • 1
  • Tagged with
  • 56
  • 56
  • 13
  • 10
  • 10
  • 9
  • 8
  • 8
  • 8
  • 8
  • 7
  • 6
  • 6
  • 6
  • 6
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
41

Volume Estimation of Airbags: A Visual Hull Approach

Anliot, Manne January 2005 (has links)
This thesis presents a complete and fully automatic method for estimating the volume of an airbag, through all stages of its inflation, with multiple synchronized high-speed cameras. Using recorded contours of the inflating airbag, its visual hull is reconstructed with a novel method: The intersections of all back-projected contours are first identified with an accelerated epipolar algorithm. These intersections, together with additional points sampled from concave surface regions of the visual hull, are then Delaunay triangulated to a connected set of tetrahedra. Finally, the visual hull is extracted by carving away the tetrahedra that are classified as inconsistent with the contours, according to a voting procedure. The volume of an airbag's visual hull is always larger than the airbag's real volume. By projecting a known synthetic model of the airbag into the cameras, this volume offset is computed, and an accurate estimate of the real airbag volume is extracted. Even though volume estimates can be computed for all camera setups, the cameras should be specially posed to achieve optimal results. Such poses are uniquely found for different airbag models with a separate, fully automatic, simulated annealing algorithm. Satisfying results are presented for both synthetic and real-world data.
42

Gerador adaptativo de malhas 2-D para problemas eletromagneticos / Adaptive 2-D mesh generator for electromagnetic problems

Hsu, Atilio Claudio 28 May 2007 (has links)
Orientador: Hugo Enrique Hernandez Figueroa / Dissertação (mestrado) - Universidade Estadual de Campinas, Faculdade de Engenharia Eletrica e de Computação / Made available in DSpace on 2018-08-09T06:26:18Z (GMT). No. of bitstreams: 1 Hsu_AtilioClaudio_M.pdf: 6986419 bytes, checksum: 5d1120daa8d4126ad4a6fd84fd5ac188 (MD5) Previous issue date: 2007 / Resumo: Nesta dissertação apresenta-se o desenvolvimento e a implementação de algoritmos para geração de malhas adaptativas de elementos triangulares, visando o estudo de problemas de propagação eletromagnética em domínios bidimensionais. Foram implementadas duas técnicas de refinamento, uma baseada no método da troca de arestas denominado Delaunay e a outra baseada na técnica de suavização de Laplace. O gerador implementado leva em conta interfaces de curvatura arbitrária e, também, pode ser facilmente integrado aos programas desenvolvidos no Departamento de Microonda e Óptica (DMO) / Abstract: This work presents the development of algorithms for adaptive mesh generation of triangular elements applicable to electromagnetic propagation problems in bidimensional domains. Several techniques were used, such as edgeflip, known as Delaunay method, and the Laplacian¿s smoothing method. The program considers arbitrary curvature interfaces and can be easily integrated to the programs created in the Department of Microwaves and Optics (DMO) / Mestrado / Telecomunicações e Telemática / Mestre em Engenharia Elétrica
43

Méthodes probabilistes pour l'analyse des algorithmes sur les tesselations aléatoires / Probabilistic methods for the analysis of algorithms on random tessellations

Hemsley, Ross 16 December 2014 (has links)
Dans cette thèse, nous exploitons les outils de la théorie des probabilités et de la géométrie stochastique pour analyser des algorithmes opérant sur les tessellations. Ce travail est divisé entre deux thèmes principaux, le premier traite de la navigation dans une tessellation de Delaunay et dans son dual, le diagramme de Voronoï avec des implications pour les algorithmes de localisation spatiales et de routage dans les réseaux en ligne. Nous proposons deux nouveaux algorithmes de navigation dans la triangulation de Delaunay, que nous appelons Pivot Walk et Cone Walk. Pour Cone Walk, nous fournissons une analyse en moyenne détaillée avec des bornes explicites sur les propriétés de la pire marche possible effectuée par l'algorithme sur une triangulation de Delaunay aléatoire d'une région convexe bornée. C'est un progrès significatif car dans l'algorithme Cone Walk, les probabilités d'utiliser un triangle ou un autre au cours de la marche présentent des dépendances complexes, dépendances inexistantes dans d'autres marches. La deuxième partie de ce travail concerne l'étude des propriétés extrémales de tessellations aléatoires. En particulier, nous dérivons les premiers et derniers statistiques d'ordre pour les boules inscrites dans les cellules d'un arrangement de droites Poissonnien; ce résultat a des implications par exemple pour le hachage respectant la localité. Comme corollaire, nous montrons que les cellules minimisant l'aire sont des triangles. / In this thesis, we leverage the tools of probability theory and stochastic geometry to investigate the behavior of algorithms on geometric tessellations of space. This work is split between two main themes, the first of which is focused on the problem of navigating the Delaunay tessellation and its geometric dual, the Voronoi diagram. We explore the applications of this problem to point location using walking algorithms and the study of online routing in networks. We then propose and investigate two new algorithms which navigate the Delaunay triangulation, which we call Pivot Walk and Cone Walk. For Cone Walk, we provide a detailed average-case analysis, giving explicit bounds on the properties of the worst possible path taken by the algorithm on a random Delaunay triangulation in a bounded convex region. This analysis is a significant departure from similar results that have been obtained, due to the difficulty of dealing with the complex dependence structure of localized navigation algorithms on the Delaunay triangulation. The second part of this work is concerned with the study of extremal properties of random tessellations. In particular, we derive the first and last order-statistics for the inballs of the cells in a Poisson line tessellation. This result has implications for algorithms involving line tessellations, such as locality sensitive hashing. As a corollary, we show that the cells minimizing the area are triangles.
44

Generátor sítě konečných prvků / Mesh Generator for the Finite Element Method

Ščišľak, Tomáš January 2011 (has links)
The paper describes basic principles of the finite element method. Further, basic properties of triangulation are discussed. The thesis is primarily focused on the Delaunay and Greedy triangulation. Greedy triangulation is simple to implement, but may not produce the optimal shape of triangles. Delaunay method is used due to its robustness in a wide range of fields especially in computer graphics. This method is relatively easy to implement and provides triangles with high quality.
45

Vectorisation compacte d’images par approches stochastiques / Compact image vectorization by stochastic approaches

Favreau, Jean-Dominique 15 March 2018 (has links)
Les artistes apprécient les images vectorielles car elles sont compactes et facilement manipulables. Cependant, beaucoup d’artistes expriment leur créativité en dessinant, en peignant ou encore en prenant des photographies. Digitaliser ces contenus produit des images rasterisées. L’objectif de cette thèse est de convertir des images rasterisées en images vectorielles qui sont facilement manipulables. Nous avons formulé le problème de vectorisation comme un problème de minimisation d’énergie que nous avons défini par deux termes. Le premier terme, plutôt classique, mesure la fidélité de l’image vectorielle générée avec l’image rasterisée d’origine. La nouveauté principale est le second terme qui mesure la simplicité de l’image vectorielle générée. Le terme de simplicité est global et contient des variables discrètes, ce qui rend sa minimisation difficile. Nous avons proposé deux algorithmes de vectorisation : un pour la vectorisation de croquis et un autre pour la vectorisation multicouches d’images couleurs. Ces deux algorithmes commencent par extraire des primitives géométriques (un squelette pour les croquis et une segmentation pour les images couleurs) qu’ils assemblent ensuite pour former l’image vectorielle. Dans la dernière partie de la thèse, nous proposons un nouvel algorithme qui est capable de vectoriser des croquis sans étapes préliminaires : on extrait et assemble les primitives simultanément. Nous montrons le potentiel de ce nouvel algorithme pour une variété de problèmes de vision par ordinateur comme l’extraction de réseaux linéiques, l’extraction d’objets et la compression d’images. / Artists appreciate vector graphics for their compactness and editability. However many artists express their creativity by sketching, painting or taking photographs. Digitizing these images produces raster graphics. The goal of this thesis is to convert raster graphics into vector graphics that are easy to edit. We cast image vectorization as an energy minimization problem. Our energy is a combination of two terms. The first term measures the fidelity of the vector graphics to the input raster graphics. This term is a standard term for image reconstruction problems. The main novelty is the second term which measures the simplicity of the vector graphics. The simplicity term is global and involves discrete unknowns which makes its minimization challenging. We propose two stochastic optimizations for this formulation: one for the line drawing vectorization problem and another one for the color image vectorization problem. These optimizations start by extracting geometric primitives (skeleton for sketches and segmentation for color images) and then assembling these primitives together to form the vector graphics. In the last chapter we propose a generic optimization method for the problem of geometric shape extraction. This new algorithm does not require any preprocessing step. We show its efficiency in a variety of vision problems including line network extraction, object contouring and image compression.
46

Delaunay-based Vector Segmentation of Volumetric Medical Images / Vektorová segmentace objemových medicínských dat založená na Delaunay triangulaci

Španěl, Michal January 2011 (has links)
Image segmentation plays an important role in medical image analysis. Many segmentation algorithms exist. Most of them produce data which are more or less not suitable for further surface extraction and anatomical modeling of human tissues. In this thesis, a novel segmentation technique based on the 3D Delaunay triangulation is proposed. A modified variational tetrahedral meshing approach is used to adapt a tetrahedral mesh to the underlying CT volumetric data, so that image edges are well approximated in the mesh. In order to classify tetrahedra into regions/tissues whose characteristics are similar, three different clustering schemes are presented. Finally, several methods for improving quality of the mesh and its adaptation to the image structure are also discussed.
47

Modélisation 3D automatique d'environnements : une approche éparse à partir d'images prises par une caméra catadioptrique / Automatic 3d modeling of environments : a sparse approach from images taken by a catadioptric camera

Yu, Shuda 03 June 2013 (has links)
La modélisation 3d automatique d'un environnement à partir d'images est un sujet toujours d'actualité en vision par ordinateur. Ce problème se résout en général en trois temps : déplacer une caméra dans la scène pour prendre la séquence d'images, reconstruire la géométrie, et utiliser une méthode de stéréo dense pour obtenir une surface de la scène. La seconde étape met en correspondances des points d'intérêts dans les images puis estime simultanément les poses de la caméra et un nuage épars de points 3d de la scène correspondant aux points d'intérêts. La troisième étape utilise l'information sur l'ensemble des pixels pour reconstruire une surface de la scène, par exemple en estimant un nuage de points dense.Ici nous proposons de traiter le problème en calculant directement une surface à partir du nuage épars de points et de son information de visibilité fournis par l'estimation de la géométrie. Les avantages sont des faibles complexités en temps et en espace, ce qui est utile par exemple pour obtenir des modèles compacts de grands environnements comme une ville. Pour cela, nous présentons une méthode de reconstruction de surface du type sculpture dans une triangulation de Delaunay 3d des points reconstruits. L'information de visibilité est utilisée pour classer les tétraèdres en espace vide ou matière. Puis une surface est extraite de sorte à séparer au mieux ces tétraèdres à l'aide d'une méthode gloutonne et d'une minorité de points de Steiner. On impose sur la surface la contrainte de 2-variété pour permettre des traitements ultérieurs classiques tels que lissage, raffinement par optimisation de photo-consistance ... Cette méthode a ensuite été étendue au cas incrémental : à chaque nouvelle image clef sélectionnée dans une vidéo, de nouveaux points 3d et une nouvelle pose sont estimés, puis la surface est mise à jour. La complexité en temps est étudiée dans les deux cas (incrémental ou non). Dans les expériences, nous utilisons une caméra catadioptrique bas coût et obtenons des modèles 3d texturés pour des environnements complets incluant bâtiments, sol, végétation ... Un inconvénient de nos méthodes est que la reconstruction des éléments fins de la scène n'est pas correcte, par exemple les branches des arbres et les pylônes électriques. / The automatic 3d modeling of an environment using images is still an active topic in Computer Vision. Standard methods have three steps : moving a camera in the environment to take an image sequence, reconstructing the geometry of the environment, and applying a dense stereo method to obtain a surface model of the environment. In the second step, interest points are detected and matched in images, then camera poses and a sparse cloud of 3d points corresponding to the interest points are simultaneously estimated. In the third step, all pixels of images are used to reconstruct a surface of the environment, e.g. by estimating a dense cloud of 3d points. Here we propose to generate a surface directly from the sparse point cloud and its visibility information provided by the geometry reconstruction step. The advantages are low time and space complexities ; this is useful e.g. for obtaining compact models of large and complete environments like a city. To do so, a surface reconstruction method by sculpting 3d Delaunay triangulation of the reconstructed points is proposed.The visibility information is used to classify the tetrahedra in free-space and matter. Then a surface is extracted thanks to a greedy method and a minority of Steiner points. The 2-manifold constraint is enforced on the surface to allow standard surface post-processing such as denoising, refinement by photo-consistency optimization ... This method is also extended to the incremental case : each time a new key-frame is selected in the input video, new 3d points and camera pose are estimated, then the reconstructed surface is updated.We study the time complexity in both cases (incremental or not). In experiments, a low-cost catadioptric camera is used to generate textured 3d models for complete environments including buildings, ground, vegetation ... A drawback of our methods is that thin scene components cannot be correctly reconstructed, e.g. tree branches and electric posts.
48

Deformação harmônica da triangulação de Delaunay / Harmonic deformation of the Delaunay triangulation

Rafael de Mattos Grisi 28 August 2009 (has links)
Dado um processo de Poisson d-dimensional, construímos funções harmônicas na triangulação de Delaunay associada, com comportamento assintótico linear, como limite de um processo de harness sem ruído. Tais funções permitem que construamos uma nova imersão da triangulação de Delaunay, que denominaremos de deformação harmônica. / Given a d-dimensional Poisson point process, we construct harmonic functions on the associated Delaunay triangulation, with linear assymptotic behaviour, as the limit of a noiseless harness process. These mappings allow us to find a new embedding for the Delaunay triangulation. We call it harmonic deformation of the graph.
49

Three-dimensional hybrid grid generation with application to high Reynolds number viscous flows

Athanasiadis, Aristotelis 29 June 2004 (has links)
In this thesis, an approach is presented for the generation of grids suitable for the simulation of high Reynolds number viscous flows in complex three-dimensional geometries. The automatic and reliable generation of such grids is today on the biggest bottlenecks in the industrial CFD simulation environment.<p><p>In the proposed approach, unstructured tetrahedral grids are employed for the regions far from the viscous boundaries of the domain, while semi-structured layers of high aspect ratio prismatic and hexahedral elements are used to provide the necessary grid resolution inside the boundary layers and normal to the viscous walls. The definition of the domain model is based on the STEP ISO standard and the topological information contained in the model is used for applying the hierarchical grid generation parameters defined by the user. An efficient, high-quality and robust algorithm is presented for the generation of the unstructured simplicial (triangular of tetrahedral) part of the grid. The algorithm is based on the Delaunay triangulation and the internal grid points are created following a centroid or frontal approach. For the surface grid generation, a hybrid approach is also proposed similar to the volume.<p>Semi-structured grids are generated on the surface grid (both on the edges and faces of the domain) to improve the grid resolution around convex and concave ridges and corners, by aligning the grid elements in the directions of high solution gradients along the surface. A method is also developed for automatically setting the grid generation parameters related to the surface grid generation based on the curvature of the surface in order to obtain an accurate and smooth surface grid. Finally, a semi-structured prismatic/hexahedral grid generation algorithm is presented for the generation of the part of grid close to the viscous walls of the domain. The algorithm is further extended with improvements meant to increase the grid quality around concave and convex ridges of the domain, where the semi-structured grids are known to be inadequate.<p><p>The combined methodology is demonstrated on a variety of complex examples mainly from the automotive and aeronautical industry. / Doctorat en sciences appliquées / info:eu-repo/semantics/nonPublished
50

Single and Multiple Emitter Localization in Cognitive Radio Networks

Ureten, Suzan January 2017 (has links)
Cognitive radio (CR) is often described as a context-intelligent radio, capable of changing the transmit parameters dynamically based on the interaction with the environment it operates. The work in this thesis explores the problem of using received signal strength (RSS) measurements taken by a network of CR nodes to generate an interference map of a given geographical area and estimate the locations of multiple primary transmitters that operate simultaneously in the area. A probabilistic model of the problem is developed, and algorithms to address location estimation challenges are proposed. Three approaches are proposed to solve the localization problem. The first approach is based on estimating the locations from the generated interference map when no information about the propagation model or any of its parameters is present. The second approach is based on approximating the maximum likelihood (ML) estimate of the transmitter locations with the grid search method when the model is known and its parameters are available. The third approach also requires the knowledge of model parameters but it is actually based on generating samples from the joint posterior of the unknown location parameter with Markov chain Monte Carlo (MCMC) methods, as an alternative for the highly computationally complex grid search approach. For RF cartography generation problem, we study global and local interpolation techniques, specifically the Delaunay triangulation based techniques as the use of existing triangulation provides a computationally attractive solution. We present a comparative performance evaluation of these interpolation techniques in terms of RF field strength estimation and emitter localization. Even though the estimates obtained from the generated interference maps are less accurate compared to the ML estimator, the rough estimates are utilized to initialize a more accurate algorithm such as the MCMC technique to reduce the complexity of the algorithm. The complexity issues of ML estimators based on full grid search are also addressed by various types of iterative grid search methods. One challenge to apply the ML estimation algorithm to multiple emitter localization problem is that, it requires a pdf approximation to summands of log-normal random variables for likelihood calculations at each grid location. This inspires our investigations on sum of log-normal approximations studied in literature for selecting the appropriate approximation to our model assumptions. As a final extension of this work, we propose our own approximation based on distribution fitting to a set of simulated data and compare our approach with Fenton-Wilkinson's well-known approximation which is a simple and computational efficient approach that fits a log-normal distribution to sum of log-normals by matching the first and second central moments of random variables. We demonstrate that the location estimation accuracy of the grid search technique obtained with our proposed approximation is higher than the one obtained with Fenton-Wilkinson's in many different case scenarios.

Page generated in 0.0485 seconds