Spelling suggestions: "subject:"abject egmentation"" "subject:"abject asegmentation""
1 |
Three dimensional object analysis and tracking by digital holography microscopySchockaert, Cédric 26 February 2007 (has links)
Digital Holography Microscopy (DHM) is a new 3D measurement technique that exists since Charge Coupled Devices (or CCD cameras) allow to record numerically high resolution images. That opens a new door to the theory of holography discovered in 1949 by Gabor: the door that masked the world of digital hologram processing. A hologram is a usual image but that contains the complex amplitude of the light coded into intensities recorded by the camera. The complex amplitude of the light can be seen as the combination of the energy information (squared amplitude modulus) with the information of the propagation angle of the light (phase of the amplitude) for each point of the image. When the hologram is digital, this dual information associated with a diffractive model of the light propagation permits to numerically investigate back and front planes to the recorded plane of the imaging system. We understand that 3D information can be recorded by a CCD camera and the acquisition rate of this volume information is only limited by the acquisition rate of the unique camera. For each digital hologram, the numerical investigation of front and back regions to the recorded plane is a tool to numerically refocus objects appearing unfocused in the original plane acquired by the CCD.
This thesis aims to develop general and robust algorithms that are devoted to automate the analysis process in the 3D space and in time of objects present in a volume studied by a specific imaging system that permits to record holograms. Indeed, the manual processing of a huge amount of holograms is not realistic and has to be automated by software implementing precise algorithms. In this thesis, the imaging system that records holograms is a Mach-Zehnder interferometer working in transmission and studied objects are either of biological nature (crystals, vesicles, cancer cells) or latex particles. We propose and test focus criteria, based on an identical focus metric, for both amplitude and phase objects. These criteria allow the determination of the best focus plane of an object when the numerical investigation is performed. The precision of the best focus plane is lower than the depth of field of the microscope. From this refocus theory, we develop object detection algorithms that build a synthetic image where objects are bright on a dark background. This detection map of objects is the first step to a fully automatic analysis of objects present in one hologram. The combination of the detection algorithm and the focus criteria allow the precise measurement of the 3D position of the objects, and of other relevant characteristics like the object surface in its focus plane, or its convexity or whatever. These extra relevant measures are carried out with a segmentation algorithm adapted to the studied objects of this thesis (opaque objects, and transparent objects in a uniform refractive index environment). The last algorithm investigated in this research work is the data association in time of objects from hologram to hologram in order to extract 3D trajectories by using the predictive Kalman filtering theory.
These algorithms are the abstract bricks of two software: DHM Object Detection and Analysis software, and Kalman Tracking software. The first software is designed for both opaque and transparent objects. The term object is not defined by one other characteristic in this work, and as a consequence, the developed algorithms are very general and can be applied on various objects studied in transmission by DHM. The tracking software is adapted to the dynamic applications of the thesis, which are flows of objects. Performance and results are exposed in a specific chapter.
|
2 |
Camouflaged Object Segmentation in ImagesYan, Jinnan January 2019 (has links)
No description available.
|
3 |
Graph-based Inference with Constraints for Object Detection and SegmentationMa, Tianyang January 2013 (has links)
For many fundamental problems of computer vision, adopting a graph-based framework can be straight-forward and very effective. In this thesis, I propose several graph-based inference methods tailored for different computer vision applications. It starts from studying contour-based object detection methods. In particular, We propose a novel framework for contour based object detection, by replacing the hough-voting framework with finding dense subgraph inference. Compared to previous work, we propose a novel shape matching scheme suitable for partial matching of edge fragments. The shape descriptor has the same geometric units as shape context but our shape representation is not histogram based. The key contribution is that we formulate the grouping of partial matching hypotheses to object detection hypotheses is expressed as maximum clique inference on a weighted graph. Consequently, each detection result not only identifies the location of the target object in the image, but also provides a precise location of its contours, since we transform a complete model contour to the image. We achieve very competitive results on ETHZ dataset, obtained in a pure shape-based framework, demonstrate that our method achieves not only accurate object detection but also precise contour localization on cluttered background. Similar to the task of grouping of partial matches in the contour-based method, in many computer vision problems, we would like to discover certain pattern among a large amount of data. For instance, in the application of unsupervised video object segmentation, where we need automatically identify the primary object and segment the object out in every frame. We propose a novel formulation of selecting object region candidates simultaneously in all frames as finding a maximum weight clique in a weighted region graph. The selected regions are expected to have high objectness score (unary potential) as well as share similar appearance (binary potential). Since both unary and binary potentials are unreliable, we introduce two types of mutex (mutual exclusion) constraints on regions in the same clique: intra-frame and inter-frame constraints. Both types of constraints are expressed in a single quadratic form. An efficient algorithm is applied to compute the maximal weight cliques that satisfy the constraints. We apply our method to challenging benchmark videos and obtain very competitive results that outperform state-of-the-art methods. We also show that the same maximum weight subgraph with mutex constraints formulation can be used to solve various computer vision problems, such as points matching, solving image jigsaw puzzle, and detecting object using 3D contours. / Computer and Information Science
|
4 |
Decomposition of Morphological Structuring Elements and Segmentation of Human Objects in Video SequencesYang, Hsin-Tai 11 September 2007 (has links)
With the rapid development of image processing techniques, many unprecedented applications are emerging from all kinds of science branches, such as medicine, meteorology, astronomy, industrial control, etc. This dissertation presents an achievement of our research work related to the image processing field. Technically, the work consists of two parts. The first part concerns decomposition of morphological structuring elements while the second part explores the problems of human object segmentation in video/image data.
In the first part, an integrated method, aiming to decompose a morphological structuring element into dilations of smaller ones, is proposed. By first formulating the decomposition problem into a set of linear constraints, the integer linear programming echnique is then applied to obtain an optimal decomposition. Compared to other existing approaches, the proposed method is more general and has several advantages. Firstly, it provides a systematic way of decomposing arbitrarily shaped structuring elements. Secondly, for convex images, factors can be of any size, not restricted to 3x3. Thirdly, the candidate set can be freely assigned by the user and finally the criteria of optimality can be flexible.
In the second part, we present a three-stage system for segmentation of multiple human objects in a video stream. In the first stage, for a base frame to be segmented, we propose a hybrid self-clustering technique that incorporates the spatial concept as well as color attributes to reduce the number of small segments. In the second stage, the face shape modeled by the eight-directional convex polygons and the face features including two eyes and a mouth are extracted, parameterized, and fed to a trained neural network for detection of a human face. In the last stage, the size and orientation of the detected face region as well as the motion information among frames are used to roughly detect the corresponding body. To locate human objects more accurately, another neural network is constructed for recognizing the ambiguous regions.
|
5 |
[en] A STUDY OF THE USE OF OBJECT SEGMENTATION FOR THE APPLICATION OF VIDEO INPAINTING TECHNIQUES / [pt] UM ESTUDO DE USO DE SEGMENTAÇÃO DE OBJETOS PARA A APLICAÇÃO DE TÉCNICAS DE VIDEO INPAINTINGSUSANA DE SOUZA BOUCHARDET 23 August 2021 (has links)
[pt] Nos últimos anos tem ocorrido um notável desenvolvimento de técnicas
de Image Inpainting, entretanto transpor esse conhecimento para aplicações
em vídeo tem se mostrado um desafio. Além dos desafios inerentes a tarefa
de Video Inpainting (VI), utilizar essa técnica requer um trabalho prévio de
anotação da área que será reconstruída. Se a aplicação do método for para
remover um objeto ao longo de um vídeo, então a anotação prévia deve ser
uma máscara da área deste objeto frame a frame. A tarefa de propagar a
anotação de um objeto ao longo de um vídeo é conhecida como Video Object
Segmentation (VOS) e já existem técnicas bem desenvolvidas para solucionar
este problemas. Assim, a proposta desse trabalho é aplicar técnicas de VOS
para gerar insumo para um algoritmo de VI. Neste trabalho iremos analisar o
impacto de utilizar anotações preditas no resultado final de um modelo de VI. / [en] In recent years there has been a remarkable development of Image
Inpainting techniques, but using this knowledge in video application is still
a challenge. Besides the inherent challenges of the Video Inpainting (VI) task, applying this technique requires a previous job of labeling the area that should be reconstructed. If this method is used to remove an object from the video, then the annotation should be a mask of this object s area frame by frame. The task of propagating an object mask in a video is known as Video Object
Segmentation (VOS) and there are already well developed techniques to solve
this kind of task. Therefore, this work aims to apply VOS techniques to create
the inputs for an VI algorithm. In this work we shall analyse the impact in the
result of a VI algorithm when we use a predicted annotation as the input.
|
6 |
Zero-shot Learning for Visual Recognition ProblemsNaha, Shujon January 2016 (has links)
In this thesis we discuss different aspects of zero-shot learning and propose solutions for three challenging visual recognition problems: 1) unknown object recognition from images 2) novel action recognition from videos and 3) unseen object segmentation. In all of these three problems, we have two different sets of classes, the “known classes”, which are used in the training phase and the “unknown classes” for which there is no training instance. Our proposed approach exploits the available semantic relationships between known and unknown object classes and use them to transfer the appearance models from known object classes to unknown object classes to recognize unknown objects. We also propose an approach to recognize novel actions from videos by learning a joint model that links videos and text. Finally, we present a ranking based approach for zero-shot object segmentation. We represent each unknown object class as a semantic ranking of all the known classes and use this semantic relationship to extend the segmentation model of known classes to segment unknown class objects. / October 2016
|
7 |
The Analysis of Visual Motion: From Computational Theory to Neuronal MechanismsHildreth, Ellen C., Koch, Christof 01 December 1986 (has links)
This paper reviews a number of aspects of visual motion analysis in biological systems from a computational perspective. We illustrate the kinds of insights that have been gained through computational studies and how these observations can be integrated with experimental studies from psychology and the neurosciences to understand the particular computations used by biological systems to analyze motion. The particular areas of motion analysis that we discuss include early motion detection and measurement, the optical flow computation, motion correspondence, the detection of motion discontinuities, and the recovery of three-dimensional structure from motion.
|
8 |
The Video Object Segmentation Method for Mpeg-4Huang, Jen-Chi 23 September 2004 (has links)
In this thesis, we proposed the series methods of moving object segmentation and object application. These methods are the moving object segmentation method in wavelet domain, double change detection method, global motion estimation method, and the moving object segmentation in the motion background.
First, we proposed the Video Object Segmentation Method in Wavelet Domain. We use the Change Detection Method with the different thresholds in four wavelet sub-bands. The experiment results show that we obtain further object shape information and more accurately extracting the moving object.
In the double change detection method, we proposed the method for moving object segmentation using three successive frames. We use change detection method twice in wavelet domain. After applying the Intersect Operation, we obtain the accurately moving object edge map and further object shape information.
Besides, we proposed the global motion estimation method in motion scene. We propose a novel global motion estimation using cross point for the reconstruction of background scene in video sequences. Due to the robust character and limit number of cross points, we can get the Affine parameters of global motion in video sequences efficiency.
At last, we proposed the object segmentation method in motion scene. We use the motion estimation method to estimate the global motion between the consecutive frames. We reconstruct a wide scene background without moving objects by the consecutive frames. At last, the moving objects will be segmented easily by comparing the object frame and the relative part in wide scene background.
The Results of our proposed have good performance in the different type of video sequences. Hence, the methods of our thesis contribute to the video coding in Mpeg-4 and multimedia technology.
|
9 |
Steps towards the object semantic hierarchyXu, Changhai, 1977- 17 November 2011 (has links)
An intelligent robot must be able to perceive and reason robustly about its world in terms of objects, among other foundational concepts. The robot can draw on rich data for object perception from continuous sensory input, in contrast to the usual formulation that focuses on objects in isolated still images. Additionally, the robot needs multiple object representations to deal with different tasks and/or different classes of objects. We propose the Object Semantic Hierarchy (OSH), which consists of multiple representations with different ontologies. The OSH factors the problems of object perception so that intermediate states of knowledge about an object have natural representations, with relatively easy transitions from less structured to more structured representations. Each layer in the hierarchy builds an explanation of the sensory input stream, in terms of a stochastic model consisting of a deterministic model and an unexplained "noise" term. Each layer is constructed by identifying new invariants from the previous layer. In the final model, the scene is explained in terms of constant background and object models, and low-dimensional dynamic poses of the observer and objects.
The OSH contains two types of layers: the Object Layers and the Model Layers. The Object Layers describe how the static background and each foreground object are individuated, and the Model Layers describe how the model for the static background or each foreground object evolves from less structured to more structured representations. Each object or background model contains the following layers: (1) 2D object in 2D space (2D2D): a set of constant 2D object views, and the time-variant 2D object poses, (2) 2D object in 3D space (2D3D): a collection of constant 2D components, with their individual time-variant 3D poses, and (3) 3D object in 3D space (3D3D): the same collection of constant 2D components but with invariant relations among their 3D poses, and the time-variant 3D pose of the object as a whole.
In building 2D2D object models, a fundamental problem is to segment out foreground objects in the pixel-level sensory input from the background environment, where motion information is an important cue to perform the segmentation. Traditional approaches for moving object segmentation usually appeal to motion analysis on pure image information without exploiting the robot's motor signals. We observe, however, that the background motion (from the robot's egocentric view) has stronger correlation to the robot's motor signals than the motion of foreground objects. Based on this observation, we propose a novel approach to segmenting moving objects by learning homography and fundamental matrices from motor signals.
In building 2D3D and 3D3D object models, estimating camera motion parameters plays a key role. We propose a novel method for camera motion estimation that takes advantage of both planar features and point features and fuses constraints from both homography and essential matrices in a single probabilistic framework. Using planar features greatly improves estimation accuracy over using point features only, and with the help of point features, the solution ambiguity from a planar feature is resolved. Compared to the two classic approaches that apply the constraint of either homography or essential matrix, the proposed method gives more accurate estimation results and avoids the drawbacks of the two approaches. / text
|
10 |
Robust visual detection and tracking of complex objects : applications to space autonomous rendez-vous and proximity operations / Détection et suivi visuels robustes d'objets complexes : applications au rendezvous spatial autonomePetit, Antoine 19 December 2013 (has links)
Dans cette thèse nous étudions le fait de localiser complètement un objet connu par vision artificielle, en utilisant une caméra monoculaire, ce qui constitue un problème majeur dans des domaines comme la robotique. Une attention particulière est ici portée sur des applications de robotique spatiale, dans le but de concevoir un système de localisation visuelle pour des opérations de rendez-vous spatial autonome. Deux composantes principales du problème sont abordées: celle de la localisation initiale de l'objet ciblé, puis celle du suivi de cet objet image par image, donnant la pose complète entre la caméra et l'objet, connaissant le modèle 3D de l'objet. Pour la détection, l'estimation de pose est basée sur une segmentation de l'objet en mouvement et sur une procédure probabiliste d'appariement et d'alignement basée contours de vues synthétiques de l'objet avec une séquence d'images initiales. Pour la phase de suivi, l'estimation de pose repose sur un algorithme de suivi basé modèle 3D, pour lequel nous proposons trois différents types de primitives visuelles, dans l'idée de décrire l'objet considéré par ses contours, sa silhouette et par un ensemble de points d'intérêts. L'intégrité du système de localisation est elle évaluée en propageant l'incertitude sur les primitives visuelles. Cette incertitude est par ailleurs utilisée au sein d'un filtre de Kalman linéaire sur les paramètres de vitesse. Des tests qualitatifs et quantitatifs ont été réalisés, sur des données synthétiques et réelles, avec notamment des conditions d'image difficiles, montrant ainsi l'efficacité et les avantages des différentes contributions proposées, et leur conformité avec un contexte de rendez vous spatial. / In this thesis, we address the issue of fully localizing a known object through computer vision, using a monocular camera, what is a central problem in robotics. A particular attention is here paid on space robotics applications, with the aims of providing a unified visual localization system for autonomous navigation purposes for space rendezvous and proximity operations. Two main challenges of the problem are tackled: initially detecting the targeted object and then tracking it frame-by-frame, providing the complete pose between the camera and the object, knowing the 3D CAD model of the object. For detection, the pose estimation process is based on the segmentation of the moving object and on an efficient probabilistic edge-based matching and alignment procedure of a set of synthetic views of the object with a sequence of initial images. For the tracking phase, pose estimation is handled through a 3D model-based tracking algorithm, for which we propose three different types of visual features, pertinently representing the object with its edges, its silhouette and with a set of interest points. The reliability of the localization process is evaluated by propagating the uncertainty from the errors of the visual features. This uncertainty besides feeds a linear Kalman filter on the camera velocity parameters. Qualitative and quantitative experiments have been performed on various synthetic and real data, with challenging imaging conditions, showing the efficiency and the benefits of the different contributions, and their compliance with space rendezvous applications.
|
Page generated in 0.128 seconds