• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 8
  • 5
  • 2
  • 2
  • 1
  • Tagged with
  • 19
  • 19
  • 13
  • 11
  • 9
  • 6
  • 5
  • 5
  • 3
  • 3
  • 3
  • 3
  • 3
  • 3
  • 3
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
11

Sistema misto reconfigurável aplicado à Interface PCI para Otimização do Algoritmo Non-local Means

Marques, Daniel Soares e 31 August 2012 (has links)
Made available in DSpace on 2015-05-14T12:36:34Z (GMT). No. of bitstreams: 1 arquivototal2.pdf: 4503412 bytes, checksum: e8c898ba24436013a2e89d08737039bb (MD5) Previous issue date: 2012-08-31 / Coordenação de Aperfeiçoamento de Pessoal de Nível Superior - CAPES / The digital image processing field is continually evolving and, although the diverse application areas, the commonly problems found converge to methods capable to improve visual information for analysis and interpretation. A major limitation issue on image precision is noise, which is defined as a perturbation in the image. The Non-Local Means (NLM) method stands out as the state-of-the-art of digital image denoising filtering. However, its computational complexity is an obstacle to make it practical on general purpose computing applications. This work presents a computer system implementation, developed with parts implemented in software and hardware applied to PCI, to optimize the NLM algorithm using hardware acceleration techniques, allowing a greater efficiency than is normally provided by general use processors. The use of reconfigurable computing helped in developing the hardware system, providing the modification of the described circuit in its use environment, accelerating the project implementation. Using an FPGA prototyping kit for PCI, dedicated to perform the dedicated calculation of the Squared Weighted Euclidean Distance, the results obtained show a gain of up to 3.5 times greater than the compared optimization approaches, also maintaining the visual quality of the denoising filtering. / A área de processamento de imagens digitais está evoluindo continuamente e, embora as áreas de aplicações sejam diversas, os problemas encontrados comumente convergem para os métodos capazes de melhorar a informação visual para a análise e interpretação. Uma das principais limitações em questão de precisão de imagens é o ruído, que é definido como uma perturbação na imagem. O método Non-Local Means (NLM) destaca-se como o estado da arte de filtragem de ruído. Contudo, sua complexidade computacional é um empecilho para torná-lo prático em aplicações computacionais de uso geral. O presente trabalho apresenta a implementação de um sistema computacional, desenvolvido com partes executadas em software e em hardware aplicado à PCI, visando a otimização do algoritmo NLM através de técnicas de aceleração em hardware, permitindo uma eficiência maior do que normalmente é fornecida por processadores de uso geral. O uso da computação reconfigurável auxiliou no desenvolvimento do sistema em hardware, proporcionando a modificação do circuito descrito no ambiente de sua utilização, acelerando a implementação do projeto. Utilizando um kit PCI de prototipação FPGA, para efetuar o cálculo dedicado da Distância Euclidiana Quadrática Ponderada, os resultados obtidos nos testes exibem um ganho de tempo até 3.5 vezes maior que as abordagens de otimização comparadas, mantendo também a qualidade visual da filtragem estabilizada.
12

Odstranění šumu z obrazů kalibračních vzorků získaných elektronovým mikroskopem / Denoising of Images from Electron Microscope

Holub, Zbyněk January 2017 (has links)
Tato Diplomová práce je zaměřena na odstranění šumu ze snímků získaných pomocí Transmisního elektronového mikroskopu. V práci jsou popsány principy digitalizace výsledných snímků a popis jednotlivých šumových složek, které vznikají při digitalizaci snímků. Tyto nechtěné složky ovlivňují kvalitu výsledného snímku. Proto byly vybrány filtrační metody založené na minimalizaci totální variace, jejichž principy jsou v této práci popsány. Jako referenční filtrační metoda byla vybrána filtrace pomocí Non-local means filtru. Tento filtr byl vybrán, jelikož v dnešní dobře patří mezi nejvíce využívané metody, které mají vysokou účinnost. Pro objektivní hodnocení kvality filtrací byly použity tyto hodnotící kritéria – SNR, PSNR a SSIM. V závěru této práce, jsou všechny získané výsledky zobrazeny a jsou diskutovány účinnosti jednotlivých filtrační metod.
13

HARDI Denoising using Non-local Means on the ℝ³ x 𝕊² Manifold

Kuurstra, Alan 20 December 2011 (has links)
Magnetic resonance imaging (MRI) has long become one of the most powerful and accurate tools of medical diagnostic imaging. Central to the diagnostic capabilities of MRI is the notion of contrast, which is determined by the biochemical composition of examined tissue as well as by its morphology. Despite the importance of the prevalent T₁, T₂, and proton density contrast mechanisms to clinical diagnosis, none of them has demonstrated effectiveness in delineating the morphological structure of the white matter - the information which is known to be related to a wide spectrum of brain-related disorders. It is only with the recent advent of diffusion-weighted MRI that scientists have been able to perform quantitative measurements of the diffusivity of white matter, making possible the structural delineation of neural fibre tracts in the human brain. One diffusion imaging technique in particular, namely high angular resolution diffusion imaging (HARDI), has inspired a substantial number of processing methods capable of obtaining the orientational information of multiple fibres within a single voxel while boasting minimal acquisition requirements. HARDI characterization of fibre morphology can be enhanced by increasing spatial and angular resolutions. However, doing so drastically reduces the signal-to-noise ratio. Since pronounced measurement noise tends to obscure and distort diagnostically relevant details of diffusion-weighted MR signals, increasing spatial or angular resolution necessitates application of the efficient and reliable tools of image denoising. The aim of this work is to develop an effective framework for the filtering of HARDI measurement noise which takes into account both the manifold to which the HARDI signal belongs and the statistical nature of MRI noise. These goals are accomplished using an approach rooted in non-local means (NLM) weighted averaging. The average includes samples, and therefore dependencies, from the entire manifold and the result of the average is used to deduce an estimate of the original signal value in accordance with MRI statistics. NLM averaging weights are determined adaptively based on a neighbourhood similarity measure. The novel neighbourhood comparison proposed in this thesis is one of spherical neighbourhoods, which assigns large weights to samples with similar local orientational diffusion characteristics. Moreover, the weights are designed to be invariant to both spatial rotations as well as to the particular sampling scheme in use. This thesis provides a detailed description of the proposed filtering procedure as well as experimental results with synthetic and real-life data. It is demonstrated that the proposed filter has substantially better denoising capabilities as compared to a number of alternative methods.
14

HARDI Denoising using Non-local Means on the ℝ³ x 𝕊² Manifold

Kuurstra, Alan 20 December 2011 (has links)
Magnetic resonance imaging (MRI) has long become one of the most powerful and accurate tools of medical diagnostic imaging. Central to the diagnostic capabilities of MRI is the notion of contrast, which is determined by the biochemical composition of examined tissue as well as by its morphology. Despite the importance of the prevalent T₁, T₂, and proton density contrast mechanisms to clinical diagnosis, none of them has demonstrated effectiveness in delineating the morphological structure of the white matter - the information which is known to be related to a wide spectrum of brain-related disorders. It is only with the recent advent of diffusion-weighted MRI that scientists have been able to perform quantitative measurements of the diffusivity of white matter, making possible the structural delineation of neural fibre tracts in the human brain. One diffusion imaging technique in particular, namely high angular resolution diffusion imaging (HARDI), has inspired a substantial number of processing methods capable of obtaining the orientational information of multiple fibres within a single voxel while boasting minimal acquisition requirements. HARDI characterization of fibre morphology can be enhanced by increasing spatial and angular resolutions. However, doing so drastically reduces the signal-to-noise ratio. Since pronounced measurement noise tends to obscure and distort diagnostically relevant details of diffusion-weighted MR signals, increasing spatial or angular resolution necessitates application of the efficient and reliable tools of image denoising. The aim of this work is to develop an effective framework for the filtering of HARDI measurement noise which takes into account both the manifold to which the HARDI signal belongs and the statistical nature of MRI noise. These goals are accomplished using an approach rooted in non-local means (NLM) weighted averaging. The average includes samples, and therefore dependencies, from the entire manifold and the result of the average is used to deduce an estimate of the original signal value in accordance with MRI statistics. NLM averaging weights are determined adaptively based on a neighbourhood similarity measure. The novel neighbourhood comparison proposed in this thesis is one of spherical neighbourhoods, which assigns large weights to samples with similar local orientational diffusion characteristics. Moreover, the weights are designed to be invariant to both spatial rotations as well as to the particular sampling scheme in use. This thesis provides a detailed description of the proposed filtering procedure as well as experimental results with synthetic and real-life data. It is demonstrated that the proposed filter has substantially better denoising capabilities as compared to a number of alternative methods.
15

Novel image processing algorithms and methods for improving their robustness and operational performance

Romanenko, Ilya January 2014 (has links)
Image processing algorithms have developed rapidly in recent years. Imaging functions are becoming more common in electronic devices, demanding better image quality, and more robust image capture in challenging conditions. Increasingly more complicated algorithms are being developed in order to achieve better signal to noise characteristics, more accurate colours, and wider dynamic range, in order to approach the human visual system performance levels.
16

Performance Analysis of Non Local Means Algorithm using Hardware Accelerators

Antony, Daniel Sanju January 2016 (has links) (PDF)
Image De-noising forms an integral part of image processing. It is used as a standalone algorithm for improving the quality of the image obtained through camera as well as a starting stage for image processing applications like face recognition, super resolution etc. Non Local Means (NL-Means) and Bilateral Filter are two computationally complex de-noising algorithms which could provide good de-noising results. Due to its computational complexity, the real time applications associated with these letters are limited. In this thesis, we propose the use of hardware accelerators such as GPU (Graphics Processing Units) and FPGA (Field Programmable Gate Arrays) to speed up the filter execution and efficiently implement using them. GPU based implementation of these letters is carried out using Open Computing Language (Open CL). The basic objective of this research is to perform high speed de-noising without compromising on the quality. Here we implement a basic NL-Means filter, a Fast NL-Means filter, and Bilateral filter using Gauss Polynomial decomposition on GPU. We also propose a modification to the existing NL-Means algorithm and Gauss Polynomial Bilateral filter. Instead of Gaussian Spatial Kernel used in standard algorithm, Box Spatial kernel is introduced to improve the speed of execution of the algorithm. This research work is a step forward towards making the real time implementation of these algorithms possible. It has been found from results that the NL-Means implementation on GPU using Open CL is about 25x faster than regular CPU based implementation for larger images (1024x1024). For Fast NL-Means, GPU based implementation is about 90x faster than CPU implementation. Even with the improved execution time, the embedded system application of the NL-Means is limited due to the power and thermal restrictions of the GPU device. In order to create a low power and faster implementation, we have implemented the algorithm on FPGA. FPGAs are reconfigurable devices and enable us to create a custom architecture for the parallel execution of the algorithm. It was found that the execution time for smaller images (256x256) is about 200x faster than CPU implementation and about 25x faster than GPU execution. Moreover the power requirements of the FPGA design of the algorithm (0.53W) is much less compared to CPU(30W) and GPU(200W).
17

Zabezpečení senzorů - ověření pravosti obrazu / Sensor Security - Verification of Image Authenticity

Juráček, Ivo January 2020 (has links)
Diploma thesis is about image sensor security. Goal of the thesis was study data integrity gained from the image sensors. Proposed method is about source camera identification from noise characteristics in image sensors. Research was about influence of denoising algorithms applied to digital images, which was acquired from 15 different image sensors. Finally the statistical evaluation had been done from computed results.
18

Analýza 3D CT obrazových dat se zaměřením na detekci a klasifikaci specifických struktur tkání / Analysis of 3D CT image data aimed at detection and classification of specific tissue structures

Šalplachta, Jakub January 2017 (has links)
This thesis deals with the segmentation and classification of paraspinal muscle and subcutaneous adipose tissue in 3D CT image data in order to use them subsequently as internal calibration phantoms to measure bone mineral density of a vertebrae. Chosen methods were tested and afterwards evaluated in terms of correctness of the classification and total functionality for subsequent BMD value calculation. Algorithms were tested in programming environment Matlab® on created patient database which contains lumbar spines of twelve patients. Following sections of this thesis contain theoretical research of the issue of measuring bone mineral density, segmentation and classification methods and description of practical part of this work.
19

Blancheur du résidu pour le débruitage d'image / Residual whiteness for image denoising

Riot, Paul 06 February 2018 (has links)
Nous proposons une étude de l’utilisation avancée de l’hypothèse de blancheur du bruit pour améliorer les performances de débruitage. Nous mettons en avant l’intérêt d’évaluer la blancheur du résidu par des mesures de corrélation dans différents cadres applicatifs. Dans un premier temps, nous nous plaçons dans un cadre variationnel et nous montrons qu’un terme de contrainte sur la blancheur du résidu peut remplacer l’attache aux données L2 en améliorant significativement les performances de débruitage. Nous le complétons ensuite par des termes de contrôle de la distribution du résidu au moyen des moments bruts. Dans une seconde partie, nous proposons une alternative au rapport de vraisemblance menant, à la norme L2 dans le cas Gaussien blanc, pour mesurer la dissimilarité entre patchs. La métrique introduite, fondée sur l’autocorrélation de la différence des patchs, se révèle plus performante pour le débruitage et la reconnaissance de patchs similaires. Finalement, les problématiques d’évaluation de qualité sans oracle et de choix local de modèle sont abordées. Encore une fois, la mesure de la blancheur du résidu apporte une information pertinente pour estimer localement la fidélité du débruitage. / We propose an advanced use of the whiteness hypothesis on the noise to imrove denoising performances. We show the interest of evaluating the residual whiteness by correlation measures in multiple applications. First, in a variational denoising framework, we show that a cost function locally constraining the residual whiteness can replace the L2 norm commonly used in the white Gaussian case, while significantly improving the denoising performances. This term is then completed by cost function constraining the residual raw moments which are a mean to control the residual distribution. In the second part of our work, we propose an alternative to the likelihood ratio, leading to the L2 norm in the white Gaussian case, to evaluate the dissimilarity between noisy patches. The introduced metric, based on the autocorrelation of the patches difference, achieves better performances both for denoising and similar patches recognition. Finally, we tackle the no reference quality evaluation and the local model choice problems. Once again, the residual whiteness bring a meaningful information to locally estimate the truthfulness of the denoising.

Page generated in 0.044 seconds