• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 8
  • 2
  • 2
  • 1
  • Tagged with
  • 15
  • 15
  • 4
  • 4
  • 4
  • 4
  • 4
  • 4
  • 3
  • 3
  • 3
  • 3
  • 3
  • 3
  • 3
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
11

Komprese a hodnocení kvality signálů EKG / Compression and Quality Assessment of ECG Signals

Němcová, Andrea January 2021 (has links)
Ztrátová komprese signálů EKG je užitečná a v současnosti stále se rozvíjející oblast. Stále se vyvíjí nové a nové kompresní algoritmy. V této oblasti ale chybí standardy pro hodnocení kvality signálu po kompresi. Existuje tedy sice mnoho různých kompresních algoritmů, které ale buď nelze objektivně porovnat vůbec, nebo jen zhruba. V oblasti komprese navíc nikde není popsáno, zda mají na výkon kompresních algoritmů vliv patologie, popřípadě jaký. Tato dizertační práce poskytuje přehled všech nalezených metod pro hodnocení kvality signálů EKG po kompresi. Navíc bylo vytvořeno 10 nových metod. V rámci práce byla provedena analýza všech těchto metod a na základě jejích výsledků bylo doporučeno 12 metod vhodných pro hodnocení kvality signálu EKG po kompresi. Také je zde představen nový kompresní algoritmus „Single-Cycle Fractal-Based (SCyF)“. Algoritmus SCyF je inspirován metodou založenou na fraktálech a využívá jednoho cyklu signálu EKG jako domény. Algoritmus SCyF byl testován na čtyřech různých databázích, přičemž kvalita signálů po kompresi byla vyhodnocena 12 doporučenými metodami. Výsledky byly porovnány s velmi populárním kompresním algoritmem založeným na vlnkové transformaci, který využívá metodu „Set Partitioning in Hierarchical Trees (SPIHT)“. Postup testování zároveň slouží jako příklad, jak by měl vypadat standard hodnocení výkonu kompresních algoritmů. Dále bylo statisticky prokázáno, že existuje rozdíl mezi kompresí fyziologických a patologických signálů. Patologické signály byly komprimovány s nižší efektivitou a kvalitou než signály fyziologické.
12

Komprese obrazu pomocí vlnkové transformace / Image Compression Using the Wavelet Transform

Kaše, David January 2015 (has links)
This thesis deals with image compression using wavelet, contourlet and shearlet transformation. It starts with quick look at image compression problem a quality measurement. Next are presented basic concepts of wavelets, multiresolution analysis and scaling function and detailed look at each transform. Representatives of algorithms for coeficients coding are EZW, SPIHT and marginally EBCOT. In second part is described design and implementation of constructed library. Last part compare result of transforms with format JPEG 2000. Comparison resulted in determining type of image in which implemented contourlet and shearlet transform were more effective than wavelet. Format JPEG 2000 was not exceeded.
13

Komprese signálů EKG nasnímaných pomocí mobilního zařízení / Compression of ECG signals recorded using mobile ECG device

Had, Filip January 2017 (has links)
Signal compression is necessary part for ECG scanning, because of relatively big amount of data, which must be transmitted primarily wirelessly for analysis. Because of the wireless sending it is necessary to minimize the amount of data as much as possible. To minimize the amount of data, lossless or lossy compression algorithms are used. This work describes an algorithm SPITH and newly created experimental method, based on PNG, and their testing. This master’s thesis there is also a bank of ECG signals with parallel sensed accelerometer data. In the last part, modification of SPIHT algorithm, which uses accelerometer data, is described and realized.
14

PCA and JPEG2000-based Lossy Compression for Hyperspectral Imagery

Zhu, Wei 30 April 2011 (has links)
This dissertation develops several new algorithms to solve existing problems in practical application of the previously developed PCA+JPEG2000, which has shown superior rate-distortion performance in hyperspectral image compression. In addition, a new scheme is proposed to facilitate multi-temporal hyperspectral image compression. Specifically, the uniqueness in each algorithm is described as follows. 1. An empirical piecewise linear equation is proposed to estimate the optimal number of major principal components (PCs) used in SubPCA+JPEG2000 for AVIRIS data. Sensor-specific equations are presented with excellent fitting performance for AVIRIS, HYDICE, and HyMap data. As a conclusion, a general guideline is provided for finding sensor-specific piecewise linear equations. 2. An anomaly-removal-based hyperspectral image compression algorithm is proposed. It preserves anomalous pixels in a lossless manner, and yields the same or even improved rate-distortion performance. It is particularly useful to SubPCA+JPEG2000 when compressing data with anomalies that may reside in minor PCs. 3. A segmented PCA-based PCA+JPEG2000 compression algorithm is developed, which spectrally partitions an image based on its spectral correlation coefficients. This compression scheme greatly improves the rate-distortion performance of PCA+JPEG2000 when the spatial size of the data is relatively smaller than its spectral size, especially at low bitrates. A sensor-specific partition method is also developed for fast processing with suboptimal performance. 4. A joint multi-temporal image compression scheme is proposed. The algorithm preserves change information in a lossless fashion during the compression. It can yield perfect change detection with slightly degraded rate-distortion performance.
15

Ultra High Compression For Weather Radar Reflectivity Data

Makkapati, Vishnu Vardhan 17 November 2006 (has links)
Honeywell Technology Solutions Lab, India / Weather is a major contributing factor in aviation accidents, incidents and delays. Doppler weather radar has emerged as a potent tool to observe weather. Aircraft carry onboard radars but their range and angular resolution are limited. Networks of ground-based weather radars provide extensive coverage of weather over large geographic regions. It would be helpful if these data can be transmitted to the pilot. However, these data are highly voluminous and the bandwidth of the ground-air communication links is limited and expensive. Hence, these data have to be compressed to an extent where they are suitable for transmission over low-bandwidth links. Several methods have been developed to compress pictorial data. General-purpose schemes do not take into account the nature of data and hence do not yield high compression ratios. A scheme for extreme compression of weather radar data is developed in this thesis that does not significantly degrade the meteorological information contained in these data. The method is based on contour encoding. It approximates a contour by a set of systematically chosen ‘control points’ that preserve its fine structure up to a certain level. The contours may be obtained using a thresholding process based on NWS or custom reflectivity levels. This process may result in region and hole contours, enclosing `high' or `low' areas, which may be nested. A tag bit is used to label region and hole contours. The control point extraction method first obtains a smoothed reference contour by averaging the original contour. Then the points on the original contour with maximum deviation from the smoothed contour between the crossings of these contours are identified and are designated as control points. Additional control points are added midway between the control point and the crossing points on either side of it, if the length of the segment between the crossing points exceeds a certain length. The control points, referenced with respect to the top-left corner of each contour for compact quantification, are transmitted to the receiving end. The contour is retrieved from the control points at the receiving end using spline interpolation. The region and hole contours are identified using the tag bit. The pixels between the region and hole contours at a given threshold level are filled using the color corresponding to it. This method is repeated till all the contours for a given threshold level are exhausted, and the process is carried out for all other thresholds, thereby resulting in a composite picture of the reconstructed field. Extensive studies have been conducted by using metrics such as compression ratio, fidelity of reconstruction and visual perception. In particular the effect of the smoothing factor, the choice of the degree of spline interpolation and the choice of thresholds are studied. It has been shown that a smoothing percentage of about 10% is optimal for most data. A degree 2 of spline interpolation is found to be best suited for smooth contour reconstruction. Augmenting NWS thresholds has resulted in improved visual perception, but at the expense of a decrease in the compression ratio. Two enhancements to the basic method that include adjustments to the control points to achieve better reconstruction and bit manipulations on the control points to obtain higher compression are proposed. The spline interpolation inherently tends to move the reconstructed contour away from the control points. This has been somewhat compensated by stretching the control points away from the smoothed reference contour. The amount and direction of stretch are optimized with respect to actual data fields to yield better reconstruction. In the bit manipulation study, the effects of discarding the least significant bits of the control point addresses are analyzed in detail. Simple bit truncation introduces a bias in the contour description and reconstruction, which is removed to a great extent by employing a bias compensation mechanism. The results obtained are compared with other methods devised for encoding weather radar contours.

Page generated in 0.0405 seconds