• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 183
  • 30
  • 14
  • 10
  • 8
  • 5
  • 5
  • 4
  • 2
  • 2
  • 1
  • 1
  • 1
  • 1
  • 1
  • Tagged with
  • 309
  • 309
  • 79
  • 64
  • 58
  • 47
  • 47
  • 42
  • 40
  • 37
  • 36
  • 32
  • 31
  • 29
  • 27
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
191

Analysis And Design Of Image And Video Encryption Algorithms

Yekkala, Anil Kumar 12 1900 (has links)
The rapid growth in multimedia based Internet systems and applications like video telephony, video on demand, network based DVD recorders and IP television has created a substantial need for multimedia security. One of the important requirements for multimedia security is transmission of the digital multimedia content in a secure manner using encryption for protecting it from eavesdropping. The simplest way of encrypting multimedia content is to consider the two-dimensional/three-dimensional image/video stream as an one-dimensional stream and to encrypt the entire content using standard block ciphers like AES, DES, IDEA or RC4 or using a stream cipher. The method of encrypting the entire multimedia content is considered as a naive encryption approach. Even though the naive encryption approach provides the desired security requirements, it imposes a large overhead on the multimedia codex. This is due to the size of the multimedia content, and also due to real time requirements of transmission and rendering. Hence, lightweight encryption schemes are gaining popularity for multimedia encryption. Lightweight Encryption schemes are based on the principle “Encrypt minimal and induce maximum noise". Lightweight encryption schemes are designed to take the structure of the multimedia content into consideration. In our work we analyze some of the existing lightweight encryption schemes for digital images and video. The analysis is done based on the amount of security, scalability and effect on compression. A detailed study of some of the existing lightweight encryption schemes is also done by designing cryptanalysis schemes. The cryptanalysis schemes are designed using image noise clearing algorithms and pixel prediction techniques. The designed cryptanalysis schemes reduce the amount of noise introduced by the corresponding lightweight encryption schemes considerably. Based on our analysis of existing lightweight encryption schemes, we propose a set of more robust lightweight encryption schemes for images and video. The proposed lightweight encryption schemes are secure, scalable, and do not degrade the compression achieved. In our work, we also propose a few enhancements to JPEG image compression for achieving more compression, without compromising on the quality. The enhancements to the JPEG compression are extensions of the pixel prediction techniques used in the proposed cryptanalysis schemes.
192

Advanced wavelet image and video coding strategies for multimedia communications

Vass, Jozsef January 2000 (has links)
Thesis (Ph. D.)--University of Missouri-Columbia, 2000. / Typescript. Vita. Includes bibliographical references (leaves 202-221). Also available on the Internet.
193

Development of a new image compression technique using a grid smoothing technique.

Bashala, Jenny Mwilambwe. January 2013 (has links)
M. Tech. Electrical Engineering. / Aims to implement a lossy image compression scheme that uses a graph-based approach. On the one hand, this new method should reach high compression rates with good visual quality, while on the other hand it may lead to the following sub-problems:efficient classification of image data with the use of bilateral mesh filtering ; Transformation of the image into a graph with grid smoothing ; reduction of the graph by means of mesh decimation techniques ; reconstruction process of the reduced graph into an image and quality analysis of the reconstructed images.
194

Codec for multimedia services using wavelets and fractals.

Brijmohan, Yarish. January 2004 (has links)
Increase in technological advancements in fields of telecommunications, computers and television have prompted the need to exchange video, image and audio files between people. Transmission of such files finds numerous multimedia applications such as, internet multimedia, video conferencing, videophone, etc. However, the transmission and rece-ption of these files are limited by the available bandwidth as well as storage capacities of systems. Thus there is a need to develop compression systems, such that required multimedia applications can operate within these limited capacities. This dissertation presents two well established coding approaches that are used in modern' image and video compression systems. These are the wavelet and fractal methods. The wavelet based coder, which adopts the transform coding paradigm, performs the discrete wavelet transform on an image before any compression algorithms are implemented. The wavelet transform provides good energy compaction and decorrelating properties that make it suited for compression. Fractal compression systems on the other hand differ from the traditional transform coders. These algorithms are based on the theory of iterated function systems and take advantage of local self-similarities present in images. In this dissertation, we first review the theoretical foundations of both wavelet and fractal coders. Thereafter we evaluate different wavelet and fractal based compression algorithms, and assess the strengths and weakness in each case. Due to the short-comings of fractal based compression schemes, such as the tiling effect appearing in reconstructed images, a wavelet based analysis of fractal image compression is presented. This is the link that produces fractal coding in the wavelet domain, and presents a hybrid coding scheme called fractal-wavelet coders. We show that by using smooth wavelet basis in computing the wavelet transform, the tiling effect of fractal systems can be removed. The few wavelet-fractal coders that have been proposed in literature are discussed, showing advantages over the traditional fractal coders. This dissertation will present a new low-bit rate video compression system that is based on fractal coding in the wavelet domain. This coder makes use of the advantages of both the wavelet and fractal coders discussed in their review. The self-similarity property of fractal coders exploits the high spatial and temporal correlation between video frames. Thus the fractal coding step gives an approximate representation of the coded frame, while the wavelet technique adds detail to the frame. In this proposed scheme, each frame is decomposed using the pyramidal multi-resolution wavelet transform. Thereafter a motion detection operation is used in which the subtrees are partitioned into motion and non-motion subtrees. The nonmotion subtrees are easily coded by a binary decision, whereas the moving ones are coded using the combination of the wavelet SPIHT and fractal variable subtree size coding scheme. All intra-frame compression is performed using the SPIHT compression algorithm and inter-frame using the fractal-wavelet method described above. The proposed coder is then compared to current low bit-rate video coding standards such as the H.263+ and MPEG-4 coders through analysis and simulations. Results show that the proposed coder is competitive with the current standards, with a performance improvement been shown in video sequences that do not posses large global motion. Finally, a real-time implementation of the proposed algorithm is performed on a digital signal processor. This illustrates the suitability of the proposed coder being applied to numerous multimedia applications. / Thesis (M.Sc.Eng.)-University of KwaZulu-Natal, Durban, 2004.
195

Maximum Energy Subsampling: A General Scheme For Multi-resolution Image Representation And Analysis

Zhao, Yanjun 18 December 2014 (has links)
Image descriptors play an important role in image representation and analysis. Multi-resolution image descriptors can effectively characterize complex images and extract their hidden information. Wavelets descriptors have been widely used in multi-resolution image analysis. However, making the wavelets transform shift and rotation invariant produces redundancy and requires complex matching processes. As to other multi-resolution descriptors, they usually depend on other theories or information, such as filtering function, prior-domain knowledge, etc.; that not only increases the computation complexity, but also generates errors. We propose a novel multi-resolution scheme that is capable of transforming any kind of image descriptor into its multi-resolution structure with high computation accuracy and efficiency. Our multi-resolution scheme is based on sub-sampling an image into an odd-even image tree. Through applying image descriptors to the odd-even image tree, we get the relative multi-resolution image descriptors. Multi-resolution analysis is based on downsampling expansion with maximum energy extraction followed by upsampling reconstruction. Since the maximum energy usually retained in the lowest frequency coefficients; we do maximum energy extraction through keeping the lowest coefficients from each resolution level. Our multi-resolution scheme can analyze images recursively and effectively without introducing artifacts or changes to the original images, produce multi-resolution representations, obtain higher resolution images only using information from lower resolutions, compress data, filter noise, extract effective image features and be implemented in parallel processing.
196

Adaptive strategies and optimization techniques for JPEG-based low bit-rate image coding

Sevcenco, Ana-Maria 26 February 2010 (has links)
The field of digital image compression has been intensively explored to obtain ever improved performance corresponding to a given bit budget. The DCT-based JPEG standard remains to be one of the most popular image compression standards due to its reasonable coding performance, fast implementations, friendly low-cost architecture. flexibility and adaptivity on block level. In this thesis, we consider the problem of low bit-rate image coding and present new approaches using adaptive strategies and optimization techniques for performance enhancement, while employing the DCT block-based JPEG standard as the main framework with several pre- and post-processing , steps. We propose an adaptive coding approach which involves a variable quality factor in the quantization step of JPEG compression to make the compression more flexible in respect to bit budget requirements. We also propose an adaptive sampling approach based on variable down-'up-scaling rate and local image characteristics. In addition. we study an adaptive filtering approach in which the optimal filters coefficients are determined by making use of optimization methods and symmetric extension techniques. Simulation results are presented to demonstrate the effectiveness of the proposed techniques relative to recent works in the field of low bit-rate image coding.
197

Maximum Energy Subsampling: A General Scheme For Multi-resolution Image Representation And Analysis

Zhao, Yanjun 18 December 2014 (has links)
Image descriptors play an important role in image representation and analysis. Multi-resolution image descriptors can effectively characterize complex images and extract their hidden information. Wavelet descriptors have been widely used in multi-resolution image analysis. However, making the wavelet transform shift and rotation invariant produces redundancy and requires complex matching processes. As to other multi-resolution descriptors, they usually depend on other methods, such as filtering function, prior-domain knowledge, etc.; that not only increases the computation complexity, but also generates errors. We propose a novel multi-resolution scheme that is capable of transforming any kind of image descriptor into its multi-resolution structure with high computation accuracy and efficiency. Our multi-resolution scheme is based on sub-sampling each image into an odd-even image tree. Through applying image descriptors to the odd-even image tree, we get the relative multi-resolution image descriptors. Multi-resolution analysis is based on downsampling expansion with maximum energy extraction followed by upsampling reconstruction. Since the maximum energy usually retained in the lowest frequency coefficients; we do maximum energy extraction through keeping the lowest coefficients from each resolution level. Our multi-resolution scheme can analyze images recursively and effectively without introducing artifacts or changes to the original images, produce multi-resolution representations, obtain higher resolution images only using information from lower resolutions, compress data, filter noise, extract effective image features and be implemented in parallel processing.
198

Fixed-analysis adaptive-synthesis filter banks

Lettsome, Clyde Alphonso 07 April 2009 (has links)
Subband/Wavelet filter analysis-synthesis filters are a major component in many compression algorithms. Such compression algorithms have been applied to images, voice, and video. These algorithms have achieved high performance. Typically, the configuration for such compression algorithms involves a bank of analysis filters whose coefficients have been designed in advance to enable high quality reconstruction. The analysis system is then followed by subband quantization and decoding on the synthesis side. Decoding is performed using a corresponding set of synthesis filters and the subbands are merged together. For many years, there has been interest in improving the analysis-synthesis filters in order to achieve better coding quality. Adaptive filter banks have been explored by a number of authors where by the analysis filters and synthesis filters coefficients are changed dynamically in response to the input. A degree of performance improvement has been reported but this approach does require that the analysis system dynamically maintain synchronization with the synthesis system in order to perform reconstruction. In this thesis, we explore a variant of the adaptive filter bank idea. We will refer to this approach as fixed-analysis adaptive-synthesis filter banks. Unlike the adaptive filter banks proposed previously, there is no analysis synthesis synchronization issue involved. This implies less coder complexity and more coder flexibility. Such an approach can be compatible with existing subband wavelet encoders. The design methodology and a performance analysis are presented.
199

Adaptive image and video compression using vector quantization and self-organizing neural networks

Liu, Hui January 1994 (has links)
Thesis (Ph. D.)--University of Hawaii at Manoa, 1994. / Includes bibliographical references (leaves 199-209). / Microfiche. / xii, 209 leaves, bound ill. 29 cm
200

Error resilience in JPEG2000 /

Natu, Ambarish Shrikrishna. January 2003 (has links)
Thesis (M.E.)--University of New South Wales, 2003. / Also available online.

Page generated in 0.2695 seconds