• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 177
  • 56
  • 9
  • 9
  • 6
  • 2
  • 2
  • 2
  • 1
  • 1
  • 1
  • Tagged with
  • 283
  • 283
  • 88
  • 78
  • 76
  • 72
  • 47
  • 43
  • 41
  • 40
  • 37
  • 36
  • 35
  • 33
  • 33
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
21

Transmissão de fluxos MPEG-2 com QoS s partir de servidores multimídia em redes ATM

Melo, Patrícia Lima Seixas Vieira de January 2001 (has links)
Dissertação (mestrado) - Universidade Federal de Santa Catarina, Centro Tecnológico. Programa de Pós-Graduação em Ciência da Computação / Made available in DSpace on 2012-10-18T10:32:19Z (GMT). No. of bitstreams: 0Bitstream added on 2014-09-25T20:53:58Z : No. of bitstreams: 1 181350.pdf: 4168422 bytes, checksum: 2d241c00d2a71342000a9e91be0b2eb0 (MD5) / Os servidores de armazenamento multimídia surgiram da necessidade de manipulação das informações multimídia. Diante das características deste tipo de informação, as quais diferenciam -se amplamente das informações alfanuméricas, o conceito de Qualidade de Serviço (QoS) precisou ser incorporado ao projeto deste tipo de servidor, visando proporcionar a entraga satifatória dos serviços solicitados. Aqualidade originalmente associada ao serviço, a fim de atender as necessidades dos usuários, pode ser modificada e desta forma, passar por um processo de adaptação às novas exigências dos mesmos. O trabalho de Dissertação de Mestrado apresentado tem como objetivo propor um modelo para a transmissão de fluxos MPEG-2 com QoS, a partir de servidores de armazenamento multimídia. Esta contribuição poderá posteriormente fazer parte de uma infra-estrutura maior, haja visto que o modelo proposto trata ainda dos processos de mapeamento e negociação dos parâmetros de Qualidade e Serviço executado durante a fase de configuração destas tranmissões. O presente trabalho baseia-se especificamente em servidores de vídeo MPEG-2 tendo como meio de transporte a Rede ATM, um vez que o mesmo interessa-se somente na definição de QoS ao nível desta tecnologia. A fim de que o embasamento necessário para a apresentação desta Dissertação fosse obtido, a exposição de alguns conceitos iniciais tornou-se essencialmente importante. Portanto, este trabalho introduz também os principais conceitos relativos a Aplicações Multimídia Distribuídas, Padrão de Compressão MPEG-2, Qualidade de Serviço, Tecnologia ATM e Sistema de Gerenciamento de Banco de Dados Multimídia e Servidores de Armazenamento Multimídia.
22

Multiple transforms for video coding / Transformées multiples pour le codage vidéo

Arrufat Batalla, Adrià 11 December 2015 (has links)
Les codeurs vidéo état de l’art utilisent des transformées pour assurer une représentation compacte du signal. L’étape de transformation constitue le domaine dans lequel s’effectue la compression, pourtant peu de variabilité dans les types de transformations est constatée dans les systèmes de codage vidéo normalisés : souvent, une seule transformée est considérée, habituellement la transformée en cosinus discrète (DCT). Récemment, d’autres transformées ont commencé à être considérées en complément de la DCT. Par exemple, dans le dernier standard de compression vidéo, nommé HEVC (High Efficiency Video Coding), les blocs de taille 4x4 peuvent utiliser la transformée en sinus discrète (DST), de plus, il est également possible de ne pas les transformer. Ceci révèle un intérêt croissant pour considérer une pluralité de transformées afin d’augmenter les taux de compression. Cette thèse se concentre sur l’extension de HEVC au travers de l’utilisation de multiples transformées. Après une introduction générale au codage vidéo et au codage par transformée, une étude détaillée de deux méthodes de construction de transformations est menée : la transformée de Karhunen Loève (KLT) et une transformée optimisée en débit et distorsion sont considérées. Ces deux méthodes sont comparées entre-elles en substituant les transformées utilisées par HEVC. Une expérimentation valide la pertinence des approches. Un schéma de codage qui incorpore et augmente l’utilisation de multiples transformées est alors introduit : plusieurs transformées sont mises à disposition de l’encodeur, qui sélectionne celle qui apporte le meilleur compromis dans le plan débit distorsion. Pour ce faire, une méthode de construction qui permet de concevoir des systèmes comportant de multiples transformations est décrite. Avec ce schéma de codage, le débit est significativement réduit par rapport à HEVC, tout particulièrement lorsque les transformées sont nombreuses et complexes à mettre en oeuvre. Néanmoins, ces améliorations viennent au prix d’une complexité accrue en termes d’encodage, de décodage et de contrainte de stockage. En conséquence, des simplifications sont considérées dans la suite du document, qui ont vocation à limiter l’impact en réduction de débit. Une première approche est introduite dans laquelle des transformées incomplètes sont motivées. Les transformations de ce type utilisent un seul vecteur de base, et sont conçues pour travailler de concert avec les transformations de HEVC. Cette technique est évaluée et apporte une réduction de complexité significative par rapport au précédent système, bien que la réduction de débit soit modeste. Une méthode systématique, qui détermine les meilleurs compromis entre le nombre de transformées et l’économie de débit est alors définie. Cette méthode utilise deux types différents de transformée : basés sur des transformées orthogonales séparables et des transformées trigonométriques discrètes (DTT) en particulier. Plusieurs points d’opération sont présentés qui illustrent plusieurs compromis complexité / gain en débit. Ces systèmes révèlent l’intérêt de l’utilisation de transformations multiples pour le codage vidéo. / State of the art video codecs use transforms to ensure a compact signal representation. The transform stage is where compression takes place, however, little variety is observed in the type of transforms used for standardised video coding schemes: often, a single transform is considered, usually a Discrete Cosine Transform (DCT). Recently, other transforms have started being considered in addition to the DCT. For instance, in the latest video coding standard, High Efficiency Video Coding (HEVC), the 4x4 sized blocks can make use of the Discrete Sine Transform (DST) and, in addition, it also possible not to transform them. This fact reveals an increasing interest to consider a plurality of transforms to achieve higher compression rates. This thesis focuses on extending HEVC through the use of multiple transforms. After a general introduction to video compression and transform coding, two transform designs are studied in detail: the Karhunen Loève Transform (KLT) and a Rate-Distortion Optimised Transform are considered. These two methods are compared against each other by replacing the transforms in HEVC. This experiment validates the appropriateness of the design. A coding scheme that incorporates and boosts the use of multiple transforms is introduced: several transforms are made available to the encoder, which chooses the one that provides the best rate-distortion trade-off. Consequently, a design method for building systems using multiple transforms is also described. With this coding scheme, significant amounts of bit-rate savings are achieved over HEVC, especially when using many complex transforms. However, these improvements come at the expense of increased complexity in terms of coding, decoding and storage requirements. As a result, simplifications are considered while limiting the impact on bit-rate savings. A first approach is introduced, in which incomplete transforms are used. This kind of transforms use one single base vector and are conceived to work as companions of the HEVC transforms. This technique is evaluated and provides significant complexity reductions over the previous system, although the bit-rate savings are modest. A systematic method, which specifically determines the best trade-offs between the number of transforms and bit-rate savings, is designed. This method uses two different types of transforms based separable orthogonal transforms and Discrete Trigonometric Transforms (DTTs) in particular. Several designs are presented, allowing for different complexity and bitrate savings trade-offs. These systems reveal the interest of using multiple transforms for video coding.
23

Scalable video coding using the Discrete Wavelet Transform : Skalbar videokodning med användning av den diskreta wavelettransformen

Johansson, Gustaf January 2010 (has links)
<p>A method for constructing a highly scalable bit stream for video coding is presented in detail and implemented in a demo application with a GUI in the Windows Vista operating system.</p><p>The video codec uses the Discrete Wavelet Transform in both spatial and temporal directions together with a zerotree quantizer to achieve a highly scalable bit stream in the senses of quality, spatial resolution and frame rate.</p> / <p>I detta arbete presenteras en metod för att skapa en mycket skalbar videoström. Metoden implementeras sedan i sin helhet i programspråken C och C++ med ett grafiskt användargränssnitt på operativsystemet Windows Vista.</p><p>I metoden används den diskreta wavelettransformen i såväl de spatiella dimensionerna som tidsdimensionen tillsammans med en nollträdskvantiserare för att åstakomma en skalbar videoström i avseendena bildkvalitet, skärmupplösning och antal bildrutor per sekund.</p>
24

Scalable video coding using the Discrete Wavelet Transform : Skalbar videokodning med användning av den diskreta wavelettransformen

Johansson, Gustaf January 2010 (has links)
A method for constructing a highly scalable bit stream for video coding is presented in detail and implemented in a demo application with a GUI in the Windows Vista operating system. The video codec uses the Discrete Wavelet Transform in both spatial and temporal directions together with a zerotree quantizer to achieve a highly scalable bit stream in the senses of quality, spatial resolution and frame rate. / I detta arbete presenteras en metod för att skapa en mycket skalbar videoström. Metoden implementeras sedan i sin helhet i programspråken C och C++ med ett grafiskt användargränssnitt på operativsystemet Windows Vista. I metoden används den diskreta wavelettransformen i såväl de spatiella dimensionerna som tidsdimensionen tillsammans med en nollträdskvantiserare för att åstakomma en skalbar videoström i avseendena bildkvalitet, skärmupplösning och antal bildrutor per sekund.
25

Multiterminal Video Coding: From Theory to Application

Zhang, Yifu 2012 August 1900 (has links)
Multiterminal (MT) video coding is a practical application of the MT source coding theory. For MT source coding theory, two problems associated with achievable rate regions are well investigated into in this thesis: a new sufficient condition for BT sum-rate tightness, and the sum-rate loss for quadratic Gaussian MT source coding. Practical code design for ideal Gaussian sources with quadratic distortion measure is also achieved for cases more than two sources with minor rate loss compared to theoretical limits. However, when the theory is applied to practical applications, the performance of MT video coding has been unsatisfactory due to the difficulty to explore the correlation between different camera views. In this dissertation, we present an MT video coding scheme under the H.264/AVC framework. In this scheme, depth camera information can be optionally sent to the decoder separately as another source sequence. With the help of depth information at the decoder end, inter-view correlation can be largely improved and thus so is the compression performance. With the depth information, joint estimation from decoded frames and side information at the decoder also becomes available to improve the quality of reconstructed video frames. Experimental result shows that compared to separate encoding, up to 9.53% of the bit rate can be saved by the proposed MT scheme using decoder depth information, while up to 5.65% can be saved by the scheme without depth camera information. Comparisons to joint video coding schemes are also provided.
26

Enhancing H.26x coding for visual communications - with applications in telemedicine and television

Khire, Sourabh Mohan 14 March 2013 (has links)
In a wireless and mobile communication paradigm, distribution and sharing of video content often occurs over unfriendly network environments constrained by lack of sufficient bandwidth, and prone to jitter, delay and packet losses. The research presented in this thesis proposed an assortment of application-specific optimizations designed to enable high-quality video communication over bandwidth constrained and unreliable channels. This assortment of solutions, termed herein as the Application Specific Video Coding and Delivery (ASVCD) toolkit, comprises of content and network adaptive approaches such as Region of Interest (ROI) video coding, Multiple Representation Coding (MRC), and Multiple Representation Coding of the Region of Interest (ROI + MRC). Thus, the effectiveness of ROI based video-coding in facilitating diagnostically lossless delivery of surgical videos over very low bandwidth channels was studied in this thesis. Furthermore, to facilitate error resilient video delivery over channels prone to burst losses and signal loss intervals, the MRC scheme was presented in this thesis. Finally, the thesis proposed a scheme for unequal protection of the ROI in the video by using the MRC scheme to effectively enable a distance learning application. To summarize, the ASVCD toolkit contributed in enabling high-quality video communications applications to become seamless and pervasive.
27

Model- and image-based scene representation.

January 1999 (has links)
Lee Kam Sum. / Thesis (M.Phil.)--Chinese University of Hong Kong, 1999. / Includes bibliographical references (leaves 97-101). / Abstracts in English and Chinese. / Chapter 1 --- Introduction --- p.2 / Chapter 1.1 --- Video representation using panorama mosaic and 3D face model --- p.2 / Chapter 1.2 --- Mosaic-based Video Representation --- p.3 / Chapter 1.3 --- "3D Human Face modeling ," --- p.7 / Chapter 2 --- Background --- p.13 / Chapter 2.1 --- Video Representation using Mosaic Image --- p.13 / Chapter 2.1.1 --- Traditional Video Compression --- p.17 / Chapter 2.2 --- 3D Face model Reconstruction via Multiple Views --- p.19 / Chapter 2.2.1 --- Shape from Silhouettes --- p.19 / Chapter 2.2.2 --- Head and Face Model Reconstruction --- p.22 / Chapter 2.2.3 --- Reconstruction using Generic Model --- p.24 / Chapter 3 --- System Overview --- p.27 / Chapter 3.1 --- Panoramic Video Coding Process --- p.27 / Chapter 3.2 --- 3D Face model Reconstruction Process --- p.28 / Chapter 4 --- Panoramic Video Representation --- p.32 / Chapter 4.1 --- Mosaic Construction --- p.32 / Chapter 4.1.1 --- Cylindrical Panorama Mosaic --- p.32 / Chapter 4.1.2 --- Cylindrical Projection of Mosaic Image --- p.34 / Chapter 4.2 --- Foreground Segmentation and Registration --- p.37 / Chapter 4.2.1 --- Segmentation Using Panorama Mosaic --- p.37 / Chapter 4.2.2 --- Determination of Background by Local Processing --- p.38 / Chapter 4.2.3 --- Segmentation from Frame-Mosaic Comparison --- p.40 / Chapter 4.3 --- Compression of the Foreground Regions --- p.44 / Chapter 4.3.1 --- MPEG-1 Compression --- p.44 / Chapter 4.3.2 --- MPEG Coding Method: I/P/B Frames --- p.45 / Chapter 4.4 --- Video Stream Reconstruction --- p.48 / Chapter 5 --- Three Dimensional Human Face modeling --- p.52 / Chapter 5.1 --- Capturing Images for 3D Face modeling --- p.53 / Chapter 5.2 --- Shape Estimation and Model Deformation --- p.55 / Chapter 5.2.1 --- Head Shape Estimation and Model deformation --- p.55 / Chapter 5.2.2 --- Face organs shaping and positioning --- p.58 / Chapter 5.2.3 --- Reconstruction with both intrinsic and extrinsic parameters --- p.59 / Chapter 5.2.4 --- Reconstruction with only Intrinsic Parameter --- p.63 / Chapter 5.2.5 --- Essential Matrix --- p.65 / Chapter 5.2.6 --- Estimation of Essential Matrix --- p.66 / Chapter 5.2.7 --- Recovery of 3D Coordinates from Essential Matrix --- p.67 / Chapter 5.3 --- Integration of Head Shape and Face Organs --- p.70 / Chapter 5.4 --- Texture-Mapping --- p.71 / Chapter 6 --- Experimental Result & Discussion --- p.74 / Chapter 6.1 --- Panoramic Video Representation --- p.74 / Chapter 6.1.1 --- Compression Improvement from Foreground Extraction --- p.76 / Chapter 6.1.2 --- Video Compression Performance --- p.78 / Chapter 6.1.3 --- Quality of Reconstructed Video Sequence --- p.80 / Chapter 6.2 --- 3D Face model Reconstruction --- p.91 / Chapter 7 --- Conclusion and Future Direction --- p.94 / Bibliography --- p.101
28

Foreground/background video coding for video conferencing =: 應用於視訊會議之前景/後景視訊編碼. / 應用於視訊會議之前景/後景視訊編碼 / Foreground/background video coding for video conferencing =: Ying yong yu shi xun hui yi zhi qian jing/ hou jing shi xun bian ma. / Ying yong yu shi xun hui yi zhi qian jing/ hou jing shi xun bian ma

January 2002 (has links)
Lee Kar Kin Edwin. / Thesis (M.Phil.)--Chinese University of Hong Kong, 2002. / Includes bibliographical references (leaves 129-134). / Text in English; abstracts in English and Chinese. / Lee Kar Kin Edwin. / Acknowledgement --- p.ii / Abstract --- p.iii / Contents --- p.vii / List of Figures --- p.ix / List of Tables --- p.xiii / Chapter 1 --- Introduction --- p.1 / Chapter 1.1 --- A brief review of transform-based video coding --- p.1 / Chapter 1.2 --- A brief review of content-based video coding --- p.6 / Chapter 1.3 --- Objectives of the research work --- p.9 / Chapter 1.4 --- Thesis outline --- p.12 / Chapter 2 --- Incorporation of DC Coefficient Restoration into Foreground/Background coding --- p.13 / Chapter 2.1 --- Introduction --- p.13 / Chapter 2.2 --- A review of FB coding in H.263 sequence --- p.15 / Chapter 2.3 --- A review of DCCR --- p.18 / Chapter 2.4 --- DCCRFB coding --- p.23 / Chapter 2.4.1 --- Methodology --- p.23 / Chapter 2.4.2 --- Implementation --- p.24 / Chapter 2.4.3 --- Experimental results --- p.26 / Chapter 2.5 --- The use of block selection scheme in DCCRFB coding --- p.32 / Chapter 2.5.1 --- Introduction --- p.32 / Chapter 2.5.2 --- Experimental results --- p.34 / Chapter 2.6 --- Summary --- p.47 / Chapter 3 --- Chin contour estimation on foreground human faces --- p.48 / Chapter 3.1 --- Introduction --- p.48 / Chapter 3.2 --- Least mean square estimation of chin location --- p.50 / Chapter 3.3 --- Chin contour estimation using chin edge detector and contour modeling --- p.58 / Chapter 3.3.1 --- Face segmentation and facial organ extraction --- p.59 / Chapter 3.3.2 --- Identification of search window --- p.59 / Chapter 3.3.3 --- Edge detection using chin edge detector --- p.60 / Chapter 3.3.4 --- "Determination of C0, C1 and c2" --- p.63 / Chapter 3.3.5 --- Chin contour modeling --- p.67 / Chapter 3.4 --- Experimental results --- p.71 / Chapter 3.5 --- Summary --- p.77 / Chapter 4 --- Wire-frame model deformation and face animation using FAP --- p.78 / Chapter 4.1 --- Introduction --- p.78 / Chapter 4.2 --- Wire-frame face model deformation --- p.79 / Chapter 4.2.1 --- Introduction --- p.79 / Chapter 4.2.2 --- Wire-frame model selection and FDP generation --- p.81 / Chapter 4.2.3 --- Global deformation --- p.85 / Chapter 4.2.4 --- Local deformation --- p.87 / Chapter 4.2.5 --- Experimental results --- p.93 / Chapter 4.3 --- Face animation using FAP --- p.98 / Chapter 4.3.1 --- Introduction and methodology --- p.98 / Chapter 4.3.2 --- Experiments --- p.102 / Chapter 4.4 --- Summary --- p.112 / Chapter 5 --- Conclusions and future developments --- p.113 / Chapter 5.1 --- Contributions and conclusions --- p.113 / Chapter 5.2 --- Future developments --- p.117 / Appendix A H.263 bitstream syntax --- p.122 / Appendix B Excerpt of the FAP specification table [17] --- p.123 / Bibliography --- p.129
29

Robust and efficient techniques for automatic video segmentation.

January 1998 (has links)
by Lam Cheung Fai. / Thesis (M.Phil.)--Chinese University of Hong Kong, 1998. / Includes bibliographical references (leaves 174-179). / Abstract also in Chinese. / Chapter 1 --- Introduction --- p.1 / Chapter 1.1 --- Problem Definition --- p.2 / Chapter 1.2 --- Motivation --- p.5 / Chapter 1.3 --- Problems --- p.7 / Chapter 1.3.1 --- Illumination Changes and Motions in Videos --- p.7 / Chapter 1.3.2 --- Variations in Video Scene Characteristics --- p.8 / Chapter 1.3.3 --- High Complexity of Algorithms --- p.10 / Chapter 1.3.4 --- Heterogeneous Approaches to Video Segmentation --- p.10 / Chapter 1.4 --- Objectives and Approaches --- p.11 / Chapter 1.5 --- Organization of the Thesis --- p.13 / Chapter 2 --- Related Work --- p.15 / Chapter 2.1 --- Algorithms for Uncompressed Videos --- p.16 / Chapter 2.1.1 --- Pixel-based Method --- p.16 / Chapter 2.1.2 --- Histogram-based Method --- p.17 / Chapter 2.1.3 --- Motion-based Algorithms --- p.18 / Chapter 2.1.4 --- Color-ratio Based Algorithms --- p.18 / Chapter 2.2 --- Algorithms for Compressed Videos --- p.19 / Chapter 2.2.1 --- Algorithms based on JPEG Image Sequences --- p.19 / Chapter 2.2.2 --- Algorithms based on MPEG Videos --- p.20 / Chapter 2.2.3 --- Algorithms based on VQ Compressed Videos --- p.21 / Chapter 2.3 --- Frame Difference Analysis Methods --- p.21 / Chapter 2.3.1 --- Scene Cut Detection --- p.21 / Chapter 2.3.2 --- Gradual Transition Detection --- p.22 / Chapter 2.4 --- Speedup Techniques --- p.23 / Chapter 2.5 --- Other Approaches --- p.24 / Chapter 3 --- Analysis and Enhancement of Existing Algorithms --- p.25 / Chapter 3.1 --- Introduction --- p.25 / Chapter 3.2 --- Video Segmentation Algorithms --- p.26 / Chapter 3.2.1 --- Frame Difference Metrics --- p.26 / Chapter 3.2.2 --- Frame Difference Analysis Methods --- p.29 / Chapter 3.3 --- Analysis of Feature Extraction Algorithms --- p.30 / Chapter 3.3.1 --- Pair-wise pixel comparison --- p.30 / Chapter 3.3.2 --- Color histogram comparison --- p.34 / Chapter 3.3.3 --- Pair-wise block-based comparison of DCT coefficients --- p.38 / Chapter 3.3.4 --- Pair-wise pixel comparison of DC-images --- p.42 / Chapter 3.4 --- Analysis of Scene Change Detection Methods --- p.45 / Chapter 3.4.1 --- Global Threshold Method --- p.45 / Chapter 3.4.2 --- Sliding Window Method --- p.46 / Chapter 3.5 --- Enhancements and Modifications --- p.47 / Chapter 3.5.1 --- Histogram Equalization --- p.49 / Chapter 3.5.2 --- DD Method --- p.52 / Chapter 3.5.3 --- LA Method --- p.56 / Chapter 3.5.4 --- Modification for pair-wise pixel comparison --- p.57 / Chapter 3.5.5 --- Modification for pair-wise DCT block comparison --- p.61 / Chapter 3.6 --- Conclusion --- p.69 / Chapter 4 --- Color Difference Histogram --- p.72 / Chapter 4.1 --- Introduction --- p.72 / Chapter 4.2 --- Color Difference Histogram --- p.73 / Chapter 4.2.1 --- Definition of Color Difference Histogram --- p.73 / Chapter 4.2.2 --- Sparse Distribution of CDH --- p.76 / Chapter 4.2.3 --- Resolution of CDH --- p.77 / Chapter 4.2.4 --- CDH-based Inter-frame Similarity Measure --- p.77 / Chapter 4.2.5 --- Computational Cost and Discriminating Power --- p.80 / Chapter 4.2.6 --- Suitability in Scene Change Detection --- p.83 / Chapter 4.3 --- Insensitivity to Illumination Changes --- p.89 / Chapter 4.3.1 --- Sensitivity of CDH --- p.90 / Chapter 4.3.2 --- Comparison with other feature extraction algorithms --- p.93 / Chapter 4.4 --- Orientation and Motion Invariant --- p.96 / Chapter 4.4.1 --- Camera Movements --- p.97 / Chapter 4.4.2 --- Object Motion --- p.100 / Chapter 4.4.3 --- Comparison with other feature extraction algorithms --- p.100 / Chapter 4.5 --- Performance of Scene Cut Detection --- p.102 / Chapter 4.6 --- Time Complexity Comparison --- p.105 / Chapter 4.7 --- Extension to DCT-compressed Images --- p.106 / Chapter 4.7.1 --- Performance of scene cut detection --- p.108 / Chapter 4.8 --- Conclusion --- p.109 / Chapter 5 --- Scene Change Detection --- p.111 / Chapter 5.1 --- Introduction --- p.111 / Chapter 5.2 --- Previous Approaches --- p.112 / Chapter 5.2.1 --- Scene Cut Detection --- p.112 / Chapter 5.2.2 --- Gradual Transition Detection --- p.115 / Chapter 5.3 --- DD Method --- p.116 / Chapter 5.3.1 --- Detecting Scene Cuts --- p.117 / Chapter 5.3.2 --- Detecting 1-frame Transitions --- p.121 / Chapter 5.3.3 --- Detecting Gradual Transitions --- p.129 / Chapter 5.4 --- Local Thresholding --- p.131 / Chapter 5.5 --- Experimental Results --- p.134 / Chapter 5.5.1 --- Performance of CDH+DD and CDH+DL --- p.135 / Chapter 5.5.2 --- Performance of DD on other features --- p.144 / Chapter 5.6 --- Conclusion --- p.150 / Chapter 6 --- Motion Vector Based Approach --- p.151 / Chapter 6.1 --- Introduction --- p.151 / Chapter 6.2 --- Previous Approaches --- p.152 / Chapter 6.3 --- MPEG-I Video Stream Format --- p.153 / Chapter 6.4 --- Derivation of Frame Differences from Motion Vector Counts --- p.156 / Chapter 6.4.1 --- Types of Frame Pairs --- p.156 / Chapter 6.4.2 --- Conditions for Scene Changes --- p.157 / Chapter 6.4.3 --- Frame Difference Measure --- p.159 / Chapter 6.5 --- Experiment --- p.160 / Chapter 6.5.1 --- Performance of MV --- p.161 / Chapter 6.5.2 --- Performance Enhancement --- p.162 / Chapter 6.5.3 --- Limitations --- p.163 / Chapter 6.6 --- Conclusion --- p.164 / Chapter 7 --- Conclusion and Future Work --- p.165 / Chapter 7.1 --- Contributions --- p.165 / Chapter 7.2 --- Future Work --- p.169 / Chapter 7.3 --- Conclusion --- p.171 / Bibliography --- p.174 / Chapter A --- Sample Videos --- p.180 / Chapter B --- List of Abbreviations --- p.183
30

Error-resilient coding tools in MPEG-4.

January 1998 (has links)
by Cheng Shu Ling. / Thesis submitted in: July 1997. / Thesis (M.Phil.)--Chinese University of Hong Kong, 1998. / Includes bibliographical references (leaves 70-71). / Abstract also in Chinese. / Chapter Chapter 1 --- Introduction --- p.1 / Chapter 1.1 --- Image Coding Standard: JPEG --- p.1 / Chapter 1.2 --- Video Coding Standard: MPEG --- p.6 / Chapter 1.2.1 --- MPEG history --- p.6 / Chapter 1.2.2 --- MPEG video compression algorithm overview --- p.8 / Chapter 1.2.3 --- More MPEG features --- p.10 / Chapter 1.3 --- Summary --- p.17 / Chapter Chapter 2 --- Error Resiliency --- p.18 / Chapter 2.1 --- Introduction --- p.18 / Chapter 2.2 --- Traditional approaches --- p.19 / Chapter 2.2.1 --- Channel coding --- p.19 / Chapter 2.2.2 --- ARQ --- p.20 / Chapter 2.2.3 --- Multi-layer coding --- p.20 / Chapter 2.2.4 --- Error Concealment --- p.20 / Chapter 2.3 --- MPEG-4 work on error resilience --- p.21 / Chapter 2.3.1 --- Resynchronization --- p.21 / Chapter 2.3.2 --- Data Recovery --- p.25 / Chapter 2.3.3 --- Error Concealment --- p.28 / Chapter 2.4 --- Summary --- p.29 / Chapter Chapter 3 --- Fixed length codes --- p.30 / Chapter 3.1 --- Introduction --- p.30 / Chapter 3.2 --- Tunstall code --- p.31 / Chapter 3.3 --- Lempel-Ziv code --- p.34 / Chapter 3.3.1 --- LZ-77 --- p.35 / Chapter 3.3.2 --- LZ-78 --- p.36 / Chapter 3.4 --- Simulation --- p.38 / Chapter 3.4.1 --- Experiment Setup --- p.38 / Chapter 3.4.2 --- Results --- p.39 / Chapter 3.4.3 --- Concluding Remarks --- p.42 / Chapter Chapter 4 --- Self-Synchronizable codes --- p.44 / Chapter 4.1 --- Introduction --- p.44 / Chapter 4.2 --- Scholtz synchronizable code --- p.45 / Chapter 4.2.1 --- Definition --- p.45 / Chapter 4.2.2 --- Construction procedure --- p.45 / Chapter 4.2.3 --- Synchronizer --- p.48 / Chapter 4.2.4 --- Effects of errors --- p.51 / Chapter 4.3 --- Simulation --- p.52 / Chapter 4.3.1 --- Experiment Setup --- p.52 / Chapter 4.3.2 --- Results --- p.56 / Chapter 4.4 --- Concluding Remarks --- p.68 / Chapter Chapter 5 --- Conclusions --- p.69 / References --- p.70

Page generated in 0.0851 seconds