• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 3
  • Tagged with
  • 5
  • 5
  • 5
  • 3
  • 3
  • 3
  • 3
  • 2
  • 2
  • 2
  • 2
  • 2
  • 2
  • 2
  • 2
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
1

The equivocation of codes

Schofield, Mark January 2018 (has links)
Equivocation was introduced by Shannon in the late 1940’s in seminal papers that kick-started the whole field of information theory. Much ground has been covered on equivocation’s counterpart, channel capacity and in particular, its bounds. However, less work has been carried out on the evaluation of the equivocation of a code transmitted across a channel. The aim of the work covered in this thesis was to use a probabilistic approach to investigate and compare the equivocation of various codes across a range of channels. The probability and entropy of each output, given each input, can be used to calculate the equivocation. This gives a measure of the ambiguity and secrecy of a code when transmitted across a channel. The calculations increase exponentially in magnitude as both the message length and code length increase. In addition, the impact of factors such as erasures and deletions also serve to significantly complicate the process. In order to improve the calculation times offered by a conventional, linearly-programmed approach, an alternative strategy involving parallel processing with a CUDA-enabled (Compute Unified Device Architecture) graphical processor was employed. This enabled results to be obtained for codes of greater length than was possible with linear programming. However, the practical implementation of a CUDA driven, parallel processed solution gave rise to significant issues with both the software implementation and subsequent platform stability. By normalising equivocation results, it was possible to compare different codes under different conditions, making it possible to identify and select codes that gave a marked difference in the equivocation encountered by a legitimate receiver and an eavesdropper. The introduction of code expansion provided a novel method for enhancing equivocation differences still further. The work on parallel processing to calculate equivocation and the use of code expansion was published in the following conference: Schofield, M., Ahmed, M. & Tomlinson, M. (2015), Using parallel processing to calculate and improve equivocation, in ’IEEE Conference Publications - IEEE 16th International Conference on Communication Technology’. In addition to the novel use of a CUDA-enabled graphics process to calculated equivocation, equivocation calculations were also performed for expanded versions of the codes. Code expansion was shown to yield a dramatic increase in the achievable equivocation levels. Once methods had been developed with the Binary Symmetric Channel (BSC), they were extended to include work with intentional erasures on the BSC, intentional deletions on the BSC and work on the Binary Erasure Channel (BEC). The work on equivocation on the BSC with intentional erasures was published in: Schofield, M. et al, (2016), Intentional erasures and equivocation on the binary symmetric channel, in ’IEEE Conference Publications - International Computer Symposium’, IEEE, pp 233-235. The work on the BEC produced a novel outcome due to the erasure correction process employed. As the probability of an erasure occurring increases, the set of likely decoded outcomes diminishes. This directly impacts the output entropy of the system by decreasing it, thereby also affecting the equivocation value of the system. This aspect was something that had not been encountered previously. The work also extended to the consideration of intentional deletions on the BSC and the Binary Deletion Channel (BDC) itself. Although the methods used struggled to cope with the additional complexity brought by deletions, the use of Varshamov-Tenengolts codes on the BSC with intentional deletions showed that family of codes to be well suited to the channel arrangement as well as having the capability to be extended to enable the correction of multiple deletions.
2

Distributed Emitter Detector Design under Imperfect Communication Channel

Patra, Soumyadip 09 August 2017 (has links)
We consider the distributed detection of an emitter using multiple sensors deployed at deterministic locations. The signal from the emitter follows a signal attenuation model dependent on the distance between the sensor and the emitter. The sensors transmit their decisions to the fusion center through a parallel access Binary Symmetric Channel (BSC) with a cross-over probability. We seek to optimize the detection performance under a prescribed false alarm at the sensor level and at the system level. We consider the triangular topology structure and using the least favorable emitter range study the impact of the BSC on the system level detection fusion rules. The MAJORITY fusion rule is found to be optimal under certain conditions.
3

Error-robust coding and transformation of compressed hybered hybrid video streams for packet-switched wireless networks

Halbach, Till January 2004 (has links)
<p>This dissertation considers packet-switched wireless networks for transmission of variable-rate layered hybrid video streams. Target applications are video streaming and broadcasting services. The work can be divided into two main parts.</p><p>In the first part, a novel quality-scalable scheme based on coefficient refinement and encoder quality constraints is developed as a possible extension to the video coding standard H.264. After a technical introduction to the coding tools of H.264 with the main focus on error resilience features, various quality scalability schemes in previous research are reviewed. Based on this discussion, an encoder decoder framework is designed for an arbitrary number of quality layers, hereby also enabling region-of-interest coding. After that, the performance of the new system is exhaustively tested, showing that the bit rate increase typically encountered with scalable hybrid coding schemes is, for certain coding parameters, only small to moderate. The double- and triple-layer constellations of the framework are shown to perform superior to other systems.</p><p>The second part considers layered code streams as generated by the scheme of the first part. Various error propagation issues in hybrid streams are discussed, which leads to the definition of a decoder quality constraint and a segmentation of the code stream to transmit. A packetization scheme based on successive source rate consumption is drafted, followed by the formulation of the channel code rate optimization problem for an optimum assignment of available codes to the channel packets. Proper MSE-based error metrics are derived, incorporating the properties of the source signal, a terminate-on-error decoding strategy, error concealment, inter-packet dependencies, and the channel conditions. The Viterbi algorithm is presented as a low-complexity solution to the optimization problem, showing a great adaptivity of the joint source channel coding scheme to the channel conditions. An almost constant image qualiity is achieved, also in mismatch situations, while the overall channel code rate decreases only as little as necessary as the channel quality deteriorates. It is further shown that the variance of code distributions is only small, and that the codes are assigned irregularly to all channel packets.</p><p>A double-layer constellation of the framework clearly outperforms other schemes with a substantial margin. </p><p>Keywords — Digital lossy video compression, visual communication, variable bit rate (VBR), SNR scalability, layered image processing, quality layer, hybrid code stream, predictive coding, progressive bit stream, joint source channel coding, fidelity constraint, channel error robustness, resilience, concealment, packet-switched, mobile and wireless ATM, noisy transmission, packet loss, binary symmetric channel, streaming, broadcasting, satellite and radio links, H.264, MPEG-4 AVC, Viterbi, trellis, unequal error protection</p>
4

Error-robust coding and transformation of compressed hybered hybrid video streams for packet-switched wireless networks

Halbach, Till January 2004 (has links)
This dissertation considers packet-switched wireless networks for transmission of variable-rate layered hybrid video streams. Target applications are video streaming and broadcasting services. The work can be divided into two main parts. In the first part, a novel quality-scalable scheme based on coefficient refinement and encoder quality constraints is developed as a possible extension to the video coding standard H.264. After a technical introduction to the coding tools of H.264 with the main focus on error resilience features, various quality scalability schemes in previous research are reviewed. Based on this discussion, an encoder decoder framework is designed for an arbitrary number of quality layers, hereby also enabling region-of-interest coding. After that, the performance of the new system is exhaustively tested, showing that the bit rate increase typically encountered with scalable hybrid coding schemes is, for certain coding parameters, only small to moderate. The double- and triple-layer constellations of the framework are shown to perform superior to other systems. The second part considers layered code streams as generated by the scheme of the first part. Various error propagation issues in hybrid streams are discussed, which leads to the definition of a decoder quality constraint and a segmentation of the code stream to transmit. A packetization scheme based on successive source rate consumption is drafted, followed by the formulation of the channel code rate optimization problem for an optimum assignment of available codes to the channel packets. Proper MSE-based error metrics are derived, incorporating the properties of the source signal, a terminate-on-error decoding strategy, error concealment, inter-packet dependencies, and the channel conditions. The Viterbi algorithm is presented as a low-complexity solution to the optimization problem, showing a great adaptivity of the joint source channel coding scheme to the channel conditions. An almost constant image qualiity is achieved, also in mismatch situations, while the overall channel code rate decreases only as little as necessary as the channel quality deteriorates. It is further shown that the variance of code distributions is only small, and that the codes are assigned irregularly to all channel packets. A double-layer constellation of the framework clearly outperforms other schemes with a substantial margin. Keywords — Digital lossy video compression, visual communication, variable bit rate (VBR), SNR scalability, layered image processing, quality layer, hybrid code stream, predictive coding, progressive bit stream, joint source channel coding, fidelity constraint, channel error robustness, resilience, concealment, packet-switched, mobile and wireless ATM, noisy transmission, packet loss, binary symmetric channel, streaming, broadcasting, satellite and radio links, H.264, MPEG-4 AVC, Viterbi, trellis, unequal error protection
5

Extremal Problems of Error Exponents and Capacity of Duplication Channels

Ramezani, Mahdi Unknown Date
No description available.

Page generated in 0.0682 seconds