• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 3
  • 1
  • Tagged with
  • 7
  • 7
  • 3
  • 3
  • 2
  • 2
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
1

Design of the extended fixed-length instruction set for 32-bit X86 ISA

Lin, Jyun-Ji 04 August 2008 (has links)
In the microprocessor development, the high performance microprocessor applies the x86 complex instruction set is used widely. And the signal-core architecture towards slowly to multi-core one .But the variable-length instruction still creates the difficulties in instruction fetching and affects the whole executive- performance. There has the mechanism which supported the split-line and fetched fleetly the variable-lengths instruction. It has the problem in high time and hardware complexity, because it was accomplished with additional hardware. Accordingly, this paper proposed a fixed-length instruction set with design in compatible and extended x86 instruction set used the fixed-length instruction form to solve the difficulties in fetching the variable-length instructions. We considered the factor an overall arrangement of memory space and decided the length 4 bytes and 8 bytes to formulate the fixed-length instruction set. And we used the following six transitionary rules to complete the formulation for the coded form of the fixed-length instructions.(1)We used the auxiliary registers to save the value to decrease the data dependence between the original registers.(2)If it could use a few instructions to complete the translation with the original registers, we used the original registers to do it.(3)The complex case instructions were coded with eight bytes.(4)It did sign-extension by itself when displacement and immediate were moved to the auxiliary registers.(5)The auxiliary registers with the diacritic prefix were only coded in the r/m field or the index field.(6)One of displacement field and immediate field was moved first when its length was longer. And we considered the hardware complexity of saving memory space and fetching instructions, we analyzed the categories of instruction packages to compress the program space to decrease the space loss which the fixed-lengths of instructions created. In the case of verifiable and experimental framework, the CINT2006 was used to be benchmarks. And the function which translated the fixed-length instructions was succeeded to execute. It was successful to achieve the purpose the program space was compressed efficiently in the instruction package mechanism.
2

Robust video coding methods for next generation communication networks

Chung How, James T. H. January 2001 (has links)
No description available.
3

Multi-Symbol Codec for H.263 and the Synthesizable Verilog Code Generator Thereof

Lin, Jia-Hao 11 July 2007 (has links)
The first topic of this thesis is to carry out a multi-symbol codec (encoder-decoder) design for interfacing variable-length and fixed-length data conversion of H.263. The poor memory efficient of the variable-length can be avoided while its advantages can be reserved. The proposed codec converts variable-length symbols to fixed-length packets which can be decoded parallelly. The basic idea is to encode extra symbols in the redundant bits of the fixed-length packets. This encoding scheme relaxes the intrinsic poor compression rate of the prior fixed-length data codec. The second topic is a synthesizable Verilog code generator for the mentioned multi-symbol codec. According to different requirements and constraints of encoding bit rate, the generator can provide several different kinds of encoding modes by selecting proper parameters. Each codec generated by the generator is synthesizable by thorough simulations.
4

NTSC Digital Video Decoder and Multi-Symbol Codec

Chen, Chun-Chih 12 August 2004 (has links)
The first topic of this thesis proposes a digital video decoder for NTSC. The new fully digital design employs a DDFS (digital direct frequency synthesizer) and an adaptive digital PLL to track and lock the demodulation carrier. The complexity of the digital video decoder, hence, is drastically reduced. The overall cost of the proposed design is 6.0 mm2 (39K gates). The maximum power dissipation is 86 mW at the hightest clock rate which is 21.48 MHz. The second topic is to carry out a codec (encoder-decoder) design for interfacing variable-length and fixed-length data compression. The poor memory efficiency caused by the variable-length words converting into a fixed-length packet such that the compression can be hardwaredly and parallelly processing is significantly improved. The proposed codec is to encode more symbols in the redundant bits of the padding bits of the fixed-length packets. This novel encoding scheme relaxes the intrinsic poor bit rate of the traditional fixed-length data compression.
5

Hybrid orthogonal code sequences for high-density synchronous CDMA systems

Pramaita, Nyoman January 2014 (has links)
One of the primary tasks of the mobile system designers in order to support high density of devices in a CDMA system is to create a code sequence with a capacity for large number of spreading code sequences having low cross-correlation values between them, in order to ensure accommodation of large number of users and to minimise the effect of multiple access interference. In this research, the design for a novel hybrid orthogonal very large set (HOVLS) code sequence is proposed for high density mobile application scenarios. The design and development of both fixed and variable spreading factor code sequences are presented in this thesis. Both type of code sequences have been implemented via simulation using Matlab/Simulink. The performance of the code sequences has been evaluated and compared with that of existing code sequences. The proposed code sequences are more advantageous for high density mobile networks. The unique feature of the fixed length HOVLS code sequence is that its ACF, CCF, and BER performances are similar to that of orthogonal Gold code sequence and orthogonal m-sequence under Rayleigh flat and frequency selective fading channel conditions while having a significantly higher capacity than those orthogonal code sequences. The proposed HOVLS code sequence could support 134 different cells which is more than twice than that of orthogonal Gold code sequence and orthogonal m-sequence. To the knowledge of the author, this is the largest reported family size in the literature for an orthogonal code sequence for CDMA applications. In order to support variable data rate, fixed length HOVLS code sequence was developed into orthogonal variable spreading factor code sequence. It is shown that the proposed OVSF code sequence has slightly better CCF than those of OVSF Gold code sequence and m-sequence in terms of CM (correlation margin). The ACF of the proposed OVSF code sequence is similar to those of OVSF Gold code sequence and m-sequence. The proposed OVSF code sequence possesses comparable BER performance to those of orthogonal Gold code sequence and orthogonal m-sequence under flat fading channel condition. Whereas, the BER performance of the proposed OVSF code sequence is slightly better than that of Gold code sequence and OVSF m-sequence under frequency selective fading channel. Therefore, the proposed HOVLS code sequence is appropriate code sequence in CDMA systems than those of orthogonal Gold code sequence and orthogonal m-sequence for both fixed and variable rate high density network applications.
6

The analysis of enumerative source codes and their use in Burrows‑Wheeler compression algorithms

McDonald, Andre Martin 10 September 2010 (has links)
In the late 20th century the reliable and efficient transmission, reception and storage of information proved to be central to the most successful economies all over the world. The Internet, once a classified project accessible to a selected few, is now part of the everyday lives of a large part of the human population, and as such the efficient storage of information is an important part of the information economy. The improvement of the information storage density of optical and electronic media has been remarkable, but the elimination of redundancy in stored data and the reliable reconstruction of the original data is still a desired goal. The field of source coding is concerned with the compression of redundant data and its reliable decompression. The arithmetic source code, which was independently proposed by J. J. Rissanen and R. Pasco in 1976, revolutionized the field of source coding. Compression algorithms that use an arithmetic code to encode redundant data are typically more effective and computationally more efficient than compression algorithms that use earlier source codes such as extended Huffman codes. The arithmetic source code is also more flexible than earlier source codes, and is frequently used in adaptive compression algorithms. The arithmetic code remains the source code of choice, despite having been introduced more than 30 years ago. The problem of effectively encoding data from sources with known statistics (i.e. where the probability distribution of the source data is known) was solved with the introduction of the arithmetic code. The probability distribution of practical data is seldomly available to the source encoder, however. The source coding of data from sources with unknown statistics is a more challenging problem, and remains an active research topic. Enumerative source codes were introduced by T. J. Lynch and L. D. Davisson in the 1960s. These lossless source codes have the remarkable property that they may be used to effectively encode source sequences from certain sources without requiring any prior knowledge of the source statistics. One drawback of these source codes is the computationally complex nature of their implementations. Several years after the introduction of enumerative source codes, J. G. Cleary and I. H. Witten proved that approximate enumerative source codes may be realized by using an arithmetic code. Approximate enumerative source codes are significantly less complex than the original enumerative source codes, but are less effective than the original codes. Researchers have become more interested in arithmetic source codes than enumerative source codes since the publication of the work by Cleary and Witten. This thesis concerns the original enumerative source codes and their use in Burrows–Wheeler compression algorithms. A novel implementation of the original enumerative source code is proposed. This implementation has a significantly lower computational complexity than the direct implementation of the original enumerative source code. Several novel enumerative source codes are introduced in this thesis. These codes include optimal fixed–to–fixed length source codes with manageable computational complexity. A generalization of the original enumerative source code, which includes more complex data sources, is proposed in this thesis. The generalized source code uses the Burrows–Wheeler transform, which is a low–complexity algorithm for converting the redundancy of sequences from complex data sources to a more accessible form. The generalized source code effectively encodes the transformed sequences using the original enumerative source code. It is demonstrated and proved mathematically that this source code is universal (i.e. the code has an asymptotic normalized average redundancy of zero bits). AFRIKAANS : Die betroubare en doeltreffende versending, ontvangs en berging van inligting vorm teen die einde van die twintigste eeu die kern van die mees suksesvolle ekonomie¨e in die wˆereld. Die Internet, eens op ’n tyd ’n geheime projek en toeganklik vir slegs ’n klein groep verbruikers, is vandag deel van die alledaagse lewe van ’n groot persentasie van die mensdom, en derhalwe is die doeltreffende berging van inligting ’n belangrike deel van die inligtingsekonomie. Die verbetering van die bergingsdigteid van optiese en elektroniese media is merkwaardig, maar die uitwissing van oortolligheid in gebergde data, asook die betroubare herwinning van oorspronklike data, bly ’n doel om na te streef. Bronkodering is gemoeid met die kompressie van oortollige data, asook die betroubare dekompressie van die data. Die rekenkundige bronkode, wat onafhanklik voorgestel is deur J. J. Rissanen en R. Pasco in 1976, het ’n revolusie veroorsaak in die bronkoderingsveld. Kompressiealgoritmes wat rekenkundige bronkodes gebruik vir die kodering van oortollige data is tipies meer doeltreffend en rekenkundig meer effektief as kompressiealgoritmes wat vroe¨ere bronkodes, soos verlengde Huffman kodes, gebruik. Rekenkundige bronkodes, wat gereeld in aanpasbare kompressiealgoritmes gebruik word, is ook meer buigbaar as vroe¨ere bronkodes. Die rekenkundige bronkode bly na 30 jaar steeds die bronkode van eerste keuse. Die probleem om data wat afkomstig is van bronne met bekende statistieke (d.w.s. waar die waarskynlikheidsverspreiding van die brondata bekend is) doeltreffend te enkodeer is opgelos deur die instelling van rekenkundige bronkodes. Die bronenkodeerder het egter selde toegang tot die waarskynlikheidsverspreiding van praktiese data. Die bronkodering van data wat afkomstig is van bronne met onbekende statistieke is ’n groter uitdaging, en bly steeds ’n aktiewe navorsingsveld. T. J. Lynch and L. D. Davisson het tel–bronkodes in die 1960s voorgestel. Tel– bronkodes het die merkwaardige eienskap dat bronsekwensies van sekere bronne effektief met hierdie foutlose kodes ge¨enkodeer kan word, sonder dat die bronenkodeerder enige vooraf kennis omtrent die statistieke van die bron hoef te besit. Een nadeel van tel–bronkodes is die ho¨e rekenkompleksiteit van hul implementasies. J. G. Cleary en I. H. Witten het verskeie jare na die instelling van tel–bronkodes bewys dat benaderde tel–bronkodes gerealiseer kan word deur die gebruik van rekenkundige bronkodes. Benaderde tel–bronkodes het ’n laer rekenkompleksiteit as tel–bronkodes, maar benaderde tel–bronkodes is minder doeltreffend as die oorspronklike tel–bronkodes. Navorsers het sedert die werk van Cleary en Witten meer belangstelling getoon in rekenkundige bronkodes as tel–bronkodes. Hierdie tesis is gemoeid met die oorspronklike tel–bronkodes en die gebruik daarvan in Burrows–Wheeler kompressiealgoritmes. ’n Nuwe implementasie van die oorspronklike tel–bronkode word voorgestel. Die voorgestelde implementasie het ’n beduidende laer rekenkompleksiteit as die direkte implementasie van die oorspronklike tel–bronkode. Verskeie nuwe tel–bronkodes, insluitende optimale vaste–tot–vaste lengte tel–bronkodes met beheerbare rekenkompleksiteit, word voorgestel. ’n Veralgemening van die oorspronklike tel–bronkode, wat meer komplekse databronne insluit as die oorspronklike tel–bronkode, word voorgestel in hierdie tesis. The veralgemeende tel–bronkode maak gebruik van die Burrows–Wheeler omskakeling. Die Burrows–Wheeler omskakeling is ’n lae–kompleksiteit algoritme wat die oortolligheid van bronsekwensies wat afkomstig is van komplekse databronne omskakel na ’n meer toeganklike vorm. Die veralgemeende bronkode enkodeer die omgeskakelde sekwensies effektief deur die oorspronklike tel–bronkode te gebruik. Die universele aard van hierdie bronkode word gedemonstreer en wiskundig bewys (d.w.s. dit word bewys dat die kode ’n asimptotiese genormaliseerde gemiddelde oortolligheid van nul bisse het). Copyright / Dissertation (MEng)--University of Pretoria, 2010. / Electrical, Electronic and Computer Engineering / unrestricted
7

Analýza chování kořene kotvy v jemnozrnných zeminách / Behavior analysis of a ground anchor fixed length in fine grained soils

Chalmovský, Juraj January 2016 (has links)
Ground anchors represent an important structural element in the area of ground engineering. Despite an extensive usage of these elements, their design is usually performed using simple empirical and semi-empirical methods. An application of these procedures brings to the design a number of simplifying assumptions. The goal of the dissertation is to refine the computational design of ground anchors, analysis and quantification of selected factors significantly affecting their load displacement behavior. Firstly, the finite element method is applied. Two novel constitutive models are used: Multilaminate Constitutive Model for Stiff Soils (Schädlich, 2012) involving post peak shear softening of overconsolidated cohesive soils and Shotcrete Model (Schädlich, 2014) involving tension softening after tensile strength is reached. First constitutive model was used in order to simulate progressive decrease of skin friction along the anchor fixed length. Second constitutive model was applied for the grout material. In the next step, experimental program including several anchor load tests was carried out. The goal of the experimental program was to confirm conclusions from numerical studies and to obtain relevant data for further back – analysis. A newly developed application, in which all the findings from numerical computations and experimental measurements are incorporated, is described afterwards. The application is based on the use of so-called load-transfer functions. The program verification was conducted through series of back analysis of investigation anchor load tests realized on four different construction sites in two different types of fine-grained soils. The verification of the application is followed by series of parametric studies in which an influence of input parameters values is analyzed. Dissertation is concluded by the summarization of the most important findings.

Page generated in 0.0419 seconds