Spelling suggestions: "subject:"applikationsspecifika"" "subject:"applikationsserver""
1 |
Design of Application Specific Microcontroller / Design av applikationsspecifik mikrocontrollerMartinsson, Kristoffer January 2002 (has links)
<p>This master thesis describes the process of designing an application specific microcontroller. The microcontroller should be used in a demonstrator for a protocol processor. </p><p>The demonstrator should show the possibilities to access high speed networks with small processor cores. The demonstrator should be able to receive and playback an audio stream. Some of the tasks in the demonstrator should be performed by the microcontroller. It should handle ARP requests, buffer handling and sending audio samples to a stereo codec. Behavioral models for these applications were constructed and used to design the instruction set for the microcontroller. </p><p>An instruction set simulator was constructed. It was used to verify that the instruction set was sufficient to achieve functional coverage. </p><p>The micro architecture for the microcontroller was designed and implemented in VHDL. This implementation was verified by simulation. The test vectors used during simulation were mainly randomly generated.</p>
|
2 |
Design of Application Specific Microcontroller / Design av applikationsspecifik mikrocontrollerMartinsson, Kristoffer January 2002 (has links)
This master thesis describes the process of designing an application specific microcontroller. The microcontroller should be used in a demonstrator for a protocol processor. The demonstrator should show the possibilities to access high speed networks with small processor cores. The demonstrator should be able to receive and playback an audio stream. Some of the tasks in the demonstrator should be performed by the microcontroller. It should handle ARP requests, buffer handling and sending audio samples to a stereo codec. Behavioral models for these applications were constructed and used to design the instruction set for the microcontroller. An instruction set simulator was constructed. It was used to verify that the instruction set was sufficient to achieve functional coverage. The micro architecture for the microcontroller was designed and implemented in VHDL. This implementation was verified by simulation. The test vectors used during simulation were mainly randomly generated.
|
3 |
Programmable Address Generation Unit for Deep Neural Network AcceleratorsKhan, Muhammad Jazib January 2020 (has links)
The Convolutional Neural Networks are getting more and more popular due to their applications in revolutionary technologies like Autonomous Driving, Biomedical Imaging, and Natural Language Processing. With this increase in adoption, the complexity of underlying algorithms is also increasing. This trend entails implications for the computation platforms as well, i.e. GPUs, FPGA, or ASIC based accelerators, especially for the Address Generation Unit (AGU), which is responsible for the memory access. Existing accelerators typically have Parametrizable Datapath AGUs, which have minimal adaptability towards evolution in algorithms. Hence new hardware is required for new algorithms, which is a very inefficient approach in terms of time, resources, and reusability. In this research, six algorithms with different implications for hardware are evaluated for address generation, and a fully Programmable AGU (PAGU) is presented, which can adapt to these algorithms. These algorithms are Standard, Strided, Dilated, Upsampled and Padded convolution, and MaxPooling. The proposed AGU architecture is a Very Long Instruction Word based Application Specific Instruction Processor which has specialized components like hardware counters and zero-overhead loops and a powerful Instruction Set Architecture (ISA), which can model static and dynamic constraints and affine and non-affine Address Equations. The target has been to minimize the flexibility vs. area, power, and performance trade-off. For a working test network of Semantic Segmentation, results have shown that PAGU shows close to the ideal performance, one cycle per address, for all the algorithms under consideration excepts Upsampled Convolution for which it is 1.7 cycles per address. The area of PAGU is approx. 4.6 times larger than the Parametrizable Datapath approach, which is still reasonable considering the high flexibility benefits. The potential of PAGU is not just limited to neural network applications but also in more general digital signal processing areas, which can be explored in the future. / Convolutional Neural Networks blir mer och mer populära på grund av deras applikationer inom revolutionerande tekniker som autonom körning, biomedicinsk bildbehandling och naturligt språkbearbetning. Med denna ökning av antagandet ökar också komplexiteten hos underliggande algoritmer. Detta medför implikationer för beräkningsplattformarna såväl som GPU: er, FPGAeller ASIC-baserade acceleratorer, särskilt för Adressgenerationsenheten (AGU) som är ansvarig för minnesåtkomst. Befintliga acceleratorer har normalt Parametrizable Datapath AGU: er som har mycket begränsad anpassningsförmåga till utveckling i algoritmer. Därför krävs ny hårdvara för nya algoritmer, vilket är en mycket ineffektiv metod när det gäller tid, resurser och återanvändbarhet. I denna forskning utvärderas sex algoritmer med olika implikationer för hårdvara för adressgenerering och en helt programmerbar AGU (PAGU) presenteras som kan anpassa sig till dessa algoritmer. Dessa algoritmer är Standard, Strided, Dilated, Upsampled och Padded convolution och MaxPooling. Den föreslagna AGU-arkitekturen är en Very Long Instruction Word-baserad applikationsspecifik instruktionsprocessor som har specialiserade komponenter som hårdvara räknare och noll-overhead-slingor och en kraftfull Instruktionsuppsättning Arkitektur (ISA) som kan modellera statiska och dynamiska begränsningar och affinera och icke-affinerad adress ekvationer. Målet har varit att minimera flexibiliteten kontra avvägning av område, kraft och prestanda. För ett fungerande testnätverk av semantisk segmentering har resultaten visat att PAGU visar nära den perfekta prestanda, 1 cykel per adress, för alla algoritmer som beaktas undantar Upsampled Convolution för vilken det är 1,7 cykler per adress. Området för PAGU är ungefär 4,6 gånger större än Parametrizable Datapath-metoden, vilket fortfarande är rimligt med tanke på de stora flexibilitetsfördelarna. Potentialen för PAGU är inte bara begränsad till neurala nätverksapplikationer utan också i mer allmänna digitala signalbehandlingsområden som kan utforskas i framtiden.
|
Page generated in 0.0796 seconds