1 |
An Enhanced Dynamic Algorithm For Packet BufferRajan, Vinod 11 December 2004 (has links)
A packet buffer for the protocol processor is a large memory space that holds incoming data packets for an application. Data packets for each application are stored in the form of FIFO queues in the packet buffer. Packets are dropped when the buffer is full. An efficient buffer management algorithm is required to manage the buffer space among the different FIFO queues and to avoid heavy packet loss. This thesis develops a simulation model for the packet buffer and studies the performance of conventional buffer management algorithms when applied to packet buffer. This thesis proposes a new buffer management algorithm, Dynamic Algorithm with Different Thresholds (DADT) to improve the packet loss ratio. This algorithm takes advantage of the different packet sizes for each application and proportionally allocates buffer space for each queue. The performance of the DADT algorithm is dependent upon the packet size distribution in a network traffic load. Three different network traffic loads are considered for our simulations. For the average network traffic load, the DADT algorithm shows an improvement of 6.7 % in packet loss ratio over the conventional dynamic buffer management algorithm. For the high and actual network traffic loads, the DADT algorithm shows an improvement of 5.45 % and 3.6 % in packet loss ratio respectively. Based on the simulation results, the DADT algorithm outperforms the conventional buffer management algorithms for various network traffic loads.
|
2 |
Tools and Techniques for Efficient TransactionsPoudel, Pavan 07 September 2021 (has links)
No description available.
|
3 |
An empirically derived system for high-speed renderingRautenbach, Helperus Ritzema 25 September 2012 (has links)
This thesis focuses on 3D computer graphics and the continuous maximisation of rendering quality and performance. Its main focus is the critical analysis of numerous real-time rendering algorithms and the construction of an empirically derived system for the high-speed rendering of shader-based special effects, lighting effects, shadows, reflection and refraction, post-processing effects and the processing of physics. This critical analysis allows us to assess the relationship between rendering quality and performance. It also allows for the isolation of key algorithmic weaknesses and possible bottleneck areas. Using this performance data, gathered during the analysis of various rendering algorithms, we are able to define a selection engine to control the real-time cycling of rendering algorithms and special effects groupings based on environmental conditions. Furthermore, as a proof of concept, to balance Central Processing Unit (CPU) and Graphic Processing Unit (GPU) load for and increased speed of execution, our selection system unifies the GPU and CPU as a single computational unit for physics processing and environmental mapping. This parallel computing system enables the CPU to process cube mapping computations while the GPU can be tasked with calculations traditionally handled solely by the CPU. All analysed and benchmarked algorithms were implemented as part of a modular rendering engine. This engine offers conventional first-person perspective input control, mesh loading and support for shader model 4.0 shaders (via Microsoft’s High Level Shader Language) for effects such as high dynamic range rendering (HDR), dynamic ambient lighting, volumetric fog, specular reflections, reflective and refractive water, realistic physics, particle effects, etc. The test engine also supports the dynamic placement, movement and elimination of light sources, meshes and spatial geometry. Critical analysis was performed via scripted camera movement and object and light source additions – done not only to ensure consistent testing, but also to ease future validation and replication of results. This provided us with a scalable interactive testing environment as well as a complete solution for the rendering of computationally intensive 3D environments. As a full-fledged game engine, our rendering engine is amenable to first- and third-person shooter games, role playing games and 3D immersive environments. Evaluation criteria (identified to access the relationship between rendering quality and performance), as mentioned, allows us to effectively cycle algorithms based on empirical results and to distribute specific processing (cube mapping and physics processing) between the CPU and GPU, a unification that ensures the following: nearby effects are always of high-quality (where computational resources are available), distant effects are, under certain conditions, rendered at a lower quality and the frames per second rendering performance is always maximised. The implication of our work is clear: unifying the CPU and GPU and dynamically cycling through the most appropriate algorithms based on ever-changing environmental conditions allow for maximised rendering quality and performance and shows that it is possible to render high-quality visual effects with realism, without overburdening scarce computational resources. Immersive rendering approaches used in conjunction with AI subsystems, game networking and logic, physics processing and other special effects (such as post-processing shader effects) are immensely processor intensive and can only be successfully implemented on high-end hardware. Only by cycling and distributing algorithms based on environmental conditions and through the exploitation of algorithmic strengths can high-quality real-time special effects and highly accurate calculations become as common as texture mapping. Furthermore, in a gaming context, players often spend an inordinate amount of time fine-tuning their graphics settings to achieve the perfect balance between rendering quality and frames-per-second performance. Using this system, however, ensures that performance vs. quality is always optimised, not only for the game as a whole but also for the current scene being rendered – some scenes might, for example, require more computational power than others, resulting in noticeable slowdowns, slowdowns not experienced thanks to our system’s dynamic cycling of rendering algorithms and its proof of concept unification of the CPU and GPU. / Thesis (PhD)--University of Pretoria, 2012. / Computer Science / unrestricted
|
Page generated in 0.0513 seconds