Spelling suggestions: "subject:"superconducting"" "subject:"supercomputers""
11 |
Výstavba a programování clusteru o nízkém příkonu / Development and Programming of Low Power ClusterHradecký, Michal January 2016 (has links)
This thesis deals with the building and programming of a low power cluster composed of Hardkernel Odroid XU4 kits based on ARM Cortex A15 and Cortex A7 chips. The goal was to design a simple cluster composed of multiple kits and run a set of benchmarks to analyze performance and power consumption. The test set consisted of HPL and Stream benchmarks and various tests for the MPI interface. The overall performance of the cluster composed of four kits in HPL benchmark was measured 23~GFLOP/s in double-precision. During this test, the cluster showed power efficiency about 0.58~GFLOP/W. The work also describes the installation of PBS Torque scheduler and HPC software build and installation framework EasyBuild on 32-bit ARM platform. The comparison with Anselm supercomputer showed that Odroid cluster is as effiecient as large supercomputer but with slightly higher price.
|
12 |
Erarbeitung einer grafischen Benutzerschnittstelle fuer das Intensive ComputingSchumann, Merten 21 June 1995 (has links)
Entwicklung einer grafischen Nutzerschnittstelle auf der
Basis von WWW, um Jobs fuer das Batchsystem DQS zu
aktivieren.
|
13 |
Stabilized Explicit Time Integration for Parallel Air Quality ModelsSrivastava, Anurag 09 November 2006 (has links)
Air Quality Models are defined for prediction and simulation of air pollutant concentrations over a certain period of time. The predictions can be used in setting limits for the emission levels of industrial facilities. The input data for the air quality models are very large and encompass various environmental conditions like wind speed, turbulence, temperature and cloud density.
Most air quality models are based on advection-diffusion equations. These differential equations are moderately stiff and require appropriate techniques for fast integration over large intervals of time. Implicit time stepping techniques for solving differential equations being unconditionally stable are considered suitable for the solution. However, implicit time stepping techniques impose certain data dependencies that can cause the parallelization of air quality models to be inefficient.
The current approach uses Runge Kutta Chebyshev explicit method for solution of advection diffusion equations. It is found that even if the explicit method used is computationally more expensive in the serial execution, it takes lesser execution time when parallelized because of less complicated data dependencies presented by the explicit time-stepping. The implicit time-stepping on the other hand cannot be parallelized efficiently because of the inherent complicated data dependencies. / Master of Science
|
14 |
Mitteilungen des URZ 1/2001Grunewald, 23 February 2001 (has links)
Inhalt:
Jahresrückblick 1999/2000
MoUSe - ein Werkzeug zur Verwaltung von URZ-Geschäftsvorgängen
|
15 |
Designing Scalable and Efficient I/O Middleware for Fault-Resilient High-Performance Computing ClustersRaja Chandrasekar, Raghunath January 2014 (has links)
No description available.
|
16 |
Models and Techniques for Green High-Performance ComputingAdhinarayanan, Vignesh 01 June 2020 (has links)
High-performance computing (HPC) systems have become power limited. For instance, the U.S. Department of Energy set a power envelope of 20MW in 2008 for the first exascale supercomputer now expected to arrive in 2021--22. Toward this end, we seek to improve the greenness of HPC systems by improving their performance per watt at the allocated power budget.
In this dissertation, we develop a series of models and techniques to manage power at micro-, meso-, and macro-levels of the system hierarchy, specifically addressing data movement and heterogeneity. We target the chip interconnect at the micro-level, heterogeneous nodes at the meso-level, and a supercomputing cluster at the macro-level. Overall, our goal is to improve the greenness of HPC systems by intelligently managing power.
The first part of this dissertation focuses on measurement and modeling problems for power. First, we study how to infer chip-interconnect power by observing the system-wide power consumption. Our proposal is to design a novel micro-benchmarking methodology based on data-movement distance by which we can properly isolate the chip interconnect and measure its power. Next, we study how to develop software power meters to monitor a GPU's power consumption at runtime. Our proposal is to adapt performance counter-based models for their use at runtime via a combination of heuristics, statistical techniques, and application-specific knowledge.
In the second part of this dissertation, we focus on managing power. First, we propose to reduce the chip-interconnect power by proactively managing its dynamic voltage and frequency (DVFS) state. Toward this end, we develop a novel phase predictor that uses approximate pattern matching to forecast future requirements and in turn, proactively manage power. Second, we study the problem of applying a power cap to a heterogeneous node. Our proposal proactively manages the GPU power using phase prediction and a DVFS power model but reactively manages the CPU. The resulting hybrid approach can take advantage of the differences in the capabilities of the two devices. Third, we study how in-situ techniques can be applied to improve the greenness of HPC clusters.
Overall, in our dissertation, we demonstrate that it is possible to infer power consumption of real hardware components without directly measuring them, using the chip interconnect and GPU as examples. We also demonstrate that it is possible to build models of sufficient accuracy and apply them for intelligently managing power at many levels of the system hierarchy. / Doctor of Philosophy / Past research in green high-performance computing (HPC) mostly focused on managing the power consumed by general-purpose processors, known as central processing units (CPUs) and to a lesser extent, memory. In this dissertation, we study two increasingly important components: interconnects (predominantly focused on those inside a chip, but not limited to them) and graphics processing units (GPUs). Our contributions in this dissertation include a set of innovative measurement techniques to estimate the power consumed by the target components, statistical and analytical approaches to develop power models and their optimizations, and algorithms to manage power statically and at runtime. Experimental results show that it is possible to build models of sufficient accuracy and apply them for intelligently managing power on multiple levels of the system hierarchy: chip interconnect at the micro-level, heterogeneous nodes at the meso-level, and a supercomputing cluster at the macro-level.
|
17 |
High-Level Parallel Programming of Computation-Intensive Algorithms on Fine-Grained ArchitectureCheema, Fahad Islam January 2009 (has links)
<p>Computation-intensive algorithms require a high level of parallelism and programmability, which </p><p>make them good candidate for hardware acceleration using fine-grained processor arrays. Using </p><p>Hardware Description Language (HDL), it is very difficult to design and manage fine-grained </p><p>processing units and therefore High-Level Language (HLL) is a preferred alternative. </p><p> </p><p>This thesis analyzes HLL programming of fine-grained architecture in terms of achieved </p><p>performance and resource consumption. In a case study, highly computation-intensive algorithms </p><p>(interpolation kernels) are implemented on fine-grained architecture (FPGA) using a high-level </p><p>language (Mitrion-C). Mitrion Virtual Processor (MVP) is extracted as an application-specific </p><p>fine-grain processor array, and the Mitrion development environment translates high-level design </p><p>to hardware description (HDL). </p><p> </p><p>Performance requirements, parallelism possibilities/limitations and resource requirement for </p><p>parallelism vary from algorithm to algorithm as well as by hardware platform. By considering </p><p>parallelism at different levels, we can adjust the parallelism according to available hardware </p><p>resources and can achieve better adjustment of different tradeoffs like gates-performance and </p><p>memory-performance tradeoffs. This thesis proposes different design approaches to adjust </p><p>parallelism at different design levels. For interpolation kernels, different parallelism levels and </p><p>design variants are proposed, which can be mixed to get a well-tuned application and resource </p><p>specific design.</p>
|
18 |
High-Level Parallel Programming of Computation-Intensive Algorithms on Fine-Grained ArchitectureCheema, Fahad Islam January 2009 (has links)
Computation-intensive algorithms require a high level of parallelism and programmability, which make them good candidate for hardware acceleration using fine-grained processor arrays. Using Hardware Description Language (HDL), it is very difficult to design and manage fine-grained processing units and therefore High-Level Language (HLL) is a preferred alternative. This thesis analyzes HLL programming of fine-grained architecture in terms of achieved performance and resource consumption. In a case study, highly computation-intensive algorithms (interpolation kernels) are implemented on fine-grained architecture (FPGA) using a high-level language (Mitrion-C). Mitrion Virtual Processor (MVP) is extracted as an application-specific fine-grain processor array, and the Mitrion development environment translates high-level design to hardware description (HDL). Performance requirements, parallelism possibilities/limitations and resource requirement for parallelism vary from algorithm to algorithm as well as by hardware platform. By considering parallelism at different levels, we can adjust the parallelism according to available hardware resources and can achieve better adjustment of different tradeoffs like gates-performance and memory-performance tradeoffs. This thesis proposes different design approaches to adjust parallelism at different design levels. For interpolation kernels, different parallelism levels and design variants are proposed, which can be mixed to get a well-tuned application and resource specific design.
|
19 |
Data services: bringing I/O processing to petascaleAbbasi, Mohammad Hasan 08 July 2011 (has links)
The increasing size of high performance computing systems and the associated
increase in the volume of generated data, has resulted in an I/O bottleneck for these applications.
This bottleneck is further exacerbated by the imbalance in the growth of processing
capability compared to storage capability, due mainly to the power and cost requirements
of scaling the storage. This thesis introduces data services, a new abstraction which provides
significant benefits for data intensive applications. Data services combine low overhead
data movement with flexible placement of data manipulation operations, to address
the I/O challenges of leadership class scientific applications. The impact of asynchronous
data movement on application runtime is minimized by utilizing novel server side data
movement schedulers to avoid contention related jitter in application communication. Additionally,
the JITStager component is presented. Utilizing dynamic code generation and
flexible code placement, the JITStager allows data services to be executed as a pipeline
extending from the application to storage. It is shown in this thesis that data services can
add new functionality to the application without having an significant negative impact on
performance.
|
20 |
Mitteilungen des URZ 4/2001Becher,, Ehrig,, Fritsche,, Hübner,, Meyer,, Müller,, Pester,, Richter,, Riedel,, Ziegler, 06 December 2001 (has links)
Inhalt:
Neue Aufgaben im URZ - Änderungen im Kursprogramm;
Ein Jahr CLiC - ein Überblick;
Wissenschaftliche Rechnungen auf CLiC;
"camo - campus mobil" - das neue Funknetz im Campus;
Mailfilter - Teil 2;
HBFG-Projekt: Kommunikations- und Applikationsserver-Infrastruktur;
MONARCH-Dokumente in nichtlokalen Recherchesystemen;
SPARC III CPU-Server gulliver;
Software-News;
TeX-Stammtisch
|
Page generated in 0.0794 seconds