• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 193
  • 32
  • 31
  • 16
  • 11
  • 10
  • 6
  • 6
  • 5
  • 3
  • 3
  • 2
  • 2
  • 1
  • 1
  • Tagged with
  • 363
  • 131
  • 79
  • 72
  • 50
  • 43
  • 42
  • 40
  • 39
  • 35
  • 34
  • 33
  • 32
  • 31
  • 30
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
21

Jit4OpenCL: a compiler from Python to OpenCL

Xunhao, Li Unknown Date
No description available.
22

An application of parallelism to compilation

Kilpatrick, P. L. January 1985 (has links)
No description available.
23

Implementierung eines modifizierten Variablenmodelles zur Vermeidung Parameter-induzierter Alias-Effekte auf Basis des GNU Ada 95 Compilers

Behrendt, Clemens. January 1997 (has links)
Stuttgart, Univ., Fakultät Informatik, Diplomarb., 1997.
24

Programmanalyse des XRTL-Zwischencodes

Backes, Werner. January 2005 (has links) (PDF)
Saarbrücken, Univ., Diss., 2005. / Erscheinungsjahr an der Haupttitelstelle: 2004. Computerdatei im Fernzugriff.
25

Implementation funktionaler Programmiersprachen durch Quelltexttransformation

Maćoš, Dragan. January 1998 (has links) (PDF)
Berlin, Humboldt-Universiẗat, Diss., 1998.
26

Mapping parallelism to heterogeneous processors

Chandramohan, Kiran January 2016 (has links)
Most embedded devices are based on heterogeneous Multiprocessor System on Chips (MPSoCs). These contain a variety of processors like CPUs, micro-controllers, DSPs, GPUs and specialised accelerators. The heterogeneity of these systems helps in achieving good performance and energy efficiency but makes programming inherently difficult. There is no single programming language or runtime to program such platforms. This thesis makes three contributions to these problems. First, it presents a framework that allows code in Single Program Multiple Data (SPMD) form to be mapped to a heterogeneous platform. The mapping space is explored, and it is shown that the best mapping depends on the metric used. Next, a compiler framework is presented which bridges the gap between the high -level programming model of OpenMP and the heterogeneous resources of MPSoCs. It takes OpenMP programs and generates code which runs on all processors. It delivers programming ease while exploiting heterogeneous resources. Finally, a compiler-based approach to runtime power management for heterogeneous cores is presented. Given an externally provided budget, the approach generates heterogeneous, partitioned code that attempts to give the best performance within that budget.
27

Search-based Optimization for Compiler Machine-code Generation

Clauson, Aran 10 October 2013 (has links)
Compilation encompasses many steps. Parsing turns the input program into a more manageable syntax tree. Verification ensures that the program makes some semblance of sense. Finally, code generation transforms the internal abstract program representation into an executable program. Compilers strive to produce the best possible programs. Optimizations are applied at nearly every level of compilation. Instruction Scheduling is one of the last compilation tasks. It is part of code generation. Instruction Scheduling replaces the internal graph representation of the program with an instruction sequence. The scheduler should produce some sequence that the hardware can execute quickly. Considering that Instruction Scheduling is an NP-Complete optimization problem, it is interesting that schedules are usually generated by a greedy, heuristic algorithm called List Scheduling. Given search-based algorithms' successes in other NP-Complete optimization domains, we ask whether search-based algorithms can be applied to Instruction Scheduling to generate superior schedules without unacceptably increasing compilation time. To answer this question, we formulate a problem description that captures practical scheduling constraints. We show that this problem is NP-Complete given modest requirements on the actual hardware. We adapt three different search algorithms to Instruction Scheduling in order to show that search is an effective Instruction Scheduling technique. The schedules generated by our algorithms are generally shorter than those generated by List Scheduling. Search-based scheduling does take more time, but the increases are acceptable for some compilation domains.
28

Schopnosti překladačů vybraných strukturovaných jazyků

Zerdaloglu, Stratos January 2012 (has links)
No description available.
29

The design and implementation of a structured backend for the MCCAT C compiler

Donawa, Christopher M. January 1994 (has links)
Note:
30

A Parallelizing Compiler for Fortran

Janaki, S 08 1900 (has links)
With the advent of Distributed Memory Machines (DMMs) numerous work have been undertaken to ease the work of a programmer these systems. Data parallel languages like Fortran D, Vienna Fortran, High Performance Fortran and C+ allow the user to specify data distribution across processor with some directives, and the compiler for these language use the directives to compile the programme in to an SPMD code. There are number of old program which are still in use and rewriting them in to new data parallel languages is a costly effort. Most of the work on these parallelizing compilers concentrate on efficient data communication between the processors.With the advancement in technology, data communication time is also decreasing.This allows bigger programs to execute in the same time span.The resources of a DMM being finite puts a limit on the size of the problem that can be run. Improving the memory usage for a problem will hence allow us run bigger size problems. Further, as communication speed increases, the overhead caused by house-keeping computations like global index to local index transformation, and owner processor computation will degrade the performance of the resultant code. Hence a uniform and efficient method for these computations also becomes a necessity. We have implemented parallelizing parts of a compiler using the SUIF compiler system, which accepts programs written in Fortran77 with directives to the compiler as comments. The output of the compiler is an SPMD C program, with embedded PVM calls for message communication between the processors. We have also proposed algorithms to improve data communications,and minimizing memory usage in the output code. A uniform method for performing owner processor computations and global-to-local transformations has also been implemented.

Page generated in 0.0464 seconds