Spelling suggestions: "subject:"multiprogramming"" "subject:"autoprogramming""
1 |
Optimal organization of I/O operations in multiprogrammed systemsKho, James Wang, January 1972 (has links)
Thesis (Ph. D.)--University of Wisconsin--Madison, 1972. / Typescript. Vita. eContent provider-neutral record in process. Description based on print version record. Bibliography: leaves 161-168.
|
2 |
A kernel for distributed and shared memory communication /Rao, Ram C. January 1982 (has links)
Thesis (Ph. D.)--University of Washington, 1982. / Vita. Includes bibliographical references.
|
3 |
A mixed-strategy page replacement algorithm for a multiprogramming, virtual memory computerSpafford, Eugene Howard January 1981 (has links)
No description available.
|
4 |
The design and implementation of a multi-programming virtual memory operating system for a mini-computerParks, Lee Stephen January 1979 (has links)
Thesis (B.S.)--Massachusetts Institute of Technology, Dept. of Physics; and, (B.S.)--Massachusetts Institute of Technology, Dept. of Mathematics, 1979. / MICROFICHE COPY AVAILABLE IN ARCHIVES AND SCIENCE. / Includes bibliographical references. / by Lee Parks. / B.S.
|
5 |
Location finding algorithms for distributed systemsBernabéu-Aubán, José Manuel 12 1900 (has links)
No description available.
|
6 |
Sensitivity analysis for online management of processor power and performanceAlmoosa, Nawaf I. 22 May 2014 (has links)
The shift to manycore architectures has highlighted the need for runtime power and performance management schemes to improve the reliability, performance, and energy-efficiency of processors. However, the design of management algorithms is challenging since power and performance are strongly dependent on the workload, which cannot be determined apriori and exhibit wide and rapid runtime variations.
This dissertation seeks to show that sensitivity analysis (derivative estimation) provides runtime power and performance information that enables the design of adaptive and low-complexity management algorithms. The contributions of the dissertation include: 1) controllers that achieve rapid regulation of the power and throughput of processor cores, 2) a chip-level power control solution that maximizes the performance of manycore processors subject to the power constraints set by the cooling system, and 3) an iterative algorithm for optimizing the energy consumption of cache memories. The proposed algorithms use runtime derivative estimation to adapt to the rapid power and performance variations caused by workload, and their efficacy is demonstrated via formal analysis and simulation experiments.
|
7 |
A Study of the Performance Benefits of Controlling Parallel Asynochrous Iteractive ApplicationsJoseph, P J 09 1900 (has links)
High performance networks of workstation are becoming increasingly popular a parallel computing platform because of their lower cost. Both message passing and software distributed shared memory (DSM) programming paradigms have been developed and employed on such distributed hardware platforms. An important performance bottleneck in these systems is the effective data transmission latency, which is poorer than in high-speed parallel computer interconnection networks.
Iterative algorithms are used in a large class of applications like solution of partial algorithms are used, optimization problems, solutions to systems of linear equations, and so on. These can be parallelized in a straight-forward fashion with cad1 node computing a part of the data set and also synchronizing and exchanging data with the other nodes as required. But these synchronous version perform poorly when message transmission delays are high, as is the case in network of workstations. The asynchronous parallel version of these algorithms provide an additional degree of freedom to address large data transmission latencies. These algorithms do not synchronize, and behave correctly in the presence of losses and delays in the propagation of updates. Thus, in shared memory systems they do not synchronize accesses to shared data and they will work correctly even in the presence of delays and losses in updates. This gives synchronous algorithms a performance advantage over their synchronous counterparts since synchronization costs are avoided and further computation can be overlapped with communication.
The message generation rate of asynchronous algorithms is however greater than that of their synchronous counterparts. Uncontrolled asynchronous runs can create a large network load resulting in large queuing delays, which in turn can increase the message generation of the asynchronous algorithms. This is especially possible in lower bandwidth network like that in network of workstations. Such a positive feedback loop leads to unstable network conditions.
Recent work has tried to improve the performance of asynchronous algorithms on a distributed shared memory (DSM) system by adaptively buffering shared memory updates depending on the network load, and transmitting multiple updates together. This reduces congestion and message transmission overhead, but could still result in slow convergence since nothing is guaranteed about, the update propagation delay. Also, although adaptive throttling of message will kick in once the network gets heavily loaded, it cannot, prevent the initial flooding. Furthermore, the processor is not freed when computation with the available values does not result in much further convergence.
In this thesis we present an alternate method of controlling iterative methods and present performance results for the same. We propose a new system-supported blocking read primitive, termed Global Read that is guaranteed to return a value of acceptable age of the specified location in a DSM system. The main idea is to enforce an upper bound on the age of shared updates seen by a node in a manner visible to the underlying system (DSM). Information about processes being blocked can be used for adapting the underlying system, especially the network, towards better performance. A reading process is throttled until its Global-Read is satisfied, thus implementing program-level flow control and also freeing the processor. The Global-Read can also help in communication-based scheduling of, processes.
Performance evaluation using a benchmark from Cray, on a network of workstations and on the IBM SP2 parallel computer, showed good performance improvements. We also present results of a systematic study wherein we implement parallel code for different iterative techniques for the solution of Lap lace equation wing PVM, anti characterize when controlled asynchrony work befit. We studied the improvements in computation time and analyzed the sources of this improvement, using various input and parallelism, on both IBM SP2 and a network of workstations. We find significant performance improvements for controlling asynchrony when the traffic generated by the asynchronous algorithm becomes more than what can be sustained by the network. But we found that the scalability of the applications is limited by the software overhead for messages. Systems which have reduced software overhead will show very good scalable performance for controlled asynchrony.
|
8 |
Effective runtime management of parallelism in a functional programming context /Dermoudy, Julian Richard. January 2002 (has links)
Thesis (Ph.D.)--University of Tasmania, 2002. / Includes bibliographical references.
|
9 |
The task distribution preprocessor (TDP) /McCanna, Frank. January 1989 (has links)
Thesis (M.S.)--Rochester Institute of Technology, 1989. / Includes bibliographical references (leaves 73-77).
|
10 |
Deadlock probability prediction and detection in distributed systems /Deorukhkar, Mayuresh. January 2004 (has links)
Thesis (M.S.)--University of Missouri-Columbia, 2004. / Typescript. Includes bibliographical references (leaves 75-76). Also available on the Internet.
|
Page generated in 0.0822 seconds