• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 342
  • 54
  • 41
  • 39
  • 23
  • 16
  • 15
  • 13
  • 8
  • 8
  • 4
  • 3
  • 3
  • 3
  • 3
  • Tagged with
  • 741
  • 289
  • 279
  • 143
  • 99
  • 93
  • 90
  • 86
  • 79
  • 69
  • 64
  • 46
  • 43
  • 43
  • 38
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
91

Improving the Caching Performances in Web using Cooperative Proxies

Huang, Li-te 03 February 2009 (has links)
Nowadays, Web caching technique has been widely used and become one of the most promising ways to reduce network traffic, server load, and user-experienced latency while users surf the Web. In the context of traditional systems, caching techniques have been extensively studied. However, these techniques are not directly applicable to Web due to larger size of working set and cache storage in proxies. Many studies have presented their approaches to improving the performance of Web caching. Two of most representative approaches are hash routing [25] and directory-base digest [12]. Hash routing provides amapping fromthe URL of object to the location of proxy, which has the cached object, while directory-based digest records the pairs of proxy locations and object URLs for answering the query when local misses occur in any proxy. Hash routing can best utilize storage space by eliminating duplicated objects among proxies,while directory-based digest allows object replicas among proxies to resist proxy failures. These two conventional approaches have complementary tradeoffs. In this thesis, a comprehensive approach to cooperative caching for Web proxies, using a combination of hash routing and directory-based digest, is presented. Our approach tends to subsume these widely used approaches and thus gives a spectrum of trade-off between the overall hit ratio and its associated overhead. Through the simulations using real-life proxy traces, the performance and overhead of our proposed mechanism were evaluated. The experimental results showed that our approach outperforms the previous competitors.
92

An energy efficient cache design using spin torque transfer (STT) RAM

Rasquinha, Mitchelle 23 August 2011 (has links)
The advent of many core architectures has coincided with the energy and power limited design of modern processors. Projections for main memory clearly show widening of the processor-memory gap. Cache capacity increased to help reduce this gap will lead to increased energy and area usage and due to small growth in die size, impede performance scaling that has accompanied Moore's Law to date. Among the dominant sources of energy consumption is the on-chip memory hierar- chy, specically the L2 cache and the Last Level Cache (LLC). This work explores the use of a novel non-volatile memory technology - Spin Torque Transfer RAM (STT RAM)" for the design of the L2/LLC caches. While STTRAM is a promising memory technology, it has some limitations, particularly in terms of write energy and write latencies. The main objectives of this thesis is to use a novel cell design for a non-volatile 1T1MTJ cell and demonstrate its use at the L2 and LLC cache levels with architectural optimizations to maximize energy reduction. The proposed cache hierarchy dissipates significantly lesser energy (both leakage and dynamic) and uses less area in comparison to a conventional SRAM based cache designs.
93

Efficient Cache Organization For Application Specific And General Purpose Processors

Rajan, Kaushik 05 1900 (has links)
The performance gap between processor and memory continues to remain a major performance bottleneck in both application specific and general purpose processors. This thesis strives to ease the above bottleneck by exploiting the characteristics of the application domain to improve the cache organization for two distinct processor architectures: (1) application specific processors for packet forwarding, (2) general purpose processors. Packet forwarding algorithms make use of a trie data structure to determine the forwarding route. We observe that the locality characteristics of the nodes at various levels of such a trie are different. Nodes that are closer to the root node, especially those that are immediate children of the root node (level-one nodes), exhibit higher temporal locality than nodes lower down the trie. Based on this observation we propose a novel Heterogeneously Segmented Cache Architecture (HSCA) that uses separate caches for level-one and lower-level nodes, each with carefully chosen sizes. We also propose a new replacement policy to enhance the performance of HSCA. Performance evaluation indicates that HSCA results in up to 32% reduction in average memory access time over a unified cache that shares the same cache space among all levels of the trie. HSCA also outperforms a previously proposed results cache. The use of a large root branching factor in a forwarding trie forcefully introduces a large number of nodes at level-one. Among these, only nodes that cover prefixes from the routing table are useful while the rest, are superfluous. We find that as many as 75% of the level-one nodes are superfluous. This leads to a skewed distribution of useful nodes among the cache sets of the level-one nodes cache. We propose a novel two-level mapping framework that achieves a better nodes to cache set mapping and hence incurs fewer conflict misses. Two-level mapping first aggregates nodes into Initial Partitions (IPs) using lower order bits and then remaps them from IPs into Refined Partitions (RPs), that form sets, based on some higher order bits. It provides flexibility in placement by allowing each IP to choose a different remap function. We propose three schemes conforming to the framework. A speedup in average memory access time of as much as 16% is gained over HSCA. In general purpose processor architectures, the design objectives of caches at various levels of the hierarchy are different. To ensure low access latencies, L1 caches are small and have low associativities, making them more susceptible to conflict misses. The extent of conflict misses incurred is governed by the placement function and the memory access patterns exhibited by the program. We propose a mechanism to learn the access characteristics of the program at runtime by analyzing the repetitive phases of program. We then make use of the two-level mapping framework to dynamically adapt the placement function. Further, we elegantly incorporate two-level mapping into the cache organization without increasing the cache access latency. Performance evaluation reveals that the proposed adaptive placement mechanism eliminates 32—36% of misses on average over a range of cache sizes. To prevent expensive off-chip accesses, L2 caches are larger and have higher associativities. Hence, the replacement policy plays a significant role in determining L2 cache performance. Further, as the inherent temporal locality in memory accesses is filtered out by the L1 cache, an L2 cache using the widely prevalent LRU replacement policy incurs significantly higher misses than the optimal replacement policy (OPT). We propose to bridge this gap through a novel replacement strategy that mimics the replacement decisions of OPT. The L2 cache is logically divided into two components, a Shepherd Cache (SC) with a simple FIFO replacement and a Main Cache (MC) with an emulation of optimal replacement. The SC plays the dual role of caching lines and shepherding the replacement decisions close to optimal for MC. Our proposed organization can cover 40% of the gap between LRU and OPT, resulting in 7% overall speedup.
94

Dynamic cache-line sizes /

Van Vleet, Taylor, January 2000 (has links)
Thesis (Ph. D.)--University of Washington, 2000. / Vita. Includes bibliographical references (leaves 128-131).
95

Online searching and connecting caching /

Stafford, Matthew, January 2000 (has links)
Thesis (Ph. D.)--University of Texas at Austin, 2000. / Vita. Includes bibliographical references (leaves 125-127). Available also in a digital version from Dissertation Abstracts.
96

Caching and replication schemes on the Internet /

Loukopoulos, Athanasios. January 2002 (has links)
Thesis (Ph. D.)--Hong Kong University of Science and Technology, 2002. / Includes bibliographical references (leaves 149-163). Also available in electronic version. Access restricted to campus users.
97

Adaptive prefetching for visual data exploration

Doshi, Punit Rameshchandra. January 2003 (has links)
Thesis (M.S.)--Worcester Polytechnic Institute. / Keywords: Adaptive prefetching; Large-scale multivariate data visualization; Semantic caching; Hierarchical data exploration; Exploratory data analysis. Includes bibliographical references (p.66-70).
98

Scalable primary cache memory architectures

Agarwal, Vikas 28 August 2008 (has links)
Not available / text
99

Efficient runahead execution processors

Mutlu, Onur 28 August 2008 (has links)
Not available / text
100

Prefetch mechanisms by application memory access pattern

Agaram, Kartik Kandadai 28 August 2008 (has links)
Not available / text

Page generated in 0.0483 seconds