Techniques for Reducing the Impact of Inclusion in Shared Network Cache Multiprocessors Techniques for Reducing the Impact of Inclusion in Shared Network Cache Multiprocessors
نویسندگان
چکیده
This paper investigates design alternatives for shared network caches in clusterbased multiprocessors. Using simulation, we rst demonstrate that network caches o er several potential performance bene ts, but that the adverse impact of cache inclusion-related evictions must be mitigated for these bene ts to be fully realized. We then evaluate three network cache architectural alternatives designed to address this issue: increasing network cache associativity, adding a network victim cache, and adding a tag cache to relax inclusion requirements for clean network cache lines. Six well-known parallel scienti c applications were used in our experiments. We found that a four-way set associative network cache, or a four-entry victim cache, dramatically reduced execution time for all applications examined. Employing a tag cache for replaced clean lines improved performance for some applications, but this bene t was highly dependent upon processor cache associativity.
منابع مشابه
Cache-Affinity Scheduling for Fine Grain Multithreading
Cache utilisation is often very poor in multithreaded applications, due to the loss of data access locality incurred by frequent context switching. This problem is compounded on shared memory multiprocessors when dynamic load balancing is introduced and thread migration disrupts cache content. In this paper, we present a technique, which we refer to as ‘batching’, for reducing the negative impa...
متن کاملA performance evaluation of cache injection in bus-based shared memory multiprocessors
Bus-based shared memory multiprocessors with private caches and snooping write-invalidate cache coherence protocols are dominant form of smallto medium-scale parallel machines today. In these systems the high memory latency poses the major hurdle in achieving high performance. One way to cope with this problem is to use various techniques for tolerating high memory latency. Software-controlled ...
متن کاملCache Injection on Bus Based Multiprocessors
Software-controlled cache prefetching and data forwarding are widely used techniques for tolerating memory latency in shared memory multiprocessors. However, some previous studies show that cache prefetching is not so effective on bus-based multiprocessors, while the effectiveness of data forwarding has not been explored in this environment, yet. In this paper, a novel technique called cache in...
متن کاملToward Large Scale Shared Memory Multiprocessing
We are currently investigating two di erent approaches to scalable shared memory Munin a distributed shared memory DSM system implemented entirely in software and Willow a true shared memory multiprocessor with extensive hardware support for scalability Munin allows parallel programs written for shared memory multiprocessors to be executed e ciently on dis tributed memory multiprocessors Unlike...
متن کاملEfficient Shared Memory Multiprocessing and Object-Oriented Programming
Object-oriented techniques are shown to hold promise for addressing the growing speed gap between memory and processors on shared-memory multiprocessors. However, reducing cache misses can be at the expense of more misses from the translation lookaside buffer (TLB—a cache of recent page translations). This paper presents simulation results from use of a library called OOSH (Object-Oriented Libr...
متن کامل