Unlock instant, AI-driven research and patent intelligence for your innovation.

Bandwidth-adaptive, hybrid, cache-coherence protocol

a bandwidth adaptive, cache-coherence technology, applied in the field of cache-coherence protocol, can solve the problems of overwhelming a communication channel, invalidating all other cache memories, and complicated decisions, and achieve the effect of eliminating ambiguity

Inactive Publication Date: 2005-06-30
WISCONSIN ALUMNI RES FOUND
View PDF4 Cites 7 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Benefits of technology

The present invention provides an adaptive, hybrid protocol for communicating cache coherence messages in a multiprocessor computer system. The protocol adapts to the available bandwidth by selectively using snooping or a directory mechanism based on the evaluation of the available bandwidth. The protocol includes a method and apparatus for coordinating cache memories in a multiprocessor computer system, which can communicate cache coherence messages over a network. The invention streamlines the directory process over the process normally used in multicast snooping, and addresses possible problems of live lock and ambiguity at the receiving node. The evaluation of available bandwidth may compare it against a predetermine threshold and select between snooping and directory mechanisms accordingly. The invention provides improved control dynamics and lower latency correction mechanisms for speculatively multicasting.

Problems solved by technology

Generally, such protocols invalidate all other cache memories when one cache is written to, and updating of the main memory before a changed cache is flushed.
For large systems with many processors, however, snooping generates large numbers of messages which may overwhelm a communications channel.
While the above principals guide the system designer in selecting between snooping and directory protocols, the decision can be complicated.
Selecting one of a directory protocol or a snooping protocol will result in less than optimal performance when the same system is configured with different numbers of processors or in certain upgrade operations where more processors are added to the system.
Second, even for a fixed number of processors, the application being executed may result in a radically different demand on the cache protocol communication network for which one of the snooping or directory protocols will be preferable to the other protocol.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Bandwidth-adaptive, hybrid, cache-coherence protocol
  • Bandwidth-adaptive, hybrid, cache-coherence protocol
  • Bandwidth-adaptive, hybrid, cache-coherence protocol

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0047] Referring now to FIG. 1, a multiprocessor, shared-memory, computer system 10 includes a number of processor units 12 communicating on a network 14 with a shared memory system 16. Although the shared memory system 16 is depicted as a single unitary structure, in practice, the physical memory of the shared memory system 16 may be distributed among different processor units 12 to be shared over a network or the like. The shared memory system 16 includes a shared memory 17 of conventional architecture and storing a number of memory blocks 19, a directory 21, and a memory controller 11 as will be described below.

[0048] Each processor unit 12 includes a processor 18 connected over an internal bus 20 with a cache memory 22 and cache controller 26. Only two processor units are shown in FIG. 1, however, the present invention is applicable to architectures having an arbitrary number of processor units and is particularly well suited for multiprocessor, shared-memory, computer systems ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

A cache coordination mechanism for a multiprocessor, shared-memory computer switches between a snooping mechanism where an individual processor unit broadcasts or multicasts cache coherence messages to each other node on the system and a directory system where the individual processor unit transmits the cache control message to a directory which then identifies potential candidates to receive that message. The switching is according to the activity on the communication network used by the cache coherence messages. When network activity is high, a directory protocol is used to conserve bandwidth but when network activity is low, a snooping system is used to provide faster response.

Description

CROSS-REFERENCE TO RELATED APPLICATIONS [0001] This application is a divisional of U.S. patent application Ser. No. 10 / 037,727, filed Oct. 19, 2001, which claims the benefit of Provisional Application No. 60 / 275,743, filed Mar. 14, 2001.STATEMENT REGARDING FEDERALLY SPONSORED RESEARCH OR DEVELOPMENT [0002] This invention was made with United States government support awarded by the following agencies: [0003] NSF 9971256 [0004] The United States has certain rights in this invention. BACKGROUND OF THE INVENTION [0005] The present invention relates generally to a system for coordinating cache memories in a shared-memory computer architecture, and in particular, to a system that chooses a mechanism for communicating cache coherence messages based on the bandwidth available for transmitting such messages. [0006] Large computer software applications, such as simulators and database servers, require cost-effective computation beyond that which can be provided by a single microprocessor. Sh...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(United States)
IPC IPC(8): G06F12/08
CPCG06F12/0817G06F2212/502G06F12/0831
Inventor MARTIN, MILO M.K.SORIN, DANIEL J.HILL, MARK D.WOOD, DAVID A.
Owner WISCONSIN ALUMNI RES FOUND