Intelligent cache replacement mechanism with varying and adaptive temporal residency requirements

a technology of adaptive temporal residency and cache replacement, applied in the field of memory management, can solve the problems of costly cache miss, inconvenient for some situations, and incur delay

Inactive Publication Date: 2010-11-04
IBM CORP
View PDF8 Cites 14 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Benefits of technology

[0009]In a further aspect, the method may include resetting the value according to a varying expected access pattern for the entry. The setting may be performed by hardware, software or an operation system, or combination thereof. In one aspect, the software and/or the operating system are allowed to read the values from the cache memory control data.
[0010]In another aspect, the method may include maintaining the value assigned to the entry until the entry is replaced. In yet another aspect, the method may include decaying the value. Still in another aspect, the method may include decaying one or more values associated with an associative set of the cache memory when an entry from the associative set is selected for replacement. In another aspect, the method may include periodically decaying one or more values associated with one or more respective entries in the cache.
[0011]Further yet, the method may include decaying one or more values associated with one ...

Problems solved by technology

A cache miss is costly because the data must then be fetched from a higher-level cache, main memory, or potentially another processor's cache on a multiprocessor, which incurs a delay because accessing the other memory is slower than accessing the cache memory.
While LRU performs well in many cases, it may not be suitable for some situations.
A streaming application is an example application where LRU cache replacement may not have the desired performance.
In addition, existing cache solutions ...

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Intelligent cache replacement mechanism with varying and adaptive temporal residency requirements
  • Intelligent cache replacement mechanism with varying and adaptive temporal residency requirements
  • Intelligent cache replacement mechanism with varying and adaptive temporal residency requirements

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0019]FIG. 1 is a block diagram illustrating a cache memory with associated bit values of the present disclosure in an exemplary embodiment. Each location in a cache memory has a datum or a cache line 104. Briefly, a cache line is a unit of data that can be transferred between the main memory 110 and the cache 108. Cache line is also referred to as cache block. Different cache memories have different designs and range in size. For example, an address location 106 in the main memory 110 maps to a location in the cache memory containing the data from that main memory's location 110, which has been cached. There are several different types of mapping used in cache memory organization. Known mapping procedures include direct mapping, fully associative mapping, and set associative or n-way associative mapping. The cache is direct mapped if each entry in main memory can map to just one location in the cache. If any entry in the cache can have any entry in the main memory, the cache is ful...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

A method for replacing cached data is disclosed. The method in one aspect associates an importance value to each block of data in the cache. When a new entry needs to be stored in the cache, a cache block for replacing is selected based on the importance values associated with cache blocks. In another aspect, the importance values are set according to the hardware and/or software's knowledge of the memory access patterns. The method in one aspect may also include varying the importance value over time over different processing requirements.

Description

STATEMENT REGARDING FEDERALLY SPONSORED RESEARCH OR DEVELOPMENT[0001]This invention was made with Government support under Contract No.: NBCH020056 (DARPA) awarded by Defense, Advanced Research Projects Agency. The Government has certain rights in this invention.FIELD OF THE INVENTION[0002]The present disclosure generally relates to computer processors and particularly to memory management.BACKGROUND OF THE INVENTION[0003]Cache memory is one of the critical elements in computer processors for achieving good performance on the processors. Generally, a cache is a smaller, faster memory used by a central processing unit of a computer to reduce the average time to access its data or instructions. The cache typically stores copies of the data from the most frequently used main memory locations. The fundamental idea of cache organization is that by keeping the most frequently accessed instructions and data in the fast cache memory, the average memory access time will approach the access t...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06F12/08G06F12/00
CPCG06F12/127
Inventor SHEN, XIAOWEISINHAROY, BALARAMWISNIEWSKI, ROBERT W.
Owner IBM CORP
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products