Stream-down prefetching cache

Electrical computers and digital processing systems: memory – Storage accessing and control – Hierarchical memories

Reexamination Certificate

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

C711S134000, C711S154000

Reexamination Certificate

active

06643743

ABSTRACT:

FIELD OF THE INVENTION
The present invention relates to computer memory. More particularly, the present invention relates to stream-down prefetching data into a cache memory.
BACKGROUND OF THE INVENTION
Most modern computer system architectures include a cache. That is, a level of the memory hierarchy between the processor and the main memory. Generally, access to data stored in a cache is much faster than the time needed to retrieve data from main memory. However, cache memory is typically more expensive and smaller than main memory. A common cache analogy is that main memory is like books on shelves in a library and a cache is like a small subset of those books on a desk. The widespread use of cache memory is a clear statement that it does increase system performance, more than enough to justify added complexity.
It is possible to have multiple levels of cache, or more correctly multiple levels of memory, in a memory hierarchy. Generally, in such a system, all the data is stored at the lowest level, which is typically the largest in size and the slowest to access. Data is then copied to the higher levels, which may decrease in size and increase in speed.
Designing a memory hierarchy for a general purpose computer, including algorithms for what data should be stored in cache is a complex process that has received a lot of attention over the years. Empirically, computer programs have a tendency to reuse data that has been accessed recently (temporal locality) and also to access data located near data that has been accessed recently (spatial locality). Many memory hierarchies utilize the existence of these localities for storing data in a cache. For example, keeping recently accessed data in a cache exploits temporal locality. Retrieving blocks of data, instead of an individual word, into a cache exploits spatial locality. This disclosure will use the term “data” to include both the traditional concept of data, such as alpha-numeric constants or variables, set or read by a computer program, as well as program instructions. For the memory hierarchy, both types of data are merely values stored at a particular location and the ultimate uses of the individual stored values are irrelevant. Some architectures do, however, maintain separate caches for traditional data and instructions, and must distinguish between them.
Ideally, whenever a processor calls for data from memory, that data will be found in the cache. However, cache misses do occur. Two common causes of cache misses are cold-start misses and capacity misses. A cold-start miss occurs when a processor first accesses a particular range of data addresses. A capacity miss occurs due to limited cache sizes, such as when previously cached data is “bumped” from the cache by newer data just before it is needed. A larger cache may decrease the capacity miss rate, but have no effect on the cold-start miss rate.


REFERENCES:
patent: 5694568 (1997-12-01), Harrison, III et al.
patent: 5829028 (1998-10-01), Lynch et al.
patent: 5895485 (1999-04-01), Loechel et al.
patent: 6035375 (2000-03-01), Yanai et al.
patent: 6049850 (2000-04-01), Vishlitzky et al.
patent: 6070230 (2000-05-01), Capps
patent: 6098154 (2000-08-01), Lopez-Aguado et al.
patent: 6134633 (2000-10-01), Jacobs
patent: 6138212 (2000-10-01), Chiacchia et al.
patent: 6173392 (2001-01-01), Shinozaki
patent: 6202130 (2001-03-01), Scales, III et al.
Chen, “An Effective Programmable Prefetch Engine of On-Chip Caches”, © 1995, IEEE, p. 237-242.

LandOfFree

Say what you really think

Search LandOfFree.com for the USA inventors and patents. Rate them and share your experience with other people.

Rating

Stream-down prefetching cache does not yet have a rating. At this time, there are no reviews or comments for this patent.

If you have personal experience with Stream-down prefetching cache, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Stream-down prefetching cache will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFUS-PAI-O-3175267

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.