Address predicting apparatus and methods

Electrical computers and digital processing systems: memory – Address formation – Generating prefetch – look-ahead – jump – or predictive address

Reexamination Certificate

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

C711S137000

Reexamination Certificate

active

06785797

ABSTRACT:

FIELD OF THE INVENTION
The present invention relates to high performance memory system architectures. More specifically, the present invention relates to address predicting apparatus and methods useful in high-performance computing systems.
BACKGROUND OF THE INVENTION
The speed at which computer processors can execute instructions continues to outpace the ability of computer memory systems to supply instructions and data to the processors. Consequently, many high-performance computing systems provide a high-speed buffer storage unit, commonly called a cache or cache memory, between the working store or memory of the central processing unit (“CPU”) and the main memory.
A cache comprises one or more levels of dedicated high-speed memory holding recently accessed data, designed to speed up subsequent access to the same data. For the purposes of the present specification, unless specified otherwise data will refer to any content of memory and may include, for example, instructions, data operated on by instructions, and memory addresses. Cache technology is based on the premise that computer programs frequently reuse the same data. Generally, when data is read from main system memory, a copy of the data is saved in the cache memory, along with an index to the associated main memory. For subsequent data requests, the cache detects whether the data needed has already been stored in the cache. For each data request, if the data is stored in the cache (referred to as a “hit”), the data is delivered immediately to the processor while any attempt to fetch the data from main memory is not started or aborted if already started. On the other hand, if the requested data is not stored in the cache (referred to as a “miss”) then it is fetched from main memory and also saved in the cache for future access.
A level 1 cache (“L1”) generally refers to a memory bank built closest to the central processing unit (“CPU”) chip, typically on the same chip die. A level 2 cache (“L2”) is a secondary staging area that feeds the L1 cache. L2 may be built into the CPU chip, reside on a separate chip in a multichip package module, or be a separate bank of chips.
Address predictors are used to anticipate or predict future addresses in applications such as data prefetching or instruction scheduling. Prefetching systems and methods attempt to reduce memory latency by reducing the probability of a cache miss. The probability of a cache miss is reduced by anticipating or predicting what information will be requested before it is actually requested. Address predictors utilizing correlation prediction tables (“CPTs”) for predicting both instruction addresses and data addresses are known.
A simple correlation found in a CPT is a pair consisting of a key and a successor value. The key is used to predict the successor value. A correlated address pair (“CAP”) is built by associating two addresses that appear in an address stream. The address that appears earlier in the address stream is referred to as the key and it is paired with the currently referenced address, which is referred to as the successor value. The CAP (that is, a key and its successor value) is then stored in a CPT for later use. When an address previously selected as a key reoccurs in the address stream, it is used to query the CPT to retrieve the corresponding CAP. The successor value in the retrieved CAP is then used to predict the next address in the address stream. In sum, a goal of address predictors is to observe previous address pairs, store them, and use them as predictions in the future.
Correlations in a CPT can be built from cache miss addresses and used to predict future miss addresses. For example, when a cache miss is generated, a correlation is built between the preceding cache miss address and the current cache miss address. The next time that the key address generates a cache miss, the predictor speculates that the successor address will be the next cache miss.
One approach to improving the performance of a CPT is to implement the CPT with the ability to correlate a single key with multiple successor values. This can be achieved by entering a key with multiple successor values for each key entered. That is, each unique key is stored or entered into the CPT only once. If the key associated with a subsequent correlation to be entered is already entered into the CPT, then the successor value is entered into the same line as the already entered key, but the associated key is not entered again. Generally, when a key is correlated with more than one successor value, the successors are predicted using a most recently used (“MRU”) priority and replaced with a least recently used (“LRU”) priority.
However, this approach to improving the performance of a CPT is not without drawbacks. First, because the total table size is constant, a CPT designer is faced with undesirable tradeoffs between the maximum possible number of successors per entered key and the maximum possible number of keys entered. The maximum number of entered successors per entered key is static and cannot be changed depending on the application. Thus, a CPT designer must decide apriori how many successors can be associated with a given entered key. However, this decision can be difficult. Some addresses are highly correlated and need only one successor for their correlations. Other addresses may be followed by several different addresses at various points during program/application execution and benefit from multiple successors. Since the total size of a CPT is constant, increasing the number of entered successors per key requires that the total number of entered keys be decreased.
The second drawback to this approach relates to replacing correlations previously entered into a CPT. When a correlation is replaced, all correlations corresponding to the same key are lost, including the key and all corresponding successor values stored. The removed correlations can only be reentered into the CPT if and when the correlations reappear in the cache miss addresses. This rebuilding of a previously removed correlations can have a severe impact on the performance of the CPT.
The speed at which computer processors can execute instructions will likely continue to outpace the ability of computer memory systems to supply instructions and data to the processors. Although address predictors can be utilized to improve the performance of computer memory systems, existing address predictors have some drawbacks. Accordingly, there is still a need in the industry to improve memory system performance in computer systems by improving existing address predictors and address prediction methods.


REFERENCES:
patent: 4807110 (1989-02-01), Pomerene et al.
patent: 5817543 (1998-10-01), McAllister et al.
patent: 5850533 (1998-12-01), Panwar et al.
patent: 5895966 (1999-04-01), Penchuk
patent: 5898853 (1999-04-01), Panwar et al.
patent: 5930819 (1999-07-01), Hetherington et al.
patent: 6005503 (1999-12-01), Burrows
patent: 6052775 (2000-04-01), Panwar et al.
patent: 6055622 (2000-04-01), Spillinger
patent: 6065105 (2000-05-01), Zaidi et al.
patent: 6081873 (2000-06-01), Hetherington et al.
patent: 6122727 (2000-09-01), Witt
patent: 0335358 (1989-10-01), None
patent: 06-349678 (1994-12-01), None
Alexander, T., et al., “Disributed Prefetch-buffer/Cache Design for High Perfromance Memory Systems”,IEEE Second International Symposium on High Perfromance Computer Architecture, San Jose, CA, pp. 254-263, (Feb. 1996).
Charney, M.J., et al., “Prefetching and memory system behavior of the SPEC95 benchmark suite”,IBM Journal of Research and Development, 41(3), pp. 265-286, (May 1997).
Joseph, D., et al., “Prefetching Using Markov Predictors”,Computer Architecture News, 25(2), pp. 252-263, (Jun. 1997).

LandOfFree

Say what you really think

Search LandOfFree.com for the USA inventors and patents. Rate them and share your experience with other people.

Rating

Address predicting apparatus and methods does not yet have a rating. At this time, there are no reviews or comments for this patent.

If you have personal experience with Address predicting apparatus and methods, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Address predicting apparatus and methods will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFUS-PAI-O-3314824

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.