Method and apparatus for providing improved caching for a...

Electrical computers and digital processing systems: memory – Storage accessing and control – Specific memory composition

Reexamination Certificate

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

C711S114000, C707S793000

Reexamination Certificate

active

06269423

ABSTRACT:

BACKGROUND OF THE INVENTION
1. Field of the Invention.
This invention relates in general to data storage and processing, and more particularly to distributed storage servers.
2. Description of Related Art.
The implementation of new technology in magnetic tape products has meant that the density of data written to tape has increased by orders of magnitude in the last ten or fifteen years. The ability to record high density tapes, e.g., ten gigabytes or more on one physical volume, has led to reducing costs in physical tape hardware as well as in handling and management resources.
However, over the past five years, tape data set stacking products, i.e., software solutions to increase tape utilization, have evolved in response to the customer requirement for more efficient ways to manage the information stored on tape. To achieve increased capacity enablement, a virtual tape server (VTS) has been proposed. In a VTS, the hardware is transparent to the host and the user. The VTS requires little external management except though the library management element of the tape library into which a VTS is integrated.
In a hierarchical storage systems, such as a VTS, intensively used and frequently accessed data is stored in fast but expensive memory. One example of a fast memory is a direct access storage device (DASD). In contrast, less frequently accessed data is stored in less expensive but slower memory. Examples of slower memory are tape drives and disk drive arrays. The goal of the hierarchy is to obtain moderately priced, high-capacity storage while maintaining high-speed access to the stored information.
In the VTS system, a host data interface, a DASD file buffer, and a number of tape devices are provided. When the host writes a logical volume, or a file, to the VTS, the data is stored as a resident file on the DASD. Although the DASD provides quick access to this data, it will eventually reach full capacity and a backup or secondary storage system will be needed. An IBM 3590 tape cartridge is one example of a tape device that could be used as a backup or secondary storage system.
When the DASD fills to a predetermined threshold, the logical volume data for a selected logical volume, typically the oldest, is removed from the DASD to free space for more logical volumes. If the selected logical volume has not already been appended to a tape cartridge or a physical volume, it is appended to a tape cartridge prior being removed from the DASD. A file that has been appended to a tape and removed from the DASD is “migrated.” Optionally, any time prior to being removed from the DASD, a DASD file can be appended onto a tape cartridge with the original left on the DASD for possible cache hits. A file that has been appended to a tape cartridge while its original is left on the DASD is said to be premigrated.
When the host reads a logical volume from the VTS, a cache hit occurs if the logical volume currently resides on the DASD. If the logical volume is not on the DASD, the storage manager determines which of the physical tape volumes contains the logical volume. The corresponding physical volume is then mounted on one of the tape devices, and the data for the logical volume is transferred back to the DASD from the tape (recall).
Tape servers may use an engine to move data between the DASD and tape drives in a virtual tape server (VTS) environment. For example, the IBM Virtual Tape Server (VTS) uses the IBM Adstar Distributed Storage Manager (ADSM) as its engine to move data between the DASD and IBM 3590 tape drives on the VTS. In such a system, the VTS uses the storage manager client on the DASD, e.g., the ADSM Hierarchical Storage Manager (HSM) client, and a distributed storage manager server attached to the tape drives to provide this function.
Since recalls take a long time relative to “cache hits,” it would be preferably to have as many logical volumes as possible be cache hits. In order to accomplish this a logical volume caching method is used.
Typically the logical volumes in the cache are managed on an FIFO (first in first out) or LRU (least recently used) algorithm. However, each of these methods exhibits one or more disadvantages: the methods do not discern patterns, the methods are not adaptive, or the methods do not improve upon the cache hit rate.
It can be seen that there is a need for a method and apparatus for improving caching for a virtual tape server.
It can also be seen that there is a need for a method and apparatus for improving caching for a virtual tape server which makes assumptions to increase cache hits, but which does not under-perform a LRU algorithm when these assumptions prove to be incorrect.
SUMMARY OF THE INVENTION
To overcome the limitations in the prior art described above, and to overcome other limitations that will become apparent upon reading and understanding the present specification, the present invention discloses a method and apparatus for improving caching for a virtual tape server.
The present invention solves the above-described problems by providing a method and apparatus for improving caching for a virtual tape server which makes assumptions to increase cache hits, but which does not under-perform a LRU algorithm when these assumptions prove to be incorrect.
A method in accordance with the principles of the present invention includes the steps of correlating logical volume usage to externally determined file characteristics to derive a usage history for cached logical volumes, maintaining a record of the usage history in time bins and managing the migration of the logical volumes according to the usage history and the externally determined file characteristics of the logical volume.
Other embodiments of a system in accordance with the principles of the invention may include alternative or optional additional aspects. One such aspect of the present invention is that the step of correlating further includes the steps of associating the first logical volume with a logical volume type based on predefined characteristics, assigning a time bin for the type of the first logical volume based on the time of last access of the first logical volume therein, the time of the current access of the first logical volume, and based on the type assigned to the first logical volume at the time the first logical volume was last accessed and updating a usage count in the time bin for the type.
Another aspect of the present invention is that the method further includes the step of calculating a score for a logical volume based upon the size of the first logical volume, the age of the first logical volume, and the historic performance of the other volumes.
Another aspect of the present invention is that the time bin has data associated therewith, the data in the time bin for the first logical volume being divided by a predetermined number to prevent overrun and to maintain weighted historic data when a total for all time bins reaches a predetermined threshold.
Another aspect of the present invention is that the score is updated periodically for any cached volume as the cached volume ages.
Another aspect of the present invention is that the score is updated periodically for all cached volumes.
Yet another aspect of the present invention is that the step of updating the score further comprises the steps of determining the age and type of each logical volume, selecting an mth time bucket for the logical volume, calculating a usage score for a time bucket, computing a time-weighted usage score for the time bucket and identifying a final score for the time bucket.
Another aspect of the present invention is that the usage score comprises calculating a usage score for all nth time buckets, n>m, wherein the mth time bucket is a current time bucket, and wherein the usage score comprises a scaling constant times the sum of the time buckets for the logical volume between n and m divided by the sum of all time buckets greater than or equal to m plus a number of misses.
Another aspect of the present invention is that the usage score is proportional to the percentage of cach

LandOfFree

Say what you really think

Search LandOfFree.com for the USA inventors and patents. Rate them and share your experience with other people.

Rating

Method and apparatus for providing improved caching for a... does not yet have a rating. At this time, there are no reviews or comments for this patent.

If you have personal experience with Method and apparatus for providing improved caching for a..., we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Method and apparatus for providing improved caching for a... will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFUS-PAI-O-2544879

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.