Full track read for adaptive pre-fetching of data
First Claim
1. A method for adaptively pre-fetching data, comprising:
- collecting a first set of statistics based on a number of avoidable read-misses in which data exists that is prior to data being read;
collecting a second set of statistics based on a number of avoidable read-misses in which data exists that follows data being read;
collecting a third set of statistics based on said first and second sets of statistics;
on the basis of said second set of statistics, selecting a pre-fetch technique from a first technique that pre-fetches data following data being read and a second technique that pre-fetches data before and following the data being read, wherein the selecting of the second technique reduces a number of instances where out of order reads result in avoidable read-misses, and wherein the avoidable read-misses are read-misses that could have been averted had a sector of data at a storage location been pre-fetched with other sectors of data at the storage location that were previously fetched;
upon detection of an avoidable read-miss and selecting the first technique, pre-fetching data following data being read based on comparison of a generated random number and the first set of statistics; and
upon detection of the avoidable read-miss and selecting the second technique, pre-fetching data before and following data being read based on a comparison of a generated random number and the third set of statistics.
6 Assignments
0 Petitions

Accused Products

Abstract
Adaptively pre-fetching data includes collecting a first set of statistics based on a number of avoidable read-misses in which data exists that is prior to data being read, collecting a second set of statistics based on a number of avoidable read-misses in which data exists that follows data being read, and collecting a third set of statistics based on said first and second sets of statistics. On the basis of the second set of statistics, a pre-fetch technique is selected from a first technique that pre-fetches data following data being read and a second technique that pre-fetches data before and following the data being read. The first and third set of statistics may be used to determine when to pre-fetch data.
111 Citations
SYSTEM AND METHOD FOR TRACKING BILLING EVENTS IN A MOBILE WIRELESS NETWORK FOR A NETWORK OPERATOR | ||
Patent #
US 20110201304A1
Filed 04/28/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
TARGETED NOTIFICATION OF CONTENT AVAILABILITY TO A MOBILE DEVICE | ||
Patent #
US 20110207436A1
Filed 02/17/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Content delivery to a mobile device from a content service | ||
Patent #
US 20090157792A1
Filed 12/13/2007
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks Inc
|
Application and network-based long poll request detection and cacheability assessment therefor | ||
Patent #
US 8,166,164 B1
Filed 10/14/2011
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks Inc
|
REQUEST AND RESPONSE CHARACTERISTICS BASED ADAPTATION OF DISTRIBUTED CACHING IN A MOBILE NETWORK | ||
Patent #
US 20120110110A1
Filed 10/17/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Cache defeat detection and caching of content addressed by identifiers intended to defeat cache | ||
Patent #
US 8,190,701 B2
Filed 11/01/2011
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks Inc
|
Application and network-based long poll request detection and cacheability assessment therefor | ||
Patent #
US 8,291,076 B2
Filed 03/05/2012
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks Inc
|
Social caching for device resource sharing and management | ||
Patent #
US 8,316,098 B2
Filed 04/19/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Distributed management of keep-alive message signaling for mobile network resource conservation and optimization | ||
Patent #
US 8,326,985 B2
Filed 11/01/2011
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks Inc
|
System and method for a mobile device to use physical storage of another device for caching | ||
Patent #
US 8,356,080 B2
Filed 07/20/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Electronic-mail filtering for mobile devices | ||
Patent #
US 8,364,181 B2
Filed 12/10/2007
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Context aware data presentation | ||
Patent #
US 8,412,675 B2
Filed 02/24/2006
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Aligning data transfer to optimize connections established for transmission over a wireless network | ||
Patent #
US 8,417,823 B2
Filed 11/18/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks LLC
|
Flexible real-time inbox access | ||
Patent #
US 8,438,633 B1
Filed 12/18/2006
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Publishing data in an information community | ||
Patent #
US 8,468,126 B2
Filed 12/14/2005
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Distributed caching in a wireless network of content delivered for a mobile application over a long-held request | ||
Patent #
US 8,484,314 B2
Filed 10/14/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Provisioning applications for a mobile device | ||
Patent #
US 8,494,510 B2
Filed 12/06/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Mobile network background traffic data management with optimized polling intervals | ||
Patent #
US 8,539,040 B2
Filed 02/28/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Secure end-to-end transport through intermediary nodes | ||
Patent #
US 8,549,587 B2
Filed 02/14/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
System and method for executing commands that are non-native to the native environment of a mobile device | ||
Patent #
US 8,561,086 B2
Filed 05/17/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Detecting and preserving state for satisfying application requests in a distributed proxy and cache system | ||
Patent #
US 8,621,075 B2
Filed 04/27/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks LLC
|
Cache state management on a mobile device to preserve user experience | ||
Patent #
US 8,635,339 B2
Filed 08/22/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Polling | ||
Patent #
US 8,693,494 B2
Filed 03/31/2008
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Cache defeat detection and caching of content addressed by identifiers intended to defeat cache | ||
Patent #
US 8,700,728 B2
Filed 05/17/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Cross-platform event engine | ||
Patent #
US 8,209,709 B2
Filed 07/05/2010
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Distributed system for cache defeat detection and caching of content addressed by identifiers intended to defeat cache | ||
Patent #
US 8,204,953 B2
Filed 11/01/2011
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks Inc
|
Electronic-mail filtering for mobile devices | ||
Patent #
US 8,738,050 B2
Filed 01/07/2013
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Mobile device equipped with mobile network congestion recognition to make intelligent decisions regarding connecting to an operator network | ||
Patent #
US 8,750,123 B1
Filed 07/31/2013
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks Inc
|
Maintaining an IP connection in a mobile network | ||
Patent #
US 8,761,756 B2
Filed 09/13/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks International Oy
|
Integrated messaging | ||
Patent #
US 8,774,844 B2
Filed 04/08/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Dynamic bandwidth adjustment for browsing or streaming activity in a wireless network based on prediction of user behavior when interacting with mobile applications | ||
Patent #
US 8,775,631 B2
Filed 02/25/2013
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Timing of keep-alive messages used in a system for mobile network resource conservation and optimization | ||
Patent #
US 8,782,222 B2
Filed 09/05/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks LLC
|
Application discovery on mobile devices | ||
Patent #
US 8,787,947 B2
Filed 06/18/2008
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks Inc
|
Content delivery to a mobile device from a content service | ||
Patent #
US 8,793,305 B2
Filed 12/13/2007
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks Inc
|
System and method of a relay server for managing communications and notification between a mobile device and a web access server | ||
Patent #
US 8,799,410 B2
Filed 04/13/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Maintaining mobile terminal information for secure communications | ||
Patent #
US 8,805,334 B2
Filed 09/05/2008
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Integrated messaging | ||
Patent #
US 8,805,425 B2
Filed 01/28/2009
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Mobile device power management in data synchronization over a mobile network with or without a trigger notification | ||
Patent #
US 8,811,952 B2
Filed 05/05/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Method and system for management of a virtual network connection without heartbeat messages | ||
Patent #
US 8,812,695 B2
Filed 04/03/2013
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks Inc
|
System and method for tracking billing events in a mobile wireless network for a network operator | ||
Patent #
US 8,831,561 B2
Filed 04/28/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
System and method for making requests on behalf of a mobile device based on atomic processes for mobile network traffic relief | ||
Patent #
US 8,832,228 B2
Filed 04/26/2012
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks Inc
|
Distributed caching for resource and mobile network traffic management | ||
Patent #
US 8,838,783 B2
Filed 07/05/2011
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks Inc
|
Flexible real-time inbox access | ||
Patent #
US 8,839,412 B1
Filed 09/13/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Web-based access to data objects | ||
Patent #
US 8,838,744 B2
Filed 01/28/2009
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Mobile traffic categorization and policy for network use optimization while preserving user experience | ||
Patent #
US 8,843,153 B2
Filed 11/01/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
System for providing policy based content service in a mobile network | ||
Patent #
US 8,849,902 B2
Filed 06/24/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Mass storage system and methods of controlling resources thereof | ||
Patent #
US 8,856,484 B2
Filed 08/14/2012
|
Current Assignee
Infinidat Ltd.
|
Original Assignee
Infinidat Ltd.
|
Hierarchies and categories for management and deployment of policies for distributed wireless traffic optimization | ||
Patent #
US 8,861,354 B2
Filed 12/14/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Policy based content service | ||
Patent #
US 8,862,657 B2
Filed 01/25/2008
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
System of redundantly clustered machines to provide failover mechanisms for mobile traffic management and network resource conservation | ||
Patent #
US 8,868,753 B2
Filed 12/06/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Provisioning of e-mail settings for a mobile terminal | ||
Patent #
US 8,873,411 B2
Filed 01/12/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Signaling optimization in a wireless network for traffic utilizing proprietary and non-proprietary protocols | ||
Patent #
US 8,874,761 B2
Filed 03/15/2013
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Mobile application traffic optimization | ||
Patent #
US 8,886,176 B2
Filed 07/22/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Optimization of resource polling intervals to satisfy mobile device requests | ||
Patent #
US 8,903,954 B2
Filed 11/22/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Mobile virtual network operator | ||
Patent #
US 8,909,192 B2
Filed 08/11/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Bandwidth measurement | ||
Patent #
US 8,909,759 B2
Filed 10/12/2009
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Detection and management of user interactions with foreground applications on a mobile device in distributed caching | ||
Patent #
US 8,909,202 B2
Filed 01/07/2013
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
System and method for providing a network service in a distributed fashion to a mobile device | ||
Patent #
US 8,914,002 B2
Filed 08/11/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Optimization of mobile traffic directed to private networks and operator configurability thereof | ||
Patent #
US 8,918,503 B2
Filed 08/28/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Method and apparatus for intercepting events in a communication system | ||
Patent #
US RE45,348 E1
Filed 03/16/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Application and network-based long poll request detection and cacheability assessment therefor | ||
Patent #
US 8,966,066 B2
Filed 10/12/2012
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks Inc
|
Mobile device and method to utilize the failover mechanism for fault tolerance provided for mobile traffic management and network/device resource conservation | ||
Patent #
US 8,977,755 B2
Filed 12/06/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Monitoring mobile application activities for malicious traffic on a mobile device | ||
Patent #
US 8,984,581 B2
Filed 07/11/2012
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks Inc
|
Connection architecture for a mobile network | ||
Patent #
US 8,989,728 B2
Filed 09/07/2006
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Predictive content delivery | ||
Patent #
US 9,002,828 B2
Filed 01/02/2009
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Flexible and dynamic integration schemas of a traffic management system with various network operators for network traffic alleviation | ||
Patent #
US 9,009,250 B2
Filed 12/07/2012
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks Inc
|
Caching adapted for mobile application behavior and network conditions | ||
Patent #
US 9,021,048 B2
Filed 10/14/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Mobile network reporting and usage analytics system and method aggregated using a distributed traffic optimization system | ||
Patent #
US 9,021,021 B2
Filed 12/10/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Mobile network traffic coordination across multiple applications | ||
Patent #
US 9,043,433 B2
Filed 05/25/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Intelligent rendering of information in a limited display environment | ||
Patent #
US 9,047,142 B2
Filed 12/16/2010
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Mobile network traffic coordination across multiple applications | ||
Patent #
US 9,049,179 B2
Filed 01/20/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Location-based operations and messaging | ||
Patent #
US 9,055,102 B2
Filed 08/02/2010
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Selective data compression by a distributed traffic management system to reduce mobile data traffic and signaling traffic | ||
Patent #
US 9,060,032 B2
Filed 05/09/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Proxy server associated with a mobile carrier for enhancing mobile traffic management in a mobile network | ||
Patent #
US 9,065,765 B2
Filed 10/08/2013
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks Inc
|
Distributed implementation of dynamic wireless traffic policy | ||
Patent #
US 9,077,630 B2
Filed 07/08/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Device resources sharing for network resource conservation | ||
Patent #
US 9,084,105 B2
Filed 04/19/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Mobile network background traffic data management | ||
Patent #
US 9,100,873 B2
Filed 09/14/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Managing cache to prevent overloading of a wireless network due to user activity | ||
Patent #
US 9,131,397 B2
Filed 06/06/2013
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
ADAPTIVE CACHE PREFETCHING BASED ON COMPETING DEDICATED PREFETCH POLICIES IN DEDICATED CACHE SETS TO REDUCE CACHE POLLUTION | ||
Patent #
US 20150286571A1
Filed 04/04/2014
|
Current Assignee
Qualcomm Inc.
|
Original Assignee
Qualcomm Inc.
|
Optimized and selective management of policy deployment to mobile clients in a congested network to prevent further aggravation of network congestion | ||
Patent #
US 9,161,258 B2
Filed 03/15/2013
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks LLC
|
Radio-awareness of mobile device for sending server-side control signals using a wireless network optimized transport protocol | ||
Patent #
US 9,173,128 B2
Filed 03/06/2013
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks LLC
|
Dynamic categorization of applications for network access in a mobile network | ||
Patent #
US 9,203,864 B2
Filed 02/04/2013
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks LLC
|
Mobile device having content caching mechanisms integrated with a network operator for traffic alleviation in a wireless network and methods therefor | ||
Patent #
US 9,208,123 B2
Filed 12/07/2012
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks LLC
|
Mobile device with application or context aware fast dormancy | ||
Patent #
US 9,241,314 B2
Filed 03/15/2013
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks LLC
|
Automatic generation and distribution of policy information regarding malicious mobile traffic in a wireless network | ||
Patent #
US 9,239,800 B2
Filed 07/11/2012
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks LLC
|
Extending user relationships | ||
Patent #
US 9,251,193 B2
Filed 10/28/2007
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks LLC
|
Application or context aware fast dormancy | ||
Patent #
US 9,271,238 B2
Filed 03/15/2013
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks LLC
|
Request and response characteristics based adaptation of distributed caching in a mobile network | ||
Patent #
US 9,275,163 B2
Filed 10/17/2011
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks LLC
|
Radio-awareness of mobile device for sending server-side control signals using a wireless network optimized transport protocol | ||
Patent #
US 9,277,443 B2
Filed 12/07/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks LLC
|
System and method for a mobile device to use physical storage of another device for caching | ||
Patent #
US 9,300,719 B2
Filed 01/14/2013
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks Inc
|
Systems and methods for application management of mobile device radio state promotion and demotion | ||
Patent #
US 9,307,493 B2
Filed 03/15/2013
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks LLC
|
System and method for reduction of mobile network traffic used for domain name system (DNS) queries | ||
Patent #
US 9,325,662 B2
Filed 01/09/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks LLC
|
Method and apparatus for prefetching data | ||
Patent #
US 9,323,680 B1
Filed 09/28/2007
|
Current Assignee
Veritas Technologies LLC
|
Original Assignee
Veritas US IP Holdings LLC
|
User as an end point for profiling and optimizing the delivery of content and data in a wireless network | ||
Patent #
US 9,326,189 B2
Filed 02/04/2013
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks LLC
|
Wireless traffic management system cache optimization using http headers | ||
Patent #
US 9,330,196 B2
Filed 06/14/2012
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks LLC
|
Mobile application traffic optimization | ||
Patent #
US 9,407,713 B2
Filed 01/16/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks LLC
|
Selective data compression by a distributed traffic management system to reduce mobile data traffic and signaling traffic | ||
Patent #
US 9,432,486 B2
Filed 05/16/2015
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks LLC
|
System and method for providing a network service in a distributed fashion to a mobile device | ||
Patent #
US 9,473,914 B2
Filed 12/28/2014
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks LLC
|
Mobile device configured for communicating with another mobile device associated with an associated user | ||
Patent #
US 9,712,986 B2
Filed 03/22/2012
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks LLC
|
Operation modes for mobile traffic optimization and concurrent management of optimized and non-optimized traffic | ||
Patent #
US 9,832,095 B2
Filed 12/14/2012
|
Current Assignee
Seven Networks LLC
|
Original Assignee
Seven Networks LLC
|
Enhanced customer service for mobile carriers using real-time and historical mobile application and traffic or optimization data associated with mobile devices in a mobile network | ||
Patent #
US 10,263,899 B2
Filed 04/10/2013
|
Current Assignee
Seven Networks Inc
|
Original Assignee
Seven Networks LLC
|
Deterministic setting of replacement policy in a cache | ||
Patent #
US 6,961,824 B2
Filed 06/04/2004
|
Current Assignee
Avago Technologies International Sales Pte Limited
|
Original Assignee
Broadcom Corporation
|
Adaptive prefetching of data from a disk | ||
Patent #
US 6,529,998 B1
Filed 11/03/2000
|
Current Assignee
Emc IP Holding Company LLC
|
Original Assignee
EMC Corporation
|
Method and apparatus for enhancing the disk cache process by dynamically sizing prefetch data associated with read requests based upon the current cache hit rate | ||
Patent #
US 6,085,287 A
Filed 12/23/1998
|
Current Assignee
GENESIS ONE TECHNOLOGIES INC.
|
Original Assignee
GENESIS ONE TECHNOLOGIES INC.
|
Method and a system for determining an appropriate amount of data to cache | ||
Patent #
US 6,098,153 A
Filed 01/30/1998
|
Current Assignee
International Business Machines Corporation
|
Original Assignee
International Business Machines Corporation
|
Method and apparatus for performing point in time backup operation in a computer system | ||
Patent #
US 5,857,208 A
Filed 05/31/1996
|
Current Assignee
EMC Corporation
|
Original Assignee
EMC Corporation
|
System for using a data history table to select among multiple data prefetch algorithms | ||
Patent #
US 5,941,981 A
Filed 11/03/1997
|
Current Assignee
Advanced Micro Devices Inc.
|
Original Assignee
Advanced Micro Devices Inc.
|
Caching system and method providing aggressive prefetch | ||
Patent #
US 6,003,114 A
Filed 06/17/1997
|
Current Assignee
Emc IP Holding Company LLC
|
Original Assignee
EMC Corporation
|
Space reclamation system and method for use in connection with tape logging system | ||
Patent #
US 5,778,394 A
Filed 12/23/1996
|
Current Assignee
Emc IP Holding Company LLC
|
Original Assignee
EMC Corporation
|
Single lock command for an I/O storage system that performs both locking and I/O data operation | ||
Patent #
US 5,845,147 A
Filed 03/19/1996
|
Current Assignee
EMC Corporation
|
Original Assignee
EMC Corporation
|
System and method for disk mapping and data retrieval | ||
Patent #
US 5,206,939 A
Filed 09/24/1990
|
Current Assignee
EMC Corporation
|
Original Assignee
EMC Corporation
|
25 Claims
-
1. A method for adaptively pre-fetching data, comprising:
-
collecting a first set of statistics based on a number of avoidable read-misses in which data exists that is prior to data being read; collecting a second set of statistics based on a number of avoidable read-misses in which data exists that follows data being read; collecting a third set of statistics based on said first and second sets of statistics; on the basis of said second set of statistics, selecting a pre-fetch technique from a first technique that pre-fetches data following data being read and a second technique that pre-fetches data before and following the data being read, wherein the selecting of the second technique reduces a number of instances where out of order reads result in avoidable read-misses, and wherein the avoidable read-misses are read-misses that could have been averted had a sector of data at a storage location been pre-fetched with other sectors of data at the storage location that were previously fetched; upon detection of an avoidable read-miss and selecting the first technique, pre-fetching data following data being read based on comparison of a generated random number and the first set of statistics; and upon detection of the avoidable read-miss and selecting the second technique, pre-fetching data before and following data being read based on a comparison of a generated random number and the third set of statistics. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8)
-
-
9. Computer software, provided in a non-transitory computer-readable storage medium, that adaptively pre-fetches data, the software comprising:
-
executable code that collects a first set of statistics based on a number of avoidable read-misses in which data exists that is prior to data being read; executable code that collects a second set of statistics based on a number of avoidable read-misses in which data exists that follows data being read; executable code that collects a third set of statistics based on said first and second sets of statistics; executable code that uses said second set of statistics to select a pre-fetch technique from a first technique that pre-fetches data following data being read and a second technique that pre-fetches data before and following the data being read, wherein the selecting of the second technique reduces a number of instances where out of order reads result in avoidable read-misses, and wherein the avoidable read-misses are read-misses that could have been averted had a sector of data at a storage location been pre-fetched with other sectors of data at the storage location that were previously fetched; executable code that pre-fetches data following data being read based on comparison of a generated random number and the first set of statistics upon detection of an avoidable read-miss and selecting the first technique; and executable code that pre-fetches data before and following data being read based on a comparison of a generated random number and the third set of statistics upon detection of the avoidable read-miss and selecting the second technique. - View Dependent Claims (10, 11, 12, 13, 14, 15, 16)
-
-
17. A storage device, comprising:
-
a plurality of directors; a plurality of disk drives coupled to at least some of the directors; a global memory, coupled to the directors, wherein at least one of the directors includes software that pre-fetches data from the disks into global memory, the software including executable code that collects a first set of statistics based on a number of avoidable read-misses in which data exists that is prior to data being read, executable code that collects a second set of statistics based on a number of avoidable read-misses in which data exists that follows data being read, executable code that collects a third set of statistics based on said first and second sets of statistics, executable code that uses said second set of statistics to select a pre-fetch technique from a first technique that pre-fetches data following data being read and a second technique that pre-fetches data before and following the data being read, wherein the selecting of the second technique reduces a number of instances where out of order reads result in avoidable read-misses, and wherein the avoidable read-misses are read-misses that could have been averted had a sector of data at a storage location been pre-fetched with other sectors of data at the storage location that were previously fetched, executable code that reads data following data being read based on comparison of a generated random number and the first set of statistics upon detection of an avoidable read-miss and selecting the first technique, and executable code that reads data before and following data being read based on a comparison of a generated random number and the third set of statistics upon detection of the avoidable read-miss and selecting the second technique. - View Dependent Claims (18, 19, 20, 21)
-
-
22. A method of determining when to pre-fetch a full track of data on a disk, comprising:
-
collecting statistics based on a number of avoidable read-misses in which data exists that is in the same track as data being read; and upon detection of an avoidable read-miss, selecting and performing a pre-fetch technique for pre-fetching a full track of data based on comparison of a generated random number and the set of statistics, wherein the selecting of the pre-fetch technique reduces a number of instances where out of order reads result in avoidable read-misses, and wherein the avoidable read-misses are read-misses that could have been averted had a sector of data at a storage location been pre-fetched with other sectors of data at the storage location that were previously fetched. - View Dependent Claims (23)
-
-
24. Computer software, provided in a non-transitory computer-readable medium, that determines when to pre-fetch additional data from a disk, comprising:
-
executable code that provides for selection of one of the following modes;
a first mode for pre-fetch to end of track only, a second mode for pre-fetch entire track, a third mode where no pre-fetch is performed, and a fourth mode where pre-fetch to an end of track is performed upon second and subsequent read misses for a track, wherein the selecting of the second mode reduces a number of instances where out of order reads result in avoidable read-misses, and wherein the avoidable read-miss are read-misses that could have been averted had a sector of data at a storage location been pre-fetched with other sectors of data at the storage location that were previously fetched;executable code that collects a first set of statistics based on a number of avoidable read-misses in which data exists that is prior to data being read; executable code that collects a second set of statistics based on a number of avoidable read-misses in which data exists that follows data being read in response to selection of the second mode; executable code that collects a third set of statistics based on said first and second sets of statistics in response to selection of the second mode; executable code that pre-fetches data following data being read based on comparison of a generated random number and the first set of statistics upon detection of an avoidable read-miss and on selecting of the first mode; and executable code that pre-fetches data before and following data being read based on a comparison of a generated random number and the third set of statistics upon detection of the avoidable read-miss and selecting the second mode. - View Dependent Claims (25)
-
1 Specification
1. Technical Field
This application relates to computer storage devices, and more particularly to the field of pre-fetching certain data from a disk of a storage device to cache memory thereof.
2. Description of Related Art
Host processor systems may store and retrieve data using a storage device containing a plurality of host interface units (host adapters), disk drives, and disk interface units (disk adapters). Such storage devices are provided, for example, by EMC Corporation of Hopkinton, Mass. and disclosed in U.S. Pat. Nos. 5,206,939 to Yanai et al., 5,778,394 to Galtzur et al., U.S. Pat. No. 5,845,147 to Vishlitzky et al., and U.S. Pat. No. 5,857,208 to Ofek. The host systems access the storage device through a plurality of channels provided therewith. Host systems provide data and access control information through the channels of the storage device and the storage device provides data to the host systems also through the channels. The host systems do not address the disk drives of the storage device directly, but rather, access what appears to the host systems as a plurality of logical volumes. The logical volumes may or may not correspond to the actual disk drives.
A user who requires data from a disk initiates a complex and time-consuming sequence of events for retrieving the data. For example, in the course of retrieving data, a disk controller positions a disk arm to align a read head with the cylinder on the disk that contains that data. The disk controller then waits until the desired data begins passing under the read head. Then, when rotation of the disk finally brings the desired data under the read head, the disk controller initiates the read operation. These events introduce considerable latency into the process of satisfying the user'"'"'s request for data.
A global cache memory, which is a relatively fast memory that is separate from the disks, may be used to address some of the latency issues associated with disks. The global cache memory may contain recently fetched (requested) data. Upon receiving a request for data, the disk storage system first checks to see if the requested data is already in the global cache memory. If so, the disk storage system retrieves the data directly therefrom without having to access the disk. Such an event is referred to as a “read-hit.”A read-hit is a desirable outcome because a read from the cache slot avoids latencies associated with reading from a physical disk drive.
In some cases, the disk storage system discovers that the desired data is not in the global cache memory at all but, instead, is on a disk. In this case, the disk storage system instructs a disk controller to retrieve the desired data from an appropriate track on a disk. Such an event is referred to as a “read-miss.” A read-miss is an undesirable outcome because such an operation is afflicted with latencies associated with mechanical motion within the disk drive and possible latencies associated with data transmission between the global cache memory and the disk drive.
Data on a disk may be organized in tracks, which represent an amount of contiguous memory that may be read in a single disk drive operation. The global cache memory may be organized into slots, where each slot corresponds to a track. Whether or not to read an entire track, rather than merely the data specifically requested from that track, is a decision that requires the disk storage system to anticipate whether additional data from that track is likely to be needed in the future. U.S. Pat. No. 6,003,114 to Eitan Bachmat, the contents of which are herein incorporated by reference, describes a system where, in response to a read-miss, the requested data is fetched and stored in the global cache memory. If there is a second request for data from the same track while data for the track is still in cache, the remainder of the track is fetched.
A disadvantage of the foregoing technique is that, the second read miss to the same slot becomes unavoidable U.S. Pat. No. 6,529,998 to Yechiel Yochai and Robert Mason, which is incorporated by reference herein, addresses this issue by reading more than just the requested data based on metrics that do not depend directly on whether other data from a particular track has been accessed recently. U.S. Pat. No. 6,529,998 to Yechiel Yochai and Robert Mason discloses reading to the end of a track in instances where it appears statistically advantageous to do so based on a determination of the number of times a read miss would have been avoided had a previous read within the track been a read to the end of the track.
While the technique disclosed in 6,529,998 improves performance in many instances, it does not take into account the possibility of reads in a track being out of order. That is, an application may cause a first read to occur at one portion of a track and cause a second read to occur at a different portion of the track. If the start of the first portion is after the start of the second portion, then even if the first read is to the end of the track, the second read will still cause a read miss. Furthermore, the first read to the end of the track will not be useful. Accordingly, it is desirable to be able to reduce the number of instances where out of order reads within a track result in read misses.
According to the system described herein, adaptively pre-fetching data includes collecting a first set of statistics based on a number of avoidable read-misses in which data exists that is prior to data being read. Adaptively pre-fetching data also includes collecting a second set of statistics based on a number of avoidable read-misses in which data exists that follows data being read and collecting a third set of statistics based on said first and second sets of statistics. On the basis of said second set of statistics, a pre-fetch technique is selected from a first technique that pre-fetches data following data being read and a second technique that pre-fetches data before and following the data being read. Upon detection of a read-miss and selecting the first technique, data following data being read is pre-fetched based on comparison of a generated random number and the first set of statistics. Upon detection of a read-miss and selecting the second technique, data before and following data being read is pre-fetched based on a comparison of a generated random number and the third set of statistics. Pre-fetching data before and following data being read may include fetching an entire track from a disk. Pre-fetching data following data being read may include fetching data from the data being read to the end of a track on a disk. Selecting a pre-fetch technique may include comparing the second set of statistics with a predefined value. The predefined value may be a user settable parameter. Following selection of the first technique, data following data being read may be pre-fetched if the generated random number is less than a value based on the first set of statistics and, following selection of the second technique, data before and following data being read may be pre-fetched if the generated random number is less than a value based on the third set of statistics. Upon selection of the first technique and the generated random number being greater than the value based on the first set of statistics, the values based on the sets of statistics may be decremented. Upon selection of the second technique and the generated random number being greater than the value based on the third set of statistics, the values based on the sets of statistics may be decremented.
According further to the system described herein, computer software, provided in a computer-readable storage medium, adaptively pre-fetches data. The software includes executable code that collects a first set of statistics based on a number of avoidable read-misses in which data exists that is prior to data being read. The software also includes executable code that collects a second set of statistics based on a number of avoidable read-misses in which data exists that follows data being read and executable code that collects a third set of statistics based on said first and second sets of statistics. The software also includes executable code that uses said second set of statistics to select a pre-fetch technique from a first technique that pre-fetches data following data being read and a second technique that pre-fetches data before and following the data being read. Additional executable code pre-fetches data following data being read based on comparison of a generated random number and the first set of statistics upon detection of a read-miss and selecting the first technique. Additional executable code pre-fetches data before and following data being read based on a comparison of a generated random number and the third set of statistics upon detection of a read-miss and selecting the second technique. Pre-fetching data before and following data being read may include fetching an entire track from a disk. Pre-fetching data following data being read may include fetching data from the data being read to the end of a track on a disk. Selecting a pre-fetch technique may include comparing the second set of statistics with a predefined value. The predefined value may be a user settable parameter. Following selection of the first technique, data following data being read may be pre-fetched if the generated random number is less than a value based on the first set of statistics and, following selection of the second technique, data before and following data being read may be pre-fetched if the generated random number is less than a value based on the third set of statistics. Upon selection of the first technique and the generated random number being greater than the value based on the first set of statistics, the values based on the sets of statistics may be decremented. Upon selection of the second technique and the generated random number being greater than the value based on the third set of statistics, the values based on the sets of statistics may be decremented.
According further to the system described herein, a storage device includes a plurality of directors, a plurality of disk drives coupled to at least some of the directors, and a global memory, coupled to the directors. At least one of the directors includes software that pre-fetches data from the disks into global memory. The software includes executable code that collects a first set of statistics based on a number of avoidable read-misses in which data exists that is prior to data being read, executable code that collects a second set of statistics based on a number of avoidable read-misses in which data exists that follows data being read, and executable code that collects a third set of statistics based on said first and second sets of statistics. The software also includes executable code that uses said second set of statistics to select a pre-fetch technique from a first technique that pre-fetches data following data being read and a second technique that pre-fetches data before and following the data being read. Additional executable code reads data following data being read based on comparison of a generated random number and the first set of statistics upon detection of a read-miss and selecting the first technique. Additional executable code reads data before and following data being read based on a comparison of a generated random number and the third set of statistics upon detection of a read-miss and selecting the second technique. Pre-fetching data before and following data being read may include fetching an entire track from a disk. Pre-fetching data following data being read may include fetching data from the data being read to the end of a track on a disk. Selecting a pre-fetch technique may include comparing the second set of statistics with a predefined value. The predefined value may be a user settable parameter.
According further to the system described herein, determining when to pre-fetch a full track of data on a disk includes collecting statistics based on a number of avoidable read-misses in which data exists that is in the same track as data being read. Upon detection of a read-miss, a full track of data is pre-fetched based on comparison of a generated random number and the third set of statistics. The statistics may be based on tracking a number of avoidable read-misses in which data exists that follows data being read and a number of avoidable read-misses in which data exists that is prior to data being read.
According further to the system described herein, computer software, provided in a computer-readable medium, determines when to pre-fetch additional data from a disk. The software includes executable code that provides for selection of one of the following modes: a first mode for pre-fetch to end of track only, a second mode for pre-fetch entire track, a third mode where no pre-fetch is performed, and a fourth mode where pre-fetch to an end of track is performed upon second and subsequent read misses for a track. The software also includes executable code that collects a first set of statistics based on a number of avoidable read-misses in which data exists that is prior to data being read, executable code that collects a second set of statistics based on a number of avoidable read-misses in which data exists that follows data being read in response to selection of the second mode, and executable code that collects a third set of statistics based on said first and second sets of statistics in response to selection of the second mode. The software also includes executable code that pre-fetches data following data being read based on comparison of a generated random number and the first set of statistics upon detection of a read-miss and on selecting of the first mode and includes executable code that pre-fetches data before and following data being read based on a comparison of a generated random number and the third set of statistics upon detection of a read-miss and selecting the second mode. The second mode may also include pre-fetching to the end of track and the software may also include executable code that, in response to selection of the second mode, uses the second set of statistics to select a pre-fetch technique from a first technique that pre-fetches data following data being read and a second technique that pre-fetches data before and following the data being read and includes executable code that pre-fetches data following data being read based on selection of the second mode and the first technique.
Referring to
The storage device 24 may include one or more RDF adapter units (RA'"'"'s) 32a-32c. The RA'"'"'s 32a-32c are coupled to an RDF link 34 and are similar to the HA'"'"'s 28a-28c, but are used to transfer data between the storage device 24 and other storage devices (not shown) that are also coupled to the RDF link 34. The storage device 24 may be coupled to addition RDF links (not shown) in addition to the RDF link 34.
The storage device 24 may also include one or more disks 36a-36c, each containing a different portion of data stored on the storage device 24. Each of the disks 36a-36c may be coupled to a corresponding one of a plurality of disk adapter units (DA) 38a-38c that provides data to a corresponding one of the disks 36a-36c and receives data from a corresponding one of the disks 36a-36c. Note that, in some embodiments, it is possible for more than one disk to be serviced by a DA and that it is possible for more than one DA to service a disk.
The logical storage space in the storage device 24 that corresponds to the disks 36a-36c may be subdivided into a plurality of volumes or logical devices. The logical devices may or may not correspond to the physical storage space of the disks 36a-36c. Thus, for example, the disk 36a may contain a plurality of logical devices or, alternatively, a single logical device could span both of the disks 36a, 36b. The hosts 22a-22c may be configured to access any combination of logical devices independent of the location of the logical devices on the disks 36a-36c.
One or more internal logical data path(s) exist between the DA'"'"'s 38a-38c, the HA'"'"'s 28a-28c, the RA'"'"'s 32a-32c, and the memory 26. In some embodiments, one or more internal busses and/or communication modules may be used. In some embodiments, the memory 26 may be used to facilitate data transferred between the DA'"'"'s 38a-38c, the HA'"'"'s 28a-28c and the RA'"'"'s 32a-32c. The memory 26 may contain tasks that are to be performed by one or more of the DA'"'"'s 38a-38c, the HA'"'"'s 28a-28c and the RA'"'"'s 32a-32c, and a cache for data fetched from one or more of the disks 36a-36c. Use of the memory 26 is described in more detail hereinafter.
The storage device 24 may be provided as a stand-alone device coupled to the hosts 22a-22c as shown in
Referring to
The diagram 50 also shows an optional communication module (CM) 54 that provides an alternative communication path between the directors 52a-52c. Each of the directors 52a-52c may be coupled to the CM 54 so that any one of the directors 52a-52c may send a message and/or data to any other one of the directors 52a-52c without needing to go through the memory 26. The CM 54 may be implemented using conventional MUX/router technology where a sending one of the directors 52a-52c provides an appropriate address to cause a message and/or data to be received by an intended receiving one of the directors 52a-52c. Some or all of the functionality of the CM 54 may be implemented using one or more of the directors 52a-52c so that, for example, the directors 52a-52c may be interconnected directly with the interconnection functionality being provided on each of the directors 52a-52c. In addition, a sending one of the directors 52a-52c may be able to broadcast a message to all of the other directors 52a-52c at the same time.
In some embodiments, one or more of the directors 52a-52c may have multiple processor systems thereon and thus may be able to perform functions for multiple directors. In some embodiments, at least one of the directors 52a-52c having multiple processor systems thereon may simultaneously perform the functions of at least two different types of directors (e.g., an HA and a DA). Furthermore, in some embodiments, at least one of the directors 52a-52c having multiple processor systems thereon may simultaneously perform the functions of at least one type of director and perform other processing with the other processing system. In addition, the global memory 26 may be provided on one or more of the directors 52a-52c and shared with other ones of the directors 52a-52c.
Referring to
The system described herein assumes the existence of a relatively slow memory (e.g., the disks 36a-36c) and a relatively fast memory (e.g., the global memory 26/global cache memory 100). It is desirable to move data from the relatively slow memory to the relatively fast memory in a way that maximizes system performance by attempting to anticipate future data requests and pre-fetching appropriate data from the relatively slow memory into the relatively fast memory. It is desirable to maintain more frequently-accessed data in the global cache memory 100 while having less frequently-accessed data maintained on one of the disks 36a-36c, but not in the global cache memory 100. A request for data is serviced by first determining if the data is already in the global cache memory 100.
Referring to
Referring to
An avoidable read-miss of a sector is one that could have been averted had the sector been pre-fetched when other sectors of the same track had been previously fetched. Thus, for example, following fetching the data 122 of the slot 102, a subsequent access of data corresponding to the undefined portion 124 is an avoidable read-miss. Note that a read-miss is unavoidable if the track corresponding to the data is not already in the global cache memory 100.
An excessive number of avoidable read-misses is an indication that the system should more frequently pre-fetch additional sectors of a track in connection with fetching one of the sectors in response to a read request. Of course, all avoidable read-misses may be eliminated by fetching an entire track the first time any data for the track is fetched. However, this may be inefficient for a number of reasons. For example, pre-fetching unnecessary data may use up bandwidth in a way that degrades performance of the system, and may increase the disk queuing time. Accordingly, it is desirable to minimizes avoidable read-misses while also minimizing the amount of additional data for a track/slot that the system reads.
The system described herein classifies read-misses as being avoidable or unavoidable and uses that data to determine when and how to pre-fetch sectors of a track. Thus, when a read-miss occurs, the system updates (increases) one or more counters. Then the system uses those counters in connection with a statistical algorithm that controls whether to read more data in a track than just the requested sector.
In addition, there are two different types of avoidable read-misses: read-misses that would have been avoided if a previous read had read the full track (RM_RFT) and read misses that would have been avoided if a previous read had read from the requested sector to the end of the end of a track (RM_REOT). Of course, in any instances where an RM_REOT read miss occurs, an RM_RFT read miss also occurs. However, it is possible for an RM_RFT read-miss to occur without an RM_REOT read miss also occurring.
The system described herein may be configured to implement either no track pre-fetch mechanism at all, a track pre-fetch mechanism that pre-fetches on the second and subsequent read misses for a track, implement a track pre-fetch algorithm like that disclosed in U.S. Pat. No. 6,529,998 to Yechiel Yochai and Robert Mason (read to end of track in some cases), or to implement an track pre-fetch algorithm that, in some cases, pre-fetches to the end of a track or pre-fetches an entire track. In an embodiment herein a user (user process) may select the particular configuration. Note that other selectable configurations may also be possible.
Referring to
Processing begins at a first step 152 where it is determined if any track pre-fetch algorithms are in use. As discussed elsewhere herein, it is possible to provide an adjustable parameter that can be set so that no track pre-fetch algorithms are in use and/or only an algorithm that pre-fetches to the end of the track on second and subsequent read misses for a track. If it is determined at the test step 152 that no first I/O pre-fetching is to be performed (that no track pre-fetch algorithms are in use and/or only an algorithm that pre-fetches to the end of the track on second and subsequent read misses for a track.), then there is no need to collect any read-miss statistics. Thus, if it is determined at the test step 152 that no pre-fetch algorithms are in use, then the process of updating statistics is complete.
If it is determined at the test step 152 that a track pre-fetch algorithm is being used, then control passes from the test step 152 to a test step 154 where it is determined if previous data for the track is already in the global cache memory 100. If a previous read for the track had fetched sectors in the track that precedes the sector being read (or the first sector being read in the case of reading a number of contiguous sectors), then the current read-miss would have been avoided if the previous read for the track had read to the end of the track (an RM_REOT read-miss). Accordingly, if it is determined at the test step 154 that previous data for the track exists, then control transfers from the test step 154 to a step 156 where a Boolean variable, RM_PREV, is set to true. Otherwise, control transfers from the test step 154 to a step 158 where RM_PREV is set to false. The RM_PREV variable, as well as the RM_AHEAD variable (discussed below), are used in follow on counter calculations.
Following either the step 156 or the step 158 is a test step 162 where it is determined if a read full track pre-fetch algorithm is in use. As discussed elsewhere herein, in some instances, it is possible to provide an adjustable parameter that can be set so that only the read to end-of-track pre-fetch algorithm is in use, in which case there is no need to collect any read-miss statistics relating to the read full track pre-fetch algorithm.
If it is determined at the test step 162 that a read full track pre-fetch algorithm is being used, then control passes from the test step 162 to a test step 164 where it is determined if data for the track is already in the global cache memory 100 in a sector that is ahead (follows) the current sector being read (or the first sector being read in the case of reading a number of contiguous sectors). If a previous read for the track had fetched sectors in the track that are ahead of the sector being read, then the current read-miss would have been avoided if the previous read for the track had read the entire track (an RM_RFT read-miss). Accordingly, if it is determined at the test step 164 that previous data for the track exists in a sector that is ahead of the sector being read, then control transfers from the test step 164 to a step 166 where a Boolean variable, RM_AHEAD, is set to true. Otherwise, control transfers from the test step 164 to a step 168 where RM_AHEAD is set to false.
Following either the step 166 or the step 168 is a step 172 where counters (RFT_CNT, REOT_CNT, and RBOT_CNT) are calculated in a manner discussed in more detail below. Note that the step 172 is also reached from the test step 162 if it is determined at the test step 162 that the read full track pre-fetch algorithm is not in use. The counters are used to determine whether to read to the and of a track and/or to read a full track rather than just reading the requested sector.
Referring to
Following the test step 182 if RM_PREV is false, or following the step 184 is a test step 186 where it is determined if RM_AHEAD is true. If so, then control transfers from the test step 186 to a step 188 where the RBOT_CNT variable is incremented. Following the test step 186 if RM_AHEAD is false, or following the step 188 is a test step 192 where it is determined if either RM_PREV or RM_AHEAD are true. If not, then processing is complete. Otherwise, control transfers from the test step 192 to a step 194 where RFT_CNT is incremented. Following the step 194, processing is complete. Note that, in some embodiments, it may be possible to provide a single variable (e.g., RM_ANY) that is true whenever RM_PREV is true or RM_AHEAD is true.
Referring to
Processing begins at a first step 202 where the counter is incremented by an incremental amount, which may be a fixed amount, a variable amount, an amount that varies according to the current value of the counter being incremented, a function of a minimum acceptable value for the counter, etc. Any appropriate incremental amount may be used. The incremental amount may be made larger or smaller based on functional factors, such as observed operations of the system. Generally, the larger the incremental amount, the more the system favors reading additional sectors. In an embodiment herein, the counters range from one to 255. The incremental amount may be determined according to the methodology in U.S. Pat. No. 6,529,998 to to Yechiel Yochai and Robert Mason. Note also that setting the incremental amount and/or the ratio of the incremental amount to the decrement amount (described below) may depend on a variety of observable system parameters and may be adjusted to maximize system performance.
Following the step 202 is a test step 204 where it is determined if increases the counter has caused the counter to exceed the maximum value for the counter. If not, then processing is complete. Otherwise, control transfers from the test step 204 to a step 206 where the counter is set to its maximum value. Following the step 206, processing is complete.
Referring to
If it is determined at the test step 214 that there are not other (prior) portions of the track in global cache memory that are before the current read miss, then control transfers to a step 217 where it is determined if any track pre-fetch algorithms are enabled. As discussed elsewhere herein, the system may be configured so that no pre-fetching is performed. If it is determined at the test step 217 that no track pre-fetching is to be performed, then processing is complete and only the requested data will be bought to cache. Use of the step 214 causes the test for whether track pre-fetch algorithms are in use to only be performed on a first read miss for a track since, on second and subsequent read misses, a read to end of track is performed (at the step 216). Note, that, in some embodiments, it is possible to forgo the step 214 (and thus the step 216) altogether. This is illustrated by the dashed line in
If it is determined at the test step 217 that track pre-fetching is to be performed, then control transfers from the step 217 to a test step 218 where it is determined if the read full track pre-fetch algorithm is enabled. If so, then control transfers from the test step 218 to a test step 224 where it is determined if a predefined value (such as a user settable parameter) is less than the value of the RBOT_CNT variable. The predefined value may be any value that ranges between the minimum possible value for RBOT_CNT to the maximum possible value for RBOT_CNT. In one embodiment the value for RBOT_CNT may be set to slightly less than the maximum possible value.
If it is determined at the test step 224 that the predefined value is not less than RBOT_CNT, then control passes from the test step 224 to a step 226 where REOT algorithm processing is performed. The REOT algorithm processing performed at the step 226 is described in more detail elsewhere herein. Note that the step 226 is also reached from the step 218 if it is determined that the read full track pre-fetch algorithm is not being used. Following the step 226, processing is complete. If it is determined at the test step 224 that the predefined value is less than the RBOT_CNT value, then control passes from the test step 224 to a step 228, where the read full track pre-fetch algorithm processing is performed. The processing performed at the step 228 is described in more detail elsewhere herein. Following the step 228, processing is complete.
Note that comparison of the predefined value with RBOT_CNT is only one of many possible mechanisms that may be used to chose between using the REOT algorithm and the full track algorithm. Other appropriate techniques may be used instead, including possibly generating a random number that is compared to the value of RBOT_CNT. In addition, selection of the predefined value may be based on any appropriate criteria, including observable system performance with different values used for the predefined value.
Referring to
If it is determined at the test step 244 that the random number is not less than the value of REOT_CNT, then control transfers from the test step 244 to a step 248 where the value of REOT_CNT is decremented. Decrementing the REOT_CNT variable (and other counters) is like incrementing counters, described elsewhere herein. However, the amount used for decrementing may be different than the amount used for incrementing. In some embodiments the formula given at U.S. Pat. No. 6,529,998 can be used to determine the increasing step and the ratio between the increasing and decreasing steps. In an embodiment herein, the ratio of amount used for decrementing to the amount used for incrementing is one to nine. The relationship between the amount used for decrementing counters and the amount used for incrementing counters may correspond to the perceived cost (e.g., latency time) of performing unnecessary pre-fetch operations compared with the perceived cost of avoidable read-misses. Note also that different and independent increment and decrement amounts may be assigned to each of the counters.
Following the step 248 is a test step 252 where it is determined if the read full track algorithm is enable (discussed in detail elsewhere herein). If not, then processing is complete. Otherwise, control passes from the step 252 to a step 254 where the RFT_CNT variable is decremented. Following the step 254 is a step 256 where the RBOT_CNT variable is decremented. Following the step 256, processing is complete.
Referring to
If it is determined at the test step 264 that the random number is not less than the value of RFT_CNT, then control transfers from the test step 264 to a step 268 where the value of RFT_CNT is decremented. Decrementing the counts is described elsewhere herein. Following the step 268, is a step 272 where the REOT_CNT variable is decremented. Following the step 272 is a step 274 where the RBOT_CNT variable is decremented. Following the step 274, processing is complete.
Although the system described herein uses tracks/sectors/blocks as units of data for certain purposes, it should be understood that other units of data (including, possibly, variable length units of data) may be used. This also applies to other data structures and data units. In addition, in some instances, the order of steps in the flow charts may be modified, where appropriate. The processes described herein may be controlled by executable code stored on a computer-readable medium on one or more processors (e.g., on the directors 52a-52c) and other devices described herein that execute steps of the described processes.
While the invention has been disclosed in connection with various embodiments, modifications thereon will be readily apparent to those skilled in the art. Accordingly, the spirit and scope of the invention is set forth in the following claims.