EP1934838A2 - Accumulating access frequency and file attributes for supporting policy based storage management - Google Patents

Accumulating access frequency and file attributes for supporting policy based storage management

Info

Publication number
EP1934838A2
EP1934838A2 EP06815981A EP06815981A EP1934838A2 EP 1934838 A2 EP1934838 A2 EP 1934838A2 EP 06815981 A EP06815981 A EP 06815981A EP 06815981 A EP06815981 A EP 06815981A EP 1934838 A2 EP1934838 A2 EP 1934838A2
Authority
EP
European Patent Office
Prior art keywords
transaction
file
module
storage network
policy
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
EP06815981A
Other languages
German (de)
French (fr)
Other versions
EP1934838A4 (en
Inventor
Panagiotis Tsirigotis
Geetha Srikantan
Thomas K. Wong
Chi Ming Wong
Anand Iyengar
Rajeev Chawla
Richard Simpkins
Zuwei Liu
Gaurav Gupta
Matthew Seitz
Yu Cheong Chan
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NeoPath Networks Inc
Original Assignee
NeoPath Networks Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by NeoPath Networks Inc filed Critical NeoPath Networks Inc
Publication of EP1934838A2 publication Critical patent/EP1934838A2/en
Publication of EP1934838A4 publication Critical patent/EP1934838A4/en
Ceased legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/10File systems; File servers
    • G06F16/11File system administration, e.g. details of archiving or snapshots
    • G06F16/122File system administration, e.g. details of archiving or snapshots using management policies
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/10File systems; File servers
    • G06F16/18File system types
    • G06F16/182Distributed file systems
    • G06F16/1824Distributed file systems implemented using Network-attached Storage [NAS] architecture
    • G06F16/1827Management specifically adapted to NAS
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/10File systems; File servers
    • G06F16/18File system types
    • G06F16/1865Transactional file systems

Definitions

  • This invention relates generally to storage networks and, more specifically, to performing policy-based storage management using data related to access frequency and file attribute accumulation.
  • NAS file servers connected directly to the network provide an inexpensive and easily configurable solution for a storage network.
  • These NAS file servers are self-sufficient because they contain file systems that allow interoperability with clients running any operating system and communication using open protocols.
  • NFS Network File System
  • CIFS Common Internet File System
  • Microsoft Corp. of Redmond, Washington to access files on a NAS file server.
  • the operating system does not affect communication between the client and file server.
  • NAS file servers provide true universal file access.
  • SAN Storage Area Network
  • a SAN file system is part of the operating system or an application running on the client.
  • heterogeneous client operating systems may require additional copies of each file to be stored on the storage network to ensure compatibility on the SAN file server.
  • communication between file servers on a SAN use proprietary protocols and thus are typically provided by a common vendor.
  • NAS file servers are preferred when price and ease of use are major considerations.
  • the benefits of NAS storage networks over SAN storage networks also have drawbacks.
  • each client must maintain communication channels between each of the NFS file servers separately.
  • each client When NAS file servers are either added or removed from the storage network, each client must mount or unmount directories for the associated storage resources as appropriate. This is particularly inefficient when there are changes in hardware, but not in the particular files available on the network, such as when a failing NAS file server is swapped out for an identically configured back-up NAS file server.
  • a related drawback is that a client must be reconfigured each time a file is relocated within the storage network, such as during file migration or file replication. The client generates a NAS file handle that identifies a physical location of the directory or file object on the file server. To access the object, the client sends an object access request directly to the NAS file server. When the file is relocated to a different NAS file server, subsequent requests for access to the file require a new look-up to locate the file and generate a new NAS file handle.
  • NAS file servers can become consumed with handling I/O (Input/Output) requests associated with file manipulations and accesses.
  • I/O Input/Output
  • additional processing tasks such as queries can unduly burden the NAS file servers.
  • the file server typically walks a tree-structured directory tree in search of information requested by the query, and if there is more than one file system, each file system is individually walked. Consequentially, the file server may either become less responsive to I/O requests or have high latency in responding to the query. In some contexts, high latency will make the results stale.
  • NAS file servers can become unorganized and inefficient by, for example, storing critical data with other non-critical data.
  • large multimedia collections of MP3s used for leisure by employees can increase latency time in receiving information more germane to the productivity of an enterprise such as financial records.
  • rarely accessed files may be stored on a premium, high bandwidth file server while often accessed files may be stored on a commodity, lower bandwidth server.
  • a switch device provides transparency for transactions between a client and a storage network.
  • the transparency allows objects (e.g., files or directories) to be moved (e.g., migrated) on the storage network without affecting a reference to the object used by the client (e.g., a file handle).
  • a monitoring module generates accumulation data associated with the transactions for use in policy-based management.
  • the accumulation data can describe uses of the file such as how often certain files are accessed, modifications to files such as creations of new directories or files, and other uses.
  • One embodiment of the monitoring module comprises an I/O monitoring module to receive transaction requests processed by the switch device, and determine if the underlying transaction concerns a file system being monitored.
  • a transaction categorization module can determine whether the transaction is a type being monitored, and if so, forward to an appropriate log in a transaction logs module for storage.
  • the transaction log can be, for example, a cyclic buffer.
  • An accumulator module processes logs on a periodic basis and stores its results in a records repository.
  • a policy module manages the records repository in accordance with predetermined rules. For example, a network administrator can set the rules from a user interface. Also, the rules can be pre-loaded.
  • the policy module can include a scheduling module to invoke the policy at certain intervals, a policy execution module to form queries from the rules, and a query module to manage the queries to the records repository.
  • the queries can be, for example, SQL queries.
  • FIG. 1 is a high-level block diagram illustrating a storage network system according to one embodiment of the present invention.
  • FIG. 2 is a block diagram illustrating the NAS switch according to one embodiment of the present invention.
  • FIG. 3 is a block diagram illustrating a selective migration module according to one embodiment of the present invention.
  • FIG. 4 is a high-level flow chart illustrating a method for providing transparent file migration in a NAS storage network according to one embodiment of the present invention.
  • FIG. 5 is a flow chart illustrating a method of associating original NAS file handles with switch file handles according to one embodiment of the present invention.
  • FIG. 6 is a flow chart illustrating a method for providing selective migration in accordance with a policy according to one embodiment of the present invention.
  • FIG. 7 is a flow chart illustrating a method for providing selective migration in accordance with a policy according to another embodiment of the present invention.
  • FIG. 8 is a block diagram illustrating a monitoring module according to one embodiment of the present invention.
  • FIG. 9 is a block diagram illustrating transaction logs according to one embodiment of the present invention.
  • FIG. 10 is a block diagram illustrating a policy module according to one embodiment of the present invention.
  • FIG. 11 is a flow chart illustrating a method for logging transactions according to one embodiment of the present invention.
  • FIG. 12 is a flow chart illustrating a method for applying a policy to logged transactions according to one embodiment of the present invention.
  • a system and method for performing policy-based storage management using data related to access frequency and file attribute accumulation can include rules that establish which objects are migrated from a source file server to a destination file server. For example, large multimedia files that consume bandwidth on expensive or critical file servers, without adding much value to enterprise productivity, can be migrated to a commodity or less critical file server.
  • a NAS (Network Attached Storage) switch in the data path of a client and NAS file servers on the storage network implements selective migration using file handles that are independent of locations on the storage network. Mapping between the file handles which are static and locations which can be dynamic, allows the client to access objects after migration using file handle that persists through the migration.
  • FIG. 1 is a high-level block diagram illustrating a storage network system 100 according to one embodiment of the present invention.
  • the system 100 comprises a NAS switch 110 (or more generically, a switch device) and a client 140 coupled to a network 195.
  • the NAS switch 110, a source file server 120, and a destination file server 130, are each coupled in communication through a sub-network 196.
  • the system 100 components are implemented in, for example, a personal computer with an x86-type processor executing an operating system and/or an application program, a workstation, a specialized NAS device with an optimized operating system and/or application program, a modified server blade, etc.
  • the storage network 175 comprises a NAS using protocols such as NFS and CIFS.
  • the storage network 175 comprises a combination of NAS, SAN, and other types of storage networks.
  • the storage network 175 comprises a decentralized standard or proprietary storage system other than NAS.
  • the NAS switch 110 selectively migrates objects from a location on the source file server 120 to a location on the destination file server 130. Selective migration can determine which objects to migrate based on file attributes such as file type, file size, file access frequency, other file conditions, schedules, and the like, as determined by a policy.
  • the policy can include rules that delineate certain actions in accordance with certain file attributes or conditions.
  • the NAS switch 110 can perform a rehearsal that shows the effects of a policy in a report. The policy can be iteratively adjusted to reach desired results.
  • the NAS switch 110 provides continuous transparency to the client 140 with respect to object management. Specifically, the NAS switch 110 can off-load tasks related to physical configurations, object management, object migration, object replication, efficient storage and/or other services on the storage network 175.
  • the NAS switch 110 emulates file server processes to the client 140 and emulates client processes to the file servers 120, 130. Accordingly, the client 140 is unaware of the NAS switch 110 since the NAS switch 110 is able to redirect NAS requests intended for the source file server 120 to appropriate locations on the destination file server 130. Thus, the client 140 submits object requests, such as file writes and directory reads, directly to the NAS switch 110.
  • the file servers 120, 130 are unaware of the NAS switch 110 since the NAS switch 110 is able to resubmit requests, contained in server file handles, as if they originated from the client 140. To do so, the NAS switch 110 can use mapping, translating, bridging, packet forwarding, other network interface functionality, and other control processes to perform file handle switching, thereby relieving the client 140 of the need to track changes in a file's physical location.
  • the NAS switch 110 receives exported file system directories from the file servers 120, 130 containing NAS switch handles. To create compatibility between the client 140 and the NAS switch 110, the NAS switch 110 maps the file system directories to internal switch file systems which it sends to the client 140. To request an object, the client 140 traverses an exported switch file system and selects a switch file handle which it sends to the NAS switch 110 along with a requested operation.
  • NAS file handles uniquely identify objects, such as a directory file server, on the file servers 120, 130, such as a directory or file, as long as that object exists.
  • NAS file handles are file server specific, and are valid only to the file servers 120, 130 that issued the file handles.
  • the process of obtaining a file handle from a file name is called a look-up.
  • the NAS file handle may be formatted according to protocols such as NFS or CIFS as discussed in further detail below, e.g., with reference to Tables IA and IB.
  • a switch file handle identifies a directory or file object independent of location, making it persistent through file replications, migrations, and other data transfers.
  • the switch file handle can be a modified NAS file handle that refers to an internal system within the NAS switch 110 rather than the source file server 120. This enables the NAS switch 110 in mapping persistent file handles to a choice of alternative NAS file handles.
  • An original NAS file handle refers to an initial object location on the source file server 120.
  • a stored NAS file handle refers to a NAS file handle, stored as an object on the file servers 120, 130, which points to an alternative file location.
  • Object transaction requests handled by the NAS switch 110 include, for example, directory and/or file reads, writes, creation, deletion, moving, and copying.
  • a namespace access refers to an operation accessing or modifying the namespace such as lookup, rename, delete, or create.
  • a file access refers to an operation accessing or modifying files such as read or write.
  • An object can refer to a directory object or a file object. Directory objects can further comprise sub-directories and file objects within directory.
  • the NAS switch 110 can be configured to monitor the transaction requests to accumulate frequency and file attribute information as described in more detail below. The accumulated information can be used for various applications allowing policy enforcement.
  • One type of policy enforced on the storage network 175 can be selective migration.
  • a selective migration application can optimize the storage network 175 by moving more frequently used files to high performance servers, and by moving less frequently used files to commodity servers.
  • Another type of policy enforcement can be file clean up. For example, all non-essential files older than 2 year are deleted. Many other implementation-specific policies are possible given the disclosure herein.
  • the NAS switch 110 can be configured for other types of applications such as indexing and categorization.
  • An indexing application can build an index database for fast content-based searching. Recent transactions which modify the storage network 175 can be recorded for updating the index database without the need for walking the file systems.
  • a categorization application can maintain statistics on the types of files being stored. The list allows files of the same type to be quickly identified without walking the file systems. As a result, in one example, non-essential files such as MP3s can be quickly identified and migrated to commodity servers.
  • the client 140 accesses resources on the file servers 120, 130 by submitting a switch file handle to the NAS switch 110, intended for the source file server 120.
  • the client 140 first mounts an exported switch file system containing switch file handles.
  • the client 140 looks-up an object to obtain its file handle and submits an associated request. From the perspective of the client 140, transactions are carried out by the file servers 120, 130 having object locations that do not change.
  • the client 140 interacts with the NAS switch 110 before and after a file replication in the same manner.
  • a user of the client 140 can submit operations through a command line interface, a windows environment, a software application, or otherwise.
  • the NAS switch 110 further provides access to a storage network 175 other than a NAS storage network.
  • the source file server 120 is the default or original network file server for the client 140 before file migration.
  • the source file server 120 further comprises source objects 125, which include namespace directories and files such as enterprise data, records, database information, applications, and the like.
  • the source file server 120 can store a table of migrated directories maintained by the NAS switch 110 that correlate results from namespace migration.
  • the source file server 120 can store a file handle migration table, maintained by the NAS switch 110, denoting each migrated directory and file object.
  • the source file server 120 comprises, for example, a personal computer using an x86-type processor with an operating system and/or an application, a workstation, a specialized NAS device with an optimized operating system and/or application, a modified server blade, etc.
  • the destination file server 130 becomes the primary network file server used by the NAS switch 110 after file migration.
  • the destination file server 130 further comprises destination objects 135, which include the replicated namespace directories and source files.
  • the destination file server 130 can comprise the same hardware and/or software as described with reference to the source file server 120.
  • the source and destination file servers 120, 130 are preferably NAS file server, but can also be file servers using other decentralized protocols that do not inherently support file migration. Further embodiments of the source and destination file servers 120, 130 and related methods are described below.
  • the network 195 facilitates data transfers between connected hosts (e.g., 110,
  • the connections to the network 195 may be wired and/or wireless, packet and/or circuit switched, and use network protocols such as TCP/IP (Transmission Control Protocol/Internet Protocol), IEEE (Institute of Electrical and Electronics Engineers) 802.11, IEEE 802.3 (i.e., Ethernet), ATM (Asynchronous Transfer Mode), or the like.
  • the network, 195 comprises, for example, a LAN (Local Area Network), WAN (Wide Area Network), the Internet, and the like.
  • the NAS switch 110 acts as a gateway between the client 140, connected to the Internet, and the directory file server 120, and the shadow file servers 130, connected to a LAN.
  • the sub-network 196 is preferably a local area network providing optimal response time to the NAS switch 110. hi one embodiment, the sub- network 196 is integrated into the network 195.
  • FIG. 2 is a block diagram illustrating the NAS switch 110 according to one embodiment of the present invention.
  • the NAS switch 110 comprises a mapping module 210 and a selective migration module 220.
  • the mapping module 210 handles file handles I/O transactions associated with objects, and sends information about successful FO transactions to the selective migration module 220 for tracking updates to since the last walk of the tree-structured file servers.
  • modules are merely exemplary groupings of functionality.
  • the mapping module 210 Prior to file migration, the mapping module 210 receives a switch file handle with a request from the client 140 which it uses to find an original NAS file handle. The mapping module 210 submits the original NAS file handle with the request to the source file server 120. If the object has yet to change locations in the storage network 175, the mapping module 210 uses the original NAS file handle. The mapping module 210 can submit the switch file handle to the migration module 220 to determine if the object is part of a data migration.
  • the file server interface 220 can submit the switch file handle to the redirection module 230 to determine if the object has completed data migration, hi either case, an appropriate NAS file handle is returned for the mapping module 210 to use in forwarding the client request to the appropriate file server 120, 130 [0040]
  • the mapping module 210 looks-up switch file handles received from the client 140 in the file handle migration table. If an object has been migrated, the redirection module outputs a destination NAS file handle corresponding to a location on the destination file server 130.
  • the selective migration module 220 receives information about successfully
  • the selective migration module 220 can intercept transactions headed for the mapping module 210 before processing, or receive a duplicate of transactions sent to the mapping module 210.
  • the selective migration module 220 can update file locations in the mapping module 210.
  • FIG. 3 is a block diagram illustrating a selective migration module 220 according to one embodiment of the present invention.
  • the selective migration module 220 comprises a monitoring module 310, a policy module 320, a migration engine 330, and a records repository 340.
  • the records repository is populated using a traversal (walk) of tree-structured directories or shares, and records are updated with information about transactions with objects in the directories.
  • the policy module 320 can perform policy- based migration through query or sorting interactions with the records repository 340.
  • the monitoring module 310 comprises an I/O monitoring module 810, a transaction categorizing module 820. transaction logs 830, and an accumulator module 840.
  • the monitoring module 310 receives information relating to successful I/O transactions involving objects, and sends processed information to the records repository 340 for storage. More specifically, the I/O monitoring module 810 determines whether the transaction is associated with a file system being monitored.
  • the transaction categorizing module 820 receives transactions from the I/O monitoring module 810 to determine whether the transaction is a type of transaction being monitored.
  • the transaction logs 830 stored monitored transactions as transaction records. At certain intervals, the accumulator module 840 processes transaction records for storage in the records repository 340.
  • the transaction logs 830 comprise records for specific requests: a namespace update log 910, a data update log 920, and an access log 930; and records for general requests in a general request log 940.
  • the logs 910-940 can be part of a memory device such as a RAM (random access memory) device or other type of memory element.
  • the logs 910-940 can operate as cyclic buffers such as a FIFO (first in first out) type of buffer that replaces the oldest record with the newest record.
  • the policy module 320 comprises a scheduling module 1010, a policy execution module 1020, and a query module 1030.
  • the scheduling module 1010 initiates a process at intervals (e.g., as programmed by a system administrator).
  • the policy execution module 1020 accesses rules that form a policy, and feeds specific commands to the query engine 800.
  • the rules can depend upon the application.
  • the rules can be preconfigured or created by a network administrator.
  • the rules can be Boolean combinations of conditions. For example, FILE TYPE IS MPEG and FILE SIZE IS MORE THAN 100 MEGABYTES.
  • the query module 1030 manages the application of specific commands for application against data stored in the records repository 340 (e.g., accumulation and frequency data).
  • the migration engine 330 can migrate file identified by the policy module 320. For example, each of the files that have not been accessed in the last year can be moved to a dedicated filer server for rarely accessed files. In one embodiment, the migration engine 320 migrates the namespace associated with each object prior to migrating data associated with each object.
  • the records repository 340 can store records associated with objects in the tree-structured directories by traversing the tree-structured directories. In response to receiving logs from the monitoring module 310, the records repository 340 can update records. Periodically, the records repository 340 can synch with the directories with traversals.
  • FIG. 6 is a flow chart illustrating a method 600 for providing selective migration in accordance with a policy according to one embodiment of the present invention.
  • a NAS file handle including a location of the object on a source file server is received 610.
  • a switch file handle that is independent of the location of the object is sent as representative of the object 620.
  • the object is selectively migrated 630 from the location on the source file server to a location on a destination file server in accordance with a policy that includes rules for storing objects in the storage network.
  • An access request associated with the switch file handle is received 640.
  • the access request is sent 650 to the location on the destination file server.
  • the information can be recorded one or more logs, as discussed below with respect to FIG. 9.
  • an access log can be updated each time a file or directory is read. Logs can be maintained based on the importance of tracked transactions. For example, while all file or directory creations are tracked, in one embodiment, only the last file modification is tracked. In another optimization, an access log can count accesses over an hour without recording the time of each access.
  • the monitoring module 310 periodically sends the logs to the records repository 340 for processing (e.g., once an hour).
  • the policy module 320 can implement rules with searches on the records repository 340 to identify files meeting the conditions, hi one embodiment, a user interface (e.g., viewed in a web browser) can allow a network administrator to configure rules.
  • the policy module 320 can be triggered periodically on a per-policy basis such as once a day or once a week.
  • FIG. 7 is a flow chart illustrating a method 700 for providing selective migration in accordance with a policy according to another embodiment of the present invention.
  • a policy that includes rules for storing objects in the storage network is received 710.
  • Objects stored in a tree-structure on a source file server in the storage network are traversed 720.
  • a policy database (or records repository) is populated 730 with records representative of the tree-structure.
  • the policy database is updated 740 to reflect transactions associated with the objects.
  • the rules are executed 750 against records in the policy database to determine which objects are to be migrated, as described below with respect to FIG. 12.
  • the objects are migrated 760 from a location on the source file server to a location on a destination file server.
  • FIG. 4 is a flow chart illustrating a method 400 for migrating objects according to one embodiment of the present invention.
  • the NAS switch 110 Prior to migration, the NAS switch 110 associates 410 original NAS file handles with switch file handles 175. As detailed in FIG. 5, the exported file system 510 containing the original file handles is mounted from the source file server. The switch file handle is generated 520 independent of object locations. The switch file system containing the switch file handles is then exported 50 to the client 140. Referring again to FIG. 4, migration from the source file server 110 to the destination file server 120 is performed 420 using namespace replication. After migration, NAS requests (i.e., transaction requests) are redirected by the mapping module 210 according to the new locations.
  • the method 400 is described in further detail in U.S. Patent No.
  • FIG. 11 is a flow chart illustrating a method 1100 for generating accumulation data according to one embodiment of the present invention.
  • the mapping module 210 of the NAS switch 110 processes 1110 a transaction request for an object in the storage network 175.
  • a copy of the transaction request is sent to the monitoring module 310.
  • the NAS switch 110 may service the transaction request prior to copying to ensure its validity. For example, if a file is requested that does not exist, the transaction request would lead to invalid accumulation data.
  • the I/O monitoring module 810 determines 1120 whether the transaction request concerns a monitored file system.
  • a network administrator can use an interface to select which file systems are monitored. If the file system has not been selected, no further processing is necessary.
  • the transaction categorizing module 820 determines 1130 whether the transaction request is a type of transaction being monitored.
  • Monitored transactions can include, for example, FDNOPS, FDAOPS, or FDUOPS. More specifically, FDNOPS include updates to a namespace of a file system such as create, rename, delete, create directory, rename directory, or remove directory.
  • FDAOPS include updates to data in a file system such as a write operations, or a change or set file size operation.
  • FDUOPS include accesses to a file system such as a read operation. Some embodiments include additional types of operations such as open or close operations.
  • the transaction logs 830 log 1140 the transaction request according to its type.
  • the transaction categorization module 820 can set a code for the type of transaction which directs it to a specific log.
  • FDNOPS are stored in the namespace update log 910
  • FDUOPS are stored in the data update log 920
  • FDAOPS are stored in the access log 920
  • others are stored in the general request log 940.
  • the accumulator module 840 processes 1150 a plurality of transactions to generate accumulation data.
  • a directory table stores records relating to the directories
  • a file table stores records relating to the files
  • an hourly frequency table stores data related to frequency.
  • processing the namespace update log 910 includes creating a new record in the file table for a file create operation, changing a file name field in the file table for a file rename operations, a removing a record in the file table for a file delete operation. Similar processing can be performed on records in the directory table for directory operations.
  • processing the data update log 920 includes determining if the next record has a matching file system ID, path name, and file name, and if so, no further processing is necessary. Otherwise, or if this is the last transaction record, a last write time is updated in a record having a matching system ID, pathname, and file name.
  • Processing the access log 930 includes creating a record in an hourly frequency table for each transaction record. If successful, a value in a count field is set to 1 and a value in a bytes read field is set to a value of a bytes read field in the transaction record. A new record is not created if one already exists. Instead, the existing record is retrieved and the value in the count field is incremented by 1. Additionally, the value in the bytes read field is incremented by the value of the bytes read field in the transaction record. The record is written back to the hourly frequency table.
  • FIG. 12 is a flow chart illustrating a method 1200 for applying a policy according to one embodiment of the present invention.
  • the scheduling module 1010 of the policy module 320 can set 1210 intervals to schedule invocation of the policy.
  • One or more policies can be scheduled for execution concurrently, or at different times.
  • the policy execution module 1020 executes 1220 rules according to the policy.
  • the policy can be a list of fields and values used to set parameters. For example, in a selective migration policy, files that have not been modified or accessed in the last week can be moved to a different server. Many other policies are possible.
  • the query module 1030 sends 1230 queries, based on the rules, to the records repository 340.
  • the queries can be formed using, for example, SQL commands.
  • a list of records can be sent back to the policy execution module 1020 which in turn, sends the list to the migration engine 330 for migration.
  • Example computing devices include enterprise servers, application servers, workstations, personal computers, network computers, network appliances, personal digital assistants, game consoles, televisions, set-top boxes, premises automation equipment, point-of-sale terminals, automobiles, and personal communications devices.
  • the program instructions can be distributed on a computer readable medium, storage volume, or the Internet. Program instructions can be in any appropriate form, such as source code, object code, or scripts.

Abstract

A system and method for performing policy-based storage management using data related to access frequency and file attribute accumulation (Figure 1). A switch device provides transparency for transactions between a client and a storage network (Figure 1). The transparency allows objects to be moved on the storage network without affecting a reference to the object used by the client (Figure 1). A monitoring module generates accumulation data associated with the transactions for use in policy-based management (Figure 1). The accumulation data can describe uses of the file such as how often certain files are accessed, modifications to files such as creations of new directories or files, and other uses (Figure 1).

Description

ACCUMULATING ACCESS FREQUENCY AND FILE ATTRIBUTES FOR SUPPORTING POLICY BASED STORAGE MANAGEMENT
Inventors: Panagiotis Tsirigotis, Geetha Srikantan, Thomas K. Wong, Chi Ming Wong, Anand Iyengar, Rajeev Chawla, Richard Simpkins,
Zuwei Liu, Gaurav Gupta, Matthew Seitz and Yu Cheong Chan
CROSS-REFERENCE TO RELATED APPLICATIONS
[0001] This application claims priority under 35 U.S.C. § 119(e) to U.S. Provisional Patent Application No. 60/722,465, filed on September 30, 2005, entitled "AN ACCESS FREQUENCY / FILE ATTRIBUTES ACCUMULATOR FOR SUPPORTING POLICY BASED STORAGE MANAGEMENT," by Wong et al.; this application claims priority under 35 U.S.C. § 120 as a continuation-in-part to U.S. Patent Application No. 11/241,554, filed on September 30,2005, entitled "STORAGE POLICY MONITORING FOR A STORAGE NETWORK," by Wong et al; and this application is related to U.S. Patent Application No.10/831,376, filed on April 23, 2004, entitled "TRANSPARENT FILE MIGRATION USING NAMESPACE REPLICATION," by Thomas K. Wong et al, and to U.S. Patent Application No.10/831,701, filed on April 23, 2004 entitled "TRANSPARENT FILE REPLICATION USING NAMESPACE REPLICATION," by Thomas K. Wong et al, each of which applications are herein incorporated by reference in their entirety.
BACKGROUND OF THE INVENTION
FIELD OF THE INVENTION
[0002] This invention relates generally to storage networks and, more specifically, to performing policy-based storage management using data related to access frequency and file attribute accumulation.
DESCRIPTION OF RELATED ART
[0003] hi a computer network, NAS (Network Attached Storage) file servers connected directly to the network provide an inexpensive and easily configurable solution for a storage network. These NAS file servers are self-sufficient because they contain file systems that allow interoperability with clients running any operating system and communication using open protocols. For example, a Unix-based client can use the NFS (Network File System) protocol by Sun Microsystems, Inc. of Santa Clara, California and a Windows-based client can use CIFS (Common Internet File System) by Microsoft Corp. of Redmond, Washington to access files on a NAS file server. However, the operating system does not affect communication between the client and file server. Thus, NAS file servers provide true universal file access.
[0004] By contrast, more expensive and powerful SAN (Storage Area Network) file servers use resources connected by Fibre Channel on a back-end, or a dedicated network. A SAN file system is part of the operating system or an application running on the client. But heterogeneous client operating systems may require additional copies of each file to be stored on the storage network to ensure compatibility on the SAN file server. Additionally, communication between file servers on a SAN use proprietary protocols and thus are typically provided by a common vendor. As a result, NAS file servers are preferred when price and ease of use are major considerations. However, the benefits of NAS storage networks over SAN storage networks also have drawbacks.
[0005] One drawback with NAS file servers is that there is no centralized control.
Accordingly, each client must maintain communication channels between each of the NFS file servers separately. When NAS file servers are either added or removed from the storage network, each client must mount or unmount directories for the associated storage resources as appropriate. This is particularly inefficient when there are changes in hardware, but not in the particular files available on the network, such as when a failing NAS file server is swapped out for an identically configured back-up NAS file server. [0006] A related drawback is that a client must be reconfigured each time a file is relocated within the storage network, such as during file migration or file replication. The client generates a NAS file handle that identifies a physical location of the directory or file object on the file server. To access the object, the client sends an object access request directly to the NAS file server. When the file is relocated to a different NAS file server, subsequent requests for access to the file require a new look-up to locate the file and generate a new NAS file handle.
[0007] An additional drawback is that NAS file servers can become consumed with handling I/O (Input/Output) requests associated with file manipulations and accesses. As a result, additional processing tasks such as queries can unduly burden the NAS file servers. The file server typically walks a tree-structured directory tree in search of information requested by the query, and if there is more than one file system, each file system is individually walked. Consequentially, the file server may either become less responsive to I/O requests or have high latency in responding to the query. In some contexts, high latency will make the results stale.
[0008] Furthermore, NAS file servers can become unorganized and inefficient by, for example, storing critical data with other non-critical data. For example, large multimedia collections of MP3s used for leisure by employees can increase latency time in receiving information more germane to the productivity of an enterprise such as financial records. In another example, rarely accessed files may be stored on a premium, high bandwidth file server while often accessed files may be stored on a commodity, lower bandwidth server.
[0009] Therefore, what is needed is a network device to provide selectively migrate objects between file servers on a storage network. Furthermore, there is a need for identifying files to be migrated without burdening the file servers in, for example, servicing I/O requests.
SUMMARY
[0010] A system and method for performing policy-based storage management using data related to access frequency and file attribute accumulation. In one embodiment, a switch device provides transparency for transactions between a client and a storage network. The transparency allows objects (e.g., files or directories) to be moved (e.g., migrated) on the storage network without affecting a reference to the object used by the client (e.g., a file handle). In one embodiment of the switch device, a monitoring module generates accumulation data associated with the transactions for use in policy-based management. The accumulation data can describe uses of the file such as how often certain files are accessed, modifications to files such as creations of new directories or files, and other uses.
[0011] One embodiment of the monitoring module comprises an I/O monitoring module to receive transaction requests processed by the switch device, and determine if the underlying transaction concerns a file system being monitored. A transaction categorization module can determine whether the transaction is a type being monitored, and if so, forward to an appropriate log in a transaction logs module for storage. The transaction log can be, for example, a cyclic buffer. An accumulator module processes logs on a periodic basis and stores its results in a records repository.
[0012] In another embodiment, a policy module manages the records repository in accordance with predetermined rules. For example, a network administrator can set the rules from a user interface. Also, the rules can be pre-loaded. The policy module can include a scheduling module to invoke the policy at certain intervals, a policy execution module to form queries from the rules, and a query module to manage the queries to the records repository. The queries can be, for example, SQL queries.
BRIEF DESCRIPTION OF THE DRAWINGS
[0013] FIG. 1 is a high-level block diagram illustrating a storage network system according to one embodiment of the present invention.
[0014] FIG. 2 is a block diagram illustrating the NAS switch according to one embodiment of the present invention.
[0015] FIG. 3 is a block diagram illustrating a selective migration module according to one embodiment of the present invention.
[0016] FIG. 4 is a high-level flow chart illustrating a method for providing transparent file migration in a NAS storage network according to one embodiment of the present invention.
[0017] FIG. 5 is a flow chart illustrating a method of associating original NAS file handles with switch file handles according to one embodiment of the present invention.
[0018] FIG. 6 is a flow chart illustrating a method for providing selective migration in accordance with a policy according to one embodiment of the present invention. [0019] FIG. 7 is a flow chart illustrating a method for providing selective migration in accordance with a policy according to another embodiment of the present invention.
[0020] FIG. 8 is a block diagram illustrating a monitoring module according to one embodiment of the present invention.
[0021] FIG. 9 is a block diagram illustrating transaction logs according to one embodiment of the present invention. [0022] FIG. 10 is a block diagram illustrating a policy module according to one embodiment of the present invention.
[0023] FIG. 11 is a flow chart illustrating a method for logging transactions according to one embodiment of the present invention. [0024] FIG. 12 is a flow chart illustrating a method for applying a policy to logged transactions according to one embodiment of the present invention.
DETAILED DESCRIPTION OF THE INVENTION
[0025] A system and method for performing policy-based storage management using data related to access frequency and file attribute accumulation. A policy can include rules that establish which objects are migrated from a source file server to a destination file server. For example, large multimedia files that consume bandwidth on expensive or critical file servers, without adding much value to enterprise productivity, can be migrated to a commodity or less critical file server. In one embodiment, a NAS (Network Attached Storage) switch in the data path of a client and NAS file servers on the storage network, implements selective migration using file handles that are independent of locations on the storage network. Mapping between the file handles which are static and locations which can be dynamic, allows the client to access objects after migration using file handle that persists through the migration. [0026] FIG. 1 is a high-level block diagram illustrating a storage network system 100 according to one embodiment of the present invention. The system 100 comprises a NAS switch 110 (or more generically, a switch device) and a client 140 coupled to a network 195. The NAS switch 110, a source file server 120, and a destination file server 130, are each coupled in communication through a sub-network 196. Note that there can be various configurations of the system 100, such as embodiments including additional clients 140, additional source and/or destination file servers 120, 130, and additional NAS switches 110. The system 100 components are implemented in, for example, a personal computer with an x86-type processor executing an operating system and/or an application program, a workstation, a specialized NAS device with an optimized operating system and/or application program, a modified server blade, etc. In one embodiment, the storage network 175 comprises a NAS using protocols such as NFS and CIFS. In another embodiment, the storage network 175 comprises a combination of NAS, SAN, and other types of storage networks. In yet another embodiment the storage network 175 comprises a decentralized standard or proprietary storage system other than NAS.
[0027] The NAS switch 110 selectively migrates objects from a location on the source file server 120 to a location on the destination file server 130. Selective migration can determine which objects to migrate based on file attributes such as file type, file size, file access frequency, other file conditions, schedules, and the like, as determined by a policy. The policy can include rules that delineate certain actions in accordance with certain file attributes or conditions. In one embodiment, the NAS switch 110 can perform a rehearsal that shows the effects of a policy in a report. The policy can be iteratively adjusted to reach desired results.
[0028] The NAS switch 110 provides continuous transparency to the client 140 with respect to object management. Specifically, the NAS switch 110 can off-load tasks related to physical configurations, object management, object migration, object replication, efficient storage and/or other services on the storage network 175. Preferably, the NAS switch 110 emulates file server processes to the client 140 and emulates client processes to the file servers 120, 130. Accordingly, the client 140 is unaware of the NAS switch 110 since the NAS switch 110 is able to redirect NAS requests intended for the source file server 120 to appropriate locations on the destination file server 130. Thus, the client 140 submits object requests, such as file writes and directory reads, directly to the NAS switch 110. Likewise, the file servers 120, 130 are unaware of the NAS switch 110 since the NAS switch 110 is able to resubmit requests, contained in server file handles, as if they originated from the client 140. To do so, the NAS switch 110 can use mapping, translating, bridging, packet forwarding, other network interface functionality, and other control processes to perform file handle switching, thereby relieving the client 140 of the need to track changes in a file's physical location.
[0029] In one embodiment, the NAS switch 110 receives exported file system directories from the file servers 120, 130 containing NAS switch handles. To create compatibility between the client 140 and the NAS switch 110, the NAS switch 110 maps the file system directories to internal switch file systems which it sends to the client 140. To request an object, the client 140 traverses an exported switch file system and selects a switch file handle which it sends to the NAS switch 110 along with a requested operation.
[0030] In general, NAS file handles uniquely identify objects, such as a directory file server, on the file servers 120, 130, such as a directory or file, as long as that object exists. NAS file handles are file server specific, and are valid only to the file servers 120, 130 that issued the file handles. The process of obtaining a file handle from a file name is called a look-up. The NAS file handle may be formatted according to protocols such as NFS or CIFS as discussed in further detail below, e.g., with reference to Tables IA and IB. By contrast, a switch file handle identifies a directory or file object independent of location, making it persistent through file replications, migrations, and other data transfers. The switch file handle can be a modified NAS file handle that refers to an internal system within the NAS switch 110 rather than the source file server 120. This enables the NAS switch 110 in mapping persistent file handles to a choice of alternative NAS file handles. An original NAS file handle refers to an initial object location on the source file server 120. A stored NAS file handle refers to a NAS file handle, stored as an object on the file servers 120, 130, which points to an alternative file location.
[0031] Object transaction requests handled by the NAS switch 110 include, for example, directory and/or file reads, writes, creation, deletion, moving, and copying. A namespace access refers to an operation accessing or modifying the namespace such as lookup, rename, delete, or create. A file access refers to an operation accessing or modifying files such as read or write. An object can refer to a directory object or a file object. Directory objects can further comprise sub-directories and file objects within directory. As used herein, various terms are used synonymously to refer to a location of an object prior to migration (e.g., "primary"; "source"; "original"; and "first") and various terms are used to refer to a location of the same object after migration (e.g., "replica"; "destination"; "substitute"; and "second"). Further embodiments of the NAS switch 110 and methods operating therein are described below. [0032] In one embodiment, the NAS switch 110 can be configured to monitor the transaction requests to accumulate frequency and file attribute information as described in more detail below. The accumulated information can be used for various applications allowing policy enforcement. One type of policy enforced on the storage network 175 can be selective migration. For example, a selective migration application can optimize the storage network 175 by moving more frequently used files to high performance servers, and by moving less frequently used files to commodity servers. Another type of policy enforcement can be file clean up. For example, all non-essential files older than 2 year are deleted. Many other implementation-specific policies are possible given the disclosure herein. [0033] The NAS switch 110 can be configured for other types of applications such as indexing and categorization. An indexing application can build an index database for fast content-based searching. Recent transactions which modify the storage network 175 can be recorded for updating the index database without the need for walking the file systems. A categorization application can maintain statistics on the types of files being stored. The list allows files of the same type to be quickly identified without walking the file systems. As a result, in one example, non-essential files such as MP3s can be quickly identified and migrated to commodity servers.
[0034] In a typical transaction, the client 140 accesses resources on the file servers 120, 130 by submitting a switch file handle to the NAS switch 110, intended for the source file server 120. To find the switch handle, the client 140 first mounts an exported switch file system containing switch file handles. The client 140 looks-up an object to obtain its file handle and submits an associated request. From the perspective of the client 140, transactions are carried out by the file servers 120, 130 having object locations that do not change. Thus, the client 140 interacts with the NAS switch 110 before and after a file replication in the same manner. A user of the client 140 can submit operations through a command line interface, a windows environment, a software application, or otherwise. In one embodiment, the NAS switch 110 further provides access to a storage network 175 other than a NAS storage network. [0035] The source file server 120 is the default or original network file server for the client 140 before file migration. The source file server 120 further comprises source objects 125, which include namespace directories and files such as enterprise data, records, database information, applications, and the like. The source file server 120 can store a table of migrated directories maintained by the NAS switch 110 that correlate results from namespace migration. Moreover, the source file server 120 can store a file handle migration table, maintained by the NAS switch 110, denoting each migrated directory and file object. The source file server 120 comprises, for example, a personal computer using an x86-type processor with an operating system and/or an application, a workstation, a specialized NAS device with an optimized operating system and/or application, a modified server blade, etc. [0036] The destination file server 130 becomes the primary network file server used by the NAS switch 110 after file migration. The destination file server 130 further comprises destination objects 135, which include the replicated namespace directories and source files. The destination file server 130 can comprise the same hardware and/or software as described with reference to the source file server 120. The source and destination file servers 120, 130 are preferably NAS file server, but can also be file servers using other decentralized protocols that do not inherently support file migration. Further embodiments of the source and destination file servers 120, 130 and related methods are described below. [0037] The network 195 facilitates data transfers between connected hosts (e.g., 110,
140). The connections to the network 195 may be wired and/or wireless, packet and/or circuit switched, and use network protocols such as TCP/IP (Transmission Control Protocol/Internet Protocol), IEEE (Institute of Electrical and Electronics Engineers) 802.11, IEEE 802.3 (i.e., Ethernet), ATM (Asynchronous Transfer Mode), or the like. The network, 195 comprises, for example, a LAN (Local Area Network), WAN (Wide Area Network), the Internet, and the like. In one embodiment, the NAS switch 110 acts as a gateway between the client 140, connected to the Internet, and the directory file server 120, and the shadow file servers 130, connected to a LAN. The sub-network 196 is preferably a local area network providing optimal response time to the NAS switch 110. hi one embodiment, the sub- network 196 is integrated into the network 195.
[0038] FIG. 2 is a block diagram illustrating the NAS switch 110 according to one embodiment of the present invention. The NAS switch 110 comprises a mapping module 210 and a selective migration module 220. Generally, the mapping module 210 handles file handles I/O transactions associated with objects, and sends information about successful FO transactions to the selective migration module 220 for tracking updates to since the last walk of the tree-structured file servers. Note that modules are merely exemplary groupings of functionality.
[0039] Prior to file migration, the mapping module 210 receives a switch file handle with a request from the client 140 which it uses to find an original NAS file handle. The mapping module 210 submits the original NAS file handle with the request to the source file server 120. If the object has yet to change locations in the storage network 175, the mapping module 210 uses the original NAS file handle. The mapping module 210 can submit the switch file handle to the migration module 220 to determine if the object is part of a data migration. Also, the file server interface 220 can submit the switch file handle to the redirection module 230 to determine if the object has completed data migration, hi either case, an appropriate NAS file handle is returned for the mapping module 210 to use in forwarding the client request to the appropriate file server 120, 130 [0040] After file migration, the mapping module 210 looks-up switch file handles received from the client 140 in the file handle migration table. If an object has been migrated, the redirection module outputs a destination NAS file handle corresponding to a location on the destination file server 130. [0041] The selective migration module 220 receives information about successfully
I/O transactions from the mapping module 210. In other embodiments, the selective migration module 220 can intercept transactions headed for the mapping module 210 before processing, or receive a duplicate of transactions sent to the mapping module 210. Upon executing a policy to migrate objects, the selective migration module 220 can update file locations in the mapping module 210.
[0042] FIG. 3 is a block diagram illustrating a selective migration module 220 according to one embodiment of the present invention. The selective migration module 220 comprises a monitoring module 310, a policy module 320, a migration engine 330, and a records repository 340. Generally, the records repository is populated using a traversal (walk) of tree-structured directories or shares, and records are updated with information about transactions with objects in the directories. The policy module 320 can perform policy- based migration through query or sorting interactions with the records repository 340.
[0043] The monitoring module 310, as shown in more detail with respect to FIG. 8, comprises an I/O monitoring module 810, a transaction categorizing module 820. transaction logs 830, and an accumulator module 840. Generally, the monitoring module 310 receives information relating to successful I/O transactions involving objects, and sends processed information to the records repository 340 for storage. More specifically, the I/O monitoring module 810 determines whether the transaction is associated with a file system being monitored. The transaction categorizing module 820 receives transactions from the I/O monitoring module 810 to determine whether the transaction is a type of transaction being monitored. The transaction logs 830 stored monitored transactions as transaction records. At certain intervals, the accumulator module 840 processes transaction records for storage in the records repository 340.
[0044] An embodiment of the transaction logs 830 is shown in FIG. 9. The transaction logs 830 comprise records for specific requests: a namespace update log 910, a data update log 920, and an access log 930; and records for general requests in a general request log 940. The logs 910-940 can be part of a memory device such as a RAM (random access memory) device or other type of memory element. The logs 910-940 can operate as cyclic buffers such as a FIFO (first in first out) type of buffer that replaces the oldest record with the newest record.
[0045] The policy module 320, as shown in more detail in FIG. 10, comprises a scheduling module 1010, a policy execution module 1020, and a query module 1030. In one embodiment, the scheduling module 1010 initiates a process at intervals (e.g., as programmed by a system administrator). The policy execution module 1020 accesses rules that form a policy, and feeds specific commands to the query engine 800. The rules can depend upon the application. The rules can be preconfigured or created by a network administrator. The rules can be Boolean combinations of conditions. For example, FILE TYPE IS MPEG and FILE SIZE IS MORE THAN 100 MEGABYTES. The query module 1030 manages the application of specific commands for application against data stored in the records repository 340 (e.g., accumulation and frequency data).
[0046] The migration engine 330 can migrate file identified by the policy module 320. For example, each of the files that have not been accessed in the last year can be moved to a dedicated filer server for rarely accessed files. In one embodiment, the migration engine 320 migrates the namespace associated with each object prior to migrating data associated with each object.
[0047] The records repository 340 can store records associated with objects in the tree-structured directories by traversing the tree-structured directories. In response to receiving logs from the monitoring module 310, the records repository 340 can update records. Periodically, the records repository 340 can synch with the directories with traversals.
[0048] FIG. 6 is a flow chart illustrating a method 600 for providing selective migration in accordance with a policy according to one embodiment of the present invention.
A NAS file handle including a location of the object on a source file server is received 610.
A switch file handle that is independent of the location of the object is sent as representative of the object 620. The object is selectively migrated 630 from the location on the source file server to a location on a destination file server in accordance with a policy that includes rules for storing objects in the storage network. An access request associated with the switch file handle is received 640. The access request is sent 650 to the location on the destination file server. [0049] The information can be recorded one or more logs, as discussed below with respect to FIG. 9. For example, an access log can be updated each time a file or directory is read. Logs can be maintained based on the importance of tracked transactions. For example, while all file or directory creations are tracked, in one embodiment, only the last file modification is tracked. In another optimization, an access log can count accesses over an hour without recording the time of each access. The monitoring module 310 periodically sends the logs to the records repository 340 for processing (e.g., once an hour).
[0050] The policy module 320 can implement rules with searches on the records repository 340 to identify files meeting the conditions, hi one embodiment, a user interface (e.g., viewed in a web browser) can allow a network administrator to configure rules. The policy module 320 can be triggered periodically on a per-policy basis such as once a day or once a week.
[0051] FIG. 7 is a flow chart illustrating a method 700 for providing selective migration in accordance with a policy according to another embodiment of the present invention. A policy that includes rules for storing objects in the storage network is received 710. Objects stored in a tree-structure on a source file server in the storage network are traversed 720. A policy database (or records repository) is populated 730 with records representative of the tree-structure. The policy database is updated 740 to reflect transactions associated with the objects. The rules are executed 750 against records in the policy database to determine which objects are to be migrated, as described below with respect to FIG. 12. The objects are migrated 760 from a location on the source file server to a location on a destination file server.
[0052] FIG. 4 is a flow chart illustrating a method 400 for migrating objects according to one embodiment of the present invention. Prior to migration, the NAS switch 110 associates 410 original NAS file handles with switch file handles 175. As detailed in FIG. 5, the exported file system 510 containing the original file handles is mounted from the source file server. The switch file handle is generated 520 independent of object locations. The switch file system containing the switch file handles is then exported 50 to the client 140. Referring again to FIG. 4, migration from the source file server 110 to the destination file server 120 is performed 420 using namespace replication. After migration, NAS requests (i.e., transaction requests) are redirected by the mapping module 210 according to the new locations. The method 400 is described in further detail in U.S. Patent No. [attorney to insert migration patent no.]. [0053] FIG. 11 is a flow chart illustrating a method 1100 for generating accumulation data according to one embodiment of the present invention. The mapping module 210 of the NAS switch 110 processes 1110 a transaction request for an object in the storage network 175. In one embodiment, once the transaction request is serviced and a response is being sent back to the client 140, a copy of the transaction request is sent to the monitoring module 310. The NAS switch 110 may service the transaction request prior to copying to ensure its validity. For example, if a file is requested that does not exist, the transaction request would lead to invalid accumulation data.
[0054] Once received, the I/O monitoring module 810 determines 1120 whether the transaction request concerns a monitored file system. A network administrator can use an interface to select which file systems are monitored. If the file system has not been selected, no further processing is necessary. The transaction categorizing module 820 determines 1130 whether the transaction request is a type of transaction being monitored. Monitored transactions can include, for example, FDNOPS, FDAOPS, or FDUOPS. More specifically, FDNOPS include updates to a namespace of a file system such as create, rename, delete, create directory, rename directory, or remove directory. FDAOPS include updates to data in a file system such as a write operations, or a change or set file size operation. FDUOPS include accesses to a file system such as a read operation. Some embodiments include additional types of operations such as open or close operations. [0055] The transaction logs 830 log 1140 the transaction request according to its type.
For example, the transaction categorization module 820 can set a code for the type of transaction which directs it to a specific log. In one embodiment, FDNOPS are stored in the namespace update log 910, FDUOPS are stored in the data update log 920, FDAOPS are stored in the access log 920, and others are stored in the general request log 940. [0056] At certain intervals (e.g., when a cyclic buffer storing a log is full), the accumulator module 840 processes 1150 a plurality of transactions to generate accumulation data. A directory table stores records relating to the directories, a file table stores records relating to the files, and an hourly frequency table stores data related to frequency. In one embodiment, processing the namespace update log 910 includes creating a new record in the file table for a file create operation, changing a file name field in the file table for a file rename operations, a removing a record in the file table for a file delete operation. Similar processing can be performed on records in the directory table for directory operations. [0057] In one embodiment, processing the data update log 920 includes determining if the next record has a matching file system ID, path name, and file name, and if so, no further processing is necessary. Otherwise, or if this is the last transaction record, a last write time is updated in a record having a matching system ID, pathname, and file name. [0058] Processing the access log 930 includes creating a record in an hourly frequency table for each transaction record. If successful, a value in a count field is set to 1 and a value in a bytes read field is set to a value of a bytes read field in the transaction record. A new record is not created if one already exists. Instead, the existing record is retrieved and the value in the count field is incremented by 1. Additionally, the value in the bytes read field is incremented by the value of the bytes read field in the transaction record. The record is written back to the hourly frequency table.
[0059] FIG. 12 is a flow chart illustrating a method 1200 for applying a policy according to one embodiment of the present invention. The scheduling module 1010 of the policy module 320 can set 1210 intervals to schedule invocation of the policy. One or more policies can be scheduled for execution concurrently, or at different times. The policy execution module 1020 executes 1220 rules according to the policy. The policy can be a list of fields and values used to set parameters. For example, in a selective migration policy, files that have not been modified or accessed in the last week can be moved to a different server. Many other policies are possible. The query module 1030 sends 1230 queries, based on the rules, to the records repository 340. The queries can be formed using, for example, SQL commands. In the selected migration example, a list of records can be sent back to the policy execution module 1020 which in turn, sends the list to the migration engine 330 for migration.
[0060] The accompanying description is for the purpose of providing a thorough explanation with numerous specific details. Of course, the field of storage networking is such that many different variations of the illustrated and described features of the invention are possible. Those skilled in the art will thus undoubtedly appreciate that the invention can be practiced without some specific details described below, and indeed will see that many other variations and embodiments of the invention can be practiced while still satisfying its teachings and spirit. For example, although the present invention is described with reference to storage networks operating under the NAS protocol, it can similarly be embodied in future protocols for decentralized storage networks other than NAS, or in mixed protocol networks. Accordingly, the present invention should not be understood as being limited to the specific implementations described below, but only by the claims that follow.
[0061] The processes, features, or functions of the present invention can be implemented by program instructions that execute in an appropriate computing device. Example computing devices include enterprise servers, application servers, workstations, personal computers, network computers, network appliances, personal digital assistants, game consoles, televisions, set-top boxes, premises automation equipment, point-of-sale terminals, automobiles, and personal communications devices. The program instructions can be distributed on a computer readable medium, storage volume, or the Internet. Program instructions can be in any appropriate form, such as source code, object code, or scripts.

Claims

WE CLAIM:
1. In a switch device coupled between a client and a storage network to process transaction requests submitted by the client, a method for monitoring transactions related to objects in a storage network, comprising: receiving a transaction request concerning an object in the storage network; determining that the transaction request is directed to a file system on the storage monitor that is being monitored and that the transaction is a type of transaction request being monitored; logging the transaction according to the type of transaction; periodically processing a plurality of logged transactions to generate accumulation data associated with an object, the accumulation data describing a use of an object; and applying a policy to the objects based on the accumulation data.
2. The method of claim 1, further comprises: categorizing the transaction; and logging the transaction according to the categorization.
3. The method of claim 1 , wherein periodically processing the plurality of transactions comprises: incrementing a count associated with a record for each transaction concerning the record.
4. The method of claim 1 , wherein periodically processing the plurality of transactions comprises: incrementing a type count associated with a record for each type transaction concerning the record.
5. The method of claim 1 , wherein periodically processing the plurality of transactions comprises: incrementing a number of bytes to reflect a total number of bytes for each transaction concerning an object.
6. The method of claim 1 , wherein applying the policy comprises : mapping rules of the policy to commands; and querying the objects according to the commands
7. The method of claim 1 , wherein logging the transaction comprises logging the transaction in a .cyclic buffer, and wherein periodically processing the logged transactions comprises periodically processing the logged transactions responsive to the cyclic buffer completing cycle.
8. The method of claim 1, wherein the transaction request comprises a NAS file handle.
9. The method of claim 1 , wherein the transaction request comprises an update to a namespace of a file system on the storage network.
10. The method of claim 1, wherein the transaction request comprises an update to data of a file on the storage network.
11. The method of claim 1 , wherein the transaction request comprises an access to one of a directory or file on the storage network.
12. The method of claim 1, further comprising: processing the transaction request by sending a command to the storage network and forwarding results of the command to the client.
13. A switch device coupled between a client and a storage network to process transaction requests submitted by the client, comprising: an VO monitoring module to receive a transaction request concerning an object in the storage network, and determine whether the transaction request is directed to a file system on the storage monitor that is being monitored; a transaction categorizing module to determine whether the transaction is a type of transaction request being monitored; transaction logs to log the transaction according to the type of transaction; an accumulator module to periodically process a plurality of logged transactions to generate accumulation data associated with an object, the accumulation data describing a use of an object; and a policy module to apply a policy to the objects based on the accumulation data.
14. The switch device of claim 13 , wherein the category transaction module categorizes the transaction, and the transaction logs log the transaction according to the categorization.
15. The switch device of claim 13, wherein the accumulator module increments a count associated with a record for each transaction concerning the record.
16. The switch device of claim 13 , wherein the accumulator module increments a type count associated with a record for each type transaction concerning the record.
17. The switch device of claim 13 , wherein the accumulator module periodically increments a number of bytes to reflect a total number of bytes for each transaction concerning an object.
18. The switch device of claim 13 , further comprising a policy module to map rules of the policy to commands, and query the objects according to the commands
19. The switch device of claim 13, wherein the transaction logs store the transaction in a cyclic buffer, and wherein the accumulator module periodically processing the logged transactions responsive to the cyclic buffer completing cycle.
20. The switch device of claim 13, wherein the transaction request comprises a NAS file handle.
21. The switch device of claim 13, wherein the transaction request comprises an update to a namespace of a file system on the storage network.
22. The switch device of claim 13, wherein the transaction request comprises an update to data of a file on the storage network.
23. The switch device of claim 13 , wherein the transaction request comprises an access to one of a directory or file on the storage network.
24. The switch device of claim 13 , further comprising a mapping module to process the transaction request by sending a command to the storage network and forwarding results of the command to the client.
EP06815981A 2005-09-30 2006-10-02 Accumulating access frequency and file attributes for supporting policy based storage management Ceased EP1934838A4 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US72246505P 2005-09-30 2005-09-30
PCT/US2006/038361 WO2007041456A2 (en) 2005-09-30 2006-10-02 Accumulating access frequency and file attributes for supporting policy based storage management

Publications (2)

Publication Number Publication Date
EP1934838A2 true EP1934838A2 (en) 2008-06-25
EP1934838A4 EP1934838A4 (en) 2010-07-07

Family

ID=37906783

Family Applications (1)

Application Number Title Priority Date Filing Date
EP06815981A Ceased EP1934838A4 (en) 2005-09-30 2006-10-02 Accumulating access frequency and file attributes for supporting policy based storage management

Country Status (3)

Country Link
EP (1) EP1934838A4 (en)
CN (1) CN101263494B (en)
WO (1) WO2007041456A2 (en)

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7346664B2 (en) 2003-04-24 2008-03-18 Neopath Networks, Inc. Transparent file migration using namespace replication
US7587422B2 (en) 2003-04-24 2009-09-08 Neopath Networks, Inc. Transparent file replication using namespace replication
US7831641B2 (en) 2003-04-24 2010-11-09 Neopath Networks, Inc. Large file support for a network file server
US8539081B2 (en) 2003-09-15 2013-09-17 Neopath Networks, Inc. Enabling proxy services using referral mechanisms
US7720796B2 (en) 2004-04-23 2010-05-18 Neopath Networks, Inc. Directory and file mirroring for migration, snapshot, and replication
US8195627B2 (en) 2004-04-23 2012-06-05 Neopath Networks, Inc. Storage policy monitoring for a storage network
US8190741B2 (en) 2004-04-23 2012-05-29 Neopath Networks, Inc. Customizing a namespace in a decentralized storage environment
WO2007002855A2 (en) 2005-06-29 2007-01-04 Neopath Networks, Inc. Parallel filesystem traversal for transparent mirroring of directories and files
CN102065283B (en) * 2010-12-23 2013-10-02 浙江宇视科技有限公司 Storage management method and device of video monitoring data
CN102215268A (en) * 2011-07-14 2011-10-12 北京飞杰信息技术有限公司 Method and device for transferring file data
CN102708158B (en) * 2012-04-23 2014-03-12 杭州梵艺科技有限公司 PostgreSQL (postgres structured query language) cloud storage filing and scheduling system
CN103198119A (en) * 2013-04-02 2013-07-10 浪潮电子信息产业股份有限公司 Method for fast searching all chained files having same repeating data deleting identification
US10692012B2 (en) * 2016-05-29 2020-06-23 Microsoft Technology Licensing, Llc Classifying transactions at network accessible storage
CN109242428A (en) * 2018-08-31 2019-01-18 邹平县供电公司 A kind of business prompting system and method based on intelligent mobile terminal
CN112115102B (en) * 2020-08-25 2021-05-14 广州锦行网络科技有限公司 Method for cleaning remote login log records
CN113721854A (en) * 2021-08-31 2021-11-30 中国建设银行股份有限公司 Data storage method and device

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2004053677A2 (en) * 2002-12-06 2004-06-24 Cisco Technology, Inc. Apparatus and method for a scalable network attach storage system
US20040133577A1 (en) * 2001-01-11 2004-07-08 Z-Force Communications, Inc. Rule based aggregation of files and transactions in a switched file system
WO2004097624A1 (en) * 2003-04-24 2004-11-11 Neopath Networks, Inc. Large file support for a network file server

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6314460B1 (en) * 1998-10-30 2001-11-06 International Business Machines Corporation Method and apparatus for analyzing a storage network based on incomplete information from multiple respective controllers
US6976060B2 (en) * 2000-12-05 2005-12-13 Agami Sytems, Inc. Symmetric shared file storage system
WO2002056181A2 (en) * 2001-01-11 2002-07-18 Force Communications Inc Z File switch and switched file system
CN100438424C (en) * 2002-12-28 2008-11-26 鸿富锦精密工业(深圳)有限公司 Network equipment state information display system and method
US7237021B2 (en) * 2003-04-04 2007-06-26 Bluearc Uk Limited Network-attached storage system, device, and method supporting multiple storage device types
US7346664B2 (en) * 2003-04-24 2008-03-18 Neopath Networks, Inc. Transparent file migration using namespace replication

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040133577A1 (en) * 2001-01-11 2004-07-08 Z-Force Communications, Inc. Rule based aggregation of files and transactions in a switched file system
WO2004053677A2 (en) * 2002-12-06 2004-06-24 Cisco Technology, Inc. Apparatus and method for a scalable network attach storage system
WO2004097624A1 (en) * 2003-04-24 2004-11-11 Neopath Networks, Inc. Large file support for a network file server

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of WO2007041456A2 *

Also Published As

Publication number Publication date
CN101263494A (en) 2008-09-10
WO2007041456A2 (en) 2007-04-12
CN101263494B (en) 2010-12-22
EP1934838A4 (en) 2010-07-07
WO2007041456A3 (en) 2007-11-15

Similar Documents

Publication Publication Date Title
US8131689B2 (en) Accumulating access frequency and file attributes for supporting policy based storage management
US8195627B2 (en) Storage policy monitoring for a storage network
EP1805665B1 (en) Storage policy monitoring for a storage network
WO2007041456A2 (en) Accumulating access frequency and file attributes for supporting policy based storage management
US11086531B2 (en) Scaling events for hosting hierarchical data structures
US7072917B2 (en) Extended storage capacity for a network file server
US7587422B2 (en) Transparent file replication using namespace replication
US7599941B2 (en) Transparent redirection and load-balancing in a storage network
US7346664B2 (en) Transparent file migration using namespace replication
US7720796B2 (en) Directory and file mirroring for migration, snapshot, and replication
US7571206B2 (en) Transparent request routing for a partitioned application service
US20220114064A1 (en) Online restore for database engines
US20050138162A1 (en) System and method for managing usage quotas
US9959286B1 (en) Method, System and apparatus for synchronizing changes in a directory service
US10579597B1 (en) Data-tiering service with multiple cold tier quality of service levels
US20140259123A1 (en) Aliasing of exported paths in a storage system
US10909143B1 (en) Shared pages for database copies
WO2004097686A1 (en) Transparent file replication using namespace replication
US20070174360A1 (en) Storage system embedding database
US11455305B1 (en) Selecting alternate portions of a query plan for processing partial results generated separate from a query engine
US11106667B1 (en) Transactional scanning of portions of a database
WO2004097571A2 (en) Extended storage capacity for a network file server
US11816073B1 (en) Asynchronously forwarding database commands
KHAN et al. Implementation of Distributed Metadata Management for Large Cluster-Based Storage Applications

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20080131

AK Designated contracting states

Kind code of ref document: A2

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LI LT LU LV MC NL PL PT RO SE SI SK TR

A4 Supplementary search report drawn up and despatched

Effective date: 20100604

DAX Request for extension of the european patent (deleted)
17Q First examination report despatched

Effective date: 20180315

REG Reference to a national code

Ref country code: DE

Ref legal event code: R003

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION HAS BEEN REFUSED

18R Application refused

Effective date: 20200218