US20070288792A1 - Storage controller redundancy using packet-based protocol to transmit buffer data over reflective memory channel - Google Patents

Storage controller redundancy using packet-based protocol to transmit buffer data over reflective memory channel Download PDF

Info

Publication number
US20070288792A1
US20070288792A1 US11/835,942 US83594207A US2007288792A1 US 20070288792 A1 US20070288792 A1 US 20070288792A1 US 83594207 A US83594207 A US 83594207A US 2007288792 A1 US2007288792 A1 US 2007288792A1
Authority
US
United States
Prior art keywords
storage controller
buffer
data
storage
reflective memory
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/835,942
Inventor
Roger Thorpe
Erasmo Brenes
Stephen O'Neil
Alec Shen
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Promise Technology Inc USA
Original Assignee
iStor Networks Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by iStor Networks Inc filed Critical iStor Networks Inc
Priority to US11/835,942 priority Critical patent/US20070288792A1/en
Publication of US20070288792A1 publication Critical patent/US20070288792A1/en
Assigned to PROMISE TECHNOLOGY, INC. reassignment PROMISE TECHNOLOGY, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ISTOR NETWORKS, INC.
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0646Horizontal data movement in storage systems, i.e. moving data in between storage devices or systems
    • G06F3/065Replication mechanisms
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/08Error detection or correction by redundancy in data representation, e.g. by using checking codes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/16Error detection or correction of the data by redundancy in hardware
    • G06F11/20Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
    • G06F11/2053Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant
    • G06F11/2056Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant by mirroring
    • G06F11/2071Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant by mirroring using a plurality of controllers
    • G06F11/2074Asynchronous techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/16Error detection or correction of the data by redundancy in hardware
    • G06F11/20Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
    • G06F11/2053Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant
    • G06F11/2056Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant by mirroring
    • G06F11/2071Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant by mirroring using a plurality of controllers
    • G06F11/2079Bidirectional techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/16Error detection or correction of the data by redundancy in hardware
    • G06F11/20Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
    • G06F11/2053Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant
    • G06F11/2089Redundant storage control functionality
    • G06F11/2092Techniques of failing over between control units
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/061Improving I/O performance
    • G06F3/0613Improving I/O performance in relation to throughput
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/067Distributed or networked storage systems, e.g. storage area networks [SAN], network attached storage [NAS]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04BTRANSMISSION
    • H04B1/00Details of transmission systems, not covered by a single one of groups H04B3/00 - H04B13/00; Details of transmission systems not characterised by the medium used for transmission
    • H04B1/74Details of transmission systems, not covered by a single one of groups H04B3/00 - H04B13/00; Details of transmission systems not characterised by the medium used for transmission for increasing reliability, e.g. using redundant or spare channels or apparatus
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L1/00Arrangements for detecting or preventing errors in the information received
    • H04L1/22Arrangements for detecting or preventing errors in the information received using redundant apparatus to increase reliability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L49/00Packet switching elements
    • H04L49/90Buffering arrangements
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99951File or database maintenance
    • Y10S707/99952Coherency, e.g. same view to multiple users
    • Y10S707/99953Recoverability

Definitions

  • the present invention relates to network-based storage systems, and more particularly, to techniques for providing storage controller redundancy within a network-based storage system.
  • a variety of network-attached and SAN (Storage Area Network) based storage systems exist for allowing data to be stored on an Ethernet or other IP based network.
  • these systems include one or more storage controllers, each of which controls and provides network-based access to a respective array of disk drives.
  • Each storage controller typically includes a buffer or cache memory that is used to temporarily store data as it is transferred between the network and that controller's disk drives. For example, incoming data packets containing I/O (input/output) write data may be maintained in a buffer of the storage controller until successfully written to the appropriate disk drives.
  • I/O input/output
  • Some storage systems implement a storage controller failover mechanism to protect against the possible failure of a storage controller.
  • two storage controllers may be paired for purposes of providing controller redundancy. When one of these paired storage controllers detects a failure by the other, the non-failing storage controller may take control of the failing controller's disk drives, allowing these disk drives to be accessed via the network while the failing storage controller is replaced.
  • one storage controller may maintain or have access to a mirrored copy of the other storage controller's cache and configuration data. This allows the non-failing storage controller to effectively pick up the workload of the failing controller where the failing controller left off. Upon replacement of the failing controller, a synchronization or “rebind” operation may be performed between the non-failing and new storage controllers to copy over the cache and configuration data needed to bring the new storage controller on line.
  • the mechanism used to provide storage controller redundancy typically adversely affects or limits the performance of the storage system.
  • the mechanism used to maintain a redundant copy of a storage controller's cache data limits the rate at which the storage controller can process network traffic and perform input/output operations.
  • the respective memories of two separate controllers are updated synchronously (in lock step); as a result, if a write operation to one of these memories cannot immediately be performed, the corresponding write operation to the other memory generally must also be postponed.
  • the present invention makes use of a bi-directional reflective memory channel between a pair of storage controllers to allow each such storage controller to maintain a mirrored copy of the other's buffer contents in real time.
  • buffer write operations that fall within a reflective memory segment of one storage controller are automatically reflected across this channel to the other storage controller for execution, and vice versa.
  • the corresponding write operations are executed asynchronously by the two controllers, such that the postponement of one write operation does not require postponement of the other.
  • the write operations are preferably transmitted across the reflective memory channel and processed according to an asynchronous reflective memory protocol that provides for error checking, acknowledgements, and retransmissions.
  • This protocol is preferably implemented entirely in automated circuitry, so that the mirrored copies are maintained without any CPU (central processing unit) intervention during error-free operation.
  • the surviving storage controller uses the mirrored copy of the failed storage controller's native buffer contents to assume control over, and provide network-based access to, the failed storage controller's disk drives. Failover arbitration and control messages are preferably passed between the two controllers over a channel that is separate from the reflective memory channel, so that these messages do not interfere with reflective memory operations.
  • each storage controller preferably includes an automated reflective memory controller (RMC) that monitors write operations to that storage controller's buffer.
  • RMC automated reflective memory controller
  • Write operations falling within the local storage controller's reflective memory segment are packetized by the local RMC for transmission to the remote RMC.
  • the remote RMC executes each reflected write operation by writing the associated block of write data to the same destination address as specified within the original write operation.
  • the original and reflected write operations are executed asynchronously, meaning that one may be completed by its respective RMC before the other.
  • Packet transfers between the RMCs occur over the reflective memory channel according to the asynchronous reflective memory protocol which, as mentioned above, is preferably implemented entirely within automated circuitry.
  • the packets are sufficiently small in the preferred embodiment to allow rapid generation of packet CRC values.
  • the rapid generation of CRC values, and the low overhead of the protocol allow the reflected data to be transferred with very low latency.
  • Data packets transmitted across the reflective memory channel are preferably error checked and acknowledged by the receiving RMC. If the receiving RMC returns a negative acknowledgement or fails to return an acknowledgement, the sending RMC preferably retries the packet transmission.
  • Each RMC also preferably supports a “rebind” mode in which buffer read operations are reflected across the channel.
  • the rebind mode allows a mirrored copy of a failed storage controller's buffer data to be copied from a surviving storage controller to a replacement storage controller without the need to write data to the buffer of the surviving storage controller. Because this data need not be read from and then written back to the surviving controller's buffer (as would be the case if only write operations were reflected), the surviving controller can continue to process ordinary write operations without risk of mirroring overwritten data (i.e., buffer coherency is not a concern). As a result, during such rebind processing, the surviving storage controller can continue to provide network-based access to its own disk drives and to those of the failed storage controller.
  • FIG. 2 illustrates how native and reflected buffer data is stored in the respective buffer memories of the storage controllers of FIG. 1 .
  • FIG. 3 illustrates a process by which buffer metadata is reflected bi-directionally between the storage controllers.
  • FIG. 4 illustrates the design of the storage controllers of FIG. 1 , and particularly the reflective memory controllers thereof, in greater detail according to one embodiment.
  • FIG. 5 illustrates the design of the fast path and system FIFOs (first-in-first-out buffers) of FIG. 4 .
  • FIG. 6 illustrates the format of a data packet used to transfer data between the storage controllers.
  • FIG. 7 illustrates the flow of data between and within a paired set of storage controllers according to one embodiment.
  • FIG. 8 illustrates components of the reflective memory controller of FIG. 4 in further detail.
  • a storage system that uses a hardware-implemented, bi-directional reflective memory channel to provide storage controller redundancy will now be described in detail.
  • FIG. 1 illustrates a storage system that includes a pair of storage controllers 30 arranged as peers in a redundant configuration.
  • Each storage controller may, for example, be in the form of a card or board that plugs into a backplane of a larger storage system.
  • the storage controllers 30 implement a bi-directional reflective memory channel and an associated protocol that allow each storage controller to maintain a real time copy of the other's native buffer and configuration data.
  • each LAN interface 34 includes one ten-gigabit Ethernet (10 GE) port and eight one-gigabit Ethernet (1 GE) ports, as depicted in FIG. 1 for one of the two storage controllers.
  • each storage controller can be configured to operate in one of two modes: a 1 GE mode in which the storage controller connects to the LAN 36 (or multiple LANs) via any number of the eight 1-GE ports (but not the 10-GE port); and a 10 GE mode in which the storage controller connects to the LAN via the 10 GE port only.
  • a 1 GE mode in which the storage controller connects to the LAN 36 (or multiple LANs) via any number of the eight 1-GE ports (but not the 10-GE port)
  • a 10 GE mode in which the storage controller connects to the LAN via the 10 GE port only.
  • each storage controller 30 is assigned to or “owns” a respective set or array of disk drives 38 or other storage devices.
  • storage controller A owns the disk drives 38 of array A
  • storage controller B owns the disk drives of array B.
  • the disk drives 38 are shown as residing within the enclosure 32 , but may alternatively reside outside the enclosure.
  • the disk drives may, for example, be serial ATA (Advanced Technology Attachment) disk drives, although SCSI, non-serial ATA, and other types of disk drives may be used.
  • Other types of storage devices may be used in place of some or all of the disk drives, including but not limited to optical drives and solid state storage devices.
  • Each storage controller 30 operates generally as a disk array controller, and may operate its respective disk array as a RAID (Redundant Array of Inexpensive Disks) array, JBOD (Just a Bunch of Disks) array, or other configuration.
  • each storage controller 30 exclusively controls the disk drives 38 it owns. As depicted by dashed lines in FIG. 1 , each storage controller is also physically connected to, and is capable of assuming control of, the other storage controller's disk drives 38 . For example, if storage controller A detects a failure by storage controller B, storage controller A may initiate a failover process in which storage controller B is taken off line, and in which storage controller A assumes control of the disk drives owned by storage controller B. A selector switch (not shown), such as a conventional drive connector multiplexer, may be connected to each disk drive 38 to enable such transfers of control. Following the transfer of control, storage controller A provides network-based access to both arrays of drives, A and B, allowing storage controller B to be replaced. As described in detail below, this failover process makes use of redundant or mirrored copies of buffer data maintained via the reflective memory channel.
  • each storage controller 30 includes a buffer memory 50 , referred to herein as a “buffer.”
  • the buffers 50 are used to store incoming packet data received from the network, and to accumulate incoming and outgoing I/O data.
  • the incoming packet data stored in the buffers 50 consists primarily of packet payloads containing I/O write data.
  • each buffer 50 is capable of receiving and storing incoming packet data from the network at a rate of ten gigabits per second (Gb/s), which corresponds to the ten Gb/s maximum transfer rate supported by each LAN interface 34 .
  • the buffers 50 are preferably implemented as random access memory (RAM), such as SDRAM (synchronous dynamic RAM).
  • RAM random access memory
  • Each buffer 50 is managed by software executed by the respective controller's CPUs to provide a storage cache.
  • the cache management software stores data in the buffer 50 in cache lines.
  • the system memories 52 also store various types of configuration and state information.
  • This information preferably includes buffer metadata descriptive of the data currently stored in the respective storage controller's buffer 50 .
  • the buffer metadata is in the form of cache line headers (CLHs) descriptive of the buffer data stored within each cache line of the software-implemented storage cache. As described below, this metadata may periodically be copied from the system memory 52 to the corresponding buffer 50 to allow each storage controller to maintain a redundant copy of its peer's buffer metadata.
  • CLHs cache line headers
  • the RMCs 60 communicate over each channel 62 A, 62 B according to XGMII (10-Gigabit Medium Independent Interface), and can thus communicate at a rate of ten gigabits per second in each direction.
  • XGMII signals generated by the RMC's are converted to XAUI (10-gigabit Attachment Unit Interface) signals for transmission over a backplane that interconnects the two storage controllers 30 .
  • the conversions between the XGMII and XAUI interfaces are performed by commercially-available devices mounted on the respective boards of the storage controllers 30 , externally to the respective ASICs (Application Specific Integrated Circuits) in which the RMCs reside.
  • the XAUI signals are converted back to XGMII for delivery to the receiving ASIC/RMC.
  • the bi-directional reflective memory channel 62 can alternatively be implemented using other interfaces, and can be implemented using an alternative medium such as a cable or (in future implementations) a wireless link.
  • the protocol used to transfer buffer data over the bi-directional channel 62 is preferably implemented entirely in hardware, allowing the reflected data to be transferred at high bit rates and with low latency.
  • the protocol is referred to herein as an “asynchronous reflective memory protocol,” as it allows the RMCs to execute corresponding write operations asynchronously to one another. This is accomplished in part by providing a mechanism through which each RMC can eventually confirm that the write operations reflected to its peer were successfully received.
  • each RMC 60 operates generally by monitoring write operations to its respective buffer 50 , and by reflecting (transmitting to the other RMC) those write operations that fall within a particular address range.
  • Each such write operation in the preferred embodiment is generally in the form of a destination buffer address and a block of contiguous write data.
  • the receiving RMC forwards the reflected write operation to the buffer interface of its own buffer 50 for execution (i.e., for storage of the write data at the specified destination buffer address), preferably after performing error checking as described below.
  • each write operation that falls within a buffer's reflective memory area or “segment” is replicated, and is executed by both storage controllers 30 .
  • each storage controller maintains a mirrored copy of the other storage controller's buffer contents.
  • the storage controllers and particularly the RMCs 60 , execute the corresponding write operations asynchronously to one another (i.e., one storage controller may execute a given write operation before the other).
  • one storage controller the “source” executes a write operation and reflects that operation to the other storage controller (the “target”), the source may continue processing subsequent write operations (and/or read operations) without waiting for the target to complete the reflected operation.
  • the asynchronous reflective memory protocol implemented by the RMCs allows the source to determine whether the reflected operation is ultimately received successfully by the target, and to re-try the operation (or invoke an appropriate error correction routine) if it is not.
  • FIG. 2 depicts the buffer contents of two storage controllers 30 that are in a “bound” state.
  • the arrows in this drawing represent buffer data that is being reflected to the other storage controller's buffer. For example, when a block of data is written to the reflective memory segment of storage controller A's buffer 50 , it is also written to the buffer 50 of storage controller B at the same destination address.
  • each storage controller thus stores within its own buffer a mirrored copy 70 of its peer's native or local buffer data 72 , at the same buffer addresses as that peer.
  • the address ranges used for storing local versus mirrored buffer data need not be contiguous, and are preferably defined using a set of configuration registers within each storage controller.
  • each RMC reflects buffer write operations to its peer RMC, receives and processes write operations reflected by that peer RMC, and implements associated error checking, acknowledgement and retransmission protocol tasks (as discussed below), without any CPU involvement or intervention.
  • this protocol provides for error checking, acknowledgements and retransmissions, yet has sufficiently low overhead to be implemented over a pair of 10 Gb/s channels without limiting the rate at which the storage controllers receive traffic from the 10 Gb/s network 36 .
  • the design may, however, alternatively be implemented using higher speed reflective memory channels 62 A, B and/or a greater number of such channels, in which case a less efficient reflective memory protocol may be used.
  • the RMCs 60 and the bi-directional reflective memory channel 62 advantageously support transfers and processing of reflected write operations at ten Gb/s in both directions. Because this transfer rate matches the maximum rate at which each storage controller 30 can receive data packets over the LAN 36 , the reflective memory system does not limit the rate at which each storage controller can receive data into its respective buffer 50 from the network 36 .
  • This high rate at which the reflective memory system operates is attributable in-part to the use of a protocol that allows the storage controllers/RMCs to execute corresponding write operations asynchronously relative to each other, and is also attributable to the automation of key portions of this protocol.
  • Each RMC 60 can be enabled or disabled by firmware running on the respective storage controller 30 . When disabled, an RMC does not reflect any buffer operations. An RMC may be placed in a disabled state when, for example, the storage controller is being used in a standalone or other non-redundant configuration. Each RMC can also preferably be placed in a special “rebind” mode in which the RMC reflects buffer read operations. As described below, the rebind mode may be used following a failover event to efficiently copy mirrored buffer data 70 from the buffer 50 of a surviving storage controller to the buffer 50 of a replacement storage controller.
  • the storage controllers 30 also preferably include respective management processing units or “MPUs” 66 that communicate with each other over a separate, relatively low bandwidth channel 68 .
  • This channel 68 is preferably used to pass error status information and resolve error conditions. Because this channel 68 is separate from the bi-directional reflective memory channel, these communications between the MPUs do not interfere with transfers of reflected operations.
  • the bi-directional reflective memory channel 62 is also preferably used by each storage controller 30 to maintain a copy of its peer's buffer metadata. This is preferably accomplished by having each storage controller execute a firmware program that copies its buffer metadata from the system memory 52 to the reflective memory segment of its buffer memory 50 (or equivalently, by writing the same buffer metadata to both the system memory 52 and the buffer 50 ). This copying operation is depicted in FIG. 3 by the thin, solid arrows, while the reflected write operations are depicted by wide arrows.
  • the buffer metadata will represent less than 1% of the reflected data.
  • the surviving storage controller may move the failing storage controller's buffer metadata from the buffer 50 to its system memory 52 . This is depicted by the dashed arrow in FIG. 3 for storage controller A (the survivor). Once copied to the survivor's system memory, this metadata may be used by the survivor to interpret and process the failed storage controller's ordinary buffer data. In other implementations, the buffer metadata may alternatively be copied or moved between the paired storage controllers 30 over a separate communication channel.
  • the storage controllers 30 can optionally be configured, via firmware, to use the reflective memory channel to reflect other types of data.
  • the reflective memory channel 62 may be viewed as providing a general purpose memory reflection “service” that may be used by firmware for various types of inter-controller communication.
  • a portion of each buffer 50 may be allocated exclusively to general-purpose firmware transfers of data.
  • the reflective memory channel 62 is used in the preferred embodiment exclusively to transfer fast path data, slow path data, cache line headers, and rebind data (as described below), the underlying architecture is not so limited.
  • Incoming buffer data from the PIE RX circuit 42 A represents incoming network traffic (packet data) that is placed into the buffer 50 without any CPU intervention.
  • This packet data referred to as “fast path data,” consists primarily of properly ordered, non-fragmented packet data received from the network 36 .
  • Out of sequence data received from the network 36 is processed by one or more CPUs of a cluster of CPUs 44 before being written to the buffer 50 .
  • This “slow path” data enters the BMI 80 via the system bus 82 after being properly re-sequenced by firmware.
  • the PIE RX circuit 42 A processes the fast path data.
  • This circuit and a counterpart PIE TX circuit (not shown), automate selected portions of the iSCSI and TCP/IP protocols within application-specific circuitry, significantly reducing the quantity of protocol-level processing that needs to been performed by software/firmware. Details of a particular implementation of the protocol intercept engine 42 , including the PIE RX circuit 42 A, are set forth in a U.S. provisional patent application filed on Feb. 14, 2003 titled “High Availability Integrated Storage Network Processing For iSCSI Communication,” the disclosure of which is hereby incorporated herein by reference.
  • the PIE RX circuit is configured to write data only to the buffer's reflective memory segment, and not to other portions of the buffer 50 .
  • the packet data written to the buffer 50 consists essentially of packet payload data, without the associated headers.
  • firmware executed by one or more of the CPUs 44 generates corresponding cache line headers (CLHs) within the system memory 52 ( FIG. 1 ).
  • CLHs cache line headers
  • These CLHs describe the data currently stored in the local buffer 50 , and are generated based on status signals from the PIE RX circuit 42 A.
  • the firmware also writes these CLH's to the reflective memory segment of the local buffer 50 (via the system bus 82 ), causing the CLHs to be copied to the remote storage controller's buffer 50 .
  • the hardware-implemented asynchronous reflective memory protocol gives priority to the CLHs over other types of reflected data, essentially ensuring that the necessary CLH data will be available to the surviving storage controller in the event of a failure.
  • the RMC 60 preferably includes three transmission FIFOs (first-in-first-out buffers) that are used to accumulate data to be reflected to the other storage controller: a fast path FIFO 86 and two system FIFOs 88 , 89 . All fast path data to be reflected passes through the fast path FIFO 86 . All other reflected data, including slow path data and CLH data, passes through one of the two system FIFOs 88 , 89 . As described below, the fast path and system FIFOs operate generally by accumulating buffer write data into packet bins for transmission across the reflective memory channel. The organization of the fast path and system FIFOs is depicted in FIG. 5 and is described separately below.
  • Two separate system FIFOs 88 , 89 are provided in the illustrated embodiment in order to accommodate differing transmission priority levels for different data types.
  • the circuitry (described below) for transmitting data over the outgoing channel 62 A gives priority to system_FIFO_ 1 over system_FIFO_ 2 (and also over the fast path FIFO 86 ).
  • the storage controller 30 uses system_FIFO_ 1 exclusively to transfer CLH data, and uses system_FIFO_ 2 to transfer slow path data and rebind data.
  • the firmware may alternatively be written to use the system FIFOs to reflect additional or other types of data. Further, as will be recognized, a greater or lesser number of FIFOs may be used to buffer the outgoing data.
  • Operation of the system FIFOs 88 , 89 is configurable by firmware via a set of address registers within a set of RMC registers 94 ( FIG. 4 ). Specifically, each system FIFO may be assigned to a respective, mutually exclusive buffer address range or “window.”
  • An address comparison (ADDR COMP) circuit 96 within the BMI's bus interface 100 monitors the addresses of incoming buffer write operations received from the system bus 82 to determine whether such operations fall within either window. Write operations falling within window 1 are passed to system_FIFO_ 1 , and write operations falling within window 2 are passed to system_FIFO_ 2 . All other buffer write operations emanating from the system bus 82 are ignored by the RMC, and thus are not reflected.
  • the firmware running on the source storage controller simply places the source's RMC in rebind mode, and then performs a sequence of “dummy” reads of this buffer data from its own buffer 50 .
  • Each block of data read from source's buffer 50 as the result of a dummy read operation is reflected across the reflective memory channel to the target's RMC, and is written to a corresponding address in the target's buffer 50 .
  • the surviving storage controller continues to provide network-based access to both its own disk drives 38 and the disk drives 38 of the failed storage controller 30 .
  • a multiplexer 104 selects between the two system FIFOs 88 , 89 and the fast path FIFO 86 , and passes the output of the selected FIFO to a transmission interface (TX I/F) 106 A for transmission across the outgoing channel 62 A.
  • the data is preferably sent across the channels 62 A, 62 B in packets according to an asynchronous reflective memory protocol which, as mentioned above, provides for error checking, packet acknowledgements, and packet retransmissions.
  • an asynchronous reflective memory protocol which, as mentioned above, provides for error checking, packet acknowledgements, and packet retransmissions.
  • the control logic and structures associated with the multiplexer 104 and the FIFOs 86 , 88 , 89 are depicted in FIG. 8 and are described separately below.
  • a receive interface (RX I/F) 106 B receives the data transmitted by the remote storage controller on the incoming channel 62 B.
  • incoming packet data from the remote storage controller 30 is initially error-checked by the receive interface 106 B, and when error free, is written to a receive FIFO 110 .
  • Data accumulated within this FIFO 110 is ultimately written to the buffer 50 by the multi-port memory controller 85 .
  • the receive interface 106 B provides packet status information to the transmit interface 106 A.
  • the transmit interface 106 A uses this information to send packet acknowledgement (ACK and NAK) messages to the remote storage controller via the outgoing channel 62 A, allowing the transmitting controller 30 to re-send packets that are not successfully received.
  • the ACKs and NAKs may be communicated between the RMCs using reserved control characters of the XGMII interface.
  • the receive interface 106 B also provides link status information to the transmit interface 106 A indicative of the state of the incoming link 62 B.
  • a block diagram of the link control and error checking logic incorporated into the transmit and receive interfaces 62 A, 62 B is provided in FIG. 7 and is described below.
  • Errors detected by the receive interface 106 B propagate to the RMC registers 94 via the lines labeled ERROR and INT in FIG. 4 , respectively, allowing these events to be monitored by firmware executed by the CPUs 44 .
  • Firmware accesses the various RMC registers 94 using a register bus 83 .
  • the transmission and receive interfaces 106 A, 106 B, as well as the LAN interface 34 shown in FIG. 1 are all XGMII ten gigabit Ethernet interfaces. Other types of interfaces, including but not limited to SPI4 and Utopia 3 , may alternatively be used. Further, as mentioned above, interfaces that exceed the transmission rate of the LAN interface 34 may be used to implement the reflective memory channel.
  • FIG. 5 illustrates the design of the transmission FIFOs 86 , 88 and 89 in further detail.
  • the RMC 60 arranges the fast path FIFO 86 into one of two possible configurations, depending upon whether the storage controller is operating in the 1 GE mode versus the 10 GE mode.
  • the fast path FIFO is organized into eight 512-byte packet bins, as depicted by the labels BIN 0 through BIN 7 on the left side of the drawing.
  • the fast path FIFO is organized into two 2048-byte packet bins, as shown on the right side of the drawing.
  • Each of the system FIFOs is statically arranged as a single packet bin.
  • system_FIFO_ 2 is arranged as a single, 512-byte bin
  • system_FIFO_ 1 (the higher priority system FIFO) is arranged as a single, 64-byte bin.
  • a total of ten bins (0-9) are provided, and when operating in the 10 GE mode, a total of 4 bins are provided (0-3).
  • the 10 GE configuration advantageously allows the RMCs to sustain the streaming of jumbo packets of 9018 bytes (8960 bytes of data, 20 bytes TCP information, 20 bytes of IP information, and 18 bytes of Ethernet information).
  • Each bin of the transmit FIFOs 86 , 88 , 89 operates generally by accumulating continuous buffer write data for eventual transmission on the outgoing channel 62 A within a single packet (i.e., each outgoing data packet contains payload data from a single bin).
  • each outgoing data packet contains payload data from a single bin.
  • one important benefit of subdividing the reflected data into bins is that it reduces the amount of time needed to generate the CRC portion of each packet.
  • the RMC selects between the ready bins so as to give highest priority to system FIFO_ 1 , intermediate priority to the fast path FIFO, and lowest priority to system_FIFO_ 2 .
  • the source storage controller 30 may therefore reflect buffer write operations across the channel 62 A out-of-order (i.e., in an order that is different from the order in which the source controller 30 performs these operations).
  • the TX FIFOs maintain one internal packet control structure 120 for each of the bins.
  • Each such control structure 120 includes the following fields: DADDR[ ], which stores the destination address currently associated with the packet bin; BIN_STATUS[ ], which stores the bin's current status (Active, Free, Sent, or NAKed); WORDS_IN_BIN[ ], which keeps track of the number of words stored in the bin, and BIN_TIMEOUT[ ], which is used to implement a timeout counter.
  • the BMI 80 receives write traffic from the PIE RX circuit 42 A and the system bus in 64-byte bursts, with each such burst being accompanied by a buffer destination address. Bursts emanating from the PIE RX circuit are processed as follows. When a burst is received, the beginning address of this burst is compared, in parallel, to the DADDR fields of all currently active bins. If a match occurs (i.e. the address matches the next sequential write location from the contents of DADDR), the write data is stored in the corresponding active bin, and the bin's WORDS_IN_BIN field is updated with the new count value. The effect of this operation is to aggregate bursts for transmission to the remote RMC.
  • the write data is stored in the next available bin, and the packet control structure fields of that bin are updated as follows: the bin's status is changed to Active, the beginning burst address is written to the DADDR field, and the WORDS_IN_BIN field is updated with the current number of words stored in the bin. If no bin is available (as may result if packet errors occur), the fast path FIFO asserts a “back off” signal (not shown) that causes the PIE RX circuit to slow down its operation.
  • each bin of the fast path FIFO has a capacity of eight 64-byte bursts when in the 1 GE mode, and has a capacity of thirty two 64-byte bursts when in the 10 GE mode.
  • the RMC sends the entire contents of the active bin across the outbound RMC channel 62 A when either (1) the current burst is the last burst within a burst sequence of contiguous write data, or (2) the bin becomes full (512 bytes in 1GE mode, or 2048 bytes in 10GE mode).
  • the RMC updates the bin's status to “Sent,” and activates the bin timeout counter. The bin's “Sent” status is maintained until the transmitting RMC receives a corresponding ACK or NAK message from the receiving RMC.
  • An important attribute of the asynchronous reflective memory protocol, in the preferred embodiment, is that the bins are selected without regard to LAN port number. Thus, when running in the 1 GE mode, all eight of the fast path bins are used even if less than all of the eight 1GE LAN ports used. The bandwidth associated with the fast path FIFO is thus effectively distributed among those 1 GE ports actually being used.
  • system_FIFO_ 1 When a burst is received from the system bus 82 ( FIG. 4 ), its destination address is checked to see if it falls within either system_FIFO_ 1 's address range or system_FIFO_ 2 's address range. If the burst falls within one of these address ranges, the corresponding system FIFO 88 , 89 processes the burst in the same manner as described above for the fast path FIFO 86 . Because system_FIFO_ 1 holds only a single 64 byte burst, it becomes ready for transmission each time it receives a burst from the system bus 82 .
  • the receipt of an ACK message by the transmitting RMC indicates that the receiving RMC received the associated packet without detection of any errors. It does not, however, indicate successful completion of the write operation specified by the packet. If an error occurs when the receiving RMC performs this write operation, the receiving RMC will report this error to its own CPUs. This error will thereafter be reported to the transmitting storage controller 30 via the separate MPU channel 68 .
  • FIG. 6 illustrates the format of the data packets used to transmit bin contents across the channels 62 A, 62 B according to one embodiment of the asynchronous reflective memory protocol.
  • the data payload portion 140 contains the contents of the bin currently being transmitted, and the destination address (DADDR) portion 142 contains the associated buffer destination address.
  • the status word 144 if present, contains the ACK/NAK status of up to four packets that have been received by the transmitting RMC. Accumulated packet ACK/NAK messages are thus piggybacked on transmitted data packets. If no ACK/NAK messages are available to transmit, the status word is replaced with an end-of-frame (EOF) character. If a status word is ready to transmit but no data packets are being transmitted, the status word may be sent within a special status packet (not shown) that does not include the other fields shown in FIG. 6 .
  • EEF end-of-frame
  • the TAG field 146 carries the bin ID of the bin being transmitted, and is used by the receiving RMC 60 to check for valid headers.
  • the TAG field is protected from corruption by having the bin ID encoded using three control characters spread across three XGMII transmission lanes, with enough redundancy to sustain a one-lane failure.
  • the CRC (cyclic redundancy code) field is generated from all of the other fields within the packet except the TAG field 146 .
  • the receiving RMC checks the packet's CRC, and incorporates the results of this CRC check (ACK or NAK) into the next status word 144 to be returned.
  • the ACK or NAK is added to the status word according to the order of the packet's arrival (i.e., the ACK/NAK characters follow the packet arrival order).
  • the TAG field is protected by having the bin ID encoded across the three transmission lanes as described above.
  • the receiving RMC can thus reliably identify a failed packet transmission and return a corresponding NAK message.
  • any received data behind the failed bin transmission is also dropped in order to maintain write-order.
  • the originating RMC keeps track of the bin IDs sent so that when a NAK is received, it can determine which bins need retransmission and which do not. If the re-transmitted packet is successfully received, an ACK is sent back to the originator, which in turn frees-up the associated bin.
  • FIG. 7 illustrates the flow of data between and within the RMCs 60 of two paired storage controllers 30 in further detail.
  • the CRC portion of each packet is generated as the packet is being transmitted.
  • a CRC-32 algorithm may be used for this purpose. Because each packet has a small payload (512 bytes or smaller when in 1 GE mode, and 2048 or smaller when in 10GE mode), the CRC portions can be generated rapidly, allowing data reflection at a high transfer rate and with low latency.
  • the receive interface 106 B of a storage controller When the receive interface 106 B of a storage controller receives a data packet, it checks the CRC. If the CRC is valid, the packet's data payload is pushed into the corresponding receive FIFO 110 , and is eventually written to the receiving storage controller's buffer 50 at the destination address specified in the packet. If the CRC is determined to be bad, the received packet is dumped and a NAK status is generated. The NAKs and ACKs resulting from the CRC checks are queued by the corresponding transmit interface 106 A for sending back to the originating RMC via a status word.
  • the receive interface 106 B When the receive interface 106 B receives a status word, it passes the status (ACK/NAK) information and associated tag(s) to the receiving RMC's transmit interface 106 A. The transmit interface 106 A in turn updates the associated packet control structures 120 ( FIG. 5 ) to reflect the ACK/NAK status. If a particular bin receives an ACK, the bin is made available by setting its status to “Free.”
  • the transmit interface 106 A checks bin's status to see if this is the first NAK, and if so, queues the bin for resending and updates status to “NAKed.” If the NAKed bin is already in the NAKed state (indicating that the packet transmission has failed twice), the transmitting RMC enters into a “link down” state in which its transmission interface generates a system interrupt, stops transmitting data across the outbound RMC channel 62 A, and starts sending XGMII Link Down Sequence Remote packets to the remote RMC. Although outgoing data reflection is halted, write operations to the buffer 50 of the interrupted storage controller preferably continue.
  • a transmitting RMC When a transmitting RMC receives a NAK from the receiving RMC, it stops transmitting data, and immediately starts sending a FLUSH character.
  • the FLUSH character is sent as part of a status packet using a reserved XGMII control character.
  • the receiving RMC drops/flushes all data received from the time the error was detected until the time the first FLUSH character is received.
  • the transmitting RMC continues to send the FLUSH character until it is either (1) ready to retransmit the failed packet, or (2) is ready to send a status packet with an ACK or a NAK character. Data packet transmissions in the opposite direction may continue normally throughout this process.
  • FIG. 8 illustrates some of the additional control circuits associated with the RMC's transmit and receive interfaces 106 A, 106 B.
  • the transmit input control circuit 170 (TX INPUT CTRL) is a bi-directional interface to the BMI 80 ( FIG. 4 ). This circuit 170 is responsible for determining whether incoming write data should be added to an active bin versus placed in a new bin, and for determining whether a bin is ready to send. This circuit 170 is also responsible for slowing down the PIE RX circuit 42 A when packet retransmission events occur so that transmission FIFO overruns are avoided.
  • a transmit output control circuit 180 (TX OUTPUT CTRL) pulls bin IDs from the head of the queue 172 , and based on each such ID, controls the multiplexer 104 to select between the three transmit FIFOs 86 , 88 , and 89 .
  • This circuit 180 also provides the tag IDs and transmission control information to the transmit interface 106 A.
  • the transmit output control circuit 180 appends these status characters to outgoing data packets, or else sends them via separate status packets.
  • the transmitting RMC if it receives two NAK messages for the same packet (meaning that a packet retransmission attempt has failed), it enters into a “link down” state.
  • the transmitting RMC also enters into the link down state if a timeout event occurs, meaning that the corresponding timeout counter expired before receipt of an expected ACK or NAK message.
  • the transmitting RMC Upon entering the link down state, the transmitting RMC generates a system interrupt and ceases to reflect data. Buffer write operations to the corresponding buffer 50 continue in this event.
  • firmware running on the interrupted storage controller initiates a “system pause” event.
  • a system pause event may also be initiated by firmware in response to other types of error conditions, including conditions unrelated to the reflective memory channel.
  • the MPU ( FIG. 1 ) of the interrupted storage controller 30 runs an error handling process that attempts to resolve the error condition through communications with the MPU of the peer storage controller. As described above, these communications occur over a separate channel, and thus do not rely on the existence of an operational reflective memory channel 62 .
  • the MPUs may collectively determine that a failover event is necessary to resolve the error condition. This determination may alternatively be made by only one of the MPUs, particularly if the other MPU is not responding. In either case, one of the two storage controller's is designated as the survivor, meaning that it will provide network-based access to the volumes that were previously accessible via the other, failed storage controller.

Abstract

A bi-directional reflective memory channel between a pair of storage controllers is used to maintain a mirrored copy of each storage controller's native buffer contents within the buffer of the other storage controller. To maintain such mirrored copies, buffer write operations that fall within a reflective memory segment of one storage controller are automatically reflected across this channel to the other storage controller for execution, and vice versa. The write operations are preferably transmitted across the reflective memory channel using a protocol that provides for error checking, acknowledgements, and retransmissions. This protocol is preferably implemented entirely in automated circuitry, so that the mirrored copies are maintained without any CPU intervention during error-free operation. When a failover occurs, the surviving storage controller uses the mirrored copy of the failed storage controller's native buffer contents to assume control over the failed storage controller's disk drives.

Description

    RELATED APPLICATIONS
  • This application is a division of U.S. application Ser. No. 11/119,213, filed Apr. 29, 2005, which is a division of U.S. application Ser. No. 10/370,042, filed Feb. 19, 2003, the disclosure of which is hereby incorporated by reference.
  • FIELD OF THE INVENTION
  • The present invention relates to network-based storage systems, and more particularly, to techniques for providing storage controller redundancy within a network-based storage system.
  • BACKGROUND OF THE INVENTION
  • A variety of network-attached and SAN (Storage Area Network) based storage systems exist for allowing data to be stored on an Ethernet or other IP based network. Typically, these systems include one or more storage controllers, each of which controls and provides network-based access to a respective array of disk drives. Each storage controller typically includes a buffer or cache memory that is used to temporarily store data as it is transferred between the network and that controller's disk drives. For example, incoming data packets containing I/O (input/output) write data may be maintained in a buffer of the storage controller until successfully written to the appropriate disk drives.
  • Some storage systems implement a storage controller failover mechanism to protect against the possible failure of a storage controller. For example, in some systems, two storage controllers may be paired for purposes of providing controller redundancy. When one of these paired storage controllers detects a failure by the other, the non-failing storage controller may take control of the failing controller's disk drives, allowing these disk drives to be accessed via the network while the failing storage controller is replaced.
  • To provide such redundancy, one storage controller may maintain or have access to a mirrored copy of the other storage controller's cache and configuration data. This allows the non-failing storage controller to effectively pick up the workload of the failing controller where the failing controller left off. Upon replacement of the failing controller, a synchronization or “rebind” operation may be performed between the non-failing and new storage controllers to copy over the cache and configuration data needed to bring the new storage controller on line.
  • One significant problem with existing storage system designs is that the mechanism used to provide storage controller redundancy typically adversely affects or limits the performance of the storage system. For example, in some designs, the mechanism used to maintain a redundant copy of a storage controller's cache data limits the rate at which the storage controller can process network traffic and perform input/output operations. In one such design, described in U.S. Pat. No. 5,928,367, the respective memories of two separate controllers are updated synchronously (in lock step); as a result, if a write operation to one of these memories cannot immediately be performed, the corresponding write operation to the other memory generally must also be postponed.
  • In addition, in many designs, some or all of the system's disk drives cannot be accessed while a rebind operation is being performed between the non-failing and new storage controllers. The present invention seeks to address these and other limitations in existing designs.
  • SUMMARY OF THE INVENTION
  • The present invention makes use of a bi-directional reflective memory channel between a pair of storage controllers to allow each such storage controller to maintain a mirrored copy of the other's buffer contents in real time. To maintain such mirrored copies, buffer write operations that fall within a reflective memory segment of one storage controller are automatically reflected across this channel to the other storage controller for execution, and vice versa. The corresponding write operations are executed asynchronously by the two controllers, such that the postponement of one write operation does not require postponement of the other.
  • The write operations are preferably transmitted across the reflective memory channel and processed according to an asynchronous reflective memory protocol that provides for error checking, acknowledgements, and retransmissions. This protocol is preferably implemented entirely in automated circuitry, so that the mirrored copies are maintained without any CPU (central processing unit) intervention during error-free operation. When a failover occurs, the surviving storage controller uses the mirrored copy of the failed storage controller's native buffer contents to assume control over, and provide network-based access to, the failed storage controller's disk drives. Failover arbitration and control messages are preferably passed between the two controllers over a channel that is separate from the reflective memory channel, so that these messages do not interfere with reflective memory operations.
  • In a preferred embodiment, each storage controller is capable of receiving packet data from a local area network, and storing such data in its local buffer, at a rate of 10 gigabits per second. To support this transfer rate, two 10-gigabit reflective memory channels are provided between the two storage controllers—one for carrying data in each direction.
  • To implement the reflective memory system, each storage controller preferably includes an automated reflective memory controller (RMC) that monitors write operations to that storage controller's buffer. Write operations falling within the local storage controller's reflective memory segment are packetized by the local RMC for transmission to the remote RMC. The remote RMC executes each reflected write operation by writing the associated block of write data to the same destination address as specified within the original write operation. As mentioned above, the original and reflected write operations are executed asynchronously, meaning that one may be completed by its respective RMC before the other.
  • Packet transfers between the RMCs occur over the reflective memory channel according to the asynchronous reflective memory protocol which, as mentioned above, is preferably implemented entirely within automated circuitry. The packets are sufficiently small in the preferred embodiment to allow rapid generation of packet CRC values. The rapid generation of CRC values, and the low overhead of the protocol, allow the reflected data to be transferred with very low latency. Data packets transmitted across the reflective memory channel are preferably error checked and acknowledged by the receiving RMC. If the receiving RMC returns a negative acknowledgement or fails to return an acknowledgement, the sending RMC preferably retries the packet transmission.
  • Each RMC also preferably supports a “rebind” mode in which buffer read operations are reflected across the channel. The rebind mode allows a mirrored copy of a failed storage controller's buffer data to be copied from a surviving storage controller to a replacement storage controller without the need to write data to the buffer of the surviving storage controller. Because this data need not be read from and then written back to the surviving controller's buffer (as would be the case if only write operations were reflected), the surviving controller can continue to process ordinary write operations without risk of mirroring overwritten data (i.e., buffer coherency is not a concern). As a result, during such rebind processing, the surviving storage controller can continue to provide network-based access to its own disk drives and to those of the failed storage controller.
  • In one embodiment, each RMC includes multiple first-in-first-out buffers (FIFOs) that are used to accumulate data to be transmitted across the reflective memory channel. The RMC circuitry that selects between these transmission FIFOs gives priority to at least one of these FIFOs over the other(s), allowing different transmission priority levels to be given to different types of reflected data. One of the transmission FIFOs is also preferably dedicated to “fast path traffic,” which is packet data that is received from the network and written to the buffer without any CPU intervention. The fast path traffic is preferably processed as it is received by a hardware-implemented protocol engine that automates selected portions of the iSCSI and TCP/IP protocols.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • A particular embodiment of the invention will now be described with reference to the following drawings:
  • FIG. 1 illustrates a storage system that includes a pair of storage controllers arranged as peers in a redundant configuration.
  • FIG. 2 illustrates how native and reflected buffer data is stored in the respective buffer memories of the storage controllers of FIG. 1.
  • FIG. 3 illustrates a process by which buffer metadata is reflected bi-directionally between the storage controllers.
  • FIG. 4 illustrates the design of the storage controllers of FIG. 1, and particularly the reflective memory controllers thereof, in greater detail according to one embodiment.
  • FIG. 5 illustrates the design of the fast path and system FIFOs (first-in-first-out buffers) of FIG. 4.
  • FIG. 6 illustrates the format of a data packet used to transfer data between the storage controllers.
  • FIG. 7 illustrates the flow of data between and within a paired set of storage controllers according to one embodiment.
  • FIG. 8 illustrates components of the reflective memory controller of FIG. 4 in further detail.
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
  • A storage system that uses a hardware-implemented, bi-directional reflective memory channel to provide storage controller redundancy will now be described in detail. Throughout the description, reference will be made to numerous implementation details, including but not limited to specific data transfer rates, networking and storage standards, network port configurations, packet formats, protocols, and buffer configurations. These implementation details are provided in order to fully illustrate one particular embodiment of the invention, and not to limit the scope of the invention. The invention is defined only by the appended claims.
  • I. TABLE OF ABBREVIATIONS
  • The following abbreviations and acronyms will be used throughout the detailed description of the preferred embodiment:
      • ACK/NAK: Acknowledgement and Negative Acknowledgement, respectively.
      • BMI: Buffer Memory Interface
      • CLH—Cache Line Header
      • CPU—Central Processing Unit
      • FIFO—First-in-first-out (buffer)
      • GE (as in “1 GE” or “10 GE”)—Gigabit Ethernet
      • I/F—Interface
      • LAN—Local Area Network
      • MPU—Management Processing Unit
      • PIE—Protocol Intercept Engine
      • RMC—Reflective Memory Controller
      • RX—Receive
      • TX—Transmit
    II. OVERVIEW (FIGS. 1 AND 2)
  • FIG. 1 illustrates a storage system that includes a pair of storage controllers 30 arranged as peers in a redundant configuration. Each storage controller may, for example, be in the form of a card or board that plugs into a backplane of a larger storage system. As described in detail below, the storage controllers 30 implement a bi-directional reflective memory channel and an associated protocol that allow each storage controller to maintain a real time copy of the other's native buffer and configuration data.
  • The storage controllers 30, designated by reference characters A and B, are depicted as being housed within a common storage box or enclosure 32, but may alternatively reside in separate enclosures. Each storage controller is connected by a respective network interface (I/F) 34 to a common LAN (local area network) 36, such as an Ethernet based LAN. Client computers (not shown) access the storage controllers 30 over the network 36, preferable but not necessarily using the iSCSI protocol, in order to store and retrieve data on arrays of disk drives 38 controlled by the storage controllers 30. Although a single LAN is shown, each storage controller 30 may connect to multiple, distinct LANs.
  • In the particular embodiment described herein, each LAN interface 34 includes one ten-gigabit Ethernet (10 GE) port and eight one-gigabit Ethernet (1 GE) ports, as depicted in FIG. 1 for one of the two storage controllers. In this embodiment, each storage controller can be configured to operate in one of two modes: a 1 GE mode in which the storage controller connects to the LAN 36 (or multiple LANs) via any number of the eight 1-GE ports (but not the 10-GE port); and a 10 GE mode in which the storage controller connects to the LAN via the 10 GE port only. As will be apparent, many of the implementation details of the bi-directional reflective memory channel are dependent upon this particular LAN interface design, and will thus naturally vary in embodiments that use other LAN interfaces and port configuration options.
  • As depicted in FIG. 1, each storage controller 30 is assigned to or “owns” a respective set or array of disk drives 38 or other storage devices. Specifically, storage controller A owns the disk drives 38 of array A, and storage controller B owns the disk drives of array B. The disk drives 38 are shown as residing within the enclosure 32, but may alternatively reside outside the enclosure. The disk drives may, for example, be serial ATA (Advanced Technology Attachment) disk drives, although SCSI, non-serial ATA, and other types of disk drives may be used. Other types of storage devices may be used in place of some or all of the disk drives, including but not limited to optical drives and solid state storage devices. Each storage controller 30 operates generally as a disk array controller, and may operate its respective disk array as a RAID (Redundant Array of Inexpensive Disks) array, JBOD (Just a Bunch of Disks) array, or other configuration.
  • During normal, redundant operation (no failover events), each storage controller 30 exclusively controls the disk drives 38 it owns. As depicted by dashed lines in FIG. 1, each storage controller is also physically connected to, and is capable of assuming control of, the other storage controller's disk drives 38. For example, if storage controller A detects a failure by storage controller B, storage controller A may initiate a failover process in which storage controller B is taken off line, and in which storage controller A assumes control of the disk drives owned by storage controller B. A selector switch (not shown), such as a conventional drive connector multiplexer, may be connected to each disk drive 38 to enable such transfers of control. Following the transfer of control, storage controller A provides network-based access to both arrays of drives, A and B, allowing storage controller B to be replaced. As described in detail below, this failover process makes use of redundant or mirrored copies of buffer data maintained via the reflective memory channel.
  • The two storage controllers 30 are preferably identical in design, and operate as peers to each other. Each storage controller includes a set of I/O (input/output) processors 40 that process incoming packets received from the network 36, and which generate and transmit response packets on the network. As illustrated, the I/O processors 40 of each storage controller preferably include an automated packet processor 42 and a set of central processing units (CPUs) 44. Each automated packet processor 42 preferably includes application-specific circuitry that automates selected portions of the TCP/IP and iSCSI protocols, as is desirable for providing a high level of performance (e.g. transfer rates of 10 gigabits/sec or higher per storage controller). The CPUs execute firmware modules for performing various storage related tasks, some of which relate specifically to the reflective memory channel and failover events. As discussed below, the CPUs also perform “slow path” processing on selected packets.
  • As further shown in FIG. 1, each storage controller 30 includes a buffer memory 50, referred to herein as a “buffer.” The buffers 50 are used to store incoming packet data received from the network, and to accumulate incoming and outgoing I/O data. The incoming packet data stored in the buffers 50 consists primarily of packet payloads containing I/O write data.
  • In one embodiment, each buffer 50 is capable of receiving and storing incoming packet data from the network at a rate of ten gigabits per second (Gb/s), which corresponds to the ten Gb/s maximum transfer rate supported by each LAN interface 34. The buffers 50 are preferably implemented as random access memory (RAM), such as SDRAM (synchronous dynamic RAM). Each buffer 50 is managed by software executed by the respective controller's CPUs to provide a storage cache. The cache management software stores data in the buffer 50 in cache lines.
  • With further reference to FIG. 1, each storage controller 30 further includes a system memory 52 that stores code modules executed by the associated CPUs. As with the buffers 50, the system memories 52 may be implemented using SDRAM.
  • The system memories 52 also store various types of configuration and state information. This information preferably includes buffer metadata descriptive of the data currently stored in the respective storage controller's buffer 50. In one implementation, the buffer metadata is in the form of cache line headers (CLHs) descriptive of the buffer data stored within each cache line of the software-implemented storage cache. As described below, this metadata may periodically be copied from the system memory 52 to the corresponding buffer 50 to allow each storage controller to maintain a redundant copy of its peer's buffer metadata.
  • To support failover operations, each storage controller maintains, within its respective buffer, a complete copy of the native buffer data of the other storage controller. This allows a non-failing storage controller to rapidly pick up the workload of a failing storage controller where the failing storage controller left off. The task of maintaining the redundant copies of buffer data is the primary responsibility of the reflective memory controllers (RMCs) 60 shown in FIG. 1, which communicate over a pair of like communication channels 62A, 62B. The two unidirectional channels 62A, B collectively form a bi-directional reflective memory channel 62. This bi-directional channel 62 is preferably dedicated to transfers of reflected buffer data. As discussed below, all other types of inter-controller communications preferably occur over a separate channel and associated medium.
  • Any appropriate interface and medium may be used to implement the bi-directional reflective memory channel 62. In the preferred embodiment, the RMCs 60 communicate over each channel 62A, 62B according to XGMII (10-Gigabit Medium Independent Interface), and can thus communicate at a rate of ten gigabits per second in each direction. The XGMII signals generated by the RMC's are converted to XAUI (10-gigabit Attachment Unit Interface) signals for transmission over a backplane that interconnects the two storage controllers 30. The conversions between the XGMII and XAUI interfaces are performed by commercially-available devices mounted on the respective boards of the storage controllers 30, externally to the respective ASICs (Application Specific Integrated Circuits) in which the RMCs reside. At the receiving end, the XAUI signals are converted back to XGMII for delivery to the receiving ASIC/RMC. As will be recognized, the bi-directional reflective memory channel 62 can alternatively be implemented using other interfaces, and can be implemented using an alternative medium such as a cable or (in future implementations) a wireless link.
  • As described in detail below, the protocol used to transfer buffer data over the bi-directional channel 62 is preferably implemented entirely in hardware, allowing the reflected data to be transferred at high bit rates and with low latency. The protocol is referred to herein as an “asynchronous reflective memory protocol,” as it allows the RMCs to execute corresponding write operations asynchronously to one another. This is accomplished in part by providing a mechanism through which each RMC can eventually confirm that the write operations reflected to its peer were successfully received.
  • Referring further to FIG. 1, each RMC 60 operates generally by monitoring write operations to its respective buffer 50, and by reflecting (transmitting to the other RMC) those write operations that fall within a particular address range. Each such write operation in the preferred embodiment is generally in the form of a destination buffer address and a block of contiguous write data. The receiving RMC forwards the reflected write operation to the buffer interface of its own buffer 50 for execution (i.e., for storage of the write data at the specified destination buffer address), preferably after performing error checking as described below. Thus, each write operation that falls within a buffer's reflective memory area or “segment” is replicated, and is executed by both storage controllers 30. As a result, each storage controller maintains a mirrored copy of the other storage controller's buffer contents.
  • An important aspect of the design is that the storage controllers, and particularly the RMCs 60, execute the corresponding write operations asynchronously to one another (i.e., one storage controller may execute a given write operation before the other). Thus, when one storage controller (the “source”) executes a write operation and reflects that operation to the other storage controller (the “target”), the source may continue processing subsequent write operations (and/or read operations) without waiting for the target to complete the reflected operation. As discussed below, the asynchronous reflective memory protocol implemented by the RMCs allows the source to determine whether the reflected operation is ultimately received successfully by the target, and to re-try the operation (or invoke an appropriate error correction routine) if it is not.
  • FIG. 2 depicts the buffer contents of two storage controllers 30 that are in a “bound” state. The arrows in this drawing represent buffer data that is being reflected to the other storage controller's buffer. For example, when a block of data is written to the reflective memory segment of storage controller A's buffer 50, it is also written to the buffer 50 of storage controller B at the same destination address. As illustrated, each storage controller thus stores within its own buffer a mirrored copy 70 of its peer's native or local buffer data 72, at the same buffer addresses as that peer. The address ranges used for storing local versus mirrored buffer data need not be contiguous, and are preferably defined using a set of configuration registers within each storage controller.
  • An important aspect of the reflective memory system, in the preferred embodiment, is that the memory reflection process, including the asynchronous reflective memory protocol, is fully automated within application-specific circuitry of the RMCs 60. (The term “automated,” as used herein, refers generally to a task or function that is implemented without fetching and executing software or firmware instructions.) Specifically, during error-free operation, each RMC reflects buffer write operations to its peer RMC, receives and processes write operations reflected by that peer RMC, and implements associated error checking, acknowledgement and retransmission protocol tasks (as discussed below), without any CPU involvement or intervention.
  • Another important aspect of the reflective memory system, in the preferred embodiment, is that this protocol provides for error checking, acknowledgements and retransmissions, yet has sufficiently low overhead to be implemented over a pair of 10 Gb/s channels without limiting the rate at which the storage controllers receive traffic from the 10 Gb/s network 36. The design may, however, alternatively be implemented using higher speed reflective memory channels 62A, B and/or a greater number of such channels, in which case a less efficient reflective memory protocol may be used.
  • The RMCs 60 and the bi-directional reflective memory channel 62 advantageously support transfers and processing of reflected write operations at ten Gb/s in both directions. Because this transfer rate matches the maximum rate at which each storage controller 30 can receive data packets over the LAN 36, the reflective memory system does not limit the rate at which each storage controller can receive data into its respective buffer 50 from the network 36. This high rate at which the reflective memory system operates is attributable in-part to the use of a protocol that allows the storage controllers/RMCs to execute corresponding write operations asynchronously relative to each other, and is also attributable to the automation of key portions of this protocol.
  • Each RMC 60 can be enabled or disabled by firmware running on the respective storage controller 30. When disabled, an RMC does not reflect any buffer operations. An RMC may be placed in a disabled state when, for example, the storage controller is being used in a standalone or other non-redundant configuration. Each RMC can also preferably be placed in a special “rebind” mode in which the RMC reflects buffer read operations. As described below, the rebind mode may be used following a failover event to efficiently copy mirrored buffer data 70 from the buffer 50 of a surviving storage controller to the buffer 50 of a replacement storage controller.
  • As further depicted in FIG. 1, the storage controllers 30 also preferably include respective management processing units or “MPUs” 66 that communicate with each other over a separate, relatively low bandwidth channel 68. This channel 68 is preferably used to pass error status information and resolve error conditions. Because this channel 68 is separate from the bi-directional reflective memory channel, these communications between the MPUs do not interfere with transfers of reflected operations.
  • III. MIRRORING OF BUFFER METADATA (FIG. 3)
  • As depicted in FIG. 3, the bi-directional reflective memory channel 62 is also preferably used by each storage controller 30 to maintain a copy of its peer's buffer metadata. This is preferably accomplished by having each storage controller execute a firmware program that copies its buffer metadata from the system memory 52 to the reflective memory segment of its buffer memory 50 (or equivalently, by writing the same buffer metadata to both the system memory 52 and the buffer 50). This copying operation is depicted in FIG. 3 by the thin, solid arrows, while the reflected write operations are depicted by wide arrows. Typically, the buffer metadata will represent less than 1% of the reflected data.
  • In the event of a failover, the surviving storage controller may move the failing storage controller's buffer metadata from the buffer 50 to its system memory 52. This is depicted by the dashed arrow in FIG. 3 for storage controller A (the survivor). Once copied to the survivor's system memory, this metadata may be used by the survivor to interpret and process the failed storage controller's ordinary buffer data. In other implementations, the buffer metadata may alternatively be copied or moved between the paired storage controllers 30 over a separate communication channel.
  • The storage controllers 30 can optionally be configured, via firmware, to use the reflective memory channel to reflect other types of data. In this regard, the reflective memory channel 62 may be viewed as providing a general purpose memory reflection “service” that may be used by firmware for various types of inter-controller communication. In this regard, a portion of each buffer 50 may be allocated exclusively to general-purpose firmware transfers of data. Thus, although the reflective memory channel 62 is used in the preferred embodiment exclusively to transfer fast path data, slow path data, cache line headers, and rebind data (as described below), the underlying architecture is not so limited.
  • IV. REFLECTIVE MEMORY CONTROLLER (FIGS. 4 AND 5)
  • FIG. 4 illustrates the design of the RMCs 60 and related components according to one embodiment. As illustrated, each storage controller 30 includes a buffer memory interface (BMI) 80 that processes all traffic to and from its respective buffer 50. Incoming buffer data (in the form of write operations or bursts) enters the buffer memory interface 80 from two local sources: a receive (RX) circuit 42A of an automated protocol intercept engine (PIE) 42, and a system bus 82. The BMI 80 also receives buffer write data along a receive path 84, which carries data reflected from the remote storage controller. (The term “local” is used in the present description to refer to the storage controller illustrated in FIG. 4, and “remote” is used to refer to its peer.) Data received from each of these three sources is written to the buffer 50 by a multi-port memory controller 85. Incoming buffer write operations received from the system bus 82 first pass through the BMI's bus interface 100, which checks the addresses of such operations as described below.
  • Incoming buffer data from the PIE RX circuit 42A represents incoming network traffic (packet data) that is placed into the buffer 50 without any CPU intervention. This packet data, referred to as “fast path data,” consists primarily of properly ordered, non-fragmented packet data received from the network 36. Out of sequence data received from the network 36, on the other hand, is processed by one or more CPUs of a cluster of CPUs 44 before being written to the buffer 50. This “slow path” data enters the BMI 80 via the system bus 82 after being properly re-sequenced by firmware.
  • The PIE RX circuit 42A processes the fast path data. This circuit, and a counterpart PIE TX circuit (not shown), automate selected portions of the iSCSI and TCP/IP protocols within application-specific circuitry, significantly reducing the quantity of protocol-level processing that needs to been performed by software/firmware. Details of a particular implementation of the protocol intercept engine 42, including the PIE RX circuit 42A, are set forth in a U.S. provisional patent application filed on Feb. 14, 2003 titled “High Availability Integrated Storage Network Processing For iSCSI Communication,” the disclosure of which is hereby incorporated herein by reference. The PIE RX circuit is configured to write data only to the buffer's reflective memory segment, and not to other portions of the buffer 50. Thus, when the RMC is enabled, all incoming buffer write operations from the PIE RX circuit are reflected, regardless of the associated destination addresses. Because the PIE RX circuit strips off iSCSI and TCP/IP headers of incoming packets in the preferred embodiment, the packet data written to the buffer 50 consists essentially of packet payload data, without the associated headers.
  • As packet data from the network 36 is written to the buffer 50, firmware executed by one or more of the CPUs 44 generates corresponding cache line headers (CLHs) within the system memory 52 (FIG. 1). These CLHs describe the data currently stored in the local buffer 50, and are generated based on status signals from the PIE RX circuit 42A. To allow the remote storage controller 30 to interpret its mirrored copy of buffer data, the firmware also writes these CLH's to the reflective memory segment of the local buffer 50 (via the system bus 82), causing the CLHs to be copied to the remote storage controller's buffer 50. As described below, the hardware-implemented asynchronous reflective memory protocol gives priority to the CLHs over other types of reflected data, essentially ensuring that the necessary CLH data will be available to the surviving storage controller in the event of a failure.
  • As further depicted in FIG. 4, the RMC 60 preferably includes three transmission FIFOs (first-in-first-out buffers) that are used to accumulate data to be reflected to the other storage controller: a fast path FIFO 86 and two system FIFOs 88, 89. All fast path data to be reflected passes through the fast path FIFO 86. All other reflected data, including slow path data and CLH data, passes through one of the two system FIFOs 88, 89. As described below, the fast path and system FIFOs operate generally by accumulating buffer write data into packet bins for transmission across the reflective memory channel. The organization of the fast path and system FIFOs is depicted in FIG. 5 and is described separately below.
  • Two separate system FIFOs 88, 89 are provided in the illustrated embodiment in order to accommodate differing transmission priority levels for different data types. Specifically, the circuitry (described below) for transmitting data over the outgoing channel 62A gives priority to system_FIFO_1 over system_FIFO_2 (and also over the fast path FIFO 86). In one firmware configuration, the storage controller 30 uses system_FIFO_1 exclusively to transfer CLH data, and uses system_FIFO_2 to transfer slow path data and rebind data. The firmware may alternatively be written to use the system FIFOs to reflect additional or other types of data. Further, as will be recognized, a greater or lesser number of FIFOs may be used to buffer the outgoing data.
  • Operation of the system FIFOs 88, 89 is configurable by firmware via a set of address registers within a set of RMC registers 94 (FIG. 4). Specifically, each system FIFO may be assigned to a respective, mutually exclusive buffer address range or “window.” An address comparison (ADDR COMP) circuit 96 within the BMI's bus interface 100 monitors the addresses of incoming buffer write operations received from the system bus 82 to determine whether such operations fall within either window. Write operations falling within window 1 are passed to system_FIFO_1, and write operations falling within window 2 are passed to system_FIFO_2. All other buffer write operations emanating from the system bus 82 are ignored by the RMC, and thus are not reflected.
  • As depicted in FIG. 4, a multiplexer 102 selects between two input sources to system_FIFO_2. The upper input to this multiplexer 102 is used to source system_FIFO_2 with buffer write data as such write data is passed from the BMI bus interface 100 to the multi-port memory controller 85. Data that enters system_FIFO_2 via this path consists of buffer write data emanating from the system bus 82 as the result of firmware operations.
  • The lower input to this multiplexer 102 is used to source system_FIFO_2 with data being read from the buffer 50. This path is used during rebind operations to copy the buffer data of a failed storage controller to the buffer 50 of a new, replacement storage controller. The multiplexer 102 selects this lower path when the RMC 60 is placed, via the RMC registers 94, into “rebind” mode via firmware executed by the CPUs 44. An important aspect of this feature is that it allows buffer data to be copied over from one storage controller to the other, over the reflective memory channel 62, without having to write any new data to the source buffer 50.
  • Specifically, to initiate a transfer of a block of buffer data from the buffer 50 of the source storage controller to the buffer 50 of a target storage controller, the firmware running on the source storage controller simply places the source's RMC in rebind mode, and then performs a sequence of “dummy” reads of this buffer data from its own buffer 50. Each block of data read from source's buffer 50 as the result of a dummy read operation is reflected across the reflective memory channel to the target's RMC, and is written to a corresponding address in the target's buffer 50. During this process of copying over the mirrored data, the surviving storage controller continues to provide network-based access to both its own disk drives 38 and the disk drives 38 of the failed storage controller 30.
  • With further reference to FIG. 4, a multiplexer 104 selects between the two system FIFOs 88, 89 and the fast path FIFO 86, and passes the output of the selected FIFO to a transmission interface (TX I/F) 106A for transmission across the outgoing channel 62A. The data is preferably sent across the channels 62A, 62B in packets according to an asynchronous reflective memory protocol which, as mentioned above, provides for error checking, packet acknowledgements, and packet retransmissions. As described below, when data from a transmission FIFO 86, 88, 89 is transmitted on the outgoing channel 62A within a packet, this data is maintained within that transmission FIFO until the packet has been acknowledged by the receiving RMC. The control logic and structures associated with the multiplexer 104 and the FIFOs 86, 88, 89 are depicted in FIG. 8 and are described separately below.
  • As further illustrated in FIG. 4, a receive interface (RX I/F) 106B receives the data transmitted by the remote storage controller on the incoming channel 62B. In operation, incoming packet data from the remote storage controller 30 is initially error-checked by the receive interface 106B, and when error free, is written to a receive FIFO 110. Data accumulated within this FIFO 110 is ultimately written to the buffer 50 by the multi-port memory controller 85. As depicted by the arrow labeled STA in FIG. 4, the receive interface 106B provides packet status information to the transmit interface 106A. As described below, the transmit interface 106A uses this information to send packet acknowledgement (ACK and NAK) messages to the remote storage controller via the outgoing channel 62A, allowing the transmitting controller 30 to re-send packets that are not successfully received. The ACKs and NAKs may be communicated between the RMCs using reserved control characters of the XGMII interface. The receive interface 106B also provides link status information to the transmit interface 106A indicative of the state of the incoming link 62B. A block diagram of the link control and error checking logic incorporated into the transmit and receive interfaces 62A, 62B is provided in FIG. 7 and is described below.
  • Errors detected by the receive interface 106B propagate to the RMC registers 94 via the lines labeled ERROR and INT in FIG. 4, respectively, allowing these events to be monitored by firmware executed by the CPUs 44. Firmware accesses the various RMC registers 94 using a register bus 83.
  • With the exception of the CPUs 44, all of the modules and components depicted in FIG. 4 are preferably implemented entirely within automated, application-specific circuitry (hardware state machines, etc.), allowing data reflection operations to occur at a high rate and with low latency.
  • In one implementation, the transmission and receive interfaces 106A, 106B, as well as the LAN interface 34 shown in FIG. 1, are all XGMII ten gigabit Ethernet interfaces. Other types of interfaces, including but not limited to SPI4 and Utopia 3, may alternatively be used. Further, as mentioned above, interfaces that exceed the transmission rate of the LAN interface 34 may be used to implement the reflective memory channel.
  • FIG. 5 illustrates the design of the transmission FIFOs 86, 88 and 89 in further detail. As illustrated, the RMC 60 arranges the fast path FIFO 86 into one of two possible configurations, depending upon whether the storage controller is operating in the 1 GE mode versus the 10 GE mode. When the storage controller is in the 1 GE mode (and is thus using from one to eight 1 GE LAN ports), the fast path FIFO is organized into eight 512-byte packet bins, as depicted by the labels BIN 0 through BIN 7 on the left side of the drawing. When operating in the 10 GE mode, the fast path FIFO is organized into two 2048-byte packet bins, as shown on the right side of the drawing.
  • Each of the system FIFOs is statically arranged as a single packet bin. Specifically, system_FIFO_2 is arranged as a single, 512-byte bin, and system_FIFO_1 (the higher priority system FIFO) is arranged as a single, 64-byte bin. Thus, when operating in the 1 GE mode, a total of ten bins (0-9) are provided, and when operating in the 10 GE mode, a total of 4 bins are provided (0-3). The 10 GE configuration advantageously allows the RMCs to sustain the streaming of jumbo packets of 9018 bytes (8960 bytes of data, 20 bytes TCP information, 20 bytes of IP information, and 18 bytes of Ethernet information).
  • Each bin of the transmit FIFOs 86, 88, 89 operates generally by accumulating continuous buffer write data for eventual transmission on the outgoing channel 62A within a single packet (i.e., each outgoing data packet contains payload data from a single bin). As described below, one important benefit of subdividing the reflected data into bins is that it reduces the amount of time needed to generate the CRC portion of each packet. When more than one bin is currently “ready” to be sent, the RMC selects between the ready bins so as to give highest priority to system FIFO_1, intermediate priority to the fast path FIFO, and lowest priority to system_FIFO_2. The source storage controller 30 may therefore reflect buffer write operations across the channel 62A out-of-order (i.e., in an order that is different from the order in which the source controller 30 performs these operations).
  • As depicted at the bottom of FIG. 5, the TX FIFOs maintain one internal packet control structure 120 for each of the bins. Each such control structure 120 includes the following fields: DADDR[ ], which stores the destination address currently associated with the packet bin; BIN_STATUS[ ], which stores the bin's current status (Active, Free, Sent, or NAKed); WORDS_IN_BIN[ ], which keeps track of the number of words stored in the bin, and BIN_TIMEOUT[ ], which is used to implement a timeout counter.
  • In operation, the BMI 80 (FIG. 4) receives write traffic from the PIE RX circuit 42A and the system bus in 64-byte bursts, with each such burst being accompanied by a buffer destination address. Bursts emanating from the PIE RX circuit are processed as follows. When a burst is received, the beginning address of this burst is compared, in parallel, to the DADDR fields of all currently active bins. If a match occurs (i.e. the address matches the next sequential write location from the contents of DADDR), the write data is stored in the corresponding active bin, and the bin's WORDS_IN_BIN field is updated with the new count value. The effect of this operation is to aggregate bursts for transmission to the remote RMC.
  • If there is no match, and a bin is available, the write data is stored in the next available bin, and the packet control structure fields of that bin are updated as follows: the bin's status is changed to Active, the beginning burst address is written to the DADDR field, and the WORDS_IN_BIN field is updated with the current number of words stored in the bin. If no bin is available (as may result if packet errors occur), the fast path FIFO asserts a “back off” signal (not shown) that causes the PIE RX circuit to slow down its operation.
  • As illustrated in FIG. 5, each bin of the fast path FIFO has a capacity of eight 64-byte bursts when in the 1 GE mode, and has a capacity of thirty two 64-byte bursts when in the 10 GE mode. The RMC sends the entire contents of the active bin across the outbound RMC channel 62A when either (1) the current burst is the last burst within a burst sequence of contiguous write data, or (2) the bin becomes full (512 bytes in 1GE mode, or 2048 bytes in 10GE mode). Upon sending the contents of the bin, the RMC updates the bin's status to “Sent,” and activates the bin timeout counter. The bin's “Sent” status is maintained until the transmitting RMC receives a corresponding ACK or NAK message from the receiving RMC.
  • An important attribute of the asynchronous reflective memory protocol, in the preferred embodiment, is that the bins are selected without regard to LAN port number. Thus, when running in the 1 GE mode, all eight of the fast path bins are used even if less than all of the eight 1GE LAN ports used. The bandwidth associated with the fast path FIFO is thus effectively distributed among those 1 GE ports actually being used.
  • When a burst is received from the system bus 82 (FIG. 4), its destination address is checked to see if it falls within either system_FIFO_1's address range or system_FIFO_2's address range. If the burst falls within one of these address ranges, the corresponding system FIFO 88, 89 processes the burst in the same manner as described above for the fast path FIFO 86. Because system_FIFO_1 holds only a single 64 byte burst, it becomes ready for transmission each time it receives a burst from the system bus 82.
  • The receipt of an ACK message by the transmitting RMC indicates that the receiving RMC received the associated packet without detection of any errors. It does not, however, indicate successful completion of the write operation specified by the packet. If an error occurs when the receiving RMC performs this write operation, the receiving RMC will report this error to its own CPUs. This error will thereafter be reported to the transmitting storage controller 30 via the separate MPU channel 68.
  • IV-A PACKET FORMAT AND PROTOCOL (FIGS. 6 AND 7)
  • FIG. 6 illustrates the format of the data packets used to transmit bin contents across the channels 62A, 62B according to one embodiment of the asynchronous reflective memory protocol. The data payload portion 140 contains the contents of the bin currently being transmitted, and the destination address (DADDR) portion 142 contains the associated buffer destination address. The status word 144, if present, contains the ACK/NAK status of up to four packets that have been received by the transmitting RMC. Accumulated packet ACK/NAK messages are thus piggybacked on transmitted data packets. If no ACK/NAK messages are available to transmit, the status word is replaced with an end-of-frame (EOF) character. If a status word is ready to transmit but no data packets are being transmitted, the status word may be sent within a special status packet (not shown) that does not include the other fields shown in FIG. 6.
  • The TAG field 146 carries the bin ID of the bin being transmitted, and is used by the receiving RMC 60 to check for valid headers. The TAG field is protected from corruption by having the bin ID encoded using three control characters spread across three XGMII transmission lanes, with enough redundancy to sustain a one-lane failure.
  • The CRC (cyclic redundancy code) field is generated from all of the other fields within the packet except the TAG field 146. When a packet is received, the receiving RMC checks the packet's CRC, and incorporates the results of this CRC check (ACK or NAK) into the next status word 144 to be returned. The ACK or NAK is added to the status word according to the order of the packet's arrival (i.e., the ACK/NAK characters follow the packet arrival order).
  • Given that the entire packet is in question when a CRC error occurs, the TAG field is protected by having the bin ID encoded across the three transmission lanes as described above. The receiving RMC can thus reliably identify a failed packet transmission and return a corresponding NAK message.
  • Given the pipelined nature of packet protocol, any received data behind the failed bin transmission is also dropped in order to maintain write-order. The originating RMC keeps track of the bin IDs sent so that when a NAK is received, it can determine which bins need retransmission and which do not. If the re-transmitted packet is successfully received, an ACK is sent back to the originator, which in turn frees-up the associated bin.
  • FIG. 7 illustrates the flow of data between and within the RMCs 60 of two paired storage controllers 30 in further detail. As depicted, the CRC portion of each packet is generated as the packet is being transmitted. A CRC-32 algorithm may be used for this purpose. Because each packet has a small payload (512 bytes or smaller when in 1 GE mode, and 2048 or smaller when in 10GE mode), the CRC portions can be generated rapidly, allowing data reflection at a high transfer rate and with low latency.
  • When the receive interface 106B of a storage controller receives a data packet, it checks the CRC. If the CRC is valid, the packet's data payload is pushed into the corresponding receive FIFO 110, and is eventually written to the receiving storage controller's buffer 50 at the destination address specified in the packet. If the CRC is determined to be bad, the received packet is dumped and a NAK status is generated. The NAKs and ACKs resulting from the CRC checks are queued by the corresponding transmit interface 106A for sending back to the originating RMC via a status word.
  • When the receive interface 106B receives a status word, it passes the status (ACK/NAK) information and associated tag(s) to the receiving RMC's transmit interface 106A. The transmit interface 106A in turn updates the associated packet control structures 120 (FIG. 5) to reflect the ACK/NAK status. If a particular bin receives an ACK, the bin is made available by setting its status to “Free.”
  • If a bin receives a NAK, the transmit interface 106A checks bin's status to see if this is the first NAK, and if so, queues the bin for resending and updates status to “NAKed.” If the NAKed bin is already in the NAKed state (indicating that the packet transmission has failed twice), the transmitting RMC enters into a “link down” state in which its transmission interface generates a system interrupt, stops transmitting data across the outbound RMC channel 62A, and starts sending XGMII Link Down Sequence Remote packets to the remote RMC. Although outgoing data reflection is halted, write operations to the buffer 50 of the interrupted storage controller preferably continue.
  • When a transmitting RMC receives a NAK from the receiving RMC, it stops transmitting data, and immediately starts sending a FLUSH character. The FLUSH character is sent as part of a status packet using a reserved XGMII control character. The receiving RMC drops/flushes all data received from the time the error was detected until the time the first FLUSH character is received. The transmitting RMC continues to send the FLUSH character until it is either (1) ready to retransmit the failed packet, or (2) is ready to send a status packet with an ACK or a NAK character. Data packet transmissions in the opposite direction may continue normally throughout this process.
  • IV-B. FIFO CONTROL CIRCUITS (FIG. 8)
  • FIG. 8 illustrates some of the additional control circuits associated with the RMC's transmit and receive interfaces 106A, 106B. The transmit input control circuit 170 (TX INPUT CTRL) is a bi-directional interface to the BMI 80 (FIG. 4). This circuit 170 is responsible for determining whether incoming write data should be added to an active bin versus placed in a new bin, and for determining whether a bin is ready to send. This circuit 170 is also responsible for slowing down the PIE RX circuit 42A when packet retransmission events occur so that transmission FIFO overruns are avoided.
  • As bins become ready for transmission, the transmit input control circuit 170 places the IDs of these bins in a transmit queue 172. This queue controls and keeps track of the order in which outgoing packets are transmitted. This transmission order information is used in the event of a NAK event to determine which of the transmitted packets need to be resent.
  • A transmit output control circuit 180 (TX OUTPUT CTRL) pulls bin IDs from the head of the queue 172, and based on each such ID, controls the multiplexer 104 to select between the three transmit FIFOs 86, 88, and 89. This circuit 180 also provides the tag IDs and transmission control information to the transmit interface 106A. In addition, as status characters (ACKs and NAKs) are received from the RMC's receive interface 106B, the transmit output control circuit 180 appends these status characters to outgoing data packets, or else sends them via separate status packets.
  • V. ERROR HANDLING
  • As described above, if the transmitting RMC receives two NAK messages for the same packet (meaning that a packet retransmission attempt has failed), it enters into a “link down” state. The transmitting RMC also enters into the link down state if a timeout event occurs, meaning that the corresponding timeout counter expired before receipt of an expected ACK or NAK message. Upon entering the link down state, the transmitting RMC generates a system interrupt and ceases to reflect data. Buffer write operations to the corresponding buffer 50 continue in this event.
  • In response to the system interrupt, firmware running on the interrupted storage controller initiates a “system pause” event. A system pause event may also be initiated by firmware in response to other types of error conditions, including conditions unrelated to the reflective memory channel. In response to the system pause event, the MPU (FIG. 1) of the interrupted storage controller 30 runs an error handling process that attempts to resolve the error condition through communications with the MPU of the peer storage controller. As described above, these communications occur over a separate channel, and thus do not rely on the existence of an operational reflective memory channel 62.
  • The MPUs may collectively determine that a failover event is necessary to resolve the error condition. This determination may alternatively be made by only one of the MPUs, particularly if the other MPU is not responding. In either case, one of the two storage controller's is designated as the survivor, meaning that it will provide network-based access to the volumes that were previously accessible via the other, failed storage controller.

Claims (8)

1. A method of maintaining a redundant copy of buffer contents of a first storage controller within a buffer of a second storage controller, the method comprising:
with the first storage controller:
accumulating buffer write data of first storage controller within at least one first-in-first-out buffer of the first storage controller, said buffer write data specifying storage operations;
packetizing accumulated buffer write data stored in the at least one first-in-first-out buffer to generate packets, said packets including error checking codes; and
transmitting the packets to the second storage controller over a reflective memory channel;
with the second storage controller:
receiving and error checking the packets transmitted by the first storage controller on the reflective memory channel;
transmitting packet acknowledgement data to the first storage controller over the reflective memory channel; and
writing the buffer write data contained in successfully received packets to a buffer of the second storage controller.
2. The method of claim 1, wherein the method is implemented entirely within automated circuitry of the first and second storage controllers.
3. The method of claim 1, wherein the method is performed by automated hardware without CPU intervention.
4. The method of claim 1, further comprising concurrently performing the method with roles of the first and second storage controllers reversed, such that each storage controller maintains a redundant copy of native buffer contents of the other storage controller.
5. The method of claim 1, wherein the buffer write data specifies the storage operations in accordance with the iSCSI protocol.
6. The method of claim 1, wherein the packets further include buffer metadata descriptive of reflected buffer data.
7. The method of claim 1, wherein the packets are transmitted to the second storage controller over the reflective memory channel at a rate of at least ten gigabits per second.
8. A storage system that operates according to the method of claim 1.
US11/835,942 2003-02-19 2007-08-08 Storage controller redundancy using packet-based protocol to transmit buffer data over reflective memory channel Abandoned US20070288792A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/835,942 US20070288792A1 (en) 2003-02-19 2007-08-08 Storage controller redundancy using packet-based protocol to transmit buffer data over reflective memory channel

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US10/370,042 US6941396B1 (en) 2003-02-19 2003-02-19 Storage controller redundancy using bi-directional reflective memory channel
US11/119,213 US7353306B2 (en) 2003-02-19 2005-04-29 Storage controller redundancy using bi-directional reflective memory channel
US11/835,942 US20070288792A1 (en) 2003-02-19 2007-08-08 Storage controller redundancy using packet-based protocol to transmit buffer data over reflective memory channel

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
US10/370,042 Division US6941396B1 (en) 2003-02-19 2003-02-19 Storage controller redundancy using bi-directional reflective memory channel
US11/119,213 Division US7353306B2 (en) 2003-02-19 2005-04-29 Storage controller redundancy using bi-directional reflective memory channel

Publications (1)

Publication Number Publication Date
US20070288792A1 true US20070288792A1 (en) 2007-12-13

Family

ID=34885842

Family Applications (3)

Application Number Title Priority Date Filing Date
US10/370,042 Expired - Lifetime US6941396B1 (en) 2003-02-19 2003-02-19 Storage controller redundancy using bi-directional reflective memory channel
US11/119,213 Expired - Fee Related US7353306B2 (en) 2003-02-19 2005-04-29 Storage controller redundancy using bi-directional reflective memory channel
US11/835,942 Abandoned US20070288792A1 (en) 2003-02-19 2007-08-08 Storage controller redundancy using packet-based protocol to transmit buffer data over reflective memory channel

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US10/370,042 Expired - Lifetime US6941396B1 (en) 2003-02-19 2003-02-19 Storage controller redundancy using bi-directional reflective memory channel
US11/119,213 Expired - Fee Related US7353306B2 (en) 2003-02-19 2005-04-29 Storage controller redundancy using bi-directional reflective memory channel

Country Status (1)

Country Link
US (3) US6941396B1 (en)

Cited By (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050246388A1 (en) * 2003-07-02 2005-11-03 Satoshi Yamatake Image database system
US20070055804A1 (en) * 2005-09-07 2007-03-08 Ran Hay Method and apparatus for managing multiple components
US20090187786A1 (en) * 2008-01-17 2009-07-23 Michael John Jones Parity data management system apparatus and method
US20090327481A1 (en) * 2008-06-30 2009-12-31 International Business Machines Corporation Adaptive data throttling for storage controllers
US20100229029A1 (en) * 2009-03-06 2010-09-09 Frazier Ii Robert Claude Independent and dynamic checkpointing system and method
US20110093767A1 (en) * 2009-10-15 2011-04-21 Sharp William A System and method to serially transmit vital data from two processors
US20110307736A1 (en) * 2010-04-12 2011-12-15 Johann George Recovery and replication of a flash memory-based object store
US8543774B2 (en) 2011-04-05 2013-09-24 Ansaldo Sts Usa, Inc. Programmable logic apparatus employing shared memory, vital processor and non-vital communications processor, and system including the same
US8667212B2 (en) 2007-05-30 2014-03-04 Sandisk Enterprise Ip Llc System including a fine-grained memory and a less-fine-grained memory
US8667001B2 (en) 2008-03-20 2014-03-04 Sandisk Enterprise Ip Llc Scalable database management software on a cluster of nodes using a shared-distributed flash memory
US8666939B2 (en) 2010-06-28 2014-03-04 Sandisk Enterprise Ip Llc Approaches for the replication of write sets
US8694733B2 (en) 2011-01-03 2014-04-08 Sandisk Enterprise Ip Llc Slave consistency in a synchronous replication environment
US8856593B2 (en) 2010-04-12 2014-10-07 Sandisk Enterprise Ip Llc Failure recovery using consensus replication in a distributed flash memory system
US8868487B2 (en) 2010-04-12 2014-10-21 Sandisk Enterprise Ip Llc Event processing in a flash memory-based object store
US8874515B2 (en) 2011-04-11 2014-10-28 Sandisk Enterprise Ip Llc Low level object version tracking using non-volatile memory write generations
US9047351B2 (en) 2010-04-12 2015-06-02 Sandisk Enterprise Ip Llc Cluster of processing nodes with distributed global flash memory using commodity server technology
US20150242138A1 (en) * 2014-02-21 2015-08-27 Netapp, Inc. Systems and Methods for a Storage Array-Managed Initiator Cache
US9135064B2 (en) 2012-03-07 2015-09-15 Sandisk Enterprise Ip Llc Fine grained adaptive throttling of background processes
US9164554B2 (en) 2010-04-12 2015-10-20 Sandisk Enterprise Ip Llc Non-volatile solid-state storage system supporting high bandwidth and random access
US20170242771A1 (en) * 2016-02-19 2017-08-24 Dell Products L.P. Storage controller failover system
CN107329920A (en) * 2017-07-06 2017-11-07 中国航空工业集团公司西安飞机设计研究所 A kind of common interface frame design method of reflective memory
US20190034303A1 (en) * 2017-07-27 2019-01-31 International Business Machines Corporation Transfer track format information for tracks in cache at a first processor node to a second process node to which the first processor node is failing over
US20190034302A1 (en) * 2017-07-27 2019-01-31 International Business Machines Corporation Transfer track format information for tracks in cache at a primary storage system to a secondary storage system to which tracks are mirrored to use after a failover or failback
US10579296B2 (en) 2017-08-01 2020-03-03 International Business Machines Corporation Providing track format information when mirroring updated tracks from a primary storage system to a secondary storage system
US10613951B2 (en) 2017-09-13 2020-04-07 International Business Machines Corporation Memory mirror invocation upon detecting a correctable error

Families Citing this family (67)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7328273B2 (en) * 2001-03-26 2008-02-05 Duaxes Corporation Protocol duplicating configuration and method for replicating and order of reception of packets on a first and second computers using a shared memory queue
US7433952B1 (en) 2002-04-22 2008-10-07 Cisco Technology, Inc. System and method for interconnecting a storage area network
GB2410106B (en) * 2002-09-09 2006-09-13 Commvault Systems Inc Dynamic storage device pooling in a computer system
US7460473B1 (en) * 2003-02-14 2008-12-02 Istor Networks, Inc. Network receive interface for high bandwidth hardware-accelerated packet processing
US6941396B1 (en) * 2003-02-19 2005-09-06 Istor Networks, Inc. Storage controller redundancy using bi-directional reflective memory channel
US7831736B1 (en) 2003-02-27 2010-11-09 Cisco Technology, Inc. System and method for supporting VLANs in an iSCSI
US6965956B1 (en) * 2003-02-28 2005-11-15 3Ware, Inc. Disk array controller and system with automated detection and control of both ATA and SCSI disk drives
US7526527B1 (en) 2003-03-31 2009-04-28 Cisco Technology, Inc. Storage area network interconnect server
WO2004090740A1 (en) 2003-04-03 2004-10-21 Commvault Systems, Inc. System and method for dynamically sharing media in a computer network
US7334064B2 (en) * 2003-04-23 2008-02-19 Dot Hill Systems Corporation Application server blade for embedded storage appliance
US7451208B1 (en) 2003-06-28 2008-11-11 Cisco Technology, Inc. Systems and methods for network address failover
US7360148B2 (en) * 2003-07-15 2008-04-15 Agere Systems Inc. Reduction checksum generator and a method of calculation thereof
US7103826B2 (en) * 2003-07-31 2006-09-05 Hewlett-Packard Development Company, L.P. Memory system and controller for same
US7685384B2 (en) * 2004-02-06 2010-03-23 Globalscape, Inc. System and method for replicating files in a computer network
US9178784B2 (en) 2004-04-15 2015-11-03 Raytheon Company System and method for cluster management based on HPC architecture
US8335909B2 (en) 2004-04-15 2012-12-18 Raytheon Company Coupling processors to each other for high performance computing (HPC)
US8336040B2 (en) 2004-04-15 2012-12-18 Raytheon Company System and method for topology-aware job scheduling and backfilling in an HPC environment
JP4718851B2 (en) * 2004-05-10 2011-07-06 株式会社日立製作所 Data migration in storage systems
WO2006053084A2 (en) 2004-11-05 2006-05-18 Commvault Systems, Inc. Method and system of pooling storage devices
US7490207B2 (en) 2004-11-08 2009-02-10 Commvault Systems, Inc. System and method for performing auxillary storage operations
TWI344602B (en) * 2005-01-13 2011-07-01 Infortrend Technology Inc Redundant storage virtualization computer system
US7590885B2 (en) * 2005-04-26 2009-09-15 Hewlett-Packard Development Company, L.P. Method and system of copying memory from a source processor to a target processor by duplicating memory writes
JP2007087263A (en) * 2005-09-26 2007-04-05 Nec Corp Dump method in computer system having mirror memory configuration, dump control mechanism, and dump program
GB0610474D0 (en) * 2006-05-26 2006-07-05 Ibm Storage area network controller
GB0613239D0 (en) * 2006-07-04 2006-08-09 Ibm Storage area network system
JP2008046685A (en) 2006-08-10 2008-02-28 Fujitsu Ltd Duplex system and system switching method
US8972613B2 (en) * 2006-10-31 2015-03-03 Hewlett-Packard Development Company, L.P. System and method for increasing input/output throughput in a data storage system
US20080162984A1 (en) * 2006-12-28 2008-07-03 Network Appliance, Inc. Method and apparatus for hardware assisted takeover
DE602008004088D1 (en) * 2007-01-03 2011-02-03 Raytheon Co COMPUTER STORAGE SYSTEM
US8446846B1 (en) * 2007-02-02 2013-05-21 Radisys Canada Ulc Method of passing signal events through a voice over IP audio mixer device
JP5045229B2 (en) * 2007-05-14 2012-10-10 富士ゼロックス株式会社 Storage system and storage device
US8131957B2 (en) 2007-08-16 2012-03-06 International Business Machines Corporation Splitting writes between a storage controller and replication engine
US8024534B2 (en) * 2007-08-16 2011-09-20 International Business Machines Corporation Replication engine communicating with a splitter to split writes between a storage controller and replication engine
CA2729247C (en) * 2008-06-23 2018-02-27 Sntech, Inc. Data transfer between motors
US8271995B1 (en) * 2008-11-10 2012-09-18 Google Inc. System services for native code modules
US8151051B2 (en) * 2009-04-23 2012-04-03 International Business Machines Corporation Redundant solid state disk system via interconnect cards
US20110185099A1 (en) * 2010-01-28 2011-07-28 Lsi Corporation Modular and Redundant Data-Storage Controller And a Method for Providing a Hot-Swappable and Field-Serviceable Data-Storage Controller
US8635420B2 (en) * 2010-07-22 2014-01-21 Susan Elkington Resilient mirroring utilizing peer-to-peer storage
US8402183B2 (en) * 2010-10-06 2013-03-19 Lsi Corporation System and method for coordinating control settings for hardware-automated I/O processors
US8621603B2 (en) * 2011-09-09 2013-12-31 Lsi Corporation Methods and structure for managing visibility of devices in a clustered storage system
US8689044B2 (en) * 2011-11-16 2014-04-01 Hewlett-Packard Development Company, L.P. SAS host controller cache tracking
CN102629225B (en) * 2011-12-31 2014-05-07 华为技术有限公司 Dual-controller disk array, storage system and data storage path switching method
GB2499822B (en) 2012-02-29 2020-01-08 Metaswitch Networks Ltd Failover processing
US8694826B2 (en) * 2012-02-29 2014-04-08 Hewlett-Packard Development Company, L.P. SAS host cache control
US8977922B2 (en) * 2012-03-22 2015-03-10 Broadcom Corporation ACK-NACK signaling enhancements
US9910808B2 (en) 2012-04-30 2018-03-06 Hewlett Packard Enterprise Development Lp Reflective memory bridge for external computing nodes
US10762011B2 (en) 2012-04-30 2020-09-01 Hewlett Packard Enterprise Development Lp Reflective memory bridge for external computing nodes
US9367412B2 (en) 2012-06-25 2016-06-14 Netapp, Inc. Non-disruptive controller replacement in network storage systems
DE102012210816A1 (en) * 2012-06-26 2014-01-02 Siemens Aktiengesellschaft Data packet for a bidirectional transmission of data packets in a data transmission between a first and a second communication device and method for transmitting such a data packet
WO2014158173A1 (en) 2013-03-28 2014-10-02 Hewlett-Packard Development Company, L.P. Implementing coherency with reflective memory
US10108539B2 (en) * 2013-06-13 2018-10-23 International Business Machines Corporation Allocation of distributed data structures
EP3008880A4 (en) * 2013-06-13 2017-01-11 TSX Inc. Apparatus and method for failover of device interconnect using remote memory access with segmented queue
EP2979190A1 (en) * 2013-07-30 2016-02-03 Hewlett-Packard Development Company, L.P. Recovering stranded data
US20150039796A1 (en) * 2013-08-01 2015-02-05 Lsi Corporation Acquiring resources from low priority connection requests in sas
US9239797B2 (en) * 2013-08-15 2016-01-19 Globalfoundries Inc. Implementing enhanced data caching and takeover of non-owned storage devices in dual storage device controller configuration with data in write cache
US8996908B1 (en) * 2013-09-30 2015-03-31 Hitachi, Ltd. Information system, host system and access control method
US9755986B1 (en) * 2013-12-19 2017-09-05 EMC IP Holding Company LLC Techniques for tightly-integrating an enterprise storage array into a distributed virtualized computing environment
US9606944B2 (en) 2014-03-20 2017-03-28 International Business Machines Corporation System and method for computer memory with linked paths
US9382011B2 (en) * 2014-04-10 2016-07-05 Pratt & Whitney Canada Corp. Multiple aircraft engine control system and method of communicating data therein
DE102015004580A1 (en) * 2015-04-14 2016-10-20 Airbus Defence and Space GmbH Transmission methods and devices for transmission
US9794366B1 (en) * 2016-10-19 2017-10-17 Red Hat, Inc. Persistent-memory management
US10664189B2 (en) * 2018-08-27 2020-05-26 International Business Machines Corporation Performance in synchronous data replication environments
CN111124255B (en) * 2018-10-31 2023-09-08 伊姆西Ip控股有限责任公司 Data storage method, electronic device and computer program product
US11106384B2 (en) * 2019-05-03 2021-08-31 EMC IP Holding Company, LLC Storage management system and method
US10922009B2 (en) * 2019-07-08 2021-02-16 International Business Machines Corporation Mirroring write operations across data storage devices
CN115203087A (en) * 2021-04-09 2022-10-18 株式会社日立制作所 Storage system and control method of storage system
US11593223B1 (en) 2021-09-02 2023-02-28 Commvault Systems, Inc. Using resource pool administrative entities in a data storage management system to provide shared infrastructure to tenants

Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5146607A (en) * 1986-06-30 1992-09-08 Encore Computer Corporation Method and apparatus for sharing information between a plurality of processing units
US5546535A (en) * 1992-03-13 1996-08-13 Emc Corporation Multiple controller sharing in a redundant storage array
US5761705A (en) * 1996-04-04 1998-06-02 Symbios, Inc. Methods and structure for maintaining cache consistency in a RAID controller having redundant caches
US5928367A (en) * 1995-01-06 1999-07-27 Hewlett-Packard Company Mirrored memory dual controller disk storage system
US5987496A (en) * 1996-12-03 1999-11-16 Mitsubishi Electric Information Technology Center America, Inc. (Ita) Real-time channel-based reflective memory
US6009275A (en) * 1994-04-04 1999-12-28 Hyundai Electronics America, Inc. Centralized management of resources shared by multiple processing units
US6289471B1 (en) * 1991-09-27 2001-09-11 Emc Corporation Storage device array architecture with solid-state redundancy unit
US6289585B1 (en) * 2000-03-10 2001-09-18 Adrian Staruszkiewicz Method of attaching pipes
US6295585B1 (en) * 1995-06-07 2001-09-25 Compaq Computer Corporation High-performance communication method and apparatus for write-only networks
US6330642B1 (en) * 2000-06-29 2001-12-11 Bull Hn Informatin Systems Inc. Three interconnected raid disk controller data processing system architecture
US6397293B2 (en) * 1998-06-23 2002-05-28 Hewlett-Packard Company Storage management system and auto-RAID transaction manager for coherent memory map across hot plug interface
US20030088735A1 (en) * 2001-11-08 2003-05-08 Busser Richard W. Data mirroring using shared buses
US20030101228A1 (en) * 2001-11-02 2003-05-29 Busser Richard W. Data mirroring between controllers in an active-active controller pair
US20030105830A1 (en) * 2001-12-03 2003-06-05 Duc Pham Scalable network media access controller and methods
US20030110233A1 (en) * 2001-12-10 2003-06-12 Vmic Reflective memory system and method capable of dynamically sizing data packets
US20030217211A1 (en) * 2002-05-14 2003-11-20 Rust Robert A. Controller communications over an always-on controller interconnect
US20040117580A1 (en) * 2002-12-13 2004-06-17 Wu Chia Y. System and method for efficiently and reliably performing write cache mirroring

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4984240A (en) * 1988-12-22 1991-01-08 Codex Corporation Distributed switching architecture for communication module redundancy
US6941396B1 (en) * 2003-02-19 2005-09-06 Istor Networks, Inc. Storage controller redundancy using bi-directional reflective memory channel

Patent Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5146607A (en) * 1986-06-30 1992-09-08 Encore Computer Corporation Method and apparatus for sharing information between a plurality of processing units
US6289471B1 (en) * 1991-09-27 2001-09-11 Emc Corporation Storage device array architecture with solid-state redundancy unit
US5546535A (en) * 1992-03-13 1996-08-13 Emc Corporation Multiple controller sharing in a redundant storage array
US6009275A (en) * 1994-04-04 1999-12-28 Hyundai Electronics America, Inc. Centralized management of resources shared by multiple processing units
US5928367A (en) * 1995-01-06 1999-07-27 Hewlett-Packard Company Mirrored memory dual controller disk storage system
US6295585B1 (en) * 1995-06-07 2001-09-25 Compaq Computer Corporation High-performance communication method and apparatus for write-only networks
US5761705A (en) * 1996-04-04 1998-06-02 Symbios, Inc. Methods and structure for maintaining cache consistency in a RAID controller having redundant caches
US5987496A (en) * 1996-12-03 1999-11-16 Mitsubishi Electric Information Technology Center America, Inc. (Ita) Real-time channel-based reflective memory
US6397293B2 (en) * 1998-06-23 2002-05-28 Hewlett-Packard Company Storage management system and auto-RAID transaction manager for coherent memory map across hot plug interface
US6289585B1 (en) * 2000-03-10 2001-09-18 Adrian Staruszkiewicz Method of attaching pipes
US6330642B1 (en) * 2000-06-29 2001-12-11 Bull Hn Informatin Systems Inc. Three interconnected raid disk controller data processing system architecture
US20030101228A1 (en) * 2001-11-02 2003-05-29 Busser Richard W. Data mirroring between controllers in an active-active controller pair
US20030088735A1 (en) * 2001-11-08 2003-05-08 Busser Richard W. Data mirroring using shared buses
US20030105830A1 (en) * 2001-12-03 2003-06-05 Duc Pham Scalable network media access controller and methods
US20030110233A1 (en) * 2001-12-10 2003-06-12 Vmic Reflective memory system and method capable of dynamically sizing data packets
US20030217211A1 (en) * 2002-05-14 2003-11-20 Rust Robert A. Controller communications over an always-on controller interconnect
US20040117580A1 (en) * 2002-12-13 2004-06-17 Wu Chia Y. System and method for efficiently and reliably performing write cache mirroring

Cited By (42)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050246388A1 (en) * 2003-07-02 2005-11-03 Satoshi Yamatake Image database system
US20070055804A1 (en) * 2005-09-07 2007-03-08 Ran Hay Method and apparatus for managing multiple components
US8667212B2 (en) 2007-05-30 2014-03-04 Sandisk Enterprise Ip Llc System including a fine-grained memory and a less-fine-grained memory
US7849356B2 (en) * 2008-01-17 2010-12-07 International Business Machines Corporation Parity data management system apparatus and method
US20090187786A1 (en) * 2008-01-17 2009-07-23 Michael John Jones Parity data management system apparatus and method
US8667001B2 (en) 2008-03-20 2014-03-04 Sandisk Enterprise Ip Llc Scalable database management software on a cluster of nodes using a shared-distributed flash memory
US8255562B2 (en) 2008-06-30 2012-08-28 International Business Machines Corporation Adaptive data throttling for storage controllers
US8751716B2 (en) 2008-06-30 2014-06-10 International Business Machines Corporation Adaptive data throttling for storage controllers
US20090327481A1 (en) * 2008-06-30 2009-12-31 International Business Machines Corporation Adaptive data throttling for storage controllers
US20100229029A1 (en) * 2009-03-06 2010-09-09 Frazier Ii Robert Claude Independent and dynamic checkpointing system and method
US8458581B2 (en) 2009-10-15 2013-06-04 Ansaldo Sts Usa, Inc. System and method to serially transmit vital data from two processors
US20110093767A1 (en) * 2009-10-15 2011-04-21 Sharp William A System and method to serially transmit vital data from two processors
US8856593B2 (en) 2010-04-12 2014-10-07 Sandisk Enterprise Ip Llc Failure recovery using consensus replication in a distributed flash memory system
US8677055B2 (en) 2010-04-12 2014-03-18 Sandisk Enterprises IP LLC Flexible way of specifying storage attributes in a flash memory-based object store
US8700842B2 (en) 2010-04-12 2014-04-15 Sandisk Enterprise Ip Llc Minimizing write operations to a flash memory-based object store
US9047351B2 (en) 2010-04-12 2015-06-02 Sandisk Enterprise Ip Llc Cluster of processing nodes with distributed global flash memory using commodity server technology
US8793531B2 (en) * 2010-04-12 2014-07-29 Sandisk Enterprise Ip Llc Recovery and replication of a flash memory-based object store
US20110307736A1 (en) * 2010-04-12 2011-12-15 Johann George Recovery and replication of a flash memory-based object store
US8868487B2 (en) 2010-04-12 2014-10-21 Sandisk Enterprise Ip Llc Event processing in a flash memory-based object store
US9164554B2 (en) 2010-04-12 2015-10-20 Sandisk Enterprise Ip Llc Non-volatile solid-state storage system supporting high bandwidth and random access
US8666939B2 (en) 2010-06-28 2014-03-04 Sandisk Enterprise Ip Llc Approaches for the replication of write sets
US8954385B2 (en) 2010-06-28 2015-02-10 Sandisk Enterprise Ip Llc Efficient recovery of transactional data stores
US8694733B2 (en) 2011-01-03 2014-04-08 Sandisk Enterprise Ip Llc Slave consistency in a synchronous replication environment
US8543774B2 (en) 2011-04-05 2013-09-24 Ansaldo Sts Usa, Inc. Programmable logic apparatus employing shared memory, vital processor and non-vital communications processor, and system including the same
US8874515B2 (en) 2011-04-11 2014-10-28 Sandisk Enterprise Ip Llc Low level object version tracking using non-volatile memory write generations
US9183236B2 (en) 2011-04-11 2015-11-10 Sandisk Enterprise Ip Llc Low level object version tracking using non-volatile memory write generations
US9135064B2 (en) 2012-03-07 2015-09-15 Sandisk Enterprise Ip Llc Fine grained adaptive throttling of background processes
US9348525B2 (en) * 2014-02-21 2016-05-24 Netapp, Inc. Systems and methods for a storage array-managed initiator cache
US20150242138A1 (en) * 2014-02-21 2015-08-27 Netapp, Inc. Systems and Methods for a Storage Array-Managed Initiator Cache
US10642704B2 (en) 2016-02-19 2020-05-05 Dell Products L.P. Storage controller failover system
US20170242771A1 (en) * 2016-02-19 2017-08-24 Dell Products L.P. Storage controller failover system
US9864663B2 (en) * 2016-02-19 2018-01-09 Dell Products L.P. Storage controller failover system
CN107329920A (en) * 2017-07-06 2017-11-07 中国航空工业集团公司西安飞机设计研究所 A kind of common interface frame design method of reflective memory
US20190034303A1 (en) * 2017-07-27 2019-01-31 International Business Machines Corporation Transfer track format information for tracks in cache at a first processor node to a second process node to which the first processor node is failing over
US10540246B2 (en) * 2017-07-27 2020-01-21 International Business Machines Corporation Transfer track format information for tracks in cache at a first processor node to a second process node to which the first processor node is failing over
US10572355B2 (en) * 2017-07-27 2020-02-25 International Business Machines Corporation Transfer track format information for tracks in cache at a primary storage system to a secondary storage system to which tracks are mirrored to use after a failover or failback
US20190034302A1 (en) * 2017-07-27 2019-01-31 International Business Machines Corporation Transfer track format information for tracks in cache at a primary storage system to a secondary storage system to which tracks are mirrored to use after a failover or failback
US11157376B2 (en) 2017-07-27 2021-10-26 International Business Machines Corporation Transfer track format information for tracks in cache at a primary storage system to a secondary storage system to which tracks are mirrored to use after a failover or failback
US11188431B2 (en) 2017-07-27 2021-11-30 International Business Machines Corporation Transfer track format information for tracks at a first processor node to a second processor node
US10579296B2 (en) 2017-08-01 2020-03-03 International Business Machines Corporation Providing track format information when mirroring updated tracks from a primary storage system to a secondary storage system
US11243708B2 (en) 2017-08-01 2022-02-08 International Business Machines Corporation Providing track format information when mirroring updated tracks from a primary storage system to a secondary storage system
US10613951B2 (en) 2017-09-13 2020-04-07 International Business Machines Corporation Memory mirror invocation upon detecting a correctable error

Also Published As

Publication number Publication date
US7353306B2 (en) 2008-04-01
US20050210317A1 (en) 2005-09-22
US6941396B1 (en) 2005-09-06

Similar Documents

Publication Publication Date Title
US7353306B2 (en) Storage controller redundancy using bi-directional reflective memory channel
US6493343B1 (en) System and method for implementing multi-pathing data transfers in a system area network
US6545981B1 (en) System and method for implementing error detection and recovery in a system area network
US8576843B2 (en) Packet format for a distributed system
US7734720B2 (en) Apparatus and system for distributing block data on a private network without using TCP/IP
US7886298B2 (en) Data transfer protocol for data replication between multiple pairs of storage controllers on a san fabric
US6721806B2 (en) Remote direct memory access enabled network interface controller switchover and switchback support
US6735647B2 (en) Data reordering mechanism for high performance networks
CA2483197C (en) System, method, and product for managing data transfers in a network
US20030105931A1 (en) Architecture for transparent mirroring
US20060212563A1 (en) Split socket send queue apparatus and method with efficient queue flow control, retransmission and sack support mechanisms
JPH0981487A (en) Network data transfer method
EP1899830A2 (en) Automated serial protocol target port transport layer retry mechanism
US7305605B2 (en) Storage system
KR20030075198A (en) Fault isolation through no-overhead link level crc
US11269557B2 (en) System and method for ensuring command order in a storage controller
WO2021177997A1 (en) System and method for ensuring command order in a storage controller
JPH0320094B2 (en)
WO2008057831A2 (en) Large scale multi-processor system with a link-level interconnect providing in-order packet delivery
Li Box-to-box disk mirroring using Ethernet

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: PROMISE TECHNOLOGY, INC., TAIWAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ISTOR NETWORKS, INC.;REEL/FRAME:024804/0232

Effective date: 20100802