US8838723B2 - High availability management system for stateless components in a distributed master-slave component topology - Google Patents

High availability management system for stateless components in a distributed master-slave component topology Download PDF

Info

Publication number
US8838723B2
US8838723B2 US13/617,383 US201213617383A US8838723B2 US 8838723 B2 US8838723 B2 US 8838723B2 US 201213617383 A US201213617383 A US 201213617383A US 8838723 B2 US8838723 B2 US 8838723B2
Authority
US
United States
Prior art keywords
slave
state
heartbeat
component
computer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US13/617,383
Other versions
US20130013723A1 (en
Inventor
Harish Deshmukh
Sridhar Rajagopalan
Roger C. Raphael
Chitrang Shah
Paul S. Taylor
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
International Business Machines Corp
Original Assignee
International Business Machines Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corp filed Critical International Business Machines Corp
Priority to US13/617,383 priority Critical patent/US8838723B2/en
Publication of US20130013723A1 publication Critical patent/US20130013723A1/en
Application granted granted Critical
Publication of US8838723B2 publication Critical patent/US8838723B2/en
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/08Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters
    • H04L43/0805Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters by checking availability
    • H04L43/0817Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters by checking availability by checking functioning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/0703Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation
    • G06F11/0706Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation the processing taking place on a specific hardware platform or in a specific software environment
    • G06F11/0709Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation the processing taking place on a specific hardware platform or in a specific software environment in a distributed system consisting of a plurality of standalone computer nodes, e.g. clusters, client-server systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/0703Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation
    • G06F11/0751Error or fault detection not based on redundancy
    • G06F11/0754Error or fault detection not based on redundancy by exceeding limits
    • G06F11/076Error or fault detection not based on redundancy by exceeding limits by exceeding a count or rate limit, e.g. word- or bit count limit
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/10Active monitoring, e.g. heartbeat, ping or trace-route
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/16Threshold monitoring

Definitions

  • Embodiments of the invention relate generally to distributed systems and more specifically to techniques for managing availability of one or more components in a distributed system.
  • Distributed applications include software components running on different computers and interacting with each other via a network, frequently as part of a larger distributed system.
  • Examples of distributed applications include scalable information retrieval and/or document processing applications (such as content management environments) that exploit machine-level parallelism.
  • a distributed application may be configured to detect system failures, such as failures in hardware, software, network, etc.
  • Embodiments of the invention provide a computer-implemented method, computer program product, and system for performing an operation for managing availability of a plurality of slave components executing in a distributed system.
  • the operation may generally include sending a plurality of multicast heartbeat requests to the plurality of slave components executing in the distributed system.
  • the operation may also include monitoring, for each multicast heartbeat request, whether a unicast heartbeat acknowledgment is received from the respective slave component, responsive to the respective multicast heartbeat request.
  • the operation may also include assigning, upon determining that a count of unicast heartbeat acknowledgments not received from one of the slave components exceeds a predefined threshold, an availability state to the slave component to indicate a health status of the slave component.
  • FIGS. 1A-1B are block diagrams illustrating systems for managing availability of slave components, according to one embodiment of the invention.
  • FIG. 2 illustrates a system for managing availability of slave components, according to one embodiment of the invention.
  • FIG. 3 illustrates a state machine for a master component, according to one embodiment of the invention.
  • FIG. 4 illustrates a state machine for a slave component, according to one embodiment of the invention.
  • FIG. 5 is a flowchart depicting a method for managing availability of slave components in a distributed system, according to one embodiment of the invention.
  • FIG. 6 is a flowchart depicting a method for a master component to participate in an availability protocol of a distributed system, according to one embodiment of the invention.
  • FIG. 7 is a flowchart depicting a method for a slave component to participate in an availability protocol of a distributed system, according to one embodiment of the invention.
  • FIG. 8 depicts exemplary source code defining a data structure of a multicast heartbeat acknowledgment and/or a unicast heartbeat response, according to one embodiment of the invention.
  • Embodiments of the invention provide techniques for managing availability of slave components in a distributed system.
  • One embodiment of the invention provides a master server and a heartbeat protocol.
  • the heartbeat protocol may define state machines for the master server and/or the slave components.
  • the master server and/or the slave components act in accordance with the state machines to manage availability of the slave components in the distributed system.
  • the heartbeat protocol may specify that the master server manages availability of the slave components by sending multicast heartbeat requests to the slave components and monitoring whether unicast heartbeat acknowledgments are received from the slave components. Accordingly, the master server may more effectively manage the availability of the slave components executing in the distributed system.
  • aspects of the present invention may be embodied as a system, method or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
  • the computer readable medium may be a computer readable signal medium or a computer readable storage medium.
  • a computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing.
  • a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus or device.
  • a computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof.
  • a computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
  • Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
  • Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
  • the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
  • LAN local area network
  • WAN wide area network
  • Internet Service Provider for example, AT&T, MCI, Sprint, EarthLink, MSN, GTE, etc.
  • These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
  • the computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • each block in the flowchart or block diagrams may represent a module, segment or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s).
  • the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
  • FIGS. 1A-1B are a block diagrams illustrating systems 100 , 150 for managing availability of slave components, according to one embodiment of the invention.
  • the networked system 100 includes a computer 102 .
  • the computer 102 may also be connected to other computers via the network 130 .
  • the network 130 may be a telecommunications network and/or a wide area network (WAN).
  • the network 130 is a local area network (LAN).
  • the network 130 is the Internet.
  • the computer 102 generally includes a processor 104 connected via a bus 112 to a memory 106 , a network interface device 110 , a storage 108 , an input device 114 , and an output device 116 .
  • the computer 102 is generally under the control of an operating system 146 . Examples of operating systems include UNIX, versions of the Microsoft Windows® operating system, and distributions of the Linux® operating system. More generally, any operating system supporting the functions disclosed herein may be used.
  • the processor 104 is included to be representative of a single CPU, multiple CPUs, a single CPU having multiple processing cores, and the like.
  • the memory 106 may be a random access memory.
  • the memory 106 is shown as a single identity, it should be understood that the memory 106 may comprise a plurality of modules, and that the memory 106 may exist at multiple levels, from high speed registers and caches to lower speed but larger DRAM chips.
  • the network interface device 110 may be any type of network communications device allowing the computer 102 to communicate with other computers via the network 130 .
  • the storage 108 may be a persistent storage device. Although the storage 108 is shown as a single unit, the storage 108 may be a combination of fixed and/or removable storage devices, such as fixed disc drives, floppy disc drives, tape drives, removable memory cards or optical storage.
  • the memory 106 and the storage 108 may be part of one virtual address space spanning multiple primary and secondary storage devices.
  • the input device 114 may be any device for providing input to the computer 102 .
  • a keyboard, keypad, light pen, touch-screen, track-ball, or speech recognition unit, audio/video player, and the like may be used.
  • the output device 116 may be any device for providing output to a user of the computer 102 .
  • the output device 116 may be any conventional display screen or set of speakers, along with their respective interface cards, i.e., video cards and sound cards (not shown).
  • the output device 116 and input device 114 may be combined.
  • a display screen with an integrated touch-screen, a display with an integrated keyboard, or a speech recognition unit combined with a text speech converter may be used.
  • the networked system 100 includes multiple computers connected via the network 130 and each executing a component.
  • a component refers to an application that encapsulates a set of related functions (and/or data). Further, each component may be configured to act as a standalone server. Each component may also be configured to act in accordance with a slave state machine or a master state machine. A component acting in accordance with a slave state machine may be referred to as a slave component, while a component acting in accordance with a master state machine may be referred to as a master component. Depending on the embodiment, a component may include native source code or interpreted code.
  • An example of a component is a software application that provides a service to one or more applications locally and/or via the network 130 (i.e., a server component).
  • the memory 106 of the computer 102 includes an application 150 for managing availability of slave components, according to one embodiment of the invention.
  • the application 150 may manage, via the network 130 , availability of one or more slave components and may correspond to the master component described above.
  • the application 150 may also be referred to as a master component, a master server or a master process.
  • the master component and the slave components form a distributed system. That is, the distributed system includes a set of components deployed across a set of nodes, where the set of components include at least one master component and at least one slave component.
  • the master component may provide a service of the distributed application to the outside world (e.g., one or more devices, applications and/or users).
  • interaction of the master component with the outside world may be transactional, such as where the application 150 operates in conjunction with the slave components to provide a distributed document processing or information retrieval system.
  • interaction of the master component with the outside world may be stateful—e.g., the master component may process requests from a client and persist the state from previous exchanges with the client.
  • each slave component provides a service to the master component and/or one or more other slave components.
  • the exchanges that occur between a slave component and a master component and/or another slave component may be stateless. In other words, the exchanges may be atomic and no information is necessarily carried across the exchanges.
  • FIG. 1B illustrates the networked system 150 for managing availability of slave components, according to one embodiment of the invention.
  • the computer 102 1 is connected to other computers 102 2 , 102 3 , 102 4 , via the network 130 .
  • the other computers 102 2 , 102 3 , 102 4 each execute a respective slave component 134 .
  • the computers 102 2 , 102 3 , 102 4 are shown executing only one slave component 134 , depending on the embodiment, each computer 102 1 , 102 2 , 102 3 , 102 4 , may execute multiple slave components 134 .
  • the computer 102 1 executing the master component may also execute one or more slave components 134 .
  • the master component may include at least two subcomponents (not shown): a multicast heartbeat request transmitter and a unicast heartbeat acknowledgment receiver. Further, each slave component 134 may include at least two subcomponents (not shown): a multicast heartbeat request receiver and a unicast heartbeat acknowledgment transmitter. In one embodiment, each slave component 134 is stateless. In an alternative embodiment, at least one of the slave components 134 is stateless, while the master component is stateful.
  • the application 150 communicates with the slave components 134 using multicast.
  • Multicast refers to a technique for one-to-many communication over a network.
  • Internet Protocol (IP) multicast refers to a technique for one-to-many communication over an IP infrastructure in the network.
  • IP multicast the application 150 may send a datagram once, even if the datagram is to be delivered to multiple slave components. Nodes in the network (e.g., routers or other computing devices) replicate the packet to reach the multiple slave components.
  • the IP multicast functionality upon which the master component relies, is implemented in a software driver within the master component. In an alternative embodiment, the IP multicast functionality is implemented in a network controller and/or routers that form the network 130 .
  • the application 150 may use an IP multicast group address to send datagrams.
  • the application 150 may use the IP multicast group address as the IP destination address in the datagrams.
  • the slave components may use the IP multicast group address to join the IP multicast group.
  • the slave components may use Internet Group Management Protocol (IGMP) to join an IP multicast group.
  • IGMP Internet Group Management Protocol
  • the slave component receives, via the IP multicast group address, datagrams send by the application 150 .
  • IPv4 Internet Protocol version 4
  • addresses 224.0.0.0 to 239.255.255.255 may be designated as multicast addresses.
  • the application 150 may send a single datagram (from a unicast address of the application 150 ) to a multicast address.
  • IPv4 Internet Protocol version 4
  • IPv6 Internet Protocol version 6
  • the application 150 communicates with the slave components 134 according to a heartbeat protocol to manage availability of the slave components 134 .
  • FIG. 2 illustrates a system 200 for managing availability of slave components 134 , according to one embodiment of the invention.
  • the system 200 includes the application 150 , slave components 204 , 206 , 208 , a watch dog and node agent 218 , and an administrative application 220 .
  • the master component may be a content indexing controller that is connected to a data store 216 .
  • the slave components 204 , 206 , 208 may be of different types, e.g., a key server slave, an index server slave, and a translation slave.
  • each slave component type may be designated by an administrative user as being critical or non-critical.
  • the application 150 may be configured to terminate, whereby the distributed system becomes unavailable.
  • the administrative user may designate the key server slave and the index server slave as being critical and the translation slave as being non-critical. If no instances of the key server slave component are available or if no instances of the translation slave component are available, the application 150 may terminate itself—regardless of the number of instances of the translation slave component being available.
  • an administrative user may specify the number of instances required for each critical slave component type to prevent the application 150 from terminating itself.
  • the watch dog/node agent 218 , 214 refers to functionality provided by an operating system to automatically restart a component if the component terminates.
  • SCM Service Control Manager
  • the administrative application 220 may be an administrative console configured to manage all the software components and other artifacts of the system, based on user input.
  • one slave component 208 includes a proxy server process 210 for a component 212 that executes in an address space that is closed to the rest of the distributed system.
  • the component 212 is translation server. Further, the component 212 may also be closed to modification and/or extension by a developer of the distributed application.
  • the proxy server process 210 initiates the component 212 and analyzes a state of the component 212 to determine the availability of the component 212 .
  • the proxy server process 210 may analyze the state of the component 212 using inter-process communication (IPC) (such as redirecting a shared library or using named pipes) and/or operating system primitives (such as the UNIX kill function or SIGCHLD signal).
  • IPC inter-process communication
  • SIGCHLD operating system primitives
  • the proxy server process 210 may then transmit the determined availability to the distributed system via an availability protocol of the distributed system. Accordingly, the application 150 monitors, via the proxy server process 210 , the availability of the component having the closed address space. The distributed system may also control the availability of the component via the proxy server process 210 .
  • the heartbeat protocol specifies that the application 150 and the slave components 134 act according to predefined state machines.
  • FIG. 3 illustrates a state machine 300 for the application 150 , according to one embodiment of the invention.
  • the application 150 initializes a suspect state for each slave component (states 302 and 304 ).
  • the application 150 may transition into the one of the predefined states of the state machine as the application 150 sends multicast heartbeat requests to the slave components 134 . That is, the application 150 may maintain one of the predefined states for each of the slave components 134 in the distributed system. If the application 150 receives N unicast heartbeat acknowledgments from a slave component 134 , the application 150 may transition from a suspect state 304 to a live state 306 for the slave component 134 .
  • the value of N may be a predetermined numeric value that is tunable to the conditions of the network 130 , e.g., response times, congestion attributes, etc.
  • N may be adapted autonomically using feedback such as slave component function type, message performance history and dynamic conditions of the node and/or network, etc.
  • the application 150 may also transition from the suspect state 304 to the live state 306 .
  • the threshold percentage is 75% of expected unicast heartbeat acknowledgments (i.e., up to 25% of unicast heartbeat acknowledgments are permitted to be missed).
  • the application 150 may still transition from the suspect state 304 to the live state 306 for the slave component 134 .
  • the application 150 may transition from the live state 306 to the suspect state 304 for the slave component 134 . Further, if the application 150 is in the suspect state 304 for a slave component 134 and receives fewer unicast heartbeat acknowledgments from the slave component 134 than a second threshold percentage, the application 150 may transition from the suspect state 304 to an unknown state 308 for the slave component 134 . For example, assume that the second threshold percentage is 50% of expected unicast heartbeat acknowledgments. In one embodiment, if the application 150 receives less than 5 out of 10 heartbeat acknowledgments expected from the slave component 134 , the application 150 may transition from the suspect state 304 to the unknown state 308 for the slave component 134 .
  • the application 150 may ignore any heartbeat acknowledgments from the slave component 134 and attempt to kill and restart the slave component 134 .
  • the application 150 may transition from the unknown state 308 to a kill state 310 for the slave component 134 .
  • the application 150 may attempt to terminate the slave component (state 310 ) and verify whether the attempt is successful (state 312 ). If so, the application 150 may attempt to restart the slave component (state 314 ) and check whether the attempt is successful (state 316 ). If so, the application 150 transitions from the state 316 to the live state 306 for the restarted slave component 134 .
  • the application 150 may transition to a retry state 318 for the slave component 134 , according to one embodiment.
  • the application 150 may maintain a counter associated with the retry state 318 .
  • the application 150 may increment the counter upon each transition into the retry state 318 . If the counter is less than a predefined “retry limit” threshold, the application 150 may transition from the retry state 318 back to the unknown state 308 for the slave component 134 .
  • the application 150 may transition from the retry state 318 to a lease start state 320 for the slave component, according to one embodiment.
  • the application 150 may mark a beginning of a lease for the slave component 134 .
  • a lease refers to a predefined time period before declaring that a slave component 134 is dead. The lease is further described below in conjunction with FIG. 4 .
  • the heartbeat protocol also specifies that the application 150 may be restarted without requiring any of the slave components 134 to be terminated or restarted.
  • each slave component 134 may inspect and extract a source IP address from a heartbeat request previously received from the application 150 .
  • Each slave component 134 may then respond with a heartbeat acknowledgment to a surviving master, based on the extracted source IP address.
  • the source IP address may be obtained by introspecting the underlying IP datagram headers.
  • the source IP address and port of the sender may be extracted from the payload if the individual slave components 134 include a respective IP address within the payload. Extracting the source IP address and port of the sender is further discussed below in conjunction with FIG. 8 .
  • FIG. 4 illustrates a state machine 400 for a slave component 134 , according to one embodiment of the invention.
  • the slave component 134 transitions into a “master live” state 402 .
  • the slave component transitions from the master live state 402 to a master suspect state 404 .
  • the value of M may be a predetermined numeric value that is tunable to the conditions of the network 130 , e.g., response times, congestion attributes, master load, number of slave components in the network, etc.
  • the slave component 134 when in the master suspect state 404 , the slave component 134 may start or renew a lease. Further, if the slave component 134 receives a next heartbeat request, then the slave component 134 may renew a lease and transition back to the master live state 402 . In an alternative embodiment, upon starting, the slave component 134 may transition into the master suspect state 404 (as shown by dotted lines in FIG. 4 ) rather than the master live state 402 .
  • the slave component 134 may also transition to a time drift state relative to the master component.
  • the time drift state is taken if the local, real-time clock of the processor in question has drifted beyond a predefined, acceptable threshold, relative to a local clock of the master component. Accordingly, the expiration of leases for participating nodes in the network may be managed more efficiently. If the drift is beyond a tolerable limit, the slave component 134 may then transition from the time drift state 403 to a suicide state 408 .
  • the slave component 134 may transition into a master dead state 406 , according to one embodiment.
  • the slave component 134 may then transition into the suicide state 408 .
  • the slave component 134 transitions from the master suspect state 404 to the suicide state 408 in no more than 1.5 lease periods.
  • the lease period refers to a quantity of elapsed time that is used by a slave component to form a time window within which to count how many multicast heartbeat requests have arrived.
  • the master component may also use a lease period to form a time window within which to count how many unicast heartbeat acknowledgments have been received.
  • the lease periods used by the slave components and the master component may be the same or differ and may be adapted based on node capability and/or network performance.
  • the slave component 134 terminates.
  • an external node agent may specify that the slave component 134 does not restart itself (i.e., the slave component 134 ) by marking a local persistent state to indicate a fatal-and-no-restart condition. This condition of not restarting an instance of the slave component 134 may only be cleared administratively, according to one embodiment.
  • FIG. 5 is a flowchart depicting a method 500 for managing availability of slave components 134 in a distributed system, according to one embodiment of the invention.
  • the method 500 begins at step 510 , where the application 150 sends multicast heartbeat requests to slave components 134 executing in the distributed system.
  • the application 150 enters a loop for the sent multicast heartbeat requests.
  • the application 150 enters a loop for each slave component 134 in a live state.
  • the application monitors whether a unicast heartbeat acknowledgment is received from a slave component 134 , responsive to each multicast heartbeat request.
  • the application 150 determines whether a threshold count of unicast heartbeat acknowledgments is not received from a slave component 134 in a live state from a last iteration (of the loop of the step 525 ) within a predefined time window. If so, the application 150 assigns an availability state to the slave component to indicate that the slave component may be unhealthy (step 550 ). If the application 150 determines that less than the threshold count of unicast heartbeat acknowledgments is not received from the slave component (step 540 ), or after the step 550 , the method 500 may proceed to step 555 , where the application 150 determines whether more slave components 134 in a live state remain. If so, the method 500 returns to the step 525 to process another slave component 134 .
  • the application 150 determines whether more multicast heartbeat requests remain (step 560 ). If so, the application 150 may return to the step 530 to process a unicast heartbeat acknowledgment for another multicast heartbeat request. After the step 560 , the method 500 terminates.
  • FIG. 6 is a flowchart depicting a method 600 for a master component to participate in an availability protocol of a distributed system, according to one embodiment of the invention.
  • the method 600 begins at step 602 , where the application 150 enters a loop for each slave component 134 .
  • the application 150 sets the respective slave component 134 to a suspect state.
  • the application 150 determines whether a unicast heartbeat acknowledgment is received—or is received within a threshold percentage to be missed—responsive to a multicast heartbeat request. If not, the application 150 determines whether unicast heartbeat acknowledgments have been missed beyond a predefined threshold percentage (step 608 ). If not, the method 600 returns to the step 606 .
  • the application 150 determines that a unicast heartbeat acknowledgment is received—or is received within a threshold percentage to be missed (step 606 )—the application 150 sets the respective slave component 134 to a live state (step 610 ), according to one embodiment. Whether the unicast heartbeat acknowledgment is received may be determined relative to a predefined window of time. In an alternative embodiment, the application 150 may determine whether a minimum threshold percentage of received heartbeat acknowledgements is received. The application 150 may then determine whether a unicast heartbeat acknowledgment is missed (step 612 ). If not, the method 600 returns to the step 612 . Otherwise, the method 600 returns to the step 604 .
  • the application 150 determines whether the unicast heartbeat acknowledgment is received, relative to a predefined window of time. In one embodiment, the application 150 may use different windows of time, based on the type of slave component 134 . In alternative embodiments, the application 150 may use different windows of time, based on node capability and/or network performance.
  • the application 150 determines that a unicast heartbeat acknowledgment is missed beyond a threshold percentage (step 608 )
  • the application 150 sets the respective slave component 134 to an unknown state and begins to ignore any unicast heartbeat request from the respective slave component (step 614 ), according to one embodiment.
  • the application 150 may then attempt to kill the respective slave component 134 and determine whether the attempt is successful (step 616 ). If so, the application 150 may then attempt to launch the respective slave component 134 and determine whether the attempt is successful (step 618 ). If so, the method 600 returns to step 610 , and the application 150 no longer ignores unicast heartbeat requests from the respective slave component 134 .
  • the application 150 may determine whether a retry limit has been exceeded (step 620 ), according to one embodiment. The application 150 may then mark a start of a lease for the respective slave component (step 622 ). The application 150 may then determine whether the lease has ended for the respective slave component (step 624 ). If not, the application 150 may wait on the lease for the respective slave component (step 626 ) and return to the step 624 . Otherwise, the application 150 may set the respective slave component 134 to a dead state (step 628 ). At step 630 , the application 150 may determine whether more slave components remain to be processed.
  • the method 600 may return to the step 602 , where the application 150 processes a next slave component 134 in the loop. Otherwise, the method 600 terminates. Alternatively, instead of terminating the method 600 , the application 150 may repeat the loop of the step 602 .
  • FIG. 7 is a flowchart depicting a method 700 for a slave component 134 to participate in an availability protocol of a distributed system, according to one embodiment of the invention.
  • the method 700 begins at step 702 , where the slave component 134 sets the master component to a suspect state.
  • the slave component 134 determines whether a multicast heartbeat request is received from the master component. If not, the slave component 134 determines whether a lease for the slave component 134 has expired (step 706 ). If not, the method 700 returns to the step 704 .
  • the slave component 134 may also increment a counter based on a received multicast heartbeat request. Further, depending on the embodiment, the slave components 134 and the master component may determine whether a multicast heartbeat request or unicast heartbeat acknowledgment is received based on an adaptable time window and prior to changing state.
  • the slave component may renew a lease for the slave component 134 and set the master component to a live state (step 708 ), according to one embodiment.
  • the slave component 134 may then determine whether a multicast heartbeat request is missed (step 710 ). If not, the method 700 returns to the step 710 . Otherwise, the method 700 returns to the step 702 .
  • the slave component 134 may set the master component to an unknown state (step 712 ), according to one embodiment.
  • the slave component 134 may then terminate itself After the step 714 , the method 700 terminates.
  • the method 700 may proceed to step 716 , where the slave component 134 sets the master component to a time drift state. The slave component 134 may then terminate itself (step 714 ).
  • FIG. 8 depicts exemplary source code 802 defining a data structure of a multicast heartbeat acknowledgment and/or a unicast heartbeat response, according to one embodiment of the invention.
  • the source code 802 includes a data type 804 , a variable name 806 and a description 808 .
  • the data structure includes a data gram length, a header length, a signature, a message type, a command code, a domain identifier, a component type, a component identifier, a subcomponent identifier, a source IP address, a source port, an event identifier, a timestamp, a payload length, and payload bytes.
  • the source IP address and source port of a sending component may be sent explicitly rather than extracted from TCP headers of the datagram, due to a possible lack of implementation on some operating system TCP libraries (e.g., TCP libraries that do not support packet introspection).
  • the timestamp may be a local time stamp obtained from either a local operating system on which the master component or the slave components execute.
  • the timestamp may be consolidated at the master component to enforce a de-configuration of a slave component 134 if the clock drift exceeds a predefined threshold, relative to a local clock of the master component.
  • the slave component 134 shuts itself down—pursuant to the availability protocol disclosed herein—if the slave component 134 determines that the local timestamp of the slave component is out of drift from the timestamp transmitted on the multicast heartbeat request from the master component.
  • the master component may subsequently de-configure the slave component 134 .
  • an administrative user and/or an application 150 may specify that a given slave component type as being critical for the entire distributed application.
  • the application 150 may determine to terminate itself (i.e., the application 150 ), thereby bringing down the entire distributed application within a predefined multiple of the lease period.
  • slave component types include three slave component types (e.g., key server, index server, and translation server) shown in FIG. 2 .
  • embodiments of the invention provide techniques for managing the availability of slave components executing in a distributed system.
  • One embodiment of the invention provides a master server and a heartbeat protocol.
  • the heartbeat protocol may define state machines for the master server and/or the slave components.
  • the master server and/or the slave components may act in accordance with the state machines to manage availability of the slave components in the distributed system.
  • the heartbeat protocol may specify that the master server manages availability of the slave components by sending multicast heartbeat requests to the slave components and monitoring whether unicast heartbeat acknowledgments are received from the slave components.
  • the master server may assign an “unknown” state to a slave component and attempt to kill and restart the slave component.

Abstract

Systems, methods and computer program products are disclosed for managing availability of a slave components executing in a distributed system. A master server and a heartbeat protocol may be provided. The heartbeat protocol may define state machines for the master server and/or the slave components. The master server and/or the slave components may act in accordance with the state machines to manage availability of the slave components in the distributed system.

Description

This application is a continuation of co-pending U.S. patent application Ser. No. 12/855,241, filed Aug. 12, 2010. The aforementioned related patent application is herein incorporated by reference in its entirety.
BACKGROUND
Embodiments of the invention relate generally to distributed systems and more specifically to techniques for managing availability of one or more components in a distributed system.
Distributed applications include software components running on different computers and interacting with each other via a network, frequently as part of a larger distributed system. Examples of distributed applications include scalable information retrieval and/or document processing applications (such as content management environments) that exploit machine-level parallelism. A distributed application may be configured to detect system failures, such as failures in hardware, software, network, etc.
SUMMARY
Embodiments of the invention provide a computer-implemented method, computer program product, and system for performing an operation for managing availability of a plurality of slave components executing in a distributed system. The operation may generally include sending a plurality of multicast heartbeat requests to the plurality of slave components executing in the distributed system. The operation may also include monitoring, for each multicast heartbeat request, whether a unicast heartbeat acknowledgment is received from the respective slave component, responsive to the respective multicast heartbeat request. The operation may also include assigning, upon determining that a count of unicast heartbeat acknowledgments not received from one of the slave components exceeds a predefined threshold, an availability state to the slave component to indicate a health status of the slave component.
BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWINGS
So that the manner in which the above recited features, advantages and objects of the present invention are attained and can be understood in detail, a more particular description of the invention, briefly summarized above, may be had by reference to the embodiments thereof which are illustrated in the appended drawings.
It is to be noted, however, that the appended drawings illustrate only typical embodiments of this invention and are therefore not to be considered limiting of its scope, for the invention may admit to other equally effective embodiments.
FIGS. 1A-1B are block diagrams illustrating systems for managing availability of slave components, according to one embodiment of the invention.
FIG. 2 illustrates a system for managing availability of slave components, according to one embodiment of the invention.
FIG. 3 illustrates a state machine for a master component, according to one embodiment of the invention.
FIG. 4 illustrates a state machine for a slave component, according to one embodiment of the invention.
FIG. 5 is a flowchart depicting a method for managing availability of slave components in a distributed system, according to one embodiment of the invention.
FIG. 6 is a flowchart depicting a method for a master component to participate in an availability protocol of a distributed system, according to one embodiment of the invention.
FIG. 7 is a flowchart depicting a method for a slave component to participate in an availability protocol of a distributed system, according to one embodiment of the invention.
FIG. 8 depicts exemplary source code defining a data structure of a multicast heartbeat acknowledgment and/or a unicast heartbeat response, according to one embodiment of the invention.
DETAILED DESCRIPTION
Embodiments of the invention provide techniques for managing availability of slave components in a distributed system. One embodiment of the invention provides a master server and a heartbeat protocol. The heartbeat protocol may define state machines for the master server and/or the slave components. The master server and/or the slave components act in accordance with the state machines to manage availability of the slave components in the distributed system. For example, the heartbeat protocol may specify that the master server manages availability of the slave components by sending multicast heartbeat requests to the slave components and monitoring whether unicast heartbeat acknowledgments are received from the slave components. Accordingly, the master server may more effectively manage the availability of the slave components executing in the distributed system.
In the following, reference is made to embodiments of the invention. However, it should be understood that the invention is not limited to specific described embodiments. Instead, any combination of the following features and elements, whether related to different embodiments or not, is contemplated to implement and practice the invention. Furthermore, although embodiments of the invention may achieve advantages over other possible solutions and/or over the prior art, whether or not a particular advantage is achieved by a given embodiment is not limiting of the invention. Thus, the following aspects, features, embodiments and advantages are merely illustrative and are not considered elements or limitations of the appended claims except where explicitly recited in a claim(s). Likewise, reference to “the invention” shall not be construed as a generalization of any inventive subject matter disclosed herein and shall not be considered to be an element or limitation of the appended claims except where explicitly recited in a claim(s).
As will be appreciated by one skilled in the art, aspects of the present invention may be embodied as a system, method or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
Any combination of one or more computer readable medium(s) may be utilized. The computer readable medium may be a computer readable signal medium or a computer readable storage medium. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus or device.
A computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
Aspects of the present invention are described below with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
The computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
The flowchart and block diagrams in the Figures illustrate the architecture, functionality and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
FIGS. 1A-1B are a block diagrams illustrating systems 100, 150 for managing availability of slave components, according to one embodiment of the invention. As shown in FIG. 1A, the networked system 100 includes a computer 102. The computer 102 may also be connected to other computers via the network 130. In general, the network 130 may be a telecommunications network and/or a wide area network (WAN). In a particular embodiment, the network 130 is a local area network (LAN). In an alternative embodiment, the network 130 is the Internet.
The computer 102 generally includes a processor 104 connected via a bus 112 to a memory 106, a network interface device 110, a storage 108, an input device 114, and an output device 116. The computer 102 is generally under the control of an operating system 146. Examples of operating systems include UNIX, versions of the Microsoft Windows® operating system, and distributions of the Linux® operating system. More generally, any operating system supporting the functions disclosed herein may be used. The processor 104 is included to be representative of a single CPU, multiple CPUs, a single CPU having multiple processing cores, and the like. Similarly, the memory 106 may be a random access memory. While the memory 106 is shown as a single identity, it should be understood that the memory 106 may comprise a plurality of modules, and that the memory 106 may exist at multiple levels, from high speed registers and caches to lower speed but larger DRAM chips. The network interface device 110 may be any type of network communications device allowing the computer 102 to communicate with other computers via the network 130.
The storage 108 may be a persistent storage device. Although the storage 108 is shown as a single unit, the storage 108 may be a combination of fixed and/or removable storage devices, such as fixed disc drives, floppy disc drives, tape drives, removable memory cards or optical storage. The memory 106 and the storage 108 may be part of one virtual address space spanning multiple primary and secondary storage devices.
The input device 114 may be any device for providing input to the computer 102. For example, a keyboard, keypad, light pen, touch-screen, track-ball, or speech recognition unit, audio/video player, and the like may be used. The output device 116 may be any device for providing output to a user of the computer 102. For example, the output device 116 may be any conventional display screen or set of speakers, along with their respective interface cards, i.e., video cards and sound cards (not shown). Although shown separately from the input device 114, the output device 116 and input device 114 may be combined. For example, a display screen with an integrated touch-screen, a display with an integrated keyboard, or a speech recognition unit combined with a text speech converter may be used.
In one embodiment, the networked system 100 includes multiple computers connected via the network 130 and each executing a component. In one embodiment, a component refers to an application that encapsulates a set of related functions (and/or data). Further, each component may be configured to act as a standalone server. Each component may also be configured to act in accordance with a slave state machine or a master state machine. A component acting in accordance with a slave state machine may be referred to as a slave component, while a component acting in accordance with a master state machine may be referred to as a master component. Depending on the embodiment, a component may include native source code or interpreted code. An example of a component is a software application that provides a service to one or more applications locally and/or via the network 130 (i.e., a server component).
As shown, the memory 106 of the computer 102 includes an application 150 for managing availability of slave components, according to one embodiment of the invention. The application 150 may manage, via the network 130, availability of one or more slave components and may correspond to the master component described above. As such, the application 150 may also be referred to as a master component, a master server or a master process.
In one embodiment, the master component and the slave components form a distributed system. That is, the distributed system includes a set of components deployed across a set of nodes, where the set of components include at least one master component and at least one slave component. The master component may provide a service of the distributed application to the outside world (e.g., one or more devices, applications and/or users). Depending on the embodiment, interaction of the master component with the outside world may be transactional, such as where the application 150 operates in conjunction with the slave components to provide a distributed document processing or information retrieval system. Alternatively or additionally, interaction of the master component with the outside world may be stateful—e.g., the master component may process requests from a client and persist the state from previous exchanges with the client.
In one embodiment, each slave component provides a service to the master component and/or one or more other slave components. Depending on the embodiment, the exchanges that occur between a slave component and a master component and/or another slave component may be stateless. In other words, the exchanges may be atomic and no information is necessarily carried across the exchanges.
FIG. 1B illustrates the networked system 150 for managing availability of slave components, according to one embodiment of the invention. As shown, the computer 102 1 is connected to other computers 102 2, 102 3, 102 4, via the network 130. The other computers 102 2, 102 3, 102 4 each execute a respective slave component 134. Although the computers 102 2, 102 3, 102 4 are shown executing only one slave component 134, depending on the embodiment, each computer 102 1, 102 2, 102 3, 102 4, may execute multiple slave components 134. Further, the computer 102 1 executing the master component may also execute one or more slave components 134. The master component may include at least two subcomponents (not shown): a multicast heartbeat request transmitter and a unicast heartbeat acknowledgment receiver. Further, each slave component 134 may include at least two subcomponents (not shown): a multicast heartbeat request receiver and a unicast heartbeat acknowledgment transmitter. In one embodiment, each slave component 134 is stateless. In an alternative embodiment, at least one of the slave components 134 is stateless, while the master component is stateful.
In one embodiment, the application 150 communicates with the slave components 134 using multicast. Multicast refers to a technique for one-to-many communication over a network. Internet Protocol (IP) multicast refers to a technique for one-to-many communication over an IP infrastructure in the network. Using IP multicast, the application 150 may send a datagram once, even if the datagram is to be delivered to multiple slave components. Nodes in the network (e.g., routers or other computing devices) replicate the packet to reach the multiple slave components.
In one embodiment, the IP multicast functionality, upon which the master component relies, is implemented in a software driver within the master component. In an alternative embodiment, the IP multicast functionality is implemented in a network controller and/or routers that form the network 130.
In one embodiment, the application 150 may use an IP multicast group address to send datagrams. For example, the application 150 may use the IP multicast group address as the IP destination address in the datagrams. The slave components may use the IP multicast group address to join the IP multicast group. For example, the slave components may use Internet Group Management Protocol (IGMP) to join an IP multicast group.
In one embodiment, once a slave component has joined the IP multicast group, the slave component receives, via the IP multicast group address, datagrams send by the application 150. For example, under Internet Protocol version 4 (IPv4), addresses 224.0.0.0 to 239.255.255.255 may be designated as multicast addresses. The application 150 may send a single datagram (from a unicast address of the application 150) to a multicast address.
In one embodiment, after the application 150 sends the single datagram to the multicast address, intermediary routers in the network 130 then make copies of the datagram and deliver the copies to all slave components that have registered an interest in receiving datagrams at the multicast address. In response, each slave component sends a unicast datagram to the application 150. Unicast refers to a technique for one-to-one communication over a network. And while embodiments are described herein with reference to Internet Protocol version 4 (IPv4), other embodiments, such as Internet Protocol version 6 (IPv6), are broadly contemplated.
In one embodiment, the application 150 communicates with the slave components 134 according to a heartbeat protocol to manage availability of the slave components 134. FIG. 2 illustrates a system 200 for managing availability of slave components 134, according to one embodiment of the invention. As shown, the system 200 includes the application 150, slave components 204, 206, 208, a watch dog and node agent 218, and an administrative application 220. In one embodiment, the master component may be a content indexing controller that is connected to a data store 216. The slave components 204, 206, 208 may be of different types, e.g., a key server slave, an index server slave, and a translation slave.
In one embodiment, each slave component type may be designated by an administrative user as being critical or non-critical. Upon determining that no instances of a critical component type are available, the application 150 may be configured to terminate, whereby the distributed system becomes unavailable. For example, the administrative user may designate the key server slave and the index server slave as being critical and the translation slave as being non-critical. If no instances of the key server slave component are available or if no instances of the translation slave component are available, the application 150 may terminate itself—regardless of the number of instances of the translation slave component being available. Depending on the embodiment, an administrative user may specify the number of instances required for each critical slave component type to prevent the application 150 from terminating itself.
In one embodiment, the watch dog/ node agent 218, 214 refers to functionality provided by an operating system to automatically restart a component if the component terminates. An example on the Microsoft® Windows® operating system there is the Windows services managed by the Service Control Manager (SCM). The administrative application 220 may be an administrative console configured to manage all the software components and other artifacts of the system, based on user input.
As shown in FIG. 2, one slave component 208 includes a proxy server process 210 for a component 212 that executes in an address space that is closed to the rest of the distributed system. In this particular example, the component 212 is translation server. Further, the component 212 may also be closed to modification and/or extension by a developer of the distributed application. In one embodiment, the proxy server process 210 initiates the component 212 and analyzes a state of the component 212 to determine the availability of the component 212. For example, the proxy server process 210 may analyze the state of the component 212 using inter-process communication (IPC) (such as redirecting a shared library or using named pipes) and/or operating system primitives (such as the UNIX kill function or SIGCHLD signal). The proxy server process 210 may then transmit the determined availability to the distributed system via an availability protocol of the distributed system. Accordingly, the application 150 monitors, via the proxy server process 210, the availability of the component having the closed address space. The distributed system may also control the availability of the component via the proxy server process 210.
In one embodiment, the heartbeat protocol specifies that the application 150 and the slave components 134 act according to predefined state machines. FIG. 3 illustrates a state machine 300 for the application 150, according to one embodiment of the invention. As shown, the application 150 initializes a suspect state for each slave component (states 302 and 304). The application 150 may transition into the one of the predefined states of the state machine as the application 150 sends multicast heartbeat requests to the slave components 134. That is, the application 150 may maintain one of the predefined states for each of the slave components 134 in the distributed system. If the application 150 receives N unicast heartbeat acknowledgments from a slave component 134, the application 150 may transition from a suspect state 304 to a live state 306 for the slave component 134. The value of N may be a predetermined numeric value that is tunable to the conditions of the network 130, e.g., response times, congestion attributes, etc. In one embodiment, N may be adapted autonomically using feedback such as slave component function type, message performance history and dynamic conditions of the node and/or network, etc. However, the value of N=1 has led to acceptable behavior in at least some cases.
Alternatively, in one embodiment, if the application 150 receives more unicast heartbeat acknowledgments from the slave component 134 than a threshold percentage of unicast heartbeat acknowledgments expected to be received, the application 150 may also transition from the suspect state 304 to the live state 306. For example, assume that the threshold percentage is 75% of expected unicast heartbeat acknowledgments (i.e., up to 25% of unicast heartbeat acknowledgments are permitted to be missed). In one embodiment, if the application 150 receives only 7 out of 10 heartbeat acknowledgments expected from the slave component 134, the application 150 may still transition from the suspect state 304 to the live state 306 for the slave component 134.
In one embodiment, if the application 150 is in the live state 306 for a slave component 134 and does not receive a unicast heartbeat acknowledgment from the slave component 134 (responsive to a next multicast heartbeat request), the application 150 may transition from the live state 306 to the suspect state 304 for the slave component 134. Further, if the application 150 is in the suspect state 304 for a slave component 134 and receives fewer unicast heartbeat acknowledgments from the slave component 134 than a second threshold percentage, the application 150 may transition from the suspect state 304 to an unknown state 308 for the slave component 134. For example, assume that the second threshold percentage is 50% of expected unicast heartbeat acknowledgments. In one embodiment, if the application 150 receives less than 5 out of 10 heartbeat acknowledgments expected from the slave component 134, the application 150 may transition from the suspect state 304 to the unknown state 308 for the slave component 134.
In one embodiment, once the application 150 has transitioned to the unknown state 308 for the slave component 134, the application 150 may ignore any heartbeat acknowledgments from the slave component 134 and attempt to kill and restart the slave component 134. For example, the application 150 may transition from the unknown state 308 to a kill state 310 for the slave component 134. The application 150 may attempt to terminate the slave component (state 310) and verify whether the attempt is successful (state 312). If so, the application 150 may attempt to restart the slave component (state 314) and check whether the attempt is successful (state 316). If so, the application 150 transitions from the state 316 to the live state 306 for the restarted slave component 134.
However, if the application 150 does not successfully terminate and restart with a new instance of the slave component 134, the application 150 may transition to a retry state 318 for the slave component 134, according to one embodiment. The application 150 may maintain a counter associated with the retry state 318. The application 150 may increment the counter upon each transition into the retry state 318. If the counter is less than a predefined “retry limit” threshold, the application 150 may transition from the retry state 318 back to the unknown state 308 for the slave component 134.
However, if the counter is greater than or equal to the retry limit, the application 150 may transition from the retry state 318 to a lease start state 320 for the slave component, according to one embodiment. At the lease start state 320, the application 150 may mark a beginning of a lease for the slave component 134. In one embodiment, a lease refers to a predefined time period before declaring that a slave component 134 is dead. The lease is further described below in conjunction with FIG. 4. Once the lease for the slave component 134 elapses (states 322 and 324), the application 150 may transition into a dead state 326 for the slave component 134. The application 150 may then return to the state 302 to monitor another slave component 134.
In one embodiment, the heartbeat protocol also specifies that the application 150 may be restarted without requiring any of the slave components 134 to be terminated or restarted. For example, each slave component 134 may inspect and extract a source IP address from a heartbeat request previously received from the application 150. Each slave component 134 may then respond with a heartbeat acknowledgment to a surviving master, based on the extracted source IP address. For example, the source IP address may be obtained by introspecting the underlying IP datagram headers. In an alternative embodiment, the source IP address and port of the sender may be extracted from the payload if the individual slave components 134 include a respective IP address within the payload. Extracting the source IP address and port of the sender is further discussed below in conjunction with FIG. 8.
FIG. 4 illustrates a state machine 400 for a slave component 134, according to one embodiment of the invention. As shown, the slave component 134 transitions into a “master live” state 402. Upon not receiving M heartbeat requests that are expected from the application 150, the slave component transitions from the master live state 402 to a master suspect state 404. The value of M may be a predetermined numeric value that is tunable to the conditions of the network 130, e.g., response times, congestion attributes, master load, number of slave components in the network, etc. In one embodiment, the value of M may be adapted autonomically using feedback such as a message performance history of the master component, dynamic conditions of the node and/or network, etc. However, the value of M=1 has led to acceptable behavior in at least some cases.
In one embodiment, when in the master suspect state 404, the slave component 134 may start or renew a lease. Further, if the slave component 134 receives a next heartbeat request, then the slave component 134 may renew a lease and transition back to the master live state 402. In an alternative embodiment, upon starting, the slave component 134 may transition into the master suspect state 404 (as shown by dotted lines in FIG. 4) rather than the master live state 402.
In an alternative embodiment, from the master live state 402, the slave component 134 may also transition to a time drift state relative to the master component. The time drift state is taken if the local, real-time clock of the processor in question has drifted beyond a predefined, acceptable threshold, relative to a local clock of the master component. Accordingly, the expiration of leases for participating nodes in the network may be managed more efficiently. If the drift is beyond a tolerable limit, the slave component 134 may then transition from the time drift state 403 to a suicide state 408.
However, if the lease expires while the slave component 134 is in the master suspect state 404, the slave component 134 may transition into a master dead state 406, according to one embodiment. The slave component 134 may then transition into the suicide state 408. In some embodiments, the slave component 134 transitions from the master suspect state 404 to the suicide state 408 in no more than 1.5 lease periods. In one embodiment, the lease period refers to a quantity of elapsed time that is used by a slave component to form a time window within which to count how many multicast heartbeat requests have arrived. Further, as previously described in conjunction with FIG. 3, the master component may also use a lease period to form a time window within which to count how many unicast heartbeat acknowledgments have been received. Depending on the embodiment, the lease periods used by the slave components and the master component may be the same or differ and may be adapted based on node capability and/or network performance. After transitioning into the suicide state 408, the slave component 134 terminates. Depending on the embodiment, an external node agent may specify that the slave component 134 does not restart itself (i.e., the slave component 134) by marking a local persistent state to indicate a fatal-and-no-restart condition. This condition of not restarting an instance of the slave component 134 may only be cleared administratively, according to one embodiment.
FIG. 5 is a flowchart depicting a method 500 for managing availability of slave components 134 in a distributed system, according to one embodiment of the invention. As shown, the method 500 begins at step 510, where the application 150 sends multicast heartbeat requests to slave components 134 executing in the distributed system. At step 520, the application 150 enters a loop for the sent multicast heartbeat requests. At step 525, the application 150 enters a loop for each slave component 134 in a live state. At step 530, the application monitors whether a unicast heartbeat acknowledgment is received from a slave component 134, responsive to each multicast heartbeat request.
At step 540, the application 150 determines whether a threshold count of unicast heartbeat acknowledgments is not received from a slave component 134 in a live state from a last iteration (of the loop of the step 525) within a predefined time window. If so, the application 150 assigns an availability state to the slave component to indicate that the slave component may be unhealthy (step 550). If the application 150 determines that less than the threshold count of unicast heartbeat acknowledgments is not received from the slave component (step 540), or after the step 550, the method 500 may proceed to step 555, where the application 150 determines whether more slave components 134 in a live state remain. If so, the method 500 returns to the step 525 to process another slave component 134. Otherwise, the application 150 determines whether more multicast heartbeat requests remain (step 560). If so, the application 150 may return to the step 530 to process a unicast heartbeat acknowledgment for another multicast heartbeat request. After the step 560, the method 500 terminates.
FIG. 6 is a flowchart depicting a method 600 for a master component to participate in an availability protocol of a distributed system, according to one embodiment of the invention. As shown, the method 600 begins at step 602, where the application 150 enters a loop for each slave component 134. At step 604, the application 150 sets the respective slave component 134 to a suspect state. At step 606, the application 150 determines whether a unicast heartbeat acknowledgment is received—or is received within a threshold percentage to be missed—responsive to a multicast heartbeat request. If not, the application 150 determines whether unicast heartbeat acknowledgments have been missed beyond a predefined threshold percentage (step 608). If not, the method 600 returns to the step 606.
However, if the application 150 determines that a unicast heartbeat acknowledgment is received—or is received within a threshold percentage to be missed (step 606)—the application 150 sets the respective slave component 134 to a live state (step 610), according to one embodiment. Whether the unicast heartbeat acknowledgment is received may be determined relative to a predefined window of time. In an alternative embodiment, the application 150 may determine whether a minimum threshold percentage of received heartbeat acknowledgements is received. The application 150 may then determine whether a unicast heartbeat acknowledgment is missed (step 612). If not, the method 600 returns to the step 612. Otherwise, the method 600 returns to the step 604.
As described above, in one embodiment, the application 150 determines whether the unicast heartbeat acknowledgment is received, relative to a predefined window of time. In one embodiment, the application 150 may use different windows of time, based on the type of slave component 134. In alternative embodiments, the application 150 may use different windows of time, based on node capability and/or network performance.
On the other hand, if the application 150 determines that a unicast heartbeat acknowledgment is missed beyond a threshold percentage (step 608), the application 150 sets the respective slave component 134 to an unknown state and begins to ignore any unicast heartbeat request from the respective slave component (step 614), according to one embodiment. The application 150 may then attempt to kill the respective slave component 134 and determine whether the attempt is successful (step 616). If so, the application 150 may then attempt to launch the respective slave component 134 and determine whether the attempt is successful (step 618). If so, the method 600 returns to step 610, and the application 150 no longer ignores unicast heartbeat requests from the respective slave component 134.
Alternatively, if either attempt is unsuccessful (step 616 or step 618), the application 150 may determine whether a retry limit has been exceeded (step 620), according to one embodiment. The application 150 may then mark a start of a lease for the respective slave component (step 622). The application 150 may then determine whether the lease has ended for the respective slave component (step 624). If not, the application 150 may wait on the lease for the respective slave component (step 626) and return to the step 624. Otherwise, the application 150 may set the respective slave component 134 to a dead state (step 628). At step 630, the application 150 may determine whether more slave components remain to be processed. If so, the method 600 may return to the step 602, where the application 150 processes a next slave component 134 in the loop. Otherwise, the method 600 terminates. Alternatively, instead of terminating the method 600, the application 150 may repeat the loop of the step 602.
FIG. 7 is a flowchart depicting a method 700 for a slave component 134 to participate in an availability protocol of a distributed system, according to one embodiment of the invention. As shown, the method 700 begins at step 702, where the slave component 134 sets the master component to a suspect state. At step 704, the slave component 134 determines whether a multicast heartbeat request is received from the master component. If not, the slave component 134 determines whether a lease for the slave component 134 has expired (step 706). If not, the method 700 returns to the step 704. Depending on the embodiment, the slave component 134 may also increment a counter based on a received multicast heartbeat request. Further, depending on the embodiment, the slave components 134 and the master component may determine whether a multicast heartbeat request or unicast heartbeat acknowledgment is received based on an adaptable time window and prior to changing state.
However, if a multicast heartbeat request is not received (step 704), the slave component may renew a lease for the slave component 134 and set the master component to a live state (step 708), according to one embodiment. The slave component 134 may then determine whether a multicast heartbeat request is missed (step 710). If not, the method 700 returns to the step 710. Otherwise, the method 700 returns to the step 702.
On the other hand, if the lease for the slave component 134 has expired (step 706), the slave component 134 may set the master component to an unknown state (step 712), according to one embodiment. At step 714, the slave component 134 may then terminate itself After the step 714, the method 700 terminates. In some embodiments, if a multicast heartbeat request is not received (step 704), the method 700 may proceed to step 716, where the slave component 134 sets the master component to a time drift state. The slave component 134 may then terminate itself (step 714).
FIG. 8 depicts exemplary source code 802 defining a data structure of a multicast heartbeat acknowledgment and/or a unicast heartbeat response, according to one embodiment of the invention. As shown, the source code 802 includes a data type 804, a variable name 806 and a description 808. The data structure includes a data gram length, a header length, a signature, a message type, a command code, a domain identifier, a component type, a component identifier, a subcomponent identifier, a source IP address, a source port, an event identifier, a timestamp, a payload length, and payload bytes. The source IP address and source port of a sending component may be sent explicitly rather than extracted from TCP headers of the datagram, due to a possible lack of implementation on some operating system TCP libraries (e.g., TCP libraries that do not support packet introspection). The timestamp may be a local time stamp obtained from either a local operating system on which the master component or the slave components execute. The timestamp may be consolidated at the master component to enforce a de-configuration of a slave component 134 if the clock drift exceeds a predefined threshold, relative to a local clock of the master component. This is achieved because the slave component 134 shuts itself down—pursuant to the availability protocol disclosed herein—if the slave component 134 determines that the local timestamp of the slave component is out of drift from the timestamp transmitted on the multicast heartbeat request from the master component. The master component may subsequently de-configure the slave component 134.
In one embodiment, an administrative user and/or an application 150 may specify that a given slave component type as being critical for the entire distributed application. In such cases, if the master component determines that a slave component 134 of the given slave component type has entered an unknown state, the application 150 may determine to terminate itself (i.e., the application 150), thereby bringing down the entire distributed application within a predefined multiple of the lease period. Examples of slave component types include three slave component types (e.g., key server, index server, and translation server) shown in FIG. 2.
Advantageously, embodiments of the invention provide techniques for managing the availability of slave components executing in a distributed system. One embodiment of the invention provides a master server and a heartbeat protocol. The heartbeat protocol may define state machines for the master server and/or the slave components. The master server and/or the slave components may act in accordance with the state machines to manage availability of the slave components in the distributed system. For example, the heartbeat protocol may specify that the master server manages availability of the slave components by sending multicast heartbeat requests to the slave components and monitoring whether unicast heartbeat acknowledgments are received from the slave components. For instance, the master server may assign an “unknown” state to a slave component and attempt to kill and restart the slave component.
While the foregoing is directed to embodiments of the present invention, other and further embodiments of the invention may be devised without departing from the basic scope thereof, and the scope thereof is determined by the claims that follow.

Claims (24)

What is claimed is:
1. A computer-implemented method to manage availability of a plurality of slave components executing in a distributed system by maintaining, for at least a first slave component selected from the plurality of slave components, an availability state indicating a health status of the first slave component, the availability state comprising one of, in decreasing order of health, first, second, and third availability states, the method comprising:
sending a plurality of multicast heartbeat requests to the plurality of slave components executing in the distributed system;
upon determining that a count of unicast heartbeat acknowledgments not received from the first slave component exceeds a first predefined threshold, transitioning from the first availability state to the second availability state, whereafter receipt of a unicast heartbeat acknowledgment from the first slave component causes a transition back to the first availability state; and
upon determining that the count of unicast heartbeat acknowledgments not received from the first slave component exceeds a second predefined threshold greater than the first predefined threshold, transitioning from the second availability state to the third availability state by operation of one or more computer processors, whereafter a successful restart of the first slave component causes a transition back to one of the first and second availability states, wherein the first, second, and third availability states are in decreasing order of health.
2. The computer-implemented method of claim 1, further comprising:
for each multicast heartbeat request, monitoring whether a unicast heartbeat acknowledgment is received from the respective slave component, responsive to the respective multicast heartbeat request, wherein monitoring whether the unicast heartbeat acknowledgment is received from the respective slave component comprises incrementing a counter upon receiving the unicast heartbeat acknowledgment from the respective slave component.
3. The computer-implemented method of claim 1, wherein the method further comprises:
restarting the first slave component upon determining that the count of unicast heartbeat acknowledgments not received from the first slave component exceeds the second predefined threshold.
4. The computer-implemented method of claim 1, wherein one of the plurality of slave components is designated as a critical slave component, wherein the method is performed by a master process, and wherein the method further comprises:
upon determining that the count of unicast heartbeat acknowledgments not received from the critical slave component exceeds a predefined threshold, terminating the master process, whereby the distributed system is rendered unavailable.
5. The computer-implemented method of claim 1, wherein the method is performed by a master process, and wherein the method further comprises:
upon determining that the count of unicast heartbeat acknowledgments not received from at least one of the slave components exceeds a predefined threshold, restarting the master process.
6. The computer-implemented method of claim 1, wherein the first, second, and third availability states indicate healthy, potentially unhealthy, and unhealthy statuses, respectively, wherein the first, second, and third availability states comprise a slave live state, a slave suspect state, and a slave unknown state, respectively, wherein the distributed system includes a master process, the plurality of slave components, a node agent, a data store, and a component having an address space closed to the distributed system, wherein at least one of the plurality of slave components is configured to serve as a proxy for the component, wherein the master process is user-configurable via an administrative application, wherein the master process is configured to perform the method for managing availability of the plurality of slave components.
7. The computer-implemented method of claim 6, wherein the master process is configured to operate according to a master state machine to monitor availability of the plurality of slave components, wherein the master state machine includes the slave live state, the slave suspect state, the slave unknown state, a slave kill state, a check kill state, a slave launch state, a check launch state, a retry limit state, a lease start state, a lease wait state, a lease end state, and a slave dead state;
wherein at least one slave component of the plurality of slave components is configured to operate according to a slave state machine to monitor availability of the master process, wherein the slave state machine includes a master live state, a master suspect state, a master unknown state, and a suicide state.
8. The computer-implemented method of claim 7, wherein each predefined threshold is expressed in terms of at least one of a number of acknowledgments and a percentage of acknowledgments, wherein the method further comprises:
for each multicast heartbeat request, monitoring whether a unicast heartbeat acknowledgment is received from the respective slave component, responsive to the respective multicast heartbeat request.
9. The computer-implemented method of claim 8, wherein the master state machine further transitions:
(i) from the retry limit state to the slave unknown state upon determining a retry count does not exceed a predefined retry limit, wherein the retry count is incremented upon transitioning to the retry limit state, wherein retry count is specific to the given slave component;
(ii) from the retry limit state to the lease start state upon determining that the retry count exceeds the predefined retry limit;
(iii) from the slave kill state to the check kill state upon an errorless execution of an kill operation on the given slave component;
(iv) from the check kill state to the slave launch state upon verifying that the given slave component is terminated;
(v) from the slave launch state to the check launch state upon an errorless execution of a launch operation on the given slave component;
(vi) from the check launch state to the slave live state upon verifying that the given slave component is launched;
(vii) from the slave kill state to the retry limit state upon detecting an error during execution of the kill operation on the given slave component;
(viii) from the check kill state to the retry limit state upon determining that the given slave component is not terminated;
(ix) from the slave launch state to the retry limit state upon detecting an error during execution of the launch operation on the given slave component; and
(x) from the check launch state to the retry limit state upon determining that the given slave component is not launched.
10. The computer-implemented method of claim 9, wherein the slave state machine transitions:
(i) from the master live state to the master suspect state upon non-receipt of a predefined number of multicast heartbeat requests from the master process within a second predefined duration of time;
(ii) from the master live state to the time draft state upon a time drift threshold being exceeded;
(iii) from the master suspect state to the master unknown state a lease expiration; and
(iv) from the master suspect state to the master live state upon a lease renewal.
11. The computer-implemented method of claim 10, wherein at least one of the plurality of slave components is designated as a non-critical component, wherein the master process is terminated upon determining that the count of unicast heartbeat acknowledgments not received from the critical slave component exceeds the second predefined threshold, independent of any health status of the at least one non-critical component;
wherein the proxy for the component is configured to analyze a state of the component using inter-process communication (IPC) and operating system primitives and transmit availability information of the component to the master process according to a predefined availability protocol of the distributed system.
12. The computer-implemented method of claim 11, wherein each unicast heartbeat acknowledgment and multicast heartbeat request comprises a predefined data structure containing a plurality of fields including a datagram length, a header length, a signature, a message type, a command code, a domain identifier, a component type, a component identifier, a subcomponent identifier, a source Internet Protocol (IP) address, a source port, an event identifier, a timestamp, a payload length, and a set of payload bytes, wherein the first predefined duration of time and the second predefined duration of time are distinct.
13. The computer-implemented method of claim 12, wherein monitoring whether the unicast heartbeat acknowledgment is received from the respective slave component comprises incrementing a counter upon receiving the unicast heartbeat acknowledgment from the respective slave component;
wherein the method further comprises restarting the slave component upon determining that the count of unicast heartbeat acknowledgments not received from one of the slave components exceeds the second predefined threshold;
wherein one of the plurality of slave components is designated as a critical slave component, and wherein the method further comprises, upon determining that the count of unicast heartbeat acknowledgments not received from the critical slave component exceeds a third predefined threshold, terminating the master process, whereby the distributed system is rendered unavailable.
14. The computer-implemented method of claim 13, wherein the method further comprises, upon determining that the count of unicast heartbeat acknowledgments not received from at least one of the slave components exceeds the third threshold, restarting the master process, wherein the first threshold, the second threshold, and the third threshold are distinct;
wherein the master process is restarted without having to restart any of the plurality of slave components;
wherein at least one slave component of the plurality of slave components is configured to terminate itself upon determining that a local timestamp associated with the at least one slave component is out of drift from a timestamp of one of the plurality of multicast heartbeat requests;
wherein each threshold is determined based on a slave component function type, a message performance history, a node condition, and a network condition, and wherein each threshold is distinct.
15. A computer program product for performing an operation to manage availability of a plurality of slave components executing in a distributed system by maintaining, for at least a first slave component selected from the plurality of slave components, an availability state indicating a health status of the first slave component, the availability state comprising one of, in decreasing order of health, first, second, and third availability states, the computer program product comprising:
a computer-readable storage device having computer-readable program code embodied therewith, the computer-readable program code comprising:
computer-readable program code configured to send a plurality of multicast heartbeat requests to the plurality of slave components executing in the distributed system;
computer-readable program code configured to, upon determining that a count of unicast heartbeat acknowledgments not received from the first slave component exceeds a first predefined threshold, transition from the first availability state to the second availability state, whereafter receipt of a unicast heartbeat acknowledgment from the first slave component causes a transition back to the first availability state; and
computer-readable program code configured to, by operation of one or more computer processors when executing the computer-readable program code and upon determining that the count of unicast heartbeat acknowledgments not received from the first slave component exceeds a second predefined threshold greater than the first predefined threshold, transition from the second availability state to the third availability state, whereafter a successful restart of the first slave component causes a transition back to one of the first and second availability states, wherein the first, second, and third availability states are in decreasing order of health.
16. The computer program product of claim 15, further comprising:
computer-readable program code configured to monitor, for each multicast heartbeat request, whether a unicast heartbeat acknowledgment is received from the respective slave component, responsive to the respective multicast heartbeat request, wherein the computer-readable program code configured to monitor whether the unicast heartbeat acknowledgment is received from the respective slave component comprises:
computer-readable program code configured to increment a counter upon receiving the unicast heartbeat acknowledgment from the respective slave component.
17. The computer program product of claim 15, wherein the computer-readable program code further comprises:
computer-readable program code configured to restart the first slave component upon determining that the count of unicast heartbeat acknowledgments not received from the first slave component exceeds the second predefined threshold.
18. The computer program product of claim 15, wherein one of the plurality of slave components is designated as a critical slave component, wherein the operation is performed by a master process, and wherein the computer-readable program code further comprises:
computer-readable program code configured to terminate the master process upon determining that the count of unicast heartbeat acknowledgments not received from the critical slave component exceeds a predefined threshold, whereby the distributed system is rendered unavailable.
19. The computer program product of claim 15, wherein the operation is performed by a master process, and wherein the computer-readable program code further comprises:
computer-readable program code configured to restart the master process upon determining that the count of unicast heartbeat acknowledgments not received from at least one of the slave components exceeds a predefined threshold.
20. A system to manage availability of a plurality of slave components executing in a distributed system by maintaining, for at least a first slave component selected from the plurality of slave components, an availability state indicating a health status of the first slave component, the availability state comprising one of, in decreasing order of health, first, second, and third availability states, the system comprising:
one or more computer processors; and
a memory containing a program, which when executed by the one or more computer processors is configured to perform an operation comprising:
sending a plurality of multicast heartbeat requests to the plurality of slave components executing in the distributed system;
upon determining that a count of unicast heartbeat acknowledgments not received from the first slave component exceeds a first predefined threshold, transitioning from the first availability state to the second availability state, whereafter receipt of a unicast heartbeat acknowledgment from the first slave component causes a transition back to the first availability state; and
upon determining that the count of unicast heartbeat acknowledgments not received from the first slave component exceeds a second predefined threshold greater than the first predefined threshold, transitioning from the second availability state to the third availability state, whereafter a successful restart of the first slave component causes a transition back to one of the first and second availability states, wherein the first, second, and third availability states are in decreasing order of health.
21. The system of claim 20, wherein the operation further comprises:
for each multicast heartbeat request, monitoring whether a unicast heartbeat acknowledgment is received from the respective slave component, responsive to the respective multicast heartbeat request, wherein monitoring whether the unicast heartbeat acknowledgment is received from the respective slave component comprises incrementing a counter upon receiving the unicast heartbeat acknowledgment from the respective slave component.
22. The system of claim 20, wherein the operation further comprises:
restarting the first slave component upon determining that the count of unicast heartbeat acknowledgments not received from the first slave component exceeds the second predefined threshold.
23. The system of claim 20, wherein one of the plurality of slave components is designated as a critical slave component, wherein the operation is performed by a master process, and wherein the operation further comprises:
upon determining that the count of unicast heartbeat acknowledgments not received from the critical slave component exceeds a predefined threshold, terminating the master process, whereby the distributed system is rendered unavailable.
24. The system of claim 20, wherein the operation is performed by a master process, and wherein the operation further comprises:
upon determining that the count of unicast heartbeat acknowledgments not received from at least one of the slave components exceeds a predefined threshold, restarting the master process.
US13/617,383 2010-08-12 2012-09-14 High availability management system for stateless components in a distributed master-slave component topology Expired - Fee Related US8838723B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/617,383 US8838723B2 (en) 2010-08-12 2012-09-14 High availability management system for stateless components in a distributed master-slave component topology

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US12/855,241 US8364775B2 (en) 2010-08-12 2010-08-12 High availability management system for stateless components in a distributed master-slave component topology
US13/617,383 US8838723B2 (en) 2010-08-12 2012-09-14 High availability management system for stateless components in a distributed master-slave component topology

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US12/855,241 Continuation US8364775B2 (en) 2010-08-12 2010-08-12 High availability management system for stateless components in a distributed master-slave component topology

Publications (2)

Publication Number Publication Date
US20130013723A1 US20130013723A1 (en) 2013-01-10
US8838723B2 true US8838723B2 (en) 2014-09-16

Family

ID=45565572

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/855,241 Expired - Fee Related US8364775B2 (en) 2010-08-12 2010-08-12 High availability management system for stateless components in a distributed master-slave component topology
US13/617,383 Expired - Fee Related US8838723B2 (en) 2010-08-12 2012-09-14 High availability management system for stateless components in a distributed master-slave component topology

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US12/855,241 Expired - Fee Related US8364775B2 (en) 2010-08-12 2010-08-12 High availability management system for stateless components in a distributed master-slave component topology

Country Status (1)

Country Link
US (2) US8364775B2 (en)

Families Citing this family (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8364775B2 (en) * 2010-08-12 2013-01-29 International Business Machines Corporation High availability management system for stateless components in a distributed master-slave component topology
US9092561B2 (en) * 2010-10-20 2015-07-28 Microsoft Technology Licensing, Llc Model checking for distributed application validation
US10331801B2 (en) 2011-09-23 2019-06-25 Open Invention Network, Llc System for live-migration and automated recovery of applications in a distributed system
US9477739B2 (en) 2011-09-23 2016-10-25 Hybrid Logic Ltd System for live-migration and automated recovery of applications in a distributed system
US10311027B2 (en) 2011-09-23 2019-06-04 Open Invention Network, Llc System for live-migration and automated recovery of applications in a distributed system
GB2495079A (en) 2011-09-23 2013-04-03 Hybrid Logic Ltd Live migration of applications and file systems in a distributed system
US9501543B2 (en) 2011-09-23 2016-11-22 Hybrid Logic Ltd System for live-migration and automated recovery of applications in a distributed system
US9547705B2 (en) 2011-09-23 2017-01-17 Hybrid Logic Ltd System for live-migration and automated recovery of applications in a distributed system
US9483542B2 (en) 2011-09-23 2016-11-01 Hybrid Logic Ltd System for live-migration and automated recovery of applications in a distributed system
US8873429B2 (en) * 2011-12-22 2014-10-28 Infosys Limited Method and system to dynamically detect and form a master slave network
US8868711B2 (en) 2012-02-03 2014-10-21 Microsoft Corporation Dynamic load balancing in a scalable environment
US9852010B2 (en) 2012-02-03 2017-12-26 Microsoft Technology Licensing, Llc Decoupling partitioning for scalability
US10860384B2 (en) * 2012-02-03 2020-12-08 Microsoft Technology Licensing, Llc Managing partitions in a scalable environment
WO2015013677A2 (en) * 2013-07-26 2015-01-29 The Trustees Of Columbia University In The City Of New York Total property optimization system for energy efficiency and smart buildings
EP3085011A4 (en) 2013-12-19 2017-08-16 Google, Inc. Systems, methods, and computer program products for service processing
US9936009B2 (en) * 2014-05-22 2018-04-03 Qualcomm Incorporated Systems and methods of operating a device of a data path group network
CN104331353A (en) * 2014-10-17 2015-02-04 云南远信数通科技有限公司 Method for guaranteeing software high availability
CN107181780B (en) * 2016-03-10 2020-07-14 阿里巴巴集团控股有限公司 Communication channel processing method and system
CN107294799B (en) * 2016-03-31 2020-09-01 阿里巴巴集团控股有限公司 Method and device for processing nodes in distributed system
US10901771B2 (en) 2019-01-23 2021-01-26 Vmware, Inc. Methods and systems for securely and efficiently clustering distributed processes using a consistent database
US11086846B2 (en) * 2019-01-23 2021-08-10 Vmware, Inc. Group membership and leader election coordination for distributed applications using a consistent database
CN110740064A (en) * 2019-10-25 2020-01-31 北京浪潮数据技术有限公司 Distributed cluster node fault processing method, device, equipment and storage medium
CN112825525B (en) * 2019-11-20 2023-06-16 北京百度网讯科技有限公司 Method and apparatus for processing transactions
CN110890989A (en) * 2019-12-03 2020-03-17 杭州迪普科技股份有限公司 Channel connection method and device
KR102364450B1 (en) * 2019-12-06 2022-02-17 서강대학교 산학협력단 Sensor network system based on blockchain and method for generating dynamic link key in the sensor network system

Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020007468A1 (en) 2000-05-02 2002-01-17 Sun Microsystems, Inc. Method and system for achieving high availability in a networked computer system
US20030046230A1 (en) 2001-08-30 2003-03-06 Jacobs Dean Bernard Method for maintaining account consistency
US6604207B2 (en) 1997-05-13 2003-08-05 Micron Technology, Inc. System architecture for remote access and control of environmental management
US20030229775A1 (en) 2002-06-11 2003-12-11 Schelling Todd A. System and method to filter processors by health during early firmware for split recovery architecture
US20040034822A1 (en) 2002-05-23 2004-02-19 Benoit Marchand Implementing a scalable, dynamic, fault-tolerant, multicast based file transfer and asynchronous file replication protocol
US6847993B1 (en) 2000-05-31 2005-01-25 International Business Machines Corporation Method, system and program products for managing cluster configurations
US20060041630A1 (en) * 2004-08-17 2006-02-23 Naismith Ronald H Gateway having an input/output scanner
US7043563B2 (en) 2000-04-17 2006-05-09 Circadence Corporation Method and system for redirection to arbitrary front-ends in a communication system
US20060168084A1 (en) 2004-11-29 2006-07-27 Leonid Kogan Method and apparatus for rendering load balancing and failover
US20070041327A1 (en) 2005-08-16 2007-02-22 Cisco Technology, Inc. Multicast heartbeat signaling
EP1840741A1 (en) 2006-03-29 2007-10-03 Nokia Corporation Device, method, and computer program product for accessing a non-native application executing in a virtual machine environment
US20070294563A1 (en) 2006-05-03 2007-12-20 Patrick Glen Bose Method and system to provide high availability of shared data
US20080089361A1 (en) 2005-10-06 2008-04-17 Metcalf Thomas D System and method for transferring data
US7502841B2 (en) 2004-02-11 2009-03-10 Solutioninc Limited Server, system and method for providing access to a public network through an internal network of a multi-system operator
US7518987B2 (en) 2005-07-25 2009-04-14 Cisco Technology, Inc. Mechanisms for providing connectivity in NAT redundant/fail-over scenarios in unshared address-space
US20090216910A1 (en) 2007-04-23 2009-08-27 Duchesneau David D Computing infrastructure
US20090232041A1 (en) * 2008-03-11 2009-09-17 Sherry Smith Method And System For Advertising Bluetooth Multicast Feature
US8082464B2 (en) 2009-10-13 2011-12-20 International Business Machines Corporation Managing availability of a component having a closed address space
US20120042030A1 (en) * 2010-08-12 2012-02-16 International Business Machines Corporation High availability management system for stateless components in a distributed master-slave component topology

Patent Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6604207B2 (en) 1997-05-13 2003-08-05 Micron Technology, Inc. System architecture for remote access and control of environmental management
US7043563B2 (en) 2000-04-17 2006-05-09 Circadence Corporation Method and system for redirection to arbitrary front-ends in a communication system
US20020007468A1 (en) 2000-05-02 2002-01-17 Sun Microsystems, Inc. Method and system for achieving high availability in a networked computer system
US6847993B1 (en) 2000-05-31 2005-01-25 International Business Machines Corporation Method, system and program products for managing cluster configurations
US20030046230A1 (en) 2001-08-30 2003-03-06 Jacobs Dean Bernard Method for maintaining account consistency
US20040034822A1 (en) 2002-05-23 2004-02-19 Benoit Marchand Implementing a scalable, dynamic, fault-tolerant, multicast based file transfer and asynchronous file replication protocol
US20030229775A1 (en) 2002-06-11 2003-12-11 Schelling Todd A. System and method to filter processors by health during early firmware for split recovery architecture
US7502841B2 (en) 2004-02-11 2009-03-10 Solutioninc Limited Server, system and method for providing access to a public network through an internal network of a multi-system operator
US20060041630A1 (en) * 2004-08-17 2006-02-23 Naismith Ronald H Gateway having an input/output scanner
US20060168084A1 (en) 2004-11-29 2006-07-27 Leonid Kogan Method and apparatus for rendering load balancing and failover
US7518987B2 (en) 2005-07-25 2009-04-14 Cisco Technology, Inc. Mechanisms for providing connectivity in NAT redundant/fail-over scenarios in unshared address-space
US20070041327A1 (en) 2005-08-16 2007-02-22 Cisco Technology, Inc. Multicast heartbeat signaling
US20080089361A1 (en) 2005-10-06 2008-04-17 Metcalf Thomas D System and method for transferring data
EP1840741A1 (en) 2006-03-29 2007-10-03 Nokia Corporation Device, method, and computer program product for accessing a non-native application executing in a virtual machine environment
US20070294563A1 (en) 2006-05-03 2007-12-20 Patrick Glen Bose Method and system to provide high availability of shared data
US20090216910A1 (en) 2007-04-23 2009-08-27 Duchesneau David D Computing infrastructure
US20090232041A1 (en) * 2008-03-11 2009-09-17 Sherry Smith Method And System For Advertising Bluetooth Multicast Feature
US8082464B2 (en) 2009-10-13 2011-12-20 International Business Machines Corporation Managing availability of a component having a closed address space
US20120042030A1 (en) * 2010-08-12 2012-02-16 International Business Machines Corporation High availability management system for stateless components in a distributed master-slave component topology

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
"Service Availability Forum Application Interface Specification: Availability Management Framework," AssociationVoice, Oct. 2008: pp. 1-13, 33-95 and 230-244, , retrieved Nov. 18, 2010.
"Service Availability Forum: Specifications," AssociationVoice, Oct. 2008: pp. 1-4, , retrieved Nov. 18, 2010.
"Service Availability Forum Application Interface Specification: Availability Management Framework," AssociationVoice, Oct. 2008: pp. 1-13, 33-95 and 230-244, <http://www.saforum.org/HOA/assn16627/images/SAI-AIS-AMF-B.04.01.pdf>, retrieved Nov. 18, 2010.
"Service Availability Forum: Specifications," AssociationVoice, Oct. 2008: pp. 1-4, <http://www.saforunn.org/link/linkshow.asp?link—id=222259&assn—id=16627>, retrieved Nov. 18, 2010.
International Search Report and Written Opinion for Patent Cooperation Treaty Application Serial No. PCT/EP2010/062894 dated Dec. 17, 2010.
Kostopoulos et al., "NASS-IMS bundled Authentication Study through Core Network Concepts," Proceedings of the 3rd International Conference on Mobile Multimedia Communications, ACM International Conference Proceeding Series, 2007 vol. 329(46): pp. 1-6.

Also Published As

Publication number Publication date
US8364775B2 (en) 2013-01-29
US20130013723A1 (en) 2013-01-10
US20120042030A1 (en) 2012-02-16

Similar Documents

Publication Publication Date Title
US8838723B2 (en) High availability management system for stateless components in a distributed master-slave component topology
US10728168B2 (en) Method for providing a connection of a client to an unmanaged service in a client-server remote access system
US9407585B1 (en) Scalable, real-time messaging system
US20140280398A1 (en) Distributed database management
WO2012132815A1 (en) Virtual server id management system, integrated monitoring system, virtual server id management program, and integrated monitoring program
US9253129B2 (en) Instant messaging with browser collaboration
US8607233B2 (en) Web service management
CN111414208B (en) Application program starting method, device and equipment
US10516702B2 (en) Managing mid-dialog session initiation protocol (SIP) messages
US8352617B2 (en) Method and system for data processing
US8560714B2 (en) Message processing to improve server performance
US8352623B2 (en) System for energy efficient computer management environment via tightly integrated target status and directed work sessions
US8156174B2 (en) Method and system for information exchange utilizing an asynchronous persistent store protocol
US20190238637A1 (en) Data replication in scalable messaging system
CN115361348A (en) Method for communicating with web browser performed by data acquisition device
CN111083215B (en) Session information synchronization method, device, equipment, system and storage medium
US11086742B2 (en) Task based service management platform
CN109309583B (en) Information acquisition method and device based on distributed system, electronic equipment and medium
US8473965B2 (en) Net-timer daemon
US20150081867A1 (en) Obtaining a mac address from an external source
US9813316B1 (en) Implementing scalable throttled poller
WO2013055341A1 (en) Network address protocol
CN117714731A (en) Dynamic loading method and device
CN117560419A (en) Message pushing method and system
CN113905100A (en) Method, system, device and storage medium for dynamically controlling retransmission request of client

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.)

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20180916