US20070174411A1 - Apparatus and method for efficient communication of producer/consumer buffer status - Google Patents

Apparatus and method for efficient communication of producer/consumer buffer status Download PDF

Info

Publication number
US20070174411A1
US20070174411A1 US11/340,453 US34045306A US2007174411A1 US 20070174411 A1 US20070174411 A1 US 20070174411A1 US 34045306 A US34045306 A US 34045306A US 2007174411 A1 US2007174411 A1 US 2007174411A1
Authority
US
United States
Prior art keywords
buffer region
shared buffer
channel
pointer
update
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/340,453
Inventor
Daniel Brokenshire
Charles Johns
Mark Nutter
Barry Minor
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
International Business Machines Corp
Original Assignee
International Business Machines Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corp filed Critical International Business Machines Corp
Priority to US11/340,453 priority Critical patent/US20070174411A1/en
Assigned to MACHINES CORPORATION, INTERNATIONAL BUSINESS reassignment MACHINES CORPORATION, INTERNATIONAL BUSINESS ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BROKENSHIRE, DANIEL A., JOHNS, CHARLES R., MINOR, BARRY, NUTTER, MARK RICHARD
Priority to PCT/EP2007/050182 priority patent/WO2007085522A2/en
Publication of US20070174411A1 publication Critical patent/US20070174411A1/en
Assigned to INTERNATIONAL BUSINESS MACHINES CORPORATION reassignment INTERNATIONAL BUSINESS MACHINES CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BROKENSHIRE, DANIEL A., JOHNS, CHARLES R., MINOR, BARRY L., NUTTER, MARK R.
Priority to US12/127,464 priority patent/US8275917B2/en
Priority to US13/593,030 priority patent/US9053069B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F15/00Digital computers in general; Data processing equipment in general
    • G06F15/16Combinations of two or more digital computers each having at least an arithmetic unit, a program unit and a register, e.g. for a simultaneous processing of several programs
    • G06F15/163Interprocessor communication
    • G06F15/173Interprocessor communication using an interconnection network, e.g. matrix, shuffle, pyramid, star, snowflake
    • G06F15/17337Direct connection machines, e.g. completely connected computers, point to point communication networks

Definitions

  • the present application relates generally to an improved data processing system. More specifically, the present application is directed to an apparatus and method for efficient communication of producer/consumer buffer status.
  • one processor or device typically acts as a producer by placing data into a shared buffer while another processor or device acts as a consumer by reading and processing the data placed there by the producer.
  • a producer and a consumer independently operating on processors or devices within the data processing system, have access to a shared data buffer region.
  • the shared data buffer region may be located, for example, in shared system or I/O memory and is commonly accessible by both parties.
  • two state variables are shared between the producer and consumer: a head pointer (or index) and a tail pointer (or index), as shown in FIG. 1A .
  • FIG. 1A is an exemplary diagram of a shared buffer region in accordance with a prior art system.
  • the shared buffer region 100 has an associated tail pointer 120 and head pointer 110 .
  • the head pointer 110 represents the next unused location in the shared buffer region 100 .
  • the tail pointer 120 represents the last unprocessed data location in the shared buffer region 100 .
  • the producer increments the head pointer 110 . In this way, the producer indicates to consumers that new data is available for the consumers. Likewise, as the consumer retrieves data from the shared buffer region 100 , the consumer increments the tail pointer 120 , thereby indicating that data has been read from the shared buffer region 100 . The amount of data available in the shared buffer region 100 is indicated by the difference between the head pointer 110 and tail pointer 120 . This is referred to as the “active” data region 140 in the shared buffer region 100 . The portion of the shared buffer region 100 that is behind the tail pointer 120 is referred to as the “inactive” (or old) data region 130 . An empty shared buffer region is identified by setting the head pointer 110 and tail pointer 120 to the same location in the shared buffer region 100 .
  • the head pointer 110 and tail pointer 120 may be incremented indefinitely. However, in most systems, it is more realistic to have only a fixed amount of storage space available for the shared buffer region 100 . In this case, it is possible to reclaim the inactive region 130 by implementing what is known as a circular first-in-first-out (FIFO) type shared buffer region 100 . This type of shared buffer region is illustrated in FIG. 1B .
  • FIFO first-in-first-out
  • the head pointer 110 and tail pointer 120 are incremented as described above with regard to FIG. 1A , however when the end of the shared buffer region 100 is encountered, the head and tail pointers 110 and 120 are permitted to wrap back to the beginning of the shared buffer region 100 .
  • the active data region 140 is defined as the portion of the shared buffer region 100 between the head pointer 110 and tail pointer 120 .
  • both the producer and consumer must now account for the situation where the head pointer 110 has wrapped beyond the end of the shared buffer region 100 , i.e. back to the beginning of the shared buffer region 100 , but the tail pointer 120 has not wrapped.
  • the consumer is primarily concerned with monitoring updates to the head pointer 110 and the producer is primarily concerned with monitoring updates to the tail pointer 120 .
  • the consumer Before a consumer can begin processing, it must know that the head pointer 110 and tail pointer 120 point to different entries in the shared buffer region 100 .
  • the producer Before the producer can add data to the shared buffer region 100 , the producer must ensure that the amount of data to be added will not overflow the tail pointer 120 .
  • a simple implementation may involve storing the state variables, i.e. the head pointer 110 and tail pointer 120 , in shared system memory and monitoring these state variables remotely.
  • FIG. 2A illustrates this exemplary implementation.
  • a producer 250 performs updates to the head pointer 210 in shared memory 230 and monitors the current state of tail pointer 220 in shared memory 230 when writing data to the shared buffer region 240 .
  • the consumer 260 updates the tail pointer 220 in shared memory 230 and monitors the current state of the head pointer 210 in shared memory 230 when reading data from the shared buffer region 240 .
  • the producer 250 and consumer 260 each perform updates and monitoring of the head pointer 210 and tail pointer 220 in shared memory 230 remotely. That is, the head pointer 210 and tail pointer 220 are not located in a memory local to the producer 250 or the consumer 260 . Thus, the producer 250 and consumer 260 must access these pointers 210 and 220 via a bus or other communication interface that is not local to the producer 250 or consumer 260 . Typically, the producer 250 and consumer 260 poll the remote shared memory 230 in order to obtain information regarding the head and tail pointers 210 and 220 . As a result, this updating and monitoring requires additional resources, e.g., processor cycles, bus bandwidth, etc., within the system to support these operations. Because of the need to utilized additional system resources, this solution may be considered too detrimental to overall system performance.
  • additional resources e.g., processor cycles, bus bandwidth, etc.
  • the head and tail pointers 210 and 220 are maintained in a memory or registers that are local to either the producer 250 or the consumer 260 in order to improve performance of that one party, i.e. the producer 250 or the consumer 260 , to which the pointers 210 and 220 are made local.
  • FIG. 2B illustrates such an alternative implementation in which the head pointer 210 and tail pointer 220 are stored in a local memory region 270 to one of the producer 250 and consumer 260 (in the depicted example, the local memory region 270 is local to the consumer 260 ).
  • the head pointer 210 and tail pointer 220 are stored in local memory region 270 that is local to consumer 260 .
  • This local memory region 270 is separate from shared memory 230 and shared buffer region 240 .
  • the producer 250 must remotely monitor, i.e. poll, the value of the tail pointer 220 in order to determine the amount of free space remaining in the shared buffer region 240 . This may lead to serious degradation of performance for the producer, as the latency for reading remote locations can be several orders of magnitude worse than for main shared memory storage. Of course the situation could be reversed, requiring that the consumer 260 remotely monitor the head pointer 210 in a local memory region local to the producer 250 , for example.
  • the illustrative embodiments provide such an improved mechanism by utilizing already existing communication channels established with processing elements of the data processing system to communicate this shared buffer status information.
  • synergistic processor elements are provided in a data processing system.
  • the SPEs are a simple, yet highly optimized, processor with a single instruction, multiple data (SIMD) instruction set.
  • SIMD single instruction, multiple data
  • the SPEs have a large register file, a high speed local memory, and a channel interface that serves as the primary communication pathway between an SPE and external processors or devices.
  • the channel interface is a message passing interface that overlaps input/output (I/O) with data processing and minimizes power consumed by synchronization.
  • the channel facilities of the channel interface are accessed with three types of instructions: a read channel instruction, write channel instruction, and read channel count instruction which measures channel capacity. These channels may be configured as blocking or non-blocking channels. Channels that are defined as blocking channels cause the SPE to stall when accessing a channel whose channel count is zero. In this sense, the channel count indicates the number of outstanding operations that can be issued for that channel without causing the SPE to stall. Aside from blocking on channel access, an SPE program can monitor a variety of channel events using SPE interrupt facilities.
  • the SPE channel interface supports two signal notification channels, which may be used by the SPE to receive signals originating from other processors or devices in the system.
  • the signal notification channels can be programmed to operate in over-write mode.
  • Over-write mode is a mode in which the latest notification messages written to the channel “overwrite” previous notification messages written to the channel such that the processors may access the most recent information written to the signal notification channel.
  • the over-write mode of the signal communication channels is useful in a one-to-one communications environment, or in a logical-OR communications environment, such as in a many-to-one signaling environment.
  • the producer running on an SPE, may read its signal notification channel(s) and operate in a blocking mode until the consumer updates the tail pointer information written to the signal notification channel(s).
  • This provides a low-power, low-bandwidth alternative to monitoring, i.e. polling, a remote memory or register location that is local to the consumer. Further, this solution provides low latency responses for real-time systems since the SPE is automatically awoken as soon as data becomes available on the signal notification channel(s).
  • the producer running on the SPE may use the processor's event and interrupt facilities to receive asynchronous notification of consumer's completion.
  • This solution may be of use in environments where there is other work to be done by the SPE and blocking on a channel is not an acceptable alternative.
  • these illustrative embodiments may be extended by reversing the roles and having the SPE program act as a consumer, while an external processor or device acts as a producer. In such a case, the producer would notify the SPE of buffer state updates by writing the head pointer to a signal notification register.
  • the embodiments may be further extended by having the SPE program operate as both a consumer (receiving head pointer updates from an external device or processor on a first signal notification channel, for example) and as a producer (receiving tail pointer updates from an external device or processor on a second signal notification channel, for example).
  • a method in which an operation on a shared buffer region is performed using a first device in a data processing system.
  • the method may further comprise writing a pointer update to a signal notification channel associated with a second device in the data processing system.
  • the pointer update may be an update to one of a head pointer or a tail pointer of the shared buffer region.
  • the method may further comprise updating a pointer stored in a local store of the second device based on the pointer update.
  • the signal notification channel may be a memory mapped address region of a shared memory of the data processing system.
  • the operation on the shared buffer region may be a write operation for writing data to the shared buffer region.
  • the pointer update may be an update to a head pointer of the shared buffer region.
  • the operation on the shared buffer region may be a read operation for reading data from the shared buffer region.
  • the pointer update may be an update to a tail pointer of the shared buffer region.
  • the method may further comprise determining if the shared buffer region is full and placing the second device in a low-power consumption waiting state waiting for a predetermined amount of storage space to be freed in the shared buffer region before additional data may be written to the shared buffer region.
  • Placing the second device in a low-power consumption waiting state may comprise placing the notification channel in a blocking mode of operation in which the second device's normal operation is blocked and the second device waits in a low-power consumption state.
  • the operation may be a read operation for reading data from the shared buffer region.
  • the notification channel may exit the blocking mode of operation in response to the read operation.
  • a channel interface of the second device may periodically issue a read channel count instruction to the signal notification channel to determine a count associated with the notification channel. If the count associated with the notification channel is zero, then the notification channel may remain in a blocking mode of operation. If the count associated with the notification channel is non-zero, then the notification channel may exit the blocking mode of operation.
  • the second device may be awoken by issuing a read channel instruction to the notification channel and returning results of the read channel instruction to the second device.
  • the second device may be a synergistic processing unit and the shared buffer region is part of the local store associated with the synergistic processing unit. Writing a pointer update may be performed using a channel interface of a memory flow control unit associated with the synergistic processing unit.
  • the first device and the second device may be synergistic processing units of a multiprocessor system-on-a-chip.
  • a computer program product comprising a computer usable medium having a computer readable program.
  • the computer readable program when executed on a computing device, may cause the computing device to perform, using a first device in the data processing system, an operation on a shared buffer region and write a pointer update to a signal notification channel associated with a second device in the data processing system.
  • the pointer update may be an update to one of a head pointer or a tail pointer of the shared buffer region.
  • the computer readable program may further cause the computing device to update a pointer stored in a local store of the second device based on the pointer update.
  • the computer readable program may further cause the computing device to perform the various operations outlined above with regard to the method illustrative embodiment.
  • a system comprising a first processor, a second processor coupled to the first processor, and a local storage device coupled to the second processor.
  • the first processor may perform an operation on a shared buffer region of the local storage device and may write a pointer update to a signal notification channel associated with the second processor.
  • the pointer update may be an update to one of a head pointer or a tail pointer of the shared buffer region.
  • the second processor may update a pointer stored in the local store of the second processor based on the pointer update.
  • the operation on the shared buffer region may be a write operation for writing data to the shared buffer region.
  • the pointer update may be an update to a head pointer of the shared buffer region.
  • the operation on the shared buffer region may be a read operation for reading data from the shared buffer region.
  • the pointer update may be an update to a tail pointer of the shared buffer region.
  • the second processor may determine if the shared buffer region is full and may place itself in a low-power consumption waiting state waiting for a predetermined amount of storage space to be freed in the shared buffer region before additional data may be written to the shared buffer region.
  • the second processor may place itself in a low-power consumption waiting state by placing the notification channel in a blocking mode of operation in which the second processor's normal operation is blocked and the second processor waits in a low-power consumption state.
  • the operation may be a read operation for reading data from the shared buffer region.
  • the notification channel may exit the blocking mode of operation in response to the read operation.
  • a channel interface of the second processor may periodically issue a read channel count instruction to the signal notification channel to determine a count associated with the notification channel. If the count associated with the notification channel is non-zero, then the notification channel may exit the blocking mode of operation.
  • FIG. 1A is an exemplary diagram of a shared buffer region in accordance with a known mechanism
  • FIG. 1B is an exemplary diagram of a FIFO shared buffer region in accordance with a known mechanism
  • FIG. 2A is an exemplary diagram illustrating a mechanism for storing head and tail pointers in a shared memory in accordance with a known mechanism
  • FIG. 2B is an exemplary diagram illustrating an alternative implementation in which the head pointer and tail pointer are stored in a local memory region that is local to a consumer in accordance with a known mechanism;
  • FIG. 3 is an exemplary block diagram of a data processing system in which aspects of the illustrative embodiments may be implemented;
  • FIG. 4 is an exemplary block diagram illustrating an interaction between a producer and a consumer with regard to updating state information for a shared buffer region in accordance with one illustrative embodiment where a process running on a SPE operates as a producer;
  • FIG. 5 is an exemplary block diagram illustrating an interaction between a producer and a consumer with regard to updating state information for a shared buffer region in accordance with one illustrative embodiment where a process running on a SPE operates as a consumer;
  • FIG. 6 is an exemplary block diagram illustrating an interaction between a producer and a consumer with regard to updating state information for a shared buffer region in accordance with one illustrative embodiment where a process running on a SPE operates as both a producer and a consumer;
  • FIG. 7 is a flowchart outlining an exemplary operation of an illustrative embodiment in which a process running on a SPE operates as a producer;
  • FIG. 8 is a flowchart outlining an exemplary operation of an illustrative embodiment in which a process running on a SPE operates as a consumer.
  • FIG. 3 is an exemplary block diagram of a data processing system in which aspects of the present invention may be implemented.
  • the exemplary data processing system shown in FIG. 3 is an example of the Cell Broadband Engine (CBE) data processing system. While the CBE will be used in the description of the preferred embodiments of the present invention, the present invention is not limited to such, as will be readily apparent to those of ordinary skill in the art upon reading the following description.
  • CBE Cell Broadband Engine
  • the CBE 300 includes a power processor element (PPE) 310 having a processor (PPU) 316 and its L 1 and L 2 caches 312 and 314 , and multiple synergistic processor elements (SPEs) 320 - 334 that each has its own synergistic processor unit (SPU) 340 - 354 , memory flow control 355 - 362 , local memory or store (LS) 363 - 370 , and bus interface unit (BIU unit) 380 - 394 which may be, for example, a combination direct memory access (DMA), memory management unit (MMU), and bus interface unit.
  • a high bandwidth internal element interconnect bus (EIB) 396 , a bus interface controller (BIC) 397 , and a memory interface controller (MIC) 398 are also provided.
  • the CBE 300 may be a system-on-a-chip such that each of the elements depicted in FIG. 3 may be provided on a single microprocessor chip. Moreover, the CBE 300 is a heterogeneous processing environment in which each of the SPUs may receive different instructions from each of the other SPUs in the system. Moreover, the instruction set for the SPUs is different from that of the PPU, e.g., the PPU may execute Reduced Instruction Set Computer (RISC) based instructions while the SPU execute vectorized instructions.
  • RISC Reduced Instruction Set Computer
  • the SPEs 320 - 334 are coupled to each other and to the L 2 cache 314 via the EIB 396 .
  • the SPEs 320 - 334 are coupled to MIC 398 and BIC 397 via the EIB 196 .
  • the MIC 398 provides a communication interface to shared memory 399 .
  • the BIC 397 provides a communication interface between the CBE 300 and other external buses and devices.
  • the PPE 310 is a dual threaded PPE 310 .
  • the combination of this dual threaded PPE 310 and the eight SPEs 320 - 334 makes the CBE 300 capable of handling 10 simultaneous threads and over 128 outstanding memory requests.
  • the PPE 310 acts as a controller for the other eight SPEs 320 - 334 which handle most of the computational workload.
  • the PPE 310 may be used to run conventional operating systems while the SPEs 320 - 334 perform vectorized floating point code execution, for example.
  • the SPEs 320 - 334 comprise a synergistic processing unit (SPU) 340 - 354 , memory flow control units 355 - 362 , local memory or store 363 - 370 , and an interface unit 380 - 394 .
  • the local memory or store 363 - 370 in one exemplary embodiment, comprises a 256 KB instruction and data memory which is visible to the PPE 310 and can be addressed directly by software.
  • the PPE 310 may load the SPEs 320 - 334 with small programs or threads, chaining the SPEs together to handle each step in a complex operation.
  • a set-top box incorporating the CBE 300 may load programs for reading a DVD, video and audio decoding, and display, and the data would be passed off from SPE to SPE until it finally ended up on the output display.
  • each SPE 320 - 334 gives a theoretical 32 GFLOPS of performance with the PPE 310 having a similar level of performance.
  • the memory flow control units (MFCs) 355 - 362 serve as an interface for an SPU to the rest of the system and other elements.
  • the MFCs 355 - 362 provide the primary mechanism for data transfer, protection, and synchronization between main storage and the local storages 363 - 370 .
  • Some implementations can share resources of a single MFC between multiple SPUs. In such a case, all the facilities and commands defined for the MFC must appear independent to software for each SPU. The effects of sharing an MFC are limited to implementation-dependent facilities and commands.
  • the MFC units 355 - 362 provide a channel interface and channel interface facilities through which the SPEs 320 - 334 may communicate with other SPEs, the PPE 310 , and other devices coupled to the EIB 396 .
  • the channel interface is a message passing interface that overlaps I/O with data processing and minimizes power consumed by synchronization.
  • the channel facilities of the channel interface provide a read channel and a write channel and may be accessed by a read channel instruction, a write channel instruction, and read channel count instruction which measures channel capacity. These channels may be configured as blocking or non-blocking channels. Channels that are defined as blocking channels cause the corresponding SPE 320 - 334 to stall when accessing a channel whose channel count is zero. In this sense, the channel count indicates the number of outstanding operations that can be issued for that channel without causing the SPE 320 - 334 to stall.
  • the channel facilities may monitor a variety of channel events using channel interface event and interrupt facilities.
  • the SPE's channel interface supports two signal notification channels, which may be used by the SPE 320 - 334 to receive signals originating from other processors, e.g., other SPE's, or devices in the system, which will be collectively referred to as external devices.
  • the signal notification channels can be programmed to operate in over-write mode.
  • Over-write mode is a mode in which the latest notification messages written to the channel “overwrite” previous notification messages written to the channel such that the processors of the SPEs 320 - 334 may access the most recent information written to the signal notification channel.
  • the over-write mode of the signal communication channels is useful in a one-to-one communications environment, or in a logical-OR communications environment, such as in a many-to-one signaling environment.
  • the channels themselves are essentially memory mapped address regions of shared memory 399 . That is, when a channel is establish, an address region of the shared memory 399 is allocated for use in storing the data that is to be passed on the channel between the SPE and the external device, e.g., another SPE or other type of device in the system.
  • a channel is “written to” this means that the data is written to the memory mapped address region of shared memory 399 allocated for that channel, i.e. a register in the shared memory 399 .
  • a channel is “read,” this means that the data stored in the memory mapped address region of the shared memory 399 allocated to the channel is read.
  • the channel is in “overwrite” mode, this means that the data in the mapped address region for the channel is overwritten with new data when new data is written to the channel.
  • the channel interface facilities include facilities for monitoring the memory mapped address region of the shared memory 399 that is allocated to the channels being maintained by the channel interface.
  • the channel interface facilities may “wake up” the processor of the SPE by generating an appropriate “wake up” message that is sent to the processor. In this way, the process may remain in a low power state until a “wake up” message is received. This mechanism is referred to as a blocking mode channel operation.
  • the availability of new data in the memory mapped address region may cause the channel interface facilities to generate an event or interrupt that is received and processed by an event or interrupt handler running on the processor of the corresponding SPE.
  • Cell Broadband Engine Architecture V1.0 is available at www-128.ibm.com/developerworks/power/cell/, which is hereby incorporated by reference.
  • a shared buffer region may reside, for example, in the shared memory 399 accessible via the memory interface controller 398 , or in a local store 363 - 370 associated with a SPE 320 - 334 .
  • This shared buffer region may be used to pass data between SPEs 320 - 334 so as to perform complex processing.
  • one mechanism for maintaining the head and tail pointers for a shared buffer region is to have the head and tail pointers stored in a local storage that is local to one of the consumer or the producer (see FIG. 2B ).
  • the head and tail pointers are stored in a local memory 363 - 370 associated with an SPE 320 - 334 .
  • one of the SPEs e.g., SPE 320
  • the head and tail pointers for the shared buffer region that stores this data may be stored in the local memory 363 associated with the SPE 320 , which is acting as a consumer, for example (although it could easily as well be stored in the local memory 368 associated with SPE 330 ).
  • the SPE 320 would update the tail pointer as that SPE 320 reads and processes data from the shared buffer region.
  • the SPE 330 would update the head pointer as that SPE 330 generates data and writes that data to the shared buffer region.
  • the SPEs 320 and 330 communicate the current status of the head and tail pointers for the shared buffer region via signal notification channels provided by the MFC units 355 - 362 .
  • the producer running on SPE 330 , may read its signal notification channel(s) and may operate in a blocking mode until the consumer, running on SPE 320 , updates the tail pointer information written to the signal notification channel(s).
  • the blocking mode may be used so as to ensure that overflows of the shared buffer region are not encountered. That is, as long as the shared buffer region is not full, the producer may continue writing data and updating the head pointer even though the tail pointer is not modified.
  • this “waiting” may be facilitated by the producer blocking the SPE's operation by reading the signal notification channel and waiting in a low-power state for the consumer to update the tail pointer.
  • the producer may use a write channel instruction that writes to a memory mapped address region, or register, associated with the channel of the consumer.
  • the producer may read data from the channel by using a read channel instruction that is directed to a memory mapped address region, or register, associated with the producer's signal notification channel.
  • the consumer may use a write channel instruction to write to the memory mapped address region, or register, associated with the producer's signal notification channel and a read channel instruction to read from a memory mapped address region, or register, associated with the consumer's signal notification channel.
  • the signal notification channels of the producer and/or consumer may operate in a blocking mode by using the read channel count instruction.
  • the channel interface of the SPE may periodically issue the read channel count instruction to the appropriate memory mapped address regions, or registers, associated with the channels to thereby determine if there is any data to be retrieved from these memory mapped address regions. If the count is zero, then the channel interface of the SPE leaves the processor of the producer/consumer SPE in a low power state, i.e. a “sleep” state. If the count is other than zero, then the channel interface may “wake up” the processor of the producer/consumer SPE by issuing a read channel instruction to the memory mapped address region for the channel and returning the results of the read channel instruction to the processor of the SPE.
  • the “low power state” or “sleep” state is managed entirely by hardware.
  • the hardware takes advantage of the blocked state of a SPE channel by reducing power consumption, and ramps the power back up when an external processor or device writes data to the SPE channel.
  • Programs implemented on traditional CPUs implement a “high power” busy-poll alternative. While a SPU program may likewise busy-poll by issuing “read channel count” instructions until data is available (channel count >0), such busy-poll loops burn power and require more code than a simple “read channel” instruction as used with the illustrative embodiments.
  • This mechanism of using the signal notification channel(s) provides a low-power, low-bandwidth alternative to monitoring, i.e. polling, a remote memory or register location that is local to the consumer/producer. Further, this solution provides low latency responses for real-time systems since the SPEs 320 and 330 are automatically awoken as soon as data becomes available on the signal notification channel(s).
  • the producer running on the SPE 330 , may use the channel interface's event and interrupt facilities to receive asynchronous notification of the consumer's completion of processing of data from the shared buffer region. That is, for example, the consumer may generate an interrupt or event when the consumer completes reading and processing of a portion of data from the shared buffer region. This event or interrupt may be directed to the event or interrupt facilities of the processor 350 associated with the producer's SPE 330 .
  • Each SPE is capable of receiving notification on a set of external events.
  • An SPE program can mask or unmask a set of events, and then monitor event status (such as by reading channel 0 or polling its channel count) or can choose to branch to an asynchronous interrupt handler when an event is received.
  • SPEs can enable or disable interrupts by executing the IRETE/IRETD instructions. If asynchronous interrupts are enabled and an unmasked event occurs (e.g. data present in a signal notification channel), then the SPE hardware (a) disables interrupts, (b) saves the current program counter, and (c) branches to interrupt dispatch code, which is fixed at local store offset 0. The program's interrupt handler may then read the external event status and proceed to handle the event.
  • the CBEA document incorporated above discusses these operations further under the “SPU Event Facility” section of the document.
  • the event and interrupt facilities may include an event or interrupt handler which performs special processing of the event or interrupt.
  • event or interrupt handlers are generally known in the art and thus, a detailed explanation of event/interrupt handlers is not provided herein.
  • SPU interrupts allow user-mode programs to directly access interrupt facilities. Many CPUs expose this function only to supervisor/OS code, which typically requires several layers of system calls to access. By having access to interrupt facilities, SPU applications programmers are able to improve the real-time response of their programs, and can use programming techniques that would not normally be available to them. This solution may be of use in environments where there is other work to be done by the SPE and blocking on a channel is not acceptable.
  • these embodiments may be extended by reversing the roles and having the SPE 330 act as a consumer, while the SPE 320 or other device external to the consumer acts as a producer. In such a case, the producer would notify the SPE 330 of buffer state updates by writing the head pointer to a signal notification channel of the SPE 330 .
  • the embodiments may be further extended by having the programs or threads running on the SPEs 320 and 330 operate as both a consumer (receiving head pointer updates from an external device or processor on a first signal notification channel, for example) and as a producer (receiving tail pointer updates from an external device or processor on a second signal notification channel, for example).
  • FIG. 4 is an exemplary block diagram illustrating an interaction between a producer and a consumer with regard to updating state information for a shared buffer region in accordance with one exemplary embodiment of the present invention where a process running on a SPE operates as a producer.
  • a SPE 410 operating as a producer of data to the shared buffer region 422 of the local store 414 , includes a processor 412 , the local store 414 , and a memory flow control unit (MFC) 416 .
  • the MFC unit 416 includes a channel interface 418 and its associated channel interface facilities.
  • the channel interface 418 further includes producer channel memory mapped register(s) 424 which store, among other things, a current state of the head pointer 415 and tail pointer 420 of the shared buffer region 422 .
  • the SPE 410 In writing data to the shared buffer region 422 , the SPE 410 also updates the head pointer 415 for the shared buffer region 422 and informs the consumer, external device 430 , of these updates by performing a memory flow control direct memory access (MFC-DMA) “put” request to a memory mapped register or portion of a local storage 426 , e.g., the head pointer register 419 , associated with the consumer device 430 .
  • MFC-DMA memory flow control direct memory access
  • the external device 430 which is any device that is “external” to the SPE 410 including other SPEs or devices coupled to the system bus, reads data from the shared buffer region 422 and, in response to the reading of data, updates a tail pointer 417 of the shared buffer region 422 in the local storage 426 of the external device 430 . If the external device is another SPE, the tail pointer 417 may be written to a channel interface of the consumer SPE. In order to update the tail pointer 420 on the producer SPE 410 , the external device 430 writes tail pointer update data to the tail pointer 417 in the consumer local storage 426 and performs a channel write operation to the producer channel memory mapped register(s) 424 indicating an update to the tail pointer.
  • each SPU channel is configured for the purpose of either read or write.
  • Each SPU channel has a capacity, or channel count, which indicates the number of outstanding read or write operations that may be performed against that channel.
  • each channel can be configured as either blocking or non-blocking.
  • a blocking channel will stall the SPU when there is no remaining capacity (e.g. the channel count is 0), while operations targeting a non-blocking channel always return immediately (in other words these channels always have a channel count of at least 1).
  • the signal notification channels are configured as blocking. When there is data available in the signal notification channels, the count is 1 and reads are returned immediately. When there is no data present, the count is zero, and the reads will block the SPU until data becomes available.
  • the channel interface 418 of the producer SPE 410 periodically sends a read channel count instruction to the producer channel memory mapped register(s) 424 which determines the current count of the producer channel memory mapped register(s) 424 . If this count is zero, the processor 412 is placed in a low power “sleep” state. If this count is other than zero, then-the channel interface 418 sends a read channel instruction to read the tail pointer update data from the tail pointer register 417 in the consumer's local storage 426 . The tail pointer update data may be used to update the tail pointer register 420 in the producer channel memory mapped register(s) 424 .
  • the processor 412 may be kept in a low power “sleep” state until the consumer, i.e. external device 430 , updates the tail pointer register 417 indicating that it has read data from the shared buffer region 422 and thus, additional data may be written to the shared buffer region 422 .
  • a copy (not shown) of the updated tail pointer 420 may be stored in the local store 414 of the producer SPE 410 .
  • the producer SPE 410 may reference this copy as long as there is sufficient space available in the shared buffer region 422 .
  • the producer SPE 410 may only update its local copy of the tail pointer 417 in the local store 414 when it detects that there is not sufficient space available in the shared buffer region 422 .
  • This example uses signal notification 1 (CH3) to receive tail pointer updates from the consumer.
  • This code will block on the channel read when no data is available in that channel. However if there is already data present in the channel, then the “read channel” instruction will return immediately.
  • the producer only blocks when (a) there is not sufficient space available in the shared buffer region and (b) the consumer has not already updated the tail pointer.
  • the external device 430 be another SPE or have channel interface facilities such as channel interface 418 .
  • the consumer may be any kind of processor or device that may consume data written to shared buffer region 422 . From the producer SPE's perspective, all of the devices, whether SPEs or not, are treated the same.
  • the producer SPE 410 issues an MFC-DMA “put” command to write the head pointer 415 to the external memory or register location associated with the external device 430 . From the SPE producer's point of view, the particulars of how the consumer waits for head pointer updates (and whether it blocks, polls, etc.) do not matter to the operation of the producer.
  • FIG. 5 is an exemplary block diagram illustrating an interaction between a producer and a consumer with regard to updating state information for a shared buffer region in accordance with one illustrative embodiment where a process running on a SPE operates as a consumer.
  • external device 530 writes data to the shared buffer region 522 via the channel interface 518 and, in so doing, also updates the head pointer 515 for the shared buffer region 522 in its local storage 526 .
  • the external device 530 In order to update the head pointer 520 on the consumer SPE 510 , the external device 530 writes the head pointer update data to the head pointer register 515 of local storage 526 and performs a write channel operation to the head pointer register 520 of the consumer channel memory mapped register(s) 524 indicating the update to the head pointer 515 .
  • the channel interface 518 of the SPE 510 periodically sends a read channel count instruction to the consumer channel memory mapped register(s) 524 to determine the count for the head pointer channel. If the count is zero, the processor 512 is maintained in a low power “sleep” state. If the count is other than zero, then the channel interface 518 sends a read channel instruction to the consumer channel memory mapped register(s) 524 to thereby obtain the head pointer update data from the local storage 526 of the external device 530 . This in turn causes the processor 512 to read data from the shared buffer region 522 and update the tail pointer 517 by writing the tail pointer update data to consumer channel memory mapped register(s) 524 .
  • the external device 530 may be informed of the update to the tail pointer 517 by way of an MFC-DMA “put” command in a similar manner as previously described above with regard to the update to the head pointer in FIG. 4 .
  • the external device 530 receives the notification of the update to the tail pointer 517 , may update its local copy of the tail pointer 519 , and may write data to the shared buffer region 522 based on the updated tail pointer. If the external device 530 is another SPE having similar channel interface facilities and local store, a similar operation as described above in FIG. 4 may be followed by external device 530 to thereby update a local copy of the head and tail pointers and perform write operations to the shared buffer region 522 .
  • FIG. 6 is an exemplary block diagram illustrating an interaction between a producer and a consumer with regard to updating state information for a shared buffer region in accordance with one exemplary embodiment of the present invention where a process running on a SPE operates as both a producer and a consumer.
  • a process running on a SPE operates as both a producer and a consumer.
  • the primary difference between this embodiment and the previous embodiments described in FIGS. 4 and 5 is that there are separate channels allocated to the SPE 410 that are provided for updates to the head pointer and the tail pointer.
  • a first channel e.g., SPE channel 1 memory mapped register(s) 624
  • a second channel e.g., SPE channel 2 memory mapped register(s) 625
  • the first channel is used when the SPE 610 is operating in a producer role and the second channel is used when the SPE 610 is operating in a consumer role.
  • the SPE 610 may read and write data to the shared buffer region 622 .
  • the channel interface 618 writes head pointer update data to the head pointer register 615 in the channel 1 memory mapped address region 624 and may notify the external device 630 of the update by way of a MFC-DMA “put” instruction, for example.
  • the SPE 610 may write tail pointer update data to the tail pointer register 627 in the channel 2 memory mapped address region 627 .
  • the update to the tail pointer may be notified to the external device 630 by way of a MFC-DMA “put” instruction to the tail pointer storage location 617 , for example.
  • the external device 630 may update the head pointer 626 and tail pointer 617 in local storage 626 based on operations performed by the external device 630 to either store or read data from the shared buffer region 622 .
  • the external device 630 may notify the SPE 610 of updates to the head and tail pointers using write channel instructions to respective ones of channel 1 and channel 2 memory mapped register(s) 624 and 625 .
  • the SPE 610 may update a copy of the tail pointer in local store 614 .
  • the SPE 610 may update a local copy of the head pointer in local store 614 .
  • the receipt of such updates may cause the processor 612 to exit a low power “sleep” state if the processor 612 were in such a “sleep” state in a similar manner as described above.
  • the SPE 610 may write head and tail pointer updates to the signal notification channels and receive updates to head and tail pointers via these signal notification channels. Since the broadband engine architecture allows for two signal notification channels for each SPE or external device, the SPEs or external devices may use these signal notification channels in the manner described above to facilitate notification of head and tail pointer updates for a shared buffer region.
  • FIG. 7 is a flowchart outlining an exemplary operation of the present invention in which a process running on a SPE operates as a producer. It will be understood that each block, and combination of blocks, of the flowchart illustration in FIG. 7 , and the flowchart illustrations in subsequent figures described hereafter, can be implemented by computer program instructions. These computer program instructions may be provided to a processor or other programmable data processing apparatus to produce a machine, such that the instructions which execute on the processor or other programmable data processing apparatus create means for implementing the functions specified in the flowchart block or blocks.
  • These computer program instructions may also be stored in a computer-readable memory or storage medium that can direct a processor or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory or storage medium produce an article of manufacture including instruction means which implement the functions specified in the flowchart block or blocks.
  • blocks of the flowchart illustrations support combinations of means for performing the specified functions, combinations of steps for performing the specified functions and program instruction means for performing the specified functions. It will also be understood that each block of the flowchart illustrations, and combinations of blocks in the flowchart illustrations, can be implemented by special purpose hardware-based computer systems which perform the specified functions or steps, or by combinations of special purpose hardware and computer instructions.
  • the operation starts by receiving data from the processor of the producer for writing to the shared buffer region (step 705 ).
  • the received data is written to the shared buffer region (step 710 ).
  • a head pointer for the shared buffer region is then updated (step 720 ).
  • the head pointer update data is written to a signal notification channel of a consumer device (step 730 ).
  • a determination is then made as to whether the current count on the producer's signal notification channel is zero (step 740 ). If so, the producer is placed in a low power operational mode or “sleep” state (step 750 ). If the producer is not to be placed in a low power operation mode, a determination is then made as to whether the operation should be discontinued (step 760 ).
  • step 770 a determination is made as to whether the operation should be discontinued. If not, the operation returns to step 740 . If the operation is to be discontinued from either of step 760 or step 770 , the operation terminates.
  • FIG. 8 is a flowchart outlining an exemplary operation of the present invention in which a process running on a SPE operates as a consumer.
  • the operation starts by reading data from the shared buffer region (step 810 ).
  • a tail pointer for the shared buffer region is then updated (step 820 ).
  • the tail pointer update data is written to a signal notification channel of a producer device (step 830 ).
  • a determination is then made as to whether the current count on the consumer's signal notification channel is zero (step 840 ). If so, the consumer is placed in a low power operational mode or “sleep” state (step 850 ).
  • step 860 a determination is then made as to whether the operation should be discontinued. This may involve tearing down the channel between the producer and the consumer, or other events that cause the producer/consumer relationship to be discontinued. If not, the operation returns to step 810 and the operation is repeated. If the consumer is placed in a low power operation mode, a determination is made as to whether the operation should be discontinued (step 870 ). If not, the operation returns to step 840 . If the operation is to be discontinued from either of step 860 or step 870 , the operation terminates.
  • FIGS. 7 and 8 may also be used in conjunction in a single apparatus.
  • an SPE that operates as both a consumer and a producer may perform the above operations outlined in FIGS. 7 and 8 in substantially parallel so as to both produce data for writing to the shared buffer region and read data from the shared buffer region.
  • the present invention provides a mechanism for notifying devices, such as synergistic processing elements, of updates to head and tail pointers of a shared buffer region.
  • the mechanism of the present invention avoids the overhead and bus congestion of polling head and tail pointers by providing mechanisms for writing updates to signal notification channels of the devices and using channel interface facilities of the devices to update the head and tail pointers.
  • the present invention allows a device to remain in a low power “sleep” state until an update to the head or tail pointer is received from an external device at which time the device is awoken so that it may perform processing to either produce data to or consume data from the shared buffer region. This greatly decreases the power consumption of the system as a whole.

Abstract

An apparatus and method for efficient communication of producer/consumer buffer status are provided. With the apparatus and method, devices in a data processing system notify each other of updates to head and tail pointers of a shared buffer region when the devices perform operations on the shared buffer region using signal notification channels of the devices. Thus, when a producer device that produces data to the shared buffer region writes data to the shared buffer region, an update to the head pointer is written to a signal notification channel of a consumer device. When a consumer device reads data from the shared buffer region, the consumer device writes a tail pointer update to a signal notification channel of the producer device. In addition, channels may operate in a blocking mode so that the corresponding device is kept in a low power state until an update is received over the channel.

Description

    BACKGROUND
  • 1. Technical Field
  • The present application relates generally to an improved data processing system. More specifically, the present application is directed to an apparatus and method for efficient communication of producer/consumer buffer status.
  • 2. Description of Related Art
  • In a data processing system, one processor or device typically acts as a producer by placing data into a shared buffer while another processor or device acts as a consumer by reading and processing the data placed there by the producer. For example, consider a data processing system where a producer and a consumer, independently operating on processors or devices within the data processing system, have access to a shared data buffer region. The shared data buffer region may be located, for example, in shared system or I/O memory and is commonly accessible by both parties. In addition to the shared buffer region, two state variables are shared between the producer and consumer: a head pointer (or index) and a tail pointer (or index), as shown in FIG. 1A.
  • FIG. 1A is an exemplary diagram of a shared buffer region in accordance with a prior art system. As shown in FIG. 1A, the shared buffer region 100 has an associated tail pointer 120 and head pointer 110. The head pointer 110 represents the next unused location in the shared buffer region 100. The tail pointer 120 represents the last unprocessed data location in the shared buffer region 100.
  • As a producer places data into the shared buffer region 100, the producer increments the head pointer 110. In this way, the producer indicates to consumers that new data is available for the consumers. Likewise, as the consumer retrieves data from the shared buffer region 100, the consumer increments the tail pointer 120, thereby indicating that data has been read from the shared buffer region 100. The amount of data available in the shared buffer region 100 is indicated by the difference between the head pointer 110 and tail pointer 120. This is referred to as the “active” data region 140 in the shared buffer region 100. The portion of the shared buffer region 100 that is behind the tail pointer 120 is referred to as the “inactive” (or old) data region 130. An empty shared buffer region is identified by setting the head pointer 110 and tail pointer 120 to the same location in the shared buffer region 100.
  • When there is no constraint on the size of the shared buffer region, i.e. when there is an infinite amount of storage space available for the shared buffer region 100, the head pointer 110 and tail pointer 120 may be incremented indefinitely. However, in most systems, it is more realistic to have only a fixed amount of storage space available for the shared buffer region 100. In this case, it is possible to reclaim the inactive region 130 by implementing what is known as a circular first-in-first-out (FIFO) type shared buffer region 100. This type of shared buffer region is illustrated in FIG. 1B.
  • As shown in FIG. 1B, the head pointer 110 and tail pointer 120 are incremented as described above with regard to FIG. 1A, however when the end of the shared buffer region 100 is encountered, the head and tail pointers 110 and 120 are permitted to wrap back to the beginning of the shared buffer region 100. As with the shared buffer region 100 shown in FIG. 1A, the active data region 140 is defined as the portion of the shared buffer region 100 between the head pointer 110 and tail pointer 120. However, with the FIFO shared buffer region, both the producer and consumer must now account for the situation where the head pointer 110 has wrapped beyond the end of the shared buffer region 100, i.e. back to the beginning of the shared buffer region 100, but the tail pointer 120 has not wrapped.
  • This requires efficient communication of the current values for the head and tail pointers 110 and 120 to both the producer and the consumer. The consumer is primarily concerned with monitoring updates to the head pointer 110 and the producer is primarily concerned with monitoring updates to the tail pointer 120. Before a consumer can begin processing, it must know that the head pointer 110 and tail pointer 120 point to different entries in the shared buffer region 100. Likewise, before the producer can add data to the shared buffer region 100, the producer must ensure that the amount of data to be added will not overflow the tail pointer 120.
  • A simple implementation may involve storing the state variables, i.e. the head pointer 110 and tail pointer 120, in shared system memory and monitoring these state variables remotely. FIG. 2A illustrates this exemplary implementation.
  • As shown in FIG. 2A a producer 250 performs updates to the head pointer 210 in shared memory 230 and monitors the current state of tail pointer 220 in shared memory 230 when writing data to the shared buffer region 240. Similarly, the consumer 260 updates the tail pointer 220 in shared memory 230 and monitors the current state of the head pointer 210 in shared memory 230 when reading data from the shared buffer region 240.
  • The producer 250 and consumer 260 each perform updates and monitoring of the head pointer 210 and tail pointer 220 in shared memory 230 remotely. That is, the head pointer 210 and tail pointer 220 are not located in a memory local to the producer 250 or the consumer 260. Thus, the producer 250 and consumer 260 must access these pointers 210 and 220 via a bus or other communication interface that is not local to the producer 250 or consumer 260. Typically, the producer 250 and consumer 260 poll the remote shared memory 230 in order to obtain information regarding the head and tail pointers 210 and 220. As a result, this updating and monitoring requires additional resources, e.g., processor cycles, bus bandwidth, etc., within the system to support these operations. Because of the need to utilized additional system resources, this solution may be considered too detrimental to overall system performance.
  • In an alternative approach, the head and tail pointers 210 and 220 are maintained in a memory or registers that are local to either the producer 250 or the consumer 260 in order to improve performance of that one party, i.e. the producer 250 or the consumer 260, to which the pointers 210 and 220 are made local. FIG. 2B illustrates such an alternative implementation in which the head pointer 210 and tail pointer 220 are stored in a local memory region 270 to one of the producer 250 and consumer 260 (in the depicted example, the local memory region 270 is local to the consumer 260).
  • As shown in FIG. 2B, the head pointer 210 and tail pointer 220 are stored in local memory region 270 that is local to consumer 260. This local memory region 270 is separate from shared memory 230 and shared buffer region 240. The producer 250 must remotely monitor, i.e. poll, the value of the tail pointer 220 in order to determine the amount of free space remaining in the shared buffer region 240. This may lead to serious degradation of performance for the producer, as the latency for reading remote locations can be several orders of magnitude worse than for main shared memory storage. Of course the situation could be reversed, requiring that the consumer 260 remotely monitor the head pointer 210 in a local memory region local to the producer 250, for example.
  • SUMMARY OF THE INVENTION
  • In view of the above, it would be beneficial to have an improved mechanism for efficient communication of shared buffer status between producers and consumers in a data processing system. The illustrative embodiments provide such an improved mechanism by utilizing already existing communication channels established with processing elements of the data processing system to communicate this shared buffer status information.
  • With the illustrative embodiments, synergistic processor elements (SPEs) are provided in a data processing system. The SPEs are a simple, yet highly optimized, processor with a single instruction, multiple data (SIMD) instruction set. The SPEs have a large register file, a high speed local memory, and a channel interface that serves as the primary communication pathway between an SPE and external processors or devices.
  • The channel interface is a message passing interface that overlaps input/output (I/O) with data processing and minimizes power consumed by synchronization. The channel facilities of the channel interface are accessed with three types of instructions: a read channel instruction, write channel instruction, and read channel count instruction which measures channel capacity. These channels may be configured as blocking or non-blocking channels. Channels that are defined as blocking channels cause the SPE to stall when accessing a channel whose channel count is zero. In this sense, the channel count indicates the number of outstanding operations that can be issued for that channel without causing the SPE to stall. Aside from blocking on channel access, an SPE program can monitor a variety of channel events using SPE interrupt facilities.
  • The SPE channel interface supports two signal notification channels, which may be used by the SPE to receive signals originating from other processors or devices in the system. The signal notification channels can be programmed to operate in over-write mode. Over-write mode is a mode in which the latest notification messages written to the channel “overwrite” previous notification messages written to the channel such that the processors may access the most recent information written to the signal notification channel. The over-write mode of the signal communication channels is useful in a one-to-one communications environment, or in a logical-OR communications environment, such as in a many-to-one signaling environment.
  • With the illustrative embodiments, the producer, running on an SPE, may read its signal notification channel(s) and operate in a blocking mode until the consumer updates the tail pointer information written to the signal notification channel(s). This provides a low-power, low-bandwidth alternative to monitoring, i.e. polling, a remote memory or register location that is local to the consumer. Further, this solution provides low latency responses for real-time systems since the SPE is automatically awoken as soon as data becomes available on the signal notification channel(s).
  • In another illustrative embodiment, the producer running on the SPE may use the processor's event and interrupt facilities to receive asynchronous notification of consumer's completion. This solution may be of use in environments where there is other work to be done by the SPE and blocking on a channel is not an acceptable alternative.
  • Of course these illustrative embodiments may be extended by reversing the roles and having the SPE program act as a consumer, while an external processor or device acts as a producer. In such a case, the producer would notify the SPE of buffer state updates by writing the head pointer to a signal notification register. The embodiments may be further extended by having the SPE program operate as both a consumer (receiving head pointer updates from an external device or processor on a first signal notification channel, for example) and as a producer (receiving tail pointer updates from an external device or processor on a second signal notification channel, for example).
  • In one illustrative embodiment, a method is provided in which an operation on a shared buffer region is performed using a first device in a data processing system. The method may further comprise writing a pointer update to a signal notification channel associated with a second device in the data processing system. The pointer update may be an update to one of a head pointer or a tail pointer of the shared buffer region. The method may further comprise updating a pointer stored in a local store of the second device based on the pointer update. The signal notification channel may be a memory mapped address region of a shared memory of the data processing system.
  • The operation on the shared buffer region may be a write operation for writing data to the shared buffer region. The pointer update may be an update to a head pointer of the shared buffer region.
  • Alternatively, the operation on the shared buffer region may be a read operation for reading data from the shared buffer region. The pointer update may be an update to a tail pointer of the shared buffer region.
  • The method may further comprise determining if the shared buffer region is full and placing the second device in a low-power consumption waiting state waiting for a predetermined amount of storage space to be freed in the shared buffer region before additional data may be written to the shared buffer region. Placing the second device in a low-power consumption waiting state may comprise placing the notification channel in a blocking mode of operation in which the second device's normal operation is blocked and the second device waits in a low-power consumption state.
  • The operation may be a read operation for reading data from the shared buffer region. The notification channel may exit the blocking mode of operation in response to the read operation.
  • While in the blocking mode of operation, a channel interface of the second device may periodically issue a read channel count instruction to the signal notification channel to determine a count associated with the notification channel. If the count associated with the notification channel is zero, then the notification channel may remain in a blocking mode of operation. If the count associated with the notification channel is non-zero, then the notification channel may exit the blocking mode of operation. In response to the notification channel exiting the blocking mode of operation, the second device may be awoken by issuing a read channel instruction to the notification channel and returning results of the read channel instruction to the second device.
  • The second device may be a synergistic processing unit and the shared buffer region is part of the local store associated with the synergistic processing unit. Writing a pointer update may be performed using a channel interface of a memory flow control unit associated with the synergistic processing unit. Moreover, the first device and the second device may be synergistic processing units of a multiprocessor system-on-a-chip.
  • In another illustrative embodiment, a computer program product comprising a computer usable medium having a computer readable program is provided. The computer readable program, when executed on a computing device, may cause the computing device to perform, using a first device in the data processing system, an operation on a shared buffer region and write a pointer update to a signal notification channel associated with a second device in the data processing system. The pointer update may be an update to one of a head pointer or a tail pointer of the shared buffer region. The computer readable program may further cause the computing device to update a pointer stored in a local store of the second device based on the pointer update. Moreover, the computer readable program may further cause the computing device to perform the various operations outlined above with regard to the method illustrative embodiment.
  • In yet another illustrative embodiment, a system is provided that comprises a first processor, a second processor coupled to the first processor, and a local storage device coupled to the second processor. The first processor may perform an operation on a shared buffer region of the local storage device and may write a pointer update to a signal notification channel associated with the second processor. The pointer update may be an update to one of a head pointer or a tail pointer of the shared buffer region. The second processor may update a pointer stored in the local store of the second processor based on the pointer update.
  • The operation on the shared buffer region may be a write operation for writing data to the shared buffer region. The pointer update may be an update to a head pointer of the shared buffer region.
  • Alternatively, the operation on the shared buffer region may be a read operation for reading data from the shared buffer region. The pointer update may be an update to a tail pointer of the shared buffer region.
  • The second processor may determine if the shared buffer region is full and may place itself in a low-power consumption waiting state waiting for a predetermined amount of storage space to be freed in the shared buffer region before additional data may be written to the shared buffer region. The second processor may place itself in a low-power consumption waiting state by placing the notification channel in a blocking mode of operation in which the second processor's normal operation is blocked and the second processor waits in a low-power consumption state.
  • The operation may be a read operation for reading data from the shared buffer region. The notification channel may exit the blocking mode of operation in response to the read operation. While in the blocking mode of operation, a channel interface of the second processor may periodically issue a read channel count instruction to the signal notification channel to determine a count associated with the notification channel. If the count associated with the notification channel is non-zero, then the notification channel may exit the blocking mode of operation.
  • These and other features and advantages of the present invention will be described in, or will become apparent to those of ordinary skill in the art in view of, the following detailed description of the exemplary embodiments of the present invention.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The novel features believed characteristic of the invention are set forth in the appended claims. The invention itself, however, as well as a preferred mode of use, further objectives and advantages thereof, will best be understood by reference to the following detailed description of an illustrative embodiment when read in conjunction with the accompanying drawings, wherein:
  • FIG. 1A is an exemplary diagram of a shared buffer region in accordance with a known mechanism;
  • FIG. 1B is an exemplary diagram of a FIFO shared buffer region in accordance with a known mechanism;
  • FIG. 2A is an exemplary diagram illustrating a mechanism for storing head and tail pointers in a shared memory in accordance with a known mechanism;
  • FIG. 2B is an exemplary diagram illustrating an alternative implementation in which the head pointer and tail pointer are stored in a local memory region that is local to a consumer in accordance with a known mechanism;
  • FIG. 3 is an exemplary block diagram of a data processing system in which aspects of the illustrative embodiments may be implemented;
  • FIG. 4 is an exemplary block diagram illustrating an interaction between a producer and a consumer with regard to updating state information for a shared buffer region in accordance with one illustrative embodiment where a process running on a SPE operates as a producer;
  • FIG. 5 is an exemplary block diagram illustrating an interaction between a producer and a consumer with regard to updating state information for a shared buffer region in accordance with one illustrative embodiment where a process running on a SPE operates as a consumer;
  • FIG. 6 is an exemplary block diagram illustrating an interaction between a producer and a consumer with regard to updating state information for a shared buffer region in accordance with one illustrative embodiment where a process running on a SPE operates as both a producer and a consumer;
  • FIG. 7 is a flowchart outlining an exemplary operation of an illustrative embodiment in which a process running on a SPE operates as a producer; and
  • FIG. 8 is a flowchart outlining an exemplary operation of an illustrative embodiment in which a process running on a SPE operates as a consumer.
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
  • The following is intended to provide a detailed description of illustrative embodiments and should not be taken to be limiting of the invention itself. Rather, any number of variations may fall within the scope of the invention defined in the claims following the description.
  • FIG. 3 is an exemplary block diagram of a data processing system in which aspects of the present invention may be implemented. The exemplary data processing system shown in FIG. 3 is an example of the Cell Broadband Engine (CBE) data processing system. While the CBE will be used in the description of the preferred embodiments of the present invention, the present invention is not limited to such, as will be readily apparent to those of ordinary skill in the art upon reading the following description.
  • As shown in FIG. 3, the CBE 300 includes a power processor element (PPE) 310 having a processor (PPU) 316 and its L1 and L2 caches 312 and 314, and multiple synergistic processor elements (SPEs) 320-334 that each has its own synergistic processor unit (SPU) 340-354, memory flow control 355-362, local memory or store (LS) 363-370, and bus interface unit (BIU unit) 380-394 which may be, for example, a combination direct memory access (DMA), memory management unit (MMU), and bus interface unit. A high bandwidth internal element interconnect bus (EIB) 396, a bus interface controller (BIC) 397, and a memory interface controller (MIC) 398 are also provided.
  • The CBE 300 may be a system-on-a-chip such that each of the elements depicted in FIG. 3 may be provided on a single microprocessor chip. Moreover, the CBE 300 is a heterogeneous processing environment in which each of the SPUs may receive different instructions from each of the other SPUs in the system. Moreover, the instruction set for the SPUs is different from that of the PPU, e.g., the PPU may execute Reduced Instruction Set Computer (RISC) based instructions while the SPU execute vectorized instructions.
  • The SPEs 320-334 are coupled to each other and to the L2 cache 314 via the EIB 396. In addition, the SPEs 320-334 are coupled to MIC 398 and BIC 397 via the EIB 196. The MIC 398 provides a communication interface to shared memory 399. The BIC 397 provides a communication interface between the CBE 300 and other external buses and devices.
  • The PPE 310 is a dual threaded PPE 310. The combination of this dual threaded PPE 310 and the eight SPEs 320-334 makes the CBE 300 capable of handling 10 simultaneous threads and over 128 outstanding memory requests. The PPE 310 acts as a controller for the other eight SPEs 320-334 which handle most of the computational workload. The PPE 310 may be used to run conventional operating systems while the SPEs 320-334 perform vectorized floating point code execution, for example.
  • The SPEs 320-334 comprise a synergistic processing unit (SPU) 340-354, memory flow control units 355-362, local memory or store 363-370, and an interface unit 380-394. The local memory or store 363-370, in one exemplary embodiment, comprises a 256 KB instruction and data memory which is visible to the PPE 310 and can be addressed directly by software.
  • The PPE 310 may load the SPEs 320-334 with small programs or threads, chaining the SPEs together to handle each step in a complex operation. For example, a set-top box incorporating the CBE 300 may load programs for reading a DVD, video and audio decoding, and display, and the data would be passed off from SPE to SPE until it finally ended up on the output display. At 4 GHz, each SPE 320-334 gives a theoretical 32 GFLOPS of performance with the PPE 310 having a similar level of performance.
  • The memory flow control units (MFCs) 355-362 serve as an interface for an SPU to the rest of the system and other elements. The MFCs 355-362 provide the primary mechanism for data transfer, protection, and synchronization between main storage and the local storages 363-370. There is logically an MFC for each SPU in a processor. Some implementations can share resources of a single MFC between multiple SPUs. In such a case, all the facilities and commands defined for the MFC must appear independent to software for each SPU. The effects of sharing an MFC are limited to implementation-dependent facilities and commands.
  • The MFC units 355-362 provide a channel interface and channel interface facilities through which the SPEs 320-334 may communicate with other SPEs, the PPE 310, and other devices coupled to the EIB 396. The channel interface is a message passing interface that overlaps I/O with data processing and minimizes power consumed by synchronization. The channel facilities of the channel interface provide a read channel and a write channel and may be accessed by a read channel instruction, a write channel instruction, and read channel count instruction which measures channel capacity. These channels may be configured as blocking or non-blocking channels. Channels that are defined as blocking channels cause the corresponding SPE 320-334 to stall when accessing a channel whose channel count is zero. In this sense, the channel count indicates the number of outstanding operations that can be issued for that channel without causing the SPE 320-334 to stall. Aside from blocking on channel access, the channel facilities may monitor a variety of channel events using channel interface event and interrupt facilities.
  • The SPE's channel interface supports two signal notification channels, which may be used by the SPE 320-334 to receive signals originating from other processors, e.g., other SPE's, or devices in the system, which will be collectively referred to as external devices. The signal notification channels can be programmed to operate in over-write mode. Over-write mode is a mode in which the latest notification messages written to the channel “overwrite” previous notification messages written to the channel such that the processors of the SPEs 320-334 may access the most recent information written to the signal notification channel. The over-write mode of the signal communication channels is useful in a one-to-one communications environment, or in a logical-OR communications environment, such as in a many-to-one signaling environment.
  • The channels themselves are essentially memory mapped address regions of shared memory 399. That is, when a channel is establish, an address region of the shared memory 399 is allocated for use in storing the data that is to be passed on the channel between the SPE and the external device, e.g., another SPE or other type of device in the system. Thus, when a channel is “written to” this means that the data is written to the memory mapped address region of shared memory 399 allocated for that channel, i.e. a register in the shared memory 399. Similarly, when a channel is “read,” this means that the data stored in the memory mapped address region of the shared memory 399 allocated to the channel is read. When the channel is in “overwrite” mode, this means that the data in the mapped address region for the channel is overwritten with new data when new data is written to the channel.
  • The channel interface facilities include facilities for monitoring the memory mapped address region of the shared memory 399 that is allocated to the channels being maintained by the channel interface. When new data is made available in the memory mapped address region, the channel interface facilities may “wake up” the processor of the SPE by generating an appropriate “wake up” message that is sent to the processor. In this way, the process may remain in a low power state until a “wake up” message is received. This mechanism is referred to as a blocking mode channel operation. Alternatively, the availability of new data in the memory mapped address region may cause the channel interface facilities to generate an event or interrupt that is received and processed by an event or interrupt handler running on the processor of the corresponding SPE.
  • More information regarding the Cell Broadband Engine may be found at the IBM DeveloperWorks website. Specifically, the Cell Broadband Engine architecture document “Cell Broadband Engine Architecture V1.0” is available at www-128.ibm.com/developerworks/power/cell/, which is hereby incorporated by reference.
  • Using the broadband engine described in FIG. 3 as the basis for a data processing system in which an illustrative embodiment may be implemented, it can be seen that a shared buffer region may reside, for example, in the shared memory 399 accessible via the memory interface controller 398, or in a local store 363-370 associated with a SPE 320-334. This shared buffer region may be used to pass data between SPEs 320-334 so as to perform complex processing. Thus, it is necessary for the SPEs 320-334 that share the shared buffer region to be able to access information regarding the head and tail pointers for the shared buffer region.
  • As mentioned above, one mechanism for maintaining the head and tail pointers for a shared buffer region is to have the head and tail pointers stored in a local storage that is local to one of the consumer or the producer (see FIG. 2B). Thus, in one exemplary embodiment of the present invention, the head and tail pointers are stored in a local memory 363-370 associated with an SPE 320-334. For example, in an operation involving two SPEs, e.g., SPE 320 and SPE 330, one of the SPEs, e.g., SPE 320, may run a program or thread that operates as a consumer of data produced by a program or thread running on the other SPE, e.g., SPE 330. The head and tail pointers for the shared buffer region that stores this data may be stored in the local memory 363 associated with the SPE 320, which is acting as a consumer, for example (although it could easily as well be stored in the local memory 368 associated with SPE 330). The SPE 320 would update the tail pointer as that SPE 320 reads and processes data from the shared buffer region. The SPE 330 would update the head pointer as that SPE 330 generates data and writes that data to the shared buffer region.
  • The SPEs 320 and 330 communicate the current status of the head and tail pointers for the shared buffer region via signal notification channels provided by the MFC units 355-362. With the illustrative embodiment, the producer, running on SPE 330, may read its signal notification channel(s) and may operate in a blocking mode until the consumer, running on SPE 320, updates the tail pointer information written to the signal notification channel(s). The blocking mode may be used so as to ensure that overflows of the shared buffer region are not encountered. That is, as long as the shared buffer region is not full, the producer may continue writing data and updating the head pointer even though the tail pointer is not modified. However when the producer detects that the shared buffer region is full (or may become full by writing some number of bytes to it), such as by looking at the head and tail pointer values, then the producer needs to wait for the desired amount of free space to become available in the shared buffer region. With the illustrative embodiments, this “waiting” may be facilitated by the producer blocking the SPE's operation by reading the signal notification channel and waiting in a low-power state for the consumer to update the tail pointer.
  • In writing to a notification channel, the producer may use a write channel instruction that writes to a memory mapped address region, or register, associated with the channel of the consumer. The producer may read data from the channel by using a read channel instruction that is directed to a memory mapped address region, or register, associated with the producer's signal notification channel. Similarly, the consumer may use a write channel instruction to write to the memory mapped address region, or register, associated with the producer's signal notification channel and a read channel instruction to read from a memory mapped address region, or register, associated with the consumer's signal notification channel.
  • The signal notification channels of the producer and/or consumer may operate in a blocking mode by using the read channel count instruction. The channel interface of the SPE may periodically issue the read channel count instruction to the appropriate memory mapped address regions, or registers, associated with the channels to thereby determine if there is any data to be retrieved from these memory mapped address regions. If the count is zero, then the channel interface of the SPE leaves the processor of the producer/consumer SPE in a low power state, i.e. a “sleep” state. If the count is other than zero, then the channel interface may “wake up” the processor of the producer/consumer SPE by issuing a read channel instruction to the memory mapped address region for the channel and returning the results of the read channel instruction to the processor of the SPE.
  • The “low power state” or “sleep” state is managed entirely by hardware. The hardware takes advantage of the blocked state of a SPE channel by reducing power consumption, and ramps the power back up when an external processor or device writes data to the SPE channel. Programs implemented on traditional CPUs implement a “high power” busy-poll alternative. While a SPU program may likewise busy-poll by issuing “read channel count” instructions until data is available (channel count >0), such busy-poll loops burn power and require more code than a simple “read channel” instruction as used with the illustrative embodiments.
  • This mechanism of using the signal notification channel(s) provides a low-power, low-bandwidth alternative to monitoring, i.e. polling, a remote memory or register location that is local to the consumer/producer. Further, this solution provides low latency responses for real-time systems since the SPEs 320 and 330 are automatically awoken as soon as data becomes available on the signal notification channel(s).
  • In another illustrative embodiment, the producer, running on the SPE 330, may use the channel interface's event and interrupt facilities to receive asynchronous notification of the consumer's completion of processing of data from the shared buffer region. That is, for example, the consumer may generate an interrupt or event when the consumer completes reading and processing of a portion of data from the shared buffer region. This event or interrupt may be directed to the event or interrupt facilities of the processor 350 associated with the producer's SPE 330.
  • Each SPE is capable of receiving notification on a set of external events. An SPE program can mask or unmask a set of events, and then monitor event status (such as by reading channel 0 or polling its channel count) or can choose to branch to an asynchronous interrupt handler when an event is received. SPEs can enable or disable interrupts by executing the IRETE/IRETD instructions. If asynchronous interrupts are enabled and an unmasked event occurs (e.g. data present in a signal notification channel), then the SPE hardware (a) disables interrupts, (b) saves the current program counter, and (c) branches to interrupt dispatch code, which is fixed at local store offset 0. The program's interrupt handler may then read the external event status and proceed to handle the event. The CBEA document incorporated above discusses these operations further under the “SPU Event Facility” section of the document.
  • As mentioned above, the event and interrupt facilities may include an event or interrupt handler which performs special processing of the event or interrupt. Such event or interrupt handlers are generally known in the art and thus, a detailed explanation of event/interrupt handlers is not provided herein. However, SPU interrupts allow user-mode programs to directly access interrupt facilities. Many CPUs expose this function only to supervisor/OS code, which typically requires several layers of system calls to access. By having access to interrupt facilities, SPU applications programmers are able to improve the real-time response of their programs, and can use programming techniques that would not normally be available to them. This solution may be of use in environments where there is other work to be done by the SPE and blocking on a channel is not acceptable.
  • Of course these embodiments may be extended by reversing the roles and having the SPE 330 act as a consumer, while the SPE 320 or other device external to the consumer acts as a producer. In such a case, the producer would notify the SPE 330 of buffer state updates by writing the head pointer to a signal notification channel of the SPE 330. The embodiments may be further extended by having the programs or threads running on the SPEs 320 and 330 operate as both a consumer (receiving head pointer updates from an external device or processor on a first signal notification channel, for example) and as a producer (receiving tail pointer updates from an external device or processor on a second signal notification channel, for example).
  • FIG. 4 is an exemplary block diagram illustrating an interaction between a producer and a consumer with regard to updating state information for a shared buffer region in accordance with one exemplary embodiment of the present invention where a process running on a SPE operates as a producer. As shown in FIG. 4, a SPE 410, operating as a producer of data to the shared buffer region 422 of the local store 414, includes a processor 412, the local store 414, and a memory flow control unit (MFC) 416. The MFC unit 416 includes a channel interface 418 and its associated channel interface facilities. The channel interface 418 further includes producer channel memory mapped register(s) 424 which store, among other things, a current state of the head pointer 415 and tail pointer 420 of the shared buffer region 422.
  • In writing data to the shared buffer region 422, the SPE 410 also updates the head pointer 415 for the shared buffer region 422 and informs the consumer, external device 430, of these updates by performing a memory flow control direct memory access (MFC-DMA) “put” request to a memory mapped register or portion of a local storage 426, e.g., the head pointer register 419, associated with the consumer device 430.
  • The external device 430, which is any device that is “external” to the SPE 410 including other SPEs or devices coupled to the system bus, reads data from the shared buffer region 422 and, in response to the reading of data, updates a tail pointer 417 of the shared buffer region 422 in the local storage 426 of the external device 430. If the external device is another SPE, the tail pointer 417 may be written to a channel interface of the consumer SPE. In order to update the tail pointer 420 on the producer SPE 410, the external device 430 writes tail pointer update data to the tail pointer 417 in the consumer local storage 426 and performs a channel write operation to the producer channel memory mapped register(s) 424 indicating an update to the tail pointer.
  • In response to performing the channel write operation, the channel count for the producer channel memory mapped register(s) 424 is incremented. Each SPU channel is configured for the purpose of either read or write. Each SPU channel has a capacity, or channel count, which indicates the number of outstanding read or write operations that may be performed against that channel. Further, each channel can be configured as either blocking or non-blocking. A blocking channel will stall the SPU when there is no remaining capacity (e.g. the channel count is 0), while operations targeting a non-blocking channel always return immediately (in other words these channels always have a channel count of at least 1). The signal notification channels are configured as blocking. When there is data available in the signal notification channels, the count is 1 and reads are returned immediately. When there is no data present, the count is zero, and the reads will block the SPU until data becomes available.
  • The channel interface 418 of the producer SPE 410 periodically sends a read channel count instruction to the producer channel memory mapped register(s) 424 which determines the current count of the producer channel memory mapped register(s) 424. If this count is zero, the processor 412 is placed in a low power “sleep” state. If this count is other than zero, then-the channel interface 418 sends a read channel instruction to read the tail pointer update data from the tail pointer register 417 in the consumer's local storage 426. The tail pointer update data may be used to update the tail pointer register 420 in the producer channel memory mapped register(s) 424. It should be noted that by using the read channel count instruction of the channel interface 418, the processor 412 may be kept in a low power “sleep” state until the consumer, i.e. external device 430, updates the tail pointer register 417 indicating that it has read data from the shared buffer region 422 and thus, additional data may be written to the shared buffer region 422.
  • A copy (not shown) of the updated tail pointer 420 may be stored in the local store 414 of the producer SPE 410. The producer SPE 410 may reference this copy as long as there is sufficient space available in the shared buffer region 422. The producer SPE 410 may only update its local copy of the tail pointer 417 in the local store 414 when it detects that there is not sufficient space available in the shared buffer region 422.
  • Pseudo-code for waiting for space to become available may be implemented on the SPU as follows:
    void wait_space(int nbytes)
    {
    while (space( ) < nbytes)
    tail = spu_readch(3);
    }
  • This example uses signal notification 1 (CH3) to receive tail pointer updates from the consumer. This code will block on the channel read when no data is available in that channel. However if there is already data present in the channel, then the “read channel” instruction will return immediately. Thus, the producer only blocks when (a) there is not sufficient space available in the shared buffer region and (b) the consumer has not already updated the tail pointer.
  • It should be noted that there is no requirement that the external device 430 be another SPE or have channel interface facilities such as channel interface 418. The consumer may be any kind of processor or device that may consume data written to shared buffer region 422. From the producer SPE's perspective, all of the devices, whether SPEs or not, are treated the same. The producer SPE 410 issues an MFC-DMA “put” command to write the head pointer 415 to the external memory or register location associated with the external device 430. From the SPE producer's point of view, the particulars of how the consumer waits for head pointer updates (and whether it blocks, polls, etc.) do not matter to the operation of the producer.
  • Similar mechanisms may be employed when the SPE operates as a consumer and the external device operates as a producer of data to the shared buffer region. FIG. 5 is an exemplary block diagram illustrating an interaction between a producer and a consumer with regard to updating state information for a shared buffer region in accordance with one illustrative embodiment where a process running on a SPE operates as a consumer. As shown in FIG. 5, external device 530 writes data to the shared buffer region 522 via the channel interface 518 and, in so doing, also updates the head pointer 515 for the shared buffer region 522 in its local storage 526. In order to update the head pointer 520 on the consumer SPE 510, the external device 530 writes the head pointer update data to the head pointer register 515 of local storage 526 and performs a write channel operation to the head pointer register 520 of the consumer channel memory mapped register(s) 524 indicating the update to the head pointer 515.
  • The channel interface 518 of the SPE 510, operating as a consumer, periodically sends a read channel count instruction to the consumer channel memory mapped register(s) 524 to determine the count for the head pointer channel. If the count is zero, the processor 512 is maintained in a low power “sleep” state. If the count is other than zero, then the channel interface 518 sends a read channel instruction to the consumer channel memory mapped register(s) 524 to thereby obtain the head pointer update data from the local storage 526 of the external device 530. This in turn causes the processor 512 to read data from the shared buffer region 522 and update the tail pointer 517 by writing the tail pointer update data to consumer channel memory mapped register(s) 524. The external device 530 may be informed of the update to the tail pointer 517 by way of an MFC-DMA “put” command in a similar manner as previously described above with regard to the update to the head pointer in FIG. 4.
  • The external device 530 receives the notification of the update to the tail pointer 517, may update its local copy of the tail pointer 519, and may write data to the shared buffer region 522 based on the updated tail pointer. If the external device 530 is another SPE having similar channel interface facilities and local store, a similar operation as described above in FIG. 4 may be followed by external device 530 to thereby update a local copy of the head and tail pointers and perform write operations to the shared buffer region 522.
  • FIG. 6 is an exemplary block diagram illustrating an interaction between a producer and a consumer with regard to updating state information for a shared buffer region in accordance with one exemplary embodiment of the present invention where a process running on a SPE operates as both a producer and a consumer. As shown in FIG. 6, the primary difference between this embodiment and the previous embodiments described in FIGS. 4 and 5 is that there are separate channels allocated to the SPE 410 that are provided for updates to the head pointer and the tail pointer. That is, a first channel, e.g., SPE channel 1 memory mapped register(s) 624, is provided to which the SPE 610 writes updates to the head pointer and a second channel, e.g., SPE channel 2 memory mapped register(s) 625, is provided to which the SPE 610 writes updates to the tail pointer. Thus, the first channel is used when the SPE 610 is operating in a producer role and the second channel is used when the SPE 610 is operating in a consumer role.
  • Since the SPE 610 may operate as both producer and consumer, the SPE 610 may read and write data to the shared buffer region 622. When the SPE 610 writes data to the shared buffer region 622 as a producer, the channel interface 618 writes head pointer update data to the head pointer register 615 in the channel 1 memory mapped address region 624 and may notify the external device 630 of the update by way of a MFC-DMA “put” instruction, for example. When the SPE 610 reads data from the shared buffer region 622 as a consumer, the SPE 610 may write tail pointer update data to the tail pointer register 627 in the channel 2 memory mapped address region 627. Similarly, the update to the tail pointer may be notified to the external device 630 by way of a MFC-DMA “put” instruction to the tail pointer storage location 617, for example.
  • The external device 630 may update the head pointer 626 and tail pointer 617 in local storage 626 based on operations performed by the external device 630 to either store or read data from the shared buffer region 622. The external device 630 may notify the SPE 610 of updates to the head and tail pointers using write channel instructions to respective ones of channel 1 and channel 2 memory mapped register(s) 624 and 625.
  • When the SPE 610 receives an update to the tail pointer 620 via the SPE channel 1 memory mapped address region 624, the SPE 610 may update a copy of the tail pointer in local store 614. When the SPE 610 receives an update to the head pointer 632 via the SPE channel 2 memory mapped address region 625, the SPE 610 may update a local copy of the head pointer in local store 614. The receipt of such updates may cause the processor 612 to exit a low power “sleep” state if the processor 612 were in such a “sleep” state in a similar manner as described above. Thus, the SPE 610 may write head and tail pointer updates to the signal notification channels and receive updates to head and tail pointers via these signal notification channels. Since the broadband engine architecture allows for two signal notification channels for each SPE or external device, the SPEs or external devices may use these signal notification channels in the manner described above to facilitate notification of head and tail pointer updates for a shared buffer region.
  • FIG. 7 is a flowchart outlining an exemplary operation of the present invention in which a process running on a SPE operates as a producer. It will be understood that each block, and combination of blocks, of the flowchart illustration in FIG. 7, and the flowchart illustrations in subsequent figures described hereafter, can be implemented by computer program instructions. These computer program instructions may be provided to a processor or other programmable data processing apparatus to produce a machine, such that the instructions which execute on the processor or other programmable data processing apparatus create means for implementing the functions specified in the flowchart block or blocks. These computer program instructions may also be stored in a computer-readable memory or storage medium that can direct a processor or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory or storage medium produce an article of manufacture including instruction means which implement the functions specified in the flowchart block or blocks.
  • Accordingly, blocks of the flowchart illustrations support combinations of means for performing the specified functions, combinations of steps for performing the specified functions and program instruction means for performing the specified functions. It will also be understood that each block of the flowchart illustrations, and combinations of blocks in the flowchart illustrations, can be implemented by special purpose hardware-based computer systems which perform the specified functions or steps, or by combinations of special purpose hardware and computer instructions.
  • As shown in FIG. 7, the operation starts by receiving data from the processor of the producer for writing to the shared buffer region (step 705). The received data is written to the shared buffer region (step 710). A head pointer for the shared buffer region is then updated (step 720). The head pointer update data is written to a signal notification channel of a consumer device (step 730). A determination is then made as to whether the current count on the producer's signal notification channel is zero (step 740). If so, the producer is placed in a low power operational mode or “sleep” state (step 750). If the producer is not to be placed in a low power operation mode, a determination is then made as to whether the operation should be discontinued (step 760). This may involve tearing down the channel between the producer and the consumer, or other events that cause the producer/consumer relationship to be discontinued. If not, the operation returns to step 705 and the operation is repeated. If the producer is placed in a low power operation mode, a determination is made as to whether the operation should be discontinued (step 770). If not, the operation returns to step 740. If the operation is to be discontinued from either of step 760 or step 770, the operation terminates.
  • FIG. 8 is a flowchart outlining an exemplary operation of the present invention in which a process running on a SPE operates as a consumer. As shown in FIG. 8, the operation starts by reading data from the shared buffer region (step 810). A tail pointer for the shared buffer region is then updated (step 820). The tail pointer update data is written to a signal notification channel of a producer device (step 830). A determination is then made as to whether the current count on the consumer's signal notification channel is zero (step 840). If so, the consumer is placed in a low power operational mode or “sleep” state (step 850).
  • If the consumer is not to be placed in a low power operation mode, a determination is then made as to whether the operation should be discontinued (step 860). This may involve tearing down the channel between the producer and the consumer, or other events that cause the producer/consumer relationship to be discontinued. If not, the operation returns to step 810 and the operation is repeated. If the consumer is placed in a low power operation mode, a determination is made as to whether the operation should be discontinued (step 870). If not, the operation returns to step 840. If the operation is to be discontinued from either of step 860 or step 870, the operation terminates.
  • It should be appreciated that the operations outlined in FIGS. 7 and 8 may also be used in conjunction in a single apparatus. For example, an SPE that operates as both a consumer and a producer may perform the above operations outlined in FIGS. 7 and 8 in substantially parallel so as to both produce data for writing to the shared buffer region and read data from the shared buffer region.
  • Thus, the present invention provides a mechanism for notifying devices, such as synergistic processing elements, of updates to head and tail pointers of a shared buffer region. The mechanism of the present invention avoids the overhead and bus congestion of polling head and tail pointers by providing mechanisms for writing updates to signal notification channels of the devices and using channel interface facilities of the devices to update the head and tail pointers. Moreover, the present invention allows a device to remain in a low power “sleep” state until an update to the head or tail pointer is received from an external device at which time the device is awoken so that it may perform processing to either produce data to or consume data from the shared buffer region. This greatly decreases the power consumption of the system as a whole.
  • It is important to note that while the present invention has been described in the context of a fully functioning data processing system, those of ordinary skill in the art will appreciate that the processes of the present invention are capable of being distributed in the form of a computer readable medium of instructions and a variety of forms and that the present invention applies equally regardless of the particular type of signal bearing media actually used to carry out the distribution. Examples of computer readable media include recordable-type media, such as a floppy disk, a hard disk drive, a RAM, CD-ROMs, DVD-ROMs, and transmission-type media, such as digital and analog communications links, wired or wireless communications links using transmission forms, such as, for example, radio frequency and light wave transmissions. The computer readable media may take the form of coded formats that are decoded for actual use in a particular data processing system.
  • The description of the present invention has been presented for purposes of illustration and description, and is not intended to be exhaustive or limited to the invention in the form disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art. The embodiment was chosen and described in order to best explain the principles of the invention, the practical application, and to enable others of ordinary skill in the art to understand the invention for various embodiments with various modifications as are suited to the particular use contemplated.

Claims (35)

1. A method, in a data processing system, comprising:
performing, using a first device in the data processing system, an operation on a shared buffer region;
writing a pointer update to a signal notification channel associated with a second device in the data processing system, wherein the pointer update is an update to one of a head pointer or a tail pointer of the shared buffer region; and
updating a pointer stored in a local store of the second device based on the pointer update.
2. The method of claim 1, wherein the operation on the shared buffer region is a write operation for writing data to the shared buffer region, and wherein the pointer update is an update to a head pointer of the shared buffer region.
3. The method of claim 1, wherein the operation on the shared buffer region is a read operation for reading data from the shared buffer region, and wherein the pointer update is an update to a tail pointer of the shared buffer region.
4. The method of claim 1, wherein the signal notification channel is a memory mapped address region of a shared memory of the data processing system.
5. The method of claim 1, further comprising:
determining if the shared buffer region is full; and
placing the second device in a low-power consumption waiting state waiting for a predetermined amount of storage space to be freed in the shared buffer region before additional data may be written to the shared buffer region.
6. The method of claim 5, wherein placing the second device in a low-power consumption waiting state comprises placing the notification channel in a blocking mode of operation in which the second device's normal operation is blocked and the second device waits in a low-power consumption state.
7. The method of claim 6, wherein the operation is a read operation for reading data from the shared buffer region, and wherein the notification channel exits the blocking mode of operation in response to the read operation.
8. The method of claim 6, wherein while in the blocking mode of operation, a channel interface of the second device periodically issues a read channel count instruction to the signal notification channel to determine a count associated with the notification channel.
9. The method of claim 8, wherein if the count associated with the notification channel is zero, then the notification channel remains in a blocking mode of operation.
10. The method of claim 8, wherein if the count associated with the notification channel is non-zero, then the notification channel exits the blocking mode of operation.
11. The method of claim 10, wherein, in response to the notification channel exiting the blocking mode of operation, the second device is awoken by issuing a read channel instruction to the notification channel and returning results of the read channel instruction to the second device.
12. The method of claim 1, wherein the second device is a synergistic processing unit and the shared buffer region is part of the local store associated with the synergistic processing unit.
13. The method of claim 12, wherein writing a pointer update is performed using a channel interface of a memory flow control unit associated with the synergistic processing unit.
14. The method of claim 1, wherein the first device and the second device are synergistic processing units of a multiprocessor system-on-a-chip.
15. A computer program product comprising a computer usable medium having a computer readable program, wherein the computer readable program, when executed on a computing device, causes the computing device to:
perform, using a first device in the data processing system, an operation on a shared buffer region;
write a pointer update to a signal notification channel associated with a second device in the data processing system, wherein the pointer update is an update to one of a head pointer or a tail pointer of the shared buffer region; and
update a pointer stored in a local store of the second device based on the pointer update.
16. The computer program product of claim 15, wherein the operation on the shared buffer region is a write operation for writing data to the shared buffer region, and wherein the pointer update is an update to a head pointer of the shared buffer region.
17. The computer program product of claim 15, wherein the operation on the shared buffer region is a read operation for reading data from the shared buffer region, and wherein the pointer update is an update to a tail pointer of the shared buffer region.
18. The computer program product of claim 15, wherein the signal notification channel is a memory mapped address region of a shared memory of the data processing system.
19. The computer program product of claim 15, wherein the computer readable program further causes the computing device to:
determine if the shared buffer region is full; and
place the second device in a low-power consumption waiting state waiting for a predetermined amount of storage space to be freed in the shared buffer region before additional data may be written to the shared buffer region.
20. The computer program product of claim 19, wherein the computer readable program causes the computing device to place the second device in a low-power consumption waiting state by placing the notification channel in a blocking mode of operation in which the second device's normal operation is blocked and the second device waits in a low-power consumption state.
21. The computer program product of claim 20, wherein the operation is a read operation for reading data from the shared buffer region, and wherein the notification channel exits the blocking mode of operation in response to the read operation.
22. The computer program product of claim 20, wherein while in the blocking mode of operation, a channel interface of the second device periodically issues a read channel count instruction to the signal notification channel to determine a count associated with the notification channel.
23. The computer program product of claim 22, wherein if the count associated with the notification channel is zero, then the notification channel remains in a blocking mode of operation.
24. The computer program product of claim 22, wherein if the count associated with the notification channel is non-zero, then the notification channel exits the blocking mode of operation.
25. The computer program product of claim 24, wherein, in response to the notification channel exiting the blocking mode of operation, the second device is awoken by issuing a read channel instruction to the notification channel and returning results of the read channel instruction to the second device.
26. The computer program product of claim 15, wherein the second device is a synergistic processing unit and the shared buffer region is part of the local store associated with the synergistic processing unit.
27. The computer program product of claim 26, wherein the computer readable program causes the computing device to write a pointer update by using a channel interface of a memory flow control unit associated with the synergistic processing unit.
28. The computer program product of claim 15, wherein the first device and the second device are synergistic processing units of a multiprocessor system-on-a-chip.
29. A system, comprising:
a first processor;
a second processor coupled to the first processor; and
a local storage device coupled to the second processor, wherein the first processor performs an operation on a shared buffer region of the local storage device and writes a pointer update to a signal notification channel associated with the second processor, wherein the pointer update is an update to one of a head pointer or a tail pointer of the shared buffer region, and wherein the second processor updates a pointer stored in the local store of the second processor based on the pointer update.
30. The system of claim 29, wherein the operation on the shared buffer region is a write operation for writing data to the shared buffer region, and wherein the pointer update is an update to a head pointer of the shared buffer region.
31. The system of claim 29, wherein the operation on the shared buffer region is a read operation for reading data from the shared buffer region, and wherein the pointer update is an update to a tail pointer of the shared buffer region.
32. The system of claim 29, wherein the second processor determines if the shared buffer region is full and places itself in a low-power consumption waiting state waiting for a predetermined amount of storage space to be freed in the shared buffer region before additional data may be written to the shared buffer region.
33. The system of claim 32, wherein the second processor places itself in a low-power consumption waiting state by placing the notification channel in a blocking mode of operation in which the second processor's normal operation is blocked and the second processor waits in a low-power consumption state.
34. The system of claim 33, wherein the operation is a read operation for reading data from the shared buffer region, and wherein the notification channel exits the blocking mode of operation in response to the read operation.
35. The system of claim 33, wherein while in the blocking mode of operation, a channel interface of the second processor periodically issues a read channel count instruction to the signal notification channel to determine a count associated with the notification channel, and wherein if the count associated with the notification channel is non-zero, then the notification channel exits the blocking mode of operation.
US11/340,453 2006-01-26 2006-01-26 Apparatus and method for efficient communication of producer/consumer buffer status Abandoned US20070174411A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
US11/340,453 US20070174411A1 (en) 2006-01-26 2006-01-26 Apparatus and method for efficient communication of producer/consumer buffer status
PCT/EP2007/050182 WO2007085522A2 (en) 2006-01-26 2007-01-09 Apparatus and method for efficient communication of producer/consumer buffer status
US12/127,464 US8275917B2 (en) 2006-01-26 2008-05-27 Efficient communication of producer/consumer buffer status
US13/593,030 US9053069B2 (en) 2006-01-26 2012-08-23 Efficient communication of producer/consumer buffer status

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US11/340,453 US20070174411A1 (en) 2006-01-26 2006-01-26 Apparatus and method for efficient communication of producer/consumer buffer status

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US12/127,464 Continuation US8275917B2 (en) 2006-01-26 2008-05-27 Efficient communication of producer/consumer buffer status

Publications (1)

Publication Number Publication Date
US20070174411A1 true US20070174411A1 (en) 2007-07-26

Family

ID=38198378

Family Applications (3)

Application Number Title Priority Date Filing Date
US11/340,453 Abandoned US20070174411A1 (en) 2006-01-26 2006-01-26 Apparatus and method for efficient communication of producer/consumer buffer status
US12/127,464 Expired - Fee Related US8275917B2 (en) 2006-01-26 2008-05-27 Efficient communication of producer/consumer buffer status
US13/593,030 Expired - Fee Related US9053069B2 (en) 2006-01-26 2012-08-23 Efficient communication of producer/consumer buffer status

Family Applications After (2)

Application Number Title Priority Date Filing Date
US12/127,464 Expired - Fee Related US8275917B2 (en) 2006-01-26 2008-05-27 Efficient communication of producer/consumer buffer status
US13/593,030 Expired - Fee Related US9053069B2 (en) 2006-01-26 2012-08-23 Efficient communication of producer/consumer buffer status

Country Status (2)

Country Link
US (3) US20070174411A1 (en)
WO (1) WO2007085522A2 (en)

Cited By (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070250731A1 (en) * 2006-04-22 2007-10-25 Francisco Romero Business method, system, and media providing for power-state change as a function of direction of right-to-use status change
US20080141268A1 (en) * 2006-12-12 2008-06-12 Tirumalai Partha P Utility function execution using scout threads
US20080159059A1 (en) * 2007-01-03 2008-07-03 Freescale Semiconductor, Inc. Progressive memory initialization with waitpoints
US20080162858A1 (en) * 2007-01-03 2008-07-03 Freescale Semiconductor, Inc. Hardware-based memory initialization with software support
US20090043987A1 (en) * 2007-08-06 2009-02-12 Samsung Electronics Co., Ltd. Operation distribution method and system using buffer
US20090222374A1 (en) * 2008-02-29 2009-09-03 American Express Travel Related Services Company, Inc. Total structural risk model
US20090222375A1 (en) * 2008-02-29 2009-09-03 American Express Travel Related Services Company, Inc. Total structural risk model
US20090222377A1 (en) * 2008-02-29 2009-09-03 American Express Travel Related Services Company, Inc. Total structural risk model
US20090222378A1 (en) * 2008-02-29 2009-09-03 American Express Travel Related Services Company, Inc. Total structural risk model
US20090222380A1 (en) * 2008-02-29 2009-09-03 American Express Travel Related Services Company, Inc Total structural risk model
US20090222379A1 (en) * 2008-02-29 2009-09-03 American Express Travel Related Services Company, Inc. Total structural risk model
US20090222373A1 (en) * 2008-02-29 2009-09-03 American Express Travel Related Services Company, Inc. Total structural risk model
US20090222376A1 (en) * 2008-02-29 2009-09-03 American Express Travel Related Services Company, Inc. Total structural risk model
US20090228618A1 (en) * 2008-02-15 2009-09-10 Andreas Kuehm Peripheral interface and process for data transfer, especially for laser scanning microscopes
US20100211704A1 (en) * 2009-02-13 2010-08-19 Renesas Technology Corp. Data Processing Apparatus
US20100333112A1 (en) * 2008-06-30 2010-12-30 Anand Sundaram Method and System for Secure Communication Between Processor Partitions
US20110040904A1 (en) * 2006-08-31 2011-02-17 American Megatrends, Inc. Remotely controllable switch and testing methods using same
CN101702676B (en) * 2009-11-23 2012-01-25 华为终端有限公司 Data buffering process and device
US8196147B1 (en) * 2008-10-31 2012-06-05 Netapp, Inc. Multiple-processor core optimization for producer-consumer communication
US20130024538A1 (en) * 2008-06-18 2013-01-24 Fateev Maxim E Fast sequential message store
US9158579B1 (en) 2008-11-10 2015-10-13 Netapp, Inc. System having operation queues corresponding to operation execution time
US9250968B2 (en) 2008-09-26 2016-02-02 Samsung Electronics Co., Ltd. Method and memory manager for managing a memory in a multi-processing environment
US20160313991A1 (en) * 2013-06-16 2016-10-27 President And Fellows Of Harvard College Methods and apparatus for parallel processing
US20170357445A1 (en) * 2016-06-13 2017-12-14 International Business Machines Corporation Flexible optimized data handling in systems with multiple memories
US10326713B2 (en) * 2015-07-30 2019-06-18 Huawei Technologies Co., Ltd. Data enqueuing method, data dequeuing method, and queue management circuit
US10362131B1 (en) 2008-06-18 2019-07-23 Amazon Technologies, Inc. Fault tolerant message delivery
US10417045B2 (en) 2015-04-17 2019-09-17 Amer Sports Digital Services Oy Embedded computing device
CN110874273A (en) * 2018-08-31 2020-03-10 阿里巴巴集团控股有限公司 Data processing method and device
US11218572B2 (en) * 2017-02-17 2022-01-04 Huawei Technologies Co., Ltd. Packet processing based on latency sensitivity
CN114040136A (en) * 2021-11-05 2022-02-11 北京京东乾石科技有限公司 Track inspection device, image processing method, device, equipment and medium

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8054316B2 (en) * 2008-11-14 2011-11-08 Nvidia Corporation Picture processing using a hybrid system configuration
GB2542988B (en) * 2015-04-17 2019-11-13 Suunto Oy Embedded computing device comprising processing units interfaced with a shared information space

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6625689B2 (en) * 1998-06-15 2003-09-23 Intel Corporation Multiple consumer-multiple producer rings
US20050188373A1 (en) * 2004-02-20 2005-08-25 Sony Computer Entertainment Inc. Methods and apparatus for task management in a multi-processor system
US20060236136A1 (en) * 2005-04-14 2006-10-19 Jones Darren M Apparatus and method for automatic low power mode invocation in a multi-threaded processor

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6526491B2 (en) * 2001-03-22 2003-02-25 Sony Corporation Entertainment Inc. Memory protection system and method for computer architecture for broadband networks
US7836222B2 (en) * 2003-06-26 2010-11-16 International Business Machines Corporation System and method for tracking messages between a processing unit and an external device
US20050120185A1 (en) * 2003-12-01 2005-06-02 Sony Computer Entertainment Inc. Methods and apparatus for efficient multi-tasking
US20070083870A1 (en) * 2005-07-29 2007-04-12 Tomochika Kanakogi Methods and apparatus for task sharing among a plurality of processors

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6625689B2 (en) * 1998-06-15 2003-09-23 Intel Corporation Multiple consumer-multiple producer rings
US20050188373A1 (en) * 2004-02-20 2005-08-25 Sony Computer Entertainment Inc. Methods and apparatus for task management in a multi-processor system
US20060236136A1 (en) * 2005-04-14 2006-10-19 Jones Darren M Apparatus and method for automatic low power mode invocation in a multi-threaded processor

Cited By (55)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070250731A1 (en) * 2006-04-22 2007-10-25 Francisco Romero Business method, system, and media providing for power-state change as a function of direction of right-to-use status change
US8122264B2 (en) * 2006-04-22 2012-02-21 Hewlett-Packard Development Company, L.P. Power-state change as a function of direction of right-to-use status change
US20110040904A1 (en) * 2006-08-31 2011-02-17 American Megatrends, Inc. Remotely controllable switch and testing methods using same
US20080141268A1 (en) * 2006-12-12 2008-06-12 Tirumalai Partha P Utility function execution using scout threads
US20080159059A1 (en) * 2007-01-03 2008-07-03 Freescale Semiconductor, Inc. Progressive memory initialization with waitpoints
US8725975B2 (en) 2007-01-03 2014-05-13 Freescale Semiconductor, Inc. Progressive memory initialization with waitpoints
US20080162858A1 (en) * 2007-01-03 2008-07-03 Freescale Semiconductor, Inc. Hardware-based memory initialization with software support
US20090043987A1 (en) * 2007-08-06 2009-02-12 Samsung Electronics Co., Ltd. Operation distribution method and system using buffer
US20090228618A1 (en) * 2008-02-15 2009-09-10 Andreas Kuehm Peripheral interface and process for data transfer, especially for laser scanning microscopes
US8214561B2 (en) * 2008-02-15 2012-07-03 Carl Zeiss Microimaging Gmbh Peripheral interface and process for data transfer, especially for laser scanning microscopes
US7853520B2 (en) 2008-02-29 2010-12-14 American Express Travel Related Services Company, Inc. Total structural risk model
US10019757B2 (en) 2008-02-29 2018-07-10 American Express Travel Related Services Company, Inc. Total structural risk model
US20090222376A1 (en) * 2008-02-29 2009-09-03 American Express Travel Related Services Company, Inc. Total structural risk model
US20090222379A1 (en) * 2008-02-29 2009-09-03 American Express Travel Related Services Company, Inc. Total structural risk model
US8566229B2 (en) 2008-02-29 2013-10-22 American Express Travel Related Services Company, Inc. Total structural risk model
US20090222380A1 (en) * 2008-02-29 2009-09-03 American Express Travel Related Services Company, Inc Total structural risk model
US20090222375A1 (en) * 2008-02-29 2009-09-03 American Express Travel Related Services Company, Inc. Total structural risk model
US20090222378A1 (en) * 2008-02-29 2009-09-03 American Express Travel Related Services Company, Inc. Total structural risk model
US8554667B2 (en) 2008-02-29 2013-10-08 American Express Travel Related Services Company, Inc. Total structural risk model
US8566228B2 (en) 2008-02-29 2013-10-22 American Express Travel Related Services Company, Inc. Total structural risk model
US7991690B2 (en) 2008-02-29 2011-08-02 American Express Travel Related Services Company, Inc. Total structural risk model
US20090222373A1 (en) * 2008-02-29 2009-09-03 American Express Travel Related Services Company, Inc. Total structural risk model
US8554666B2 (en) 2008-02-29 2013-10-08 American Express Travel Related Services Company, Inc. Total structural risk model
US20090222377A1 (en) * 2008-02-29 2009-09-03 American Express Travel Related Services Company, Inc. Total structural risk model
US7849004B2 (en) 2008-02-29 2010-12-07 American Express Travel Related Services Company, Inc. Total structural risk model
US20090222374A1 (en) * 2008-02-29 2009-09-03 American Express Travel Related Services Company, Inc. Total structural risk model
US7814008B2 (en) * 2008-02-29 2010-10-12 American Express Travel Related Services Company, Inc. Total structural risk model
US8620801B2 (en) 2008-02-29 2013-12-31 American Express Travel Related Services Company, Inc. Total structural risk model
US8458083B2 (en) 2008-02-29 2013-06-04 American Express Travel Related Services Company, Inc. Total structural risk model
US20130024538A1 (en) * 2008-06-18 2013-01-24 Fateev Maxim E Fast sequential message store
US9485324B2 (en) 2008-06-18 2016-11-01 Amazon Technologies, Inc. Fast sequential message store
US8763013B2 (en) * 2008-06-18 2014-06-24 Amazon Technologies, Inc. Fast sequential message store
US10362131B1 (en) 2008-06-18 2019-07-23 Amazon Technologies, Inc. Fault tolerant message delivery
US20100333112A1 (en) * 2008-06-30 2010-12-30 Anand Sundaram Method and System for Secure Communication Between Processor Partitions
US8151026B2 (en) * 2008-06-30 2012-04-03 Wind River Systems, Inc. Method and system for secure communication between processor partitions
US9250968B2 (en) 2008-09-26 2016-02-02 Samsung Electronics Co., Ltd. Method and memory manager for managing a memory in a multi-processing environment
US8196147B1 (en) * 2008-10-31 2012-06-05 Netapp, Inc. Multiple-processor core optimization for producer-consumer communication
US9430278B2 (en) 2008-11-10 2016-08-30 Netapp, Inc. System having operation queues corresponding to operation execution time
US9158579B1 (en) 2008-11-10 2015-10-13 Netapp, Inc. System having operation queues corresponding to operation execution time
US8099530B2 (en) * 2009-02-13 2012-01-17 Renesas Electronics Corporation Data processing apparatus
US8244929B2 (en) 2009-02-13 2012-08-14 Renesas Electronics Corporation Data processing apparatus
US20100211704A1 (en) * 2009-02-13 2010-08-19 Renesas Technology Corp. Data Processing Apparatus
CN101702676B (en) * 2009-11-23 2012-01-25 华为终端有限公司 Data buffering process and device
US20160313991A1 (en) * 2013-06-16 2016-10-27 President And Fellows Of Harvard College Methods and apparatus for parallel processing
US10949200B2 (en) * 2013-06-16 2021-03-16 President And Fellows Of Harvard College Methods and apparatus for executing data-dependent threads in parallel
US10417045B2 (en) 2015-04-17 2019-09-17 Amer Sports Digital Services Oy Embedded computing device
US10326713B2 (en) * 2015-07-30 2019-06-18 Huawei Technologies Co., Ltd. Data enqueuing method, data dequeuing method, and queue management circuit
US20170357445A1 (en) * 2016-06-13 2017-12-14 International Business Machines Corporation Flexible optimized data handling in systems with multiple memories
US10996989B2 (en) * 2016-06-13 2021-05-04 International Business Machines Corporation Flexible optimized data handling in systems with multiple memories
US20210208939A1 (en) * 2016-06-13 2021-07-08 International Business Machines Corporation Flexible optimized data handling in systems with multiple memories
US11687369B2 (en) * 2016-06-13 2023-06-27 International Business Machines Corporation Flexible optimized data handling in systems with multiple memories
US20230244530A1 (en) * 2016-06-13 2023-08-03 International Business Machines Corporation Flexible optimized data handling in systems with multiple memories
US11218572B2 (en) * 2017-02-17 2022-01-04 Huawei Technologies Co., Ltd. Packet processing based on latency sensitivity
CN110874273A (en) * 2018-08-31 2020-03-10 阿里巴巴集团控股有限公司 Data processing method and device
CN114040136A (en) * 2021-11-05 2022-02-11 北京京东乾石科技有限公司 Track inspection device, image processing method, device, equipment and medium

Also Published As

Publication number Publication date
WO2007085522A2 (en) 2007-08-02
US20090037620A1 (en) 2009-02-05
WO2007085522A3 (en) 2008-06-26
US8275917B2 (en) 2012-09-25
US20120317372A1 (en) 2012-12-13
US9053069B2 (en) 2015-06-09

Similar Documents

Publication Publication Date Title
US8275917B2 (en) Efficient communication of producer/consumer buffer status
US7930457B2 (en) Channel mechanisms for communicating with a processor event facility
US8024489B2 (en) System for communicating command parameters between a processor and a memory flow controller
CN109997112B (en) Data processing
US20140013021A1 (en) Processor core and multi-core processor system
US7685601B2 (en) Methods and apparatus for segmented stack management in a processor system
US7778271B2 (en) Method for communicating instructions and data between a processor and external devices
US7600076B2 (en) Method, system, apparatus, and article of manufacture for performing cacheline polling utilizing store with reserve and load when reservation lost instructions
BR112014018434B1 (en) BUFFER MANAGEMENT FOR PARALLEL GRAPHICS PROCESSING UNIT
US20120297216A1 (en) Dynamically selecting active polling or timed waits
US9256535B2 (en) Conditional notification mechanism
CN111611125A (en) Method and apparatus for improving performance data collection for high performance computing applications
US11005970B2 (en) Data storage system with processor scheduling using distributed peek-poller threads
JPH02193248A (en) Actual storage management system
WO2010087949A1 (en) Logical power throttling
US7581067B2 (en) Load when reservation lost instruction for performing cacheline polling
US20070220212A1 (en) Method, system, apparatus, and article of manufacture for performing cacheline polling utilizing a store and reserve instruction
US10671453B1 (en) Data storage system employing two-level scheduling of processing cores
US10997075B2 (en) Adaptively enabling and disabling snooping bus commands
US10127076B1 (en) Low latency thread context caching
US20140250442A1 (en) Conditional Notification Mechanism
US9411663B2 (en) Conditional notification mechanism
US8438569B2 (en) Broadcasting a condition to threads executing on a plurality of on-chip processors
US11100014B2 (en) Providing data in a shared memory
JPH0410051A (en) Shared memory control method for multiprocessor system

Legal Events

Date Code Title Description
AS Assignment

Owner name: MACHINES CORPORATION, INTERNATIONAL BUSINESS, NEW

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BROKENSHIRE, DANIEL A.;JOHNS, CHARLES R.;NUTTER, MARK RICHARD;AND OTHERS;REEL/FRAME:017268/0011

Effective date: 20060105

AS Assignment

Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BROKENSHIRE, DANIEL A.;JOHNS, CHARLES R.;NUTTER, MARK R.;AND OTHERS;REEL/FRAME:019797/0570

Effective date: 20060105

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION