Scratchpad memory
An integrated circuit includes a random access memory (RAM) storage and a controller both located on one semiconductor chip. The controller is coupled to read data from and write data to the RAM storage. The controller is programmable to perform bitwise operations on data words stored in the RAM.
Latest Patents:
- Semiconductor device comprising magnetic tunneling junctions with different distances/widths in a magnetoresistive random access memory
- Shader-based dynamic video manipulation
- Methods of forming integrated assemblies with improved charge migration impedance
- Methods and apparatus to automate receivability updates for media crediting
- Basketball hoop
Many modern computers have a parallel structure in which several identical processors are capable of performing tasks on incoming instructions and/or data. The parallel structure can handle high instruction and/or data bandwidths if the various tasks are independent. To perform some of the tasks, the processors can access shared resources. Frequently, the shared resources include a data storage device.
SUMMARY OF THE INVENTIONAccording to one aspect, a integrated circuit includes a random-access memory (RAM) storage and a controller both located on the same semiconductor chip. The controller is coupled to read data from and write data to the RAM storage. The controller is programmable to perform bitwise operations on data words stored in the RAM.
BRIEF DESCRIPTION OF THE DRAWINGS
The router 10 performs data routing in two stages. First, one of the MAC devices 14, 14′, 14″ connected to the source network 18, 18′, 18″ transmits a data packet to the parallel processor 12 via the FIFO bus 16. Second, the parallel processor 12 retransmits the data packet over the FIFO bus 18 to the MAC device 14, 14′, 14″ connected to the destination network 18, 18′, 18″. The data transmissions over the FIFO bus 16 employ 64-byte data packets and proceed via an Ethernet protocol.
The parallel processor 12 has a parallel data forwarding structure that includes an array of identical processing engines 22a-22f. Each processing engine 22a-22f has an internal structure for executing a plurality of, e.g., four, independent threads.
Referring to
The processing engines 22a-22f also process data from the queues of the SDRAM 26. This processing includes moving data packets from the queues of the SDRAM 26 to a FIFO buffer 60. The FIFO buffer 60 temporarily stores data prior to retransmission to the MAC devices 14, 14′, 14″ over the FIFO bus 16. Along with the data, associated control and destination information are stored in the FIFO buffer 60 for use in transmitting the data. The associated data is 16 bytes wide.
The SRAM 24 and SDRAM 26 couple to the processing engines 22a-22f through respective SRAM and SDRAM controllers 34, 36. The SRAM controller 34 has content addressable memory that supports look ups of identification information on the queues of the SDRAM 24. The look-ups use header data from received data packets. The SDRAM controller 36 coordinates data writes to and reads from the queues of the SDRAM 24 that store received data packets.
The parallel processor 12 has several internal busses 39, 40, 41. An S bus 39 couples the processing engines 22a-22f to a FIFO bus interface 38 (FBI) and to the SRAM controller 34. An M bus 40 couples the processing engines 22a-22f and the FBI 38 to the SDRAM controller 36 and the SDRAM 26. An AMBA bus 41 couples a processor core 44 to the processing engines 22a-22f and the FBI 38.
The FBI 38 controls data transfers on the FIFO bus 16 and collects status data on the readiness of the ports 28, 30, 32 of the MAC devices 14, 14′, 14″ to participate in data transfers over the FIFO bus 16. The ready status data is collected from the MAC devices 14, 14′, 14″ through a ready bus 42, which is also controlled by the FBI 38.
Referring again to
The parallel processor 12 includes a PCI bus interface 46 that couples to a PCI bus 48. The PCI bus 48 can support communications between the parallel processor 12 and external processors. The other processors may control and/or reprogram the processor core 44 or other components 22a-22f, 38 of the multiprocessor 12.
Referring again to
The processing engines 22a-22f include separate receive-scheduler and transmit-scheduler threads. The receive- scheduler thread schedules the processing of data received from the FIFO bus 16. The transmit-scheduler thread schedules the processing of data to be transmitted to the FIFO bus 16.
The receive-scheduler thread assigns data forwarding and header processing tasks to other threads in the processing engines 22a-22f. These tasks include sharing operation of a push engine 62 that transports data from the receive FIFO buffer 58 in the FBI 38 to one of the storage queues in the SDRAM 26.
The transmit-scheduler thread also assigns data forwarding tasks to other threads in the processing engines 22a-22f. These tasks include sharing in operation of a pull engine 64, which moves data from the storage queues in the SDRAM 26 to the transmit FIFO buffer 60. The tasks also include directing the pull engine 62 to write transmission control and MAC device 14, 14′, 14″ address information to the FIFO buffer 60. Each data packet in the transmit FIFO buffer 60 has associated address and control information that control the retransmission over the FIFO bus 16.
To control data forwarding by the push and pull engines 62, 64, the execution threads of the processing engines 22a-22f send commands signals to FIFO command queues 66, 68 via a line 70. Components of the FBI 38 can also send commands to the command queues 66, 68 of push and pull engines 62, 64. For example, the ready bus controller 50 can send a command to the queue 66 that causes the push engine 62 to transfer ready status data from the status registers 54 to the processing engines 22a-22f. An arbiter 56 controls transmission of commands from the queues 66, 68 to the push and pull engines 62, 64.
The push and pull engines 62, 64 perform several types of tasks. The push and the pull engines 62, 64 are involved in bi-directional forwarding of data packets between the FIFO buffers 58, 60 and the SDRAM controller 36. The push and pull engines 62, 64 also operate a large hardware unit 71 located in the FBI 38. The push engine 62 also forwards ready status data from the set of status registers 54 to the receive- and transmit-scheduler threads located in the processing engines 22a-22f.
The hardware unit 71 performs various operations for the execution threads of the processing engines 22a-22f and includes a hash unit 72 and a scratchpad memory 73. The execution threads operate the hardware unit 71 by sending commands to the queues 66, 68. To perform the operations, the pull engine 64 retrieves input data over the S bus 39 from output transfer registers 80a-80f of the requesting processing engine 22a-22f. The pull engine 64 moves the retrieved data and associated commands to the hardware unit 71. The hardware unit 71 forwards results from the operations to the push engine 62. The push engine 62 uses command information from the command queue 66 and/or pull engine 64 to transmit the results back over the S bus 39 to input transfer registers 78a-78f of the requesting or destination processing engine 22a-22f.
Referring to
Referring to
The various ports 28, 30, 32 of the MAC devices 14, 14′, 14″ may support different data transfer rates. The ports 28, 30 of the MAC devices 14, 14′ support transfer rates of about ten or one hundred megabits of data per second. The port 32 of the MAC device 14″ may have a transfer rate of up to about one gigabit per second.
The ready bus 42 includes control/address and data lines. The control/address lines enable selection of a transaction type and a port 28, 30, 32 of the MAC devices 14, 14′, 14″. The data line transfers receive- and transmit-ready status data to the FBI 38 in response to status queries from the control module 50 for the ready bus 42.
Referring to 5A, the registers R1, R2, R3 that store receive-ready status data are shown. The registers R1 and R2 store receive-ready status data for individual MAC ports 28, 30, 32. The readiness of each MAC port 28, 30, 32 to transmit a data packet to the FIFO bus 16 is indicated by the value of an associated bit or flag stored in one of the registers R1, R2. One logic value of the bit or flag indicates that the associated port 28, 30, 32 has a data packet ready to transmit, and the other logic value indicates the associated port 28, 30, 323 has no ready data packets. Different ports 28, 30, 32 may have data packets of different sizes, but the receive scheduler thread knows the packet size associated with each port 28, 30, 32.
The registers R2 and R3 have 32 bits each and thus, can accommodate receive-ready status data for up to 64 different MAC ports 28, 30, 32.
The register R3 stores a cyclic counter value, which acts as a time stamp for the receive-status data stored in registers R1, R2. The counter value is incremented each time new receive-status data is collected. By comparing the counter value to a previously received counter value, the scheduler thread can determine whether the present receive-status data is new or stale, i.e., whether the data has already been seen.
Referring to
The registers R4 and R5 have a total of 64 bits and thus, can report transmit ready status for up to 64 MAC ports 28, 30, 32.
Referring to
Since the FBI 38 transmits 64-byte data packets from the buffer 60 according to a FIFO scheme, the remove pointer 96 indicates which data packets are scheduled but not transmitted. The position of the pointer 96 can be used to determine which MAC ports 28, 30, 32 have been scheduled to receive a data packet but have not yet received a data packet. Such ports 28, 30, 32 may have status data in registers R4, R5 indicating an availability to receive a data packet even though the available space has already been assigned to a waiting data packet.
The transmit scheduler thread can use the position of the remove pointer 96 to interpret transmit-ready status data of the registers R4, R5. From the position of the remove pointer 96, the transmit scheduler thread identifies MAC ports 28, 30, 32 already scheduled to receive a data packet. The transmit scheduler thread does not schedule a new data packet for such ports, because the waiting and already scheduled data packet may take the available space therein.
In the multiprocessor 12, the collection of ready status data is asynchronous with respect to scheduling of data packet transfers. The asynchronous relationship enables both the collection of ready status data and the scheduling of data packets to have higher effective bandwidths. The asynchronous relationship also introduces some unpredictability into latencies associated with the transfer of newly collected ready status data to scheduler threads.
Referring to
Making transfers of new ready status data unsolicited lowers latencies for delivering such data to scheduler threads. Since latencies in delivering such data can cause scheduling errors, making the transfer of ready status data unsolicited can lower numbers of occurrences of scheduling errors. The data is transmitted from respective input and output transfer registers 78, 80. Each input and output transfer register 78, 80 is assigned to an individual execution thread. To write data to or read data from a particular execution thread, an external device accesses one of the transfer registers 78, 80 assigned to the particular thread.
Referring to
The various ports 28, 30, 32 of the MAC devices 14, 14′, 14″ may support different data transfer rates. The ports 28, 30 of the MAC devices 14, 14′ support transfer rates of about ten or one hundred megabits of data per second. The port 32 of the MAC device 14″ may have a transfer rate of up to about one gigabit per second.
The ready bus 42 includes control/address and data lines. The control/address lines enable selection of a transaction type and a port 28, 30, 32 of the MAC devices 14, 14′, 14″. The data line transfers receive- and transmit-ready status data to the FBI 38 in response to status queries from the control module 50 for the ready bus 42.
Referring to
The registers R2 and R3 have 32 bits each and thus, can accommodate receive-ready status data for up to 64 different MAC ports 28, 30, 32.
The register R3 stores a cyclic counter value, which acts as a time stamp for the receive-status data stored in registers R1, R2. The counter value is incremented each time new receive-status data is collected. By comparing the counter value to a previously received counter value, the scheduler thread can determine whether the present receive-status data is new or stale, i.e., whether the data has already been seen.
Referring to
The registers R4 and R5 have a total of 64 bits and thus, can report transmit ready status for up to 64 MAC ports 28, 30, 32.
Referring to
Since the FBI 38 transmits 64-byte data packets from the buffer 60 according to a FIFO scheme, the remove pointer 96 indicates which data packets are scheduled but not transmitted. The position of the pointer 96 can be used to determine which MAC ports 28, 30, 32 have been scheduled to receive a data packet but have not yet received a data packet. Such ports 28, 30, 32 may have status data in registers R4, R5 indicating an availability to receive a data packet even though the available space has already been assigned to a waiting data packet.
The transmit scheduler thread can use the position of the remove pointer 96 to interpret transmit-ready status data of the registers R4, R5. From the position of the remove pointer 96, the transmit scheduler thread identifies MAC ports 28, 30, 32 already scheduled to receive a data packet. The transmit scheduler thread does not schedule a new data packet for such ports, because the waiting and already scheduled data packet may take the available space therein.
In the multiprocessor 12, the collection of ready status data is asynchronous with respect to scheduling of data packet transfers. The asynchronous relationship enables both the collection of ready status data and the scheduling of data packets to have higher effective bandwidths. The asynchronous relationship also introduces some unpredictability into latencies associated with the transfer of newly collected ready status data to scheduler threads.
Referring to
Making transfers of new ready status data unsolicited lowers latencies for delivering such data to scheduler threads. Since latencies in delivering such data can cause scheduling errors, making the transfer of ready status data unsolicited can lower numbers of occurrences of scheduling errors.
Referring to
The scratchpad memory 73 interfaces with other devices through the pull and push engines 64, 62. The pull engine 64 provides input commands, data and associated storage addresses to the controller 112. The data may include data words to write to the scratchpad memory 73 and mask words that define bit locations affected by the bitwise operations in the scratchpad memory 73. The push engine 62 receives output commands, data, and associated destination addresses from the controller 112. In some embodiments, the push engine 62 also sends read commands to the controller 112, because reads only require command and address signals, i.e., no data is inputted to the scratchpad memory 73.
The pull and push engines 64, 62 enable any of the processing engines 22a-22f to use the scratchpad memory 73. The scratchpad memory 73 is a shared on-chip resource. The processing engines 22a-22f and FBI 38, including the scratchpad memory 73, are parts of one integrated circuit, which is located on a semiconductor chip 113. The processing engines 22a-22f share the on-chip scratchpad memory 73 to store data and perform bitwise operations on data. Using a shared resource to perform rare operations and provide storage can reduce overall hardware costs associated with the large array of processing engines 22a-22f.
The bitwise operations performed by the scratchpad memory 73 include set, clear, “test and set”, “test and clear”, and increment operations. Each bitwise operation uses a 32 bit mask received from the pull engine 64. The mask defines locations at which bitwise manipulations will be performed, e.g., by locations of bits with the value 1 in the mask.
Referring to
Referring to
Referring to
Referring to
Referring to
The controller 112 performs 156 a set or clear operation on a data word stored in the RAM storage 110. If the command is for a “test and set”, the set operation is performed. If the command is for a “test and clear”, the clear operation is performed. In either case, the form of the operation performed is defined by the transmitted mask word as already described.
After completing the appropriate operation, the controller 112 sends 158 both the original data word and operation completed signal to the push engine 62. The push engine 62 writes 160 the original data word and an operation completed message to input transfer registers 80a of the execution thread that requested the operation. Thus, the requesting thread can determine from the data word received whether the set or clear operation has resulted in a change to the data word.
While various embodiments have been described in the detailed description, the description is intended to illustrate and not to limit the scope of the invention, which is defined by the appended claims. Other aspects, advantages, and modifications are within the scope of the claims.
Claims
1. An integrated circuit, comprising:
- a random access memory (RAM) storage located on a semiconductor chip; and
- a controller located on the semiconductor chip and coupled to read data from and write data to the RAM storage, the controller being programmable to perform bitwise operations on data words stored in the RAM.
2. The integrated circuit of claim 1, wherein the controller can perform a bitwise operation in response to receiving a mask word, the form of the operation being defined by the mask word.
3. The integrated circuit of claim 2, wherein the controller is configured to operate on bits of the data word that correspond to bits of the mask word having a preselected value.
4. The integrated circuit of claim 3, wherein the controller is configured to perform one of bitwise set and a bitwise clear.
5. The integrated circuit of claim 2, wherein the controller is configured to perform one of bitwise “test and set” and bitwise “test and clear”.
6. The integrated circuit of claim 2, wherein the controller is configured to perform an increment operation on the data words.
7. The integrated circuit of claim 2, further comprising:
- a plurality of processing engines located on the integrated chip and coupled to program the controller.
8. The integrated circuit of claim 7, wherein the controller is coupled to receive data from and write data to the processing engines.
9. The integrated circuit of claim 8, wherein the controller is coupled to receive a mask word from one of the processing engines in response to the one of the processing engines programming the controller.
10. A method of operating on data, comprising:
- receiving a command for bitwise operation at a pull engine;
- retrieving a mask word from an engine that sent the command;
- sending the mask word and a request for the command from the pull engine to a controller; and
- performing a bitwise operation on a data word stored in a RAM storage in response to the request, the mask word defining the location of one or more bits of the data word on which the operation is performed.
11. The method of claim 10, wherein the performing operates on bits of the data word that correspond to bits of the mask word with a preselected value.
12. The method of claim 11, wherein the performing includes doing one of a bitwise set and a bitwise clear on the data word.
13. The method of claim 10, further comprising:
- sending a copy of the data word to the engine.
14. The method of claim 10, wherein the command selects the data word from among more than 500 data words stored in the RAM storage.
15. A network processor, comprising:
- a parallel set of engines for processing data packets;
- a bus interface to transmit and receive the data packets; and
- a shared scratchpad memory coupled to receive commands from the engines, the scratchpad memory providing RAM storage for the engines, and capable of performing bitwise operations on data words stored therein in response to commands and mask words received from the engines, the mask words defining bits affected by the operations.
16. The network processor of claim 15, wherein the engines and the bus interface are located in one integrated circuit.
17. The network processor of claim 15, further comprising:
- a RAM coupled to the engines and to the bus interface, the engines capable of writing received data packets from the bus interface to the RAM and of sending data packets from the RAM to the bus interface.
18. The network processor of claim 17, further comprising:
- a bus coupled to the bus interface; and
- a plurality of devices connected to the bus, each of the devices to transmit data packets between an associated network and the bus.
19. The network processor of claim 18, wherein the bitwise operations include one of a set operation and a clear operation.
20. A computer program product for processing data packets that resides on a machine readable medium and comprises instructions for causing a processing engine to:
- send a command to a pull engine, the command requesting a bitwise operation be performed on a data word stored in a scratchpad memory; and
- write a mask word for the operation to an output transfer register readable by the engine, the mask word defining the location of one or more bits of the data word on which the operation is performed.
21. The computer program product of claim 20, wherein bits of the mask word having a preselected value define bits of the data word upon which the bitwise operation operates.
22. The computer program product of claim 20, wherein the command requests one of a bitwise set and a bitwise clear on the data word.
23. The computer program product of claim 20, the instructions further cause the processing engine to: read a copy of the data word sent from the scratchpad to the engine.
24. A processor, comprising:
- a heterogeneous set of programmable units integrated on a single semiconductor die, the heterogeneous set of programmable units including a set of multiple programmable units having the same construction and a set of at least one programmable unit having a different construction, each of the set of multiple programmable units having the same construction having multiple program counters corresponding to different threads;
- a first random access memory integrated on the single semiconductor die;
- a first memory controller integrated on the single semiconductor die, the memory controller coupled to the random access memory and to the heterogeneous set of programmable units; and
- a second memory controller integrated on the single semiconductor die, the second memory controller to interface to a second random access memory not integrated on the single semiconductor die.
25. The processor of claim 24, wherein the first memory controller comprises a controller to respond to commands from the programmable units specifying bit-wise operations on locations within the first random access memory.
26. The processor of claim 25, wherein the first memory controller comprises a controller to respond to test-and-set commands received from the programmable units.
27. The processor of claim 24, wherein the set of at least one programmable unit consists of a single general purpose processor.
28. The processor of claim 24, wherein each of the set of multiple programmable units comprises a control store and an arithmetic logic unit.
29. The processor of claim 24, further comprising dedicated hash logic integrated on the single semiconductor chip, the dedicated hash logic coupled to the processing units.
30. The processor of claim 24, further comprising an interface, integrated on the same semiconductor chip, to a network component to send and receive packets.
31. The processor of claim 24,
- wherein each of the set of multiple programmable units comprise input transfer registers to buffer data from the first memory controller and output transfer registers to buffer data to be sent to the first memory controller; and
- wherein the set of multiple programmable units are coupled to the first memory controller by a push bus under control of a push engine and a pull bus under control of a pull engine, the push engine to transfer data from the first memory controller into the input transfer registers, the pull engine to transfer data from the output transfer registers to the first memory controller.
Type: Application
Filed: Feb 16, 2005
Publication Date: Jul 7, 2005
Applicant:
Inventors: Gilbert Wolrich (Framingham, MA), Debra Bernstein (Sudbury, MA), Matthew Adiletta (Worcester, MA)
Application Number: 11/058,855