Multiport Cache Patents (Class 711/131)
-
Publication number: 20130326131Abstract: A security context management system within a security accelerator that can operate with high latency memories and can provide line-rate processing on several security protocols. The method employed hides the memory latencies by having the processing engines working in a pipelined fashion. It is designed to auto-fetch security context from external memory, and will allow any number of simultaneous security connections by caching only limited contexts on-chip and fetching other contexts as needed. The module does the task of fetching and associating security context with ingress packet, and populates the security context RAM with data from the external memory.Type: ApplicationFiled: May 29, 2012Publication date: December 5, 2013Applicant: TEXAS INSTRUMENTS INCORPORATEDInventors: Amritpal Singh Mundra, Denis Beaudoin, Eric Lasmana
-
Patent number: 8595442Abstract: Methods and systems redundantly validate values that are stored in a memory arrangement. The memory arrangement includes a first port and a second port that provide coherent access to one or more caches in the memory arrangement, and the first and second ports provide this coherent access at the same priority level. An instruction processor verifies that a first expected value matches a first check value calculated from the values as read from the memory arrangement via the first port. A check circuit verifies that a second expected value matches a second check value calculated from the values as read from the memory arrangement via the second port. A recovery operation is performed in response to the first or second expected values not matching the first and second check values, respectively.Type: GrantFiled: November 16, 2010Date of Patent: November 26, 2013Assignee: XILINX, Inc.Inventors: Philip B. James-Roxby, Austin H. Lesea
-
Patent number: 8583873Abstract: A multiport data cache apparatus and a method of controlling the same are provided. The multiport data cache apparatus includes a plurality of cache banks configured to share a cache line, and a data cache controller configured to receive cache requests for the cache banks, each of which including a cache bank identifier, transfer the received cache requests to the respective cache banks according to the cache bank identifiers, and process the cache requests independently from one another.Type: GrantFiled: February 28, 2011Date of Patent: November 12, 2013Assignee: Samsung Electronics Co., Ltd.Inventors: Jae-Un Park, Ki-Seok Kwon, Suk-Jin Kim
-
Patent number: 8578097Abstract: A scatter/gather technique optimizes unstructured streaming memory accesses, providing off-chip bandwidth efficiency by accessing only useful data at a fine granularity, and off-loading memory access overhead by supporting address calculation, data shuffling, and format conversion.Type: GrantFiled: October 24, 2011Date of Patent: November 5, 2013Assignee: Intel CorporationInventors: Daehyun Kim, Christopher J. Hughes, Yen-Kuang Chen, Partha Kundu
-
Patent number: 8516196Abstract: A processor may include several processor cores, each including a respective higher-level cache; a lower-level cache including several tag units each including several controllers, where each controller corresponds to a respective cache bank configured to store data, and where the controllers are concurrently operable to access their respective cache banks; and an interconnect network configured to convey data between the cores and the lower-level cache. The controllers in a given tag unit may share access to a resource that may include one or more of an interconnect egress port coupled to the interconnect network, an interconnect ingress port coupled to the interconnect network, a test controller, or a data storage structure.Type: GrantFiled: June 1, 2012Date of Patent: August 20, 2013Assignee: Oracle America, Inc.Inventors: Prashant Jain, Yoganand Chillarige, Sandip Das, Shukur Moulali Pathan, Srinivasan R. Iyengar, Sanjay Patel
-
Patent number: 8499128Abstract: According to the disclosure, a unique and novel archiving system that provides one or more application layer partitions to archive data is disclosed. Embodiments include an active archive including a fixed storage. The active archive can create application layer partitions that associate the application layer partitions with portions of the fixed storage. Each application layer partition, in embodiments, has a separate set of controls that allow for customized storage of different data within a single archiving system. Further, embodiments of methods for ensuring storage capacity in the active archive and the application layer partitions within the active archive is also disclosed.Type: GrantFiled: February 9, 2012Date of Patent: July 30, 2013Assignee: Imation Corp.Inventors: Matthew D. Bondurant, S. Christopher Alaimo, Randy Kerns
-
Patent number: 8495303Abstract: A processor and a computing system include a processor core and a buffer memory to read word data from a memory. The read word data includes first byte data read by the processor core from the memory. The buffer memory also stores the read word data, and determines whether second byte data requested by the processor core is stored in the buffer memory.Type: GrantFiled: July 21, 2008Date of Patent: July 23, 2013Assignee: Samsung Electronics Co., Ltd.Inventors: Sang Suk Lee, Suk Jin Kim, Yeon Gon Cho
-
Patent number: 8489814Abstract: A cache controller, a method for controlling the cache controller, and a computing system comprising the same are provided. The computer system comprises a processor and a cache controller. The cache controller is electrically connected to the processor and comprises a first port, a second port, and at least one cache. The first port is configured to receive an address of a content, wherein a type of the content is one of instruction and data. The second port is configured to receive an information bit corresponding to the content, wherein the information bit indicates the type of the content. The at least one cache comprises at least one cache lines. Each of the cache lines comprises a content field and corresponding to an information field. The content and the information bit is stored in the content field of one of the cache lines and the corresponding information field respectively according to the information bit and the address. Thereby, instruction and data are separated in a unified cache.Type: GrantFiled: June 23, 2009Date of Patent: July 16, 2013Assignee: Mediatek, Inc.Inventors: Po-Hung Chen, Chang-Hsien Tai
-
Patent number: 8484421Abstract: Embodiments of the present disclosure provide a system on a chip (SOC) comprising a processing core, and a cache including a cache instruction port, a cache data port, and a port utilization circuitry configured to selectively fetch instructions through the cache instruction port and selectively pre-fetch instructions through the cache data port. Other embodiments are also described and claimed.Type: GrantFiled: November 23, 2009Date of Patent: July 9, 2013Assignee: Marvell Israel (M.I.S.L) Ltd.Inventors: Tarek Rohana, Adi Habusha, Gil Stoler
-
Patent number: 8447931Abstract: One embodiment of the present invention provides a processor that supports multiple-issue execution. This processor includes a register file, which contains an array of memory cells, wherein the memory cells contain bits for architectural registers of the processor. The register file also includes multiple read ports and multiple write ports to support multiple-issue execution. During operation, if multiple read ports simultaneously read from a given register, the register file is configured to: read each bit of the given register out of the array of memory cells through a single bitline associated with the bit; and to use a driver located outside of the array of memory cells to drive the bit to the multiple read ports. In this way, each memory cell only has to drive a single bitline (instead of multiple bitlines) during a multiple-port read operation, thereby allowing memory cells to use smaller and more power-efficient drivers for read operations.Type: GrantFiled: July 1, 2005Date of Patent: May 21, 2013Assignee: Oracle America, Inc.Inventors: Shailender Chaudhry, Paul Caprioli, Marc Tremblay
-
Publication number: 20130111143Abstract: A multi-core system includes processor cores having caches; an external input/output bus connected to the processor cores; memory accessed by the processor cores via the external input/output bus; profile information indicating the volume of a write access to the memory by tasks concurrently allocated to the processor cores and whether a cache miss will occur in a read access to the caches; and an operating system that controls clock frequency of the external input/output bus to be a first frequency, based on the volume of the write access to the memory by the tasks and the bus width of the external input/output bus when a cache miss in read access is judged to not occur in executing the tasks and that controls the clock frequency of the external input/output bus to be a second frequency higher than the first frequency when a cache miss in read access is judged.Type: ApplicationFiled: December 18, 2012Publication date: May 2, 2013Applicant: FUJITSU LIMITEDInventor: FUJITSU LIMITED
-
Patent number: 8423717Abstract: A multi-core processor chip comprises at least one shared cache having a plurality of ports and a plurality of address spaces and a plurality of processor cores. Each processor core is coupled to one of the plurality of ports such that each processor core is able to access the at least one shared cache simultaneously with another of the plurality of processor cores. Each processor core is assigned one of a unique application or a unique application task and the multi-core processor is operable to execute a partitioning operating system that temporally and spatially isolates each unique application and each unique application task such that each of the plurality of processor cores does not attempt to write to the same address space of the at least one shared cache at the same time as another of the plurality of processor cores.Type: GrantFiled: December 2, 2009Date of Patent: April 16, 2013Assignee: Honeywell International Inc.Inventors: Scott Gray, Nicholas Wilt
-
Patent number: 8412886Abstract: In such a configuration that a port unit is provided which takes a form being shared among threads and has a plurality of entries for holding access requests, and the access requests for a cache shared by a plurality of threads being executed at the same time are controlled using the port unit, the access request issued from each tread is registered on a port section of the port unit which is assigned to the tread, thereby controlling the port unit to be divided for use in accordance with the thread configuration. In selecting the access request, the access requests are selected for each thread based on the specified priority control from among the access requests issued from the threads held in the port unit, thereafter a final access request is selected in accordance with a thread selection signal from among those selected access requests.Type: GrantFiled: December 16, 2009Date of Patent: April 2, 2013Assignee: Fujitsu LimitedInventor: Naohiro Kiyota
-
Patent number: 8387147Abstract: A method and system for detecting and removing a hidden pestware file is described. One illustrative embodiment detects, using direct drive access, a file on a computer storage device; determines whether the file is also detectable by the operating system by attempting to access the file using a standard file Application-Program-Interface (API) function call of the operating system; identifies the file as a potential hidden pestware file, when the file is undetectable by the operating system; confirms through an automated pestware-signature scan of the potential hidden pestware file that the potential hidden pestware file is a hidden pestware file; and removes automatically, using direct drive access, the hidden pestware file from the storage device.Type: GrantFiled: July 18, 2011Date of Patent: February 26, 2013Assignee: Webroot Inc.Inventor: Patrick Sprowls
-
Patent number: 8374050Abstract: A memory operative to provide multi-port functionality includes multiple single-port memory cells forming a first memory array. The first memory array is organized into multiple memory banks, each of the memory banks comprising a corresponding subset of the single-port memory cells. The memory further includes a second memory array including multiple multi-port memory cells and is operative to track status information of data stored in corresponding locations in the first memory array. At least one cache memory is connected with the first memory array and is operative to store data for resolving concurrent read and write access conflicts in the first memory array.Type: GrantFiled: June 4, 2011Date of Patent: February 12, 2013Assignee: LSI CorporationInventors: Ting Zhou, Ephrem Wu, Sheng Liu, Hyuck Jin Kwon
-
Patent number: 8312216Abstract: The data processing apparatus according to an embodiment of the present invention includes: a first processor; a second processor; and an external RAM to/from which the first processor writes/reads data, the first processor including a cache memory for storing data used in the first processor in association with an address on the external RAM, and the data being written to the cache memory by the second processor not through the external RAM.Type: GrantFiled: January 11, 2012Date of Patent: November 13, 2012Assignee: Renesas Electronics CorporationInventor: Mitsunobu Tanigawa
-
Patent number: 8275942Abstract: According to one embodiment of the invention, a method is disclosed for selecting a first subset of a plurality of cache ways in a cache for storing hardware threads identified as high priority hardware threads for processing by a multi-threaded processor in communication with the cache; assigning high priority hardware threads to the selected first subset; monitoring a cache usage of a high priority hardware thread assigned to the selected first subset of plurality of cache ways; and reassigning the assigned high priority hardware thread to any cache way of the plurality of cache ways if the cache usage of the high priority hardware thread exceeds a predetermined inactive cache usage threshold value based on the monitoring.Type: GrantFiled: December 22, 2005Date of Patent: September 25, 2012Assignee: Intel CorporationInventors: Theodros Yigzaw, Geeyarpuram N. Santhanakrishnan, Mark Rowland, Ganapati Srinivasa
-
Patent number: 8261023Abstract: A data processor includes a cache memory control section which includes: a hit/miss determination section which is supplied with a request for data processing to determine whether data to be processed is present in a cache memory and outputs a cache hit/miss determination result and, if having determined that the data is not present in the cache memory, feeds a read command to make an upper memory control section read the data from the upper memory; a FIFO storage which stores the cache hit/miss determination result and the in-block read position information according to a FIFO system; and a cache memory read/write section which reads the hit/miss determination result and the in-block read position information from the FIFO storage and reads the data from the cache memory, or writes the data from the upper memory control section into the cache memory and outputs the data.Type: GrantFiled: December 15, 2009Date of Patent: September 4, 2012Assignee: Kabushiki Kaisha ToshibaInventor: Kentaro Yoshikawa
-
Publication number: 20120221797Abstract: Provided is a multi-port cache memory apparatus and a method of the multi-port cache memory apparatus. The multi-port memory apparatus may divide an address space into address regions and allocate the divided memory regions to cache banks, thereby preventing the concentration of access to a particular cache.Type: ApplicationFiled: January 31, 2012Publication date: August 30, 2012Inventors: Moo-Kyoung Chung, Soo-Jung Ryu, Ho-Young Kim, Woong Seo, Young-Chul Cho
-
Publication number: 20120221796Abstract: Systems and methods are disclosed for multi-threading computer systems. In a computer system executing multiple program threads in a processing unit, a first load/store execution unit is configured to handle instructions from a first program thread and a second load/store execution unit is configured to handle instructions from a second program thread. When the computer system executing a single program thread, the first and second load/store execution units are reconfigured to handle instructions from the single program thread, and a Level 1 (L1) data cache is reconfigured with a first port to communicate with the first load/store execution unit and a second port to communicate with the second load/store execution unit.Type: ApplicationFiled: February 28, 2011Publication date: August 30, 2012Inventor: THANG M. TRAN
-
Patent number: 8255638Abstract: A method and apparatus for supporting cache coherency in a multiprocessor computing environment having multiple processing units, each processing unit having one or more local cache memories associated and operatively connected therewith. The method comprises providing a snoop filter device associated with each processing unit, each snoop filter device having a plurality of dedicated input ports for receiving snoop requests from dedicated memory writing sources in the multiprocessor computing environment. Each snoop filter device includes a plurality of parallel operating port snoop filters in correspondence with the plurality of dedicated input ports, each port snoop filter implementing one or more parallel operating sub-filter elements that are adapted to concurrently filter snoop requests received from respective dedicated memory writing sources and forward a subset of those requests to its associated processing unit.Type: GrantFiled: May 1, 2008Date of Patent: August 28, 2012Assignee: International Business Machines CorporationInventors: Matthias A. Blumrich, Dong Chen, Alan G. Gara, Mark E. Giampapa, Philip Heidelberger, Dirk I. Hoenicke, Martin Ohmacht, Valentina Salapura, Pavlos M. Vranas
-
Patent number: 8230176Abstract: A mechanism is provided for providing an improved reconfigurable cache. The mechanism partitions a large cache into inclusive cache regions with equal-ratio size or other coarse size increase. The cache controller includes an address decoder for the large cache with a large routing structure. The cache controller includes an additional address decoder for the small cache with a smaller routing structure. The additional address decoder for the small cache reduces decode, array access, and data return latencies. When only a small cache is actively in use, the rest of the cache can be turned into low-power mode to save power.Type: GrantFiled: June 26, 2009Date of Patent: July 24, 2012Assignee: International Business Machines CorporationInventor: Jian Li
-
Patent number: 8219761Abstract: A device that includes multiple processors that are connected to multiple level-one cache units. The device also includes a multi-port high-level cache unit that includes a first modular interconnect, a second modular interconnect, multiple high-level cache paths; whereas the multiple high-level cache paths comprise multiple concurrently accessible interleaved high-level cache units. Conveniently, the device also includes at least one non-cacheable path. A method for retrieving information from a cache that includes: concurrently receiving, by a first modular interconnect of a multiple-port high-level cache unit, requests to retrieve information. The method is characterized by providing information from at least two paths out of multiple high-level cache paths if at least two high-level cache hit occurs, and providing information via a second modular interconnect if a high-level cache miss occurs.Type: GrantFiled: November 17, 2005Date of Patent: July 10, 2012Assignee: Freescale Semiconductor, Inc.Inventors: Ron Bercovich, Odi Dahan, Norman Goldstein, Yehuda Nowogrodski
-
Patent number: 8195880Abstract: An information handling system (IHS) includes a processor with a cache memory system. The processor includes a processor core with an L1 cache memory that couples to an L2 cache memory. The processor includes an arbitration mechanism that controls load and store requests to the L2 cache memory. The arbitration mechanism includes control logic that enables a load request to interrupt a store request that the L2 cache memory is currently servicing. The L2 cache memory includes dual data banks so that one bank may perform a load operation while the other bank performs a store operation. The cache system provides dual dispatch points into the data flow to the dual cache banks of the L2 cache memory.Type: GrantFiled: April 15, 2009Date of Patent: June 5, 2012Assignee: International Business Machines CorporationInventors: Sanjeev Gai, Guy Lynn Guthrie, Hugh Shen, William John Starke
-
Patent number: 8140787Abstract: According to the disclosure, a unique and novel archiving system that provides one or more application layer partitions to archive data is disclosed. Embodiments include an active archive including a fixed storage. The active archive can create application layer partitions that associate the application layer partitions with portions of the fixed storage. Each application layer partition, in embodiments, has a separate set of controls that allow for customized storage of different data within a single archiving system. Further, embodiments of methods for ensuring storage capacity in the active archive and the application layer partitions within the active archive is also disclosed.Type: GrantFiled: August 27, 2008Date of Patent: March 20, 2012Assignee: Imation Corp.Inventors: Matthew D. Bondurant, S. Christopher Alaimo, Randy Kerns
-
Patent number: 8140765Abstract: An information handling system (IHS) includes a processor with a cache memory system. The processor includes a processor core with an L1 cache memory that couples to an L2 cache memory. The processor includes an arbitration mechanism that controls load and store requests to the L2 cache memory. The arbitration mechanism includes control logic that enables a load request to interrupt a store request that the L2 cache memory is currently servicing. The L2 cache memory includes dual data banks so that one bank may perform a load operation while the other bank performs a store operation. The cache system provides a single dispatch point into the data flow to the dual cache banks of the L2 cache memory.Type: GrantFiled: April 15, 2009Date of Patent: March 20, 2012Assignee: International Business Machines CorporationInventors: Sanjeev Gai, Guy Lynn Guthrie, Hugh Shen, William John Starke
-
Patent number: 8134884Abstract: A semiconductor memory device comprises a memory unit having a first and a second port and including plural banks; a bank address conversion circuit operative to convert a first bank address fed from external into a second bank address different from the first bank address and operative to supply the first bank address to one of the first and second ports and supply the second bank address to the other of the first and second ports; and a write data conversion circuit operative to convert input data fed from external into write data different from the input data and operative to supply the input data to one of the first and second ports and supply the converted write data to the other of the first and second ports.Type: GrantFiled: May 22, 2009Date of Patent: March 13, 2012Assignee: Kabushiki Kaisha ToshibaInventor: Hitoshi Iwai
-
Patent number: 8122192Abstract: The data processing apparatus according to an embodiment of the present invention includes: a first processor; a second processor; and an external RAM to/from which the first processor writes/reads data, the first processor including a cache memory for storing data used in the first processor in association with an address on the external RAM, and the data being written to the cache memory by the second processor not through the external RAM.Type: GrantFiled: April 26, 2006Date of Patent: February 21, 2012Assignee: Renesas Electronics CorporationInventor: Mitsunobu Tanigawa
-
Patent number: 8086820Abstract: Apparatus and method for highly efficient data queries. In accordance with various embodiments, a data structure is provided in a memory space with a first portion characterized as a virtual data space storing non-sequential entries and a second portion characterized as a first data array of sequential entries. At least a first sequential entry of the data array points to a skip list, at least a second sequential entry of the data array points to a second data array, and at least a third sequential entry points to a selected non-sequential entry in the first portion.Type: GrantFiled: June 29, 2007Date of Patent: December 27, 2011Assignee: Seagate Technology LLCInventors: Clark Edward Lubbers, Randy L. Roberson
-
Patent number: 8078920Abstract: An information processing device having two processing units capable of operating in synchronization with each other, includes: a common unit capable of outputting an identical signal to the two processing units; detection units that are respectively provided for the processing units and each detects errors occurred in corresponding processing unit respectively; a comparison unit that compares outputs from the two processing units; and a control unit that controls signals from the processing units to the common unit, based on a detection result of the detection units and a comparison result of the comparison unit, and determines, if errors of an identical type are simultaneously detected by the detection units, that the errors are due to an error of the common unit.Type: GrantFiled: September 4, 2009Date of Patent: December 13, 2011Assignee: Fujitsu LimitedInventors: Atsushi Morosawa, Takaharu Ishizuka, Toshikazu Ueki, Makoto Hataida, Yuka Hosokawa, Takeshi Owaki, Takashi Yamamoto, Daisuke Itou
-
Patent number: 8074026Abstract: A scatter/gather technique optimizes unstructured streaming memory accesses, providing off-chip bandwidth efficiency by accessing only useful data at a fine granularity, and off-loading memory access overhead by supporting address calculation, data shuffling, and format conversion.Type: GrantFiled: May 10, 2006Date of Patent: December 6, 2011Assignee: Intel CorporationInventors: Daehyun Kim, Christopher J. Hughes, Yen-Kuang Chen, Partha Kundu
-
Patent number: 8060721Abstract: A method of and apparatus for arbitrating a memory access conflict to a memory array. The apparatus may include selection logic coupled with a plurality of ports and a memory array to arbitrate among a plurality of contending memory access requests and to conditionally block write data from accessing the memory array when write data arrives late in time.Type: GrantFiled: August 13, 2008Date of Patent: November 15, 2011Assignee: Cypress Semiconductor CorporationInventor: Rishi Yadav
-
Publication number: 20110225369Abstract: A multiport data cache apparatus and a method of controlling the same are provided. The multiport data cache apparatus includes a plurality of cache banks configured to share a cache line, and a data cache controller configured to receive cache requests for the cache banks, each of which including a cache bank identifier, transfer the received cache requests to the respective cache banks according to the cache bank identifiers, and process the cache requests independently from one another.Type: ApplicationFiled: February 28, 2011Publication date: September 15, 2011Inventors: Jae-Un PARK, Ki-Seok Kwon, Suk-Jin Kim
-
Publication number: 20110197013Abstract: A cache system includes a primary cache memory configured to input and output data between a computation unit, the primary cache memory includes multi-port memory units each including a storing unit that stores unit data having a first data size, a writing unit that simultaneously writes sequentially inputted plural unit data to consecutive locations of the storing unit, and an outputting unit that reads out and outputs unit data written in the storing unit, wherein when writing data having a second data size that is an arbitrary multiple of a first data size and is segmented into unit data to the primary cache memory, the data is stored in different multi-port memory units by writing the sequential unit data to a subset of the multi-port memory units, and writing the other sequential unit data to another subset of the multi-port memory units.Type: ApplicationFiled: January 25, 2011Publication date: August 11, 2011Applicant: FUJITSU LIMITEDInventor: Takahito HIRANO
-
Publication number: 20110191543Abstract: An apparatus for storing data that is being processed is disclosed. The apparatus comprises: a cache associated with a processor and for storing a local copy of data items stored in a memory for use by the processor, monitoring circuitry associated with the cache for monitoring write transaction requests to the memory initiated by a further device, the further device being configured not to store data in the cache. The monitoring circuitry is responsive to detecting a write transaction request to write a data item, a local copy of which is stored in the cache, to block a write acknowledge signal transmitted from the memory to the further device indicating the write has completed and to invalidate the stored local copy in the cache and on completion of the invalidation to send the write acknowledge signal to the further device.Type: ApplicationFiled: February 2, 2010Publication date: August 4, 2011Applicant: ARM LIMITEDInventors: Simon John Craske, Antony John Penton, Loic Pierron, Andrew Christopher Rose
-
Publication number: 20110131377Abstract: A multi-core processor chip comprises at least one shared cache having a plurality of ports and a plurality of address spaces and a plurality of processor cores. Each processor core is coupled to one of the plurality of ports such that each processor core is able to access the at least one shared cache simultaneously with another of the plurality of processor cores. Each processor core is assigned one of a unique application or a unique application task and the multi-core processor is operable to execute a partitioning operating system that temporally and spatially isolates each unique application and each unique application task such that each of the plurality of processor cores does not attempt to write to the same address space of the at least one shared cache at the same time as another of the plurality of processor cores.Type: ApplicationFiled: December 2, 2009Publication date: June 2, 2011Applicant: HONEYWELL INTERNATIONAL INC.Inventors: Scott Gray, Nicholas Wilt
-
Publication number: 20110119448Abstract: Interconnect circuitry for a data processing apparatus is disclosed. The interconnect circuitry is configured to provide data routes via which at least one initiator device may access at least one recipient device.Type: ApplicationFiled: October 5, 2010Publication date: May 19, 2011Inventors: Peter Andrew Riocreux, Bruce James Mathewson, Christopher William Laycock, Richard Roy Grisenthwaite
-
Patent number: 7934057Abstract: Embodiments of the invention are directed to systems and method for providing predictable timing for read operations in a multiport memory device. Accordingly, an embodiment is directed to a multiport memory system, comprising a single port memory core synchronized to a first clock, multiple access ports synchronized to at least a second clock, and a multiplexer logic coupled to the core memory and the plurality of access ports. The multiplexer logic arbitrates access to the memory core between multiple access ports. Each access ports includes an uncertainty detect logic that measures data path latency, and an uncertainty adjust logic that operates to selectively add data path delay to increase the measured path latency to a predictable value.Type: GrantFiled: December 24, 2003Date of Patent: April 26, 2011Assignee: Cypress Semiconductor CorporationInventor: S. Babar Raza
-
Patent number: 7904667Abstract: A static random access memory (SRAM) includes an input read register (IRR) for monitoring the state of external binary devices and an output drive register (ODR) for controlling the state of external binary devices. The SRAM can be a multi-port device for access by multiple processors or controllers. Each bit of the IRR can mirror the state of a connected external binary device. Each bit of the ODR can manipulate the state of a connected external binary device or can be read without changing the state. The memory device may include settable controlling bits and a set of controlled register bits. Setting the one or more controlling bits may define which controlled register bits are associated with the IRR and which are associated with the ODR.Type: GrantFiled: August 10, 2006Date of Patent: March 8, 2011Assignee: Integrated Device Technology, Inc.Inventors: Yunsheng Wang, Casey Springer, Tak Kwong Wong, Bill Beane
-
Patent number: 7833096Abstract: A video game system and method is described in which map scenes and battle scenes are used. A player character may move through the map scene, and upon encountering enemies to battle, an encounter area may be generated and displayed to show the user which enemies will be included in the subsequent battle scene, and which enemies will not be initially included in the battle scene.Type: GrantFiled: September 9, 2005Date of Patent: November 16, 2010Assignee: Microsoft CorporationInventors: Hironobu Sakaguchi, Takehiro Kaminagayoshi
-
Patent number: 7827441Abstract: A disk-less quorum device in a clustered storage system includes non-volatile memory to store status information regarding the cluster and each storage controller in the cluster. The quorum device maintains a bitmap, shared by the controllers in the cluster, in the non-volatile memory. The bitmap indicates the status of a write operation to any data block or parity block. A “dirty” data unit in the bitmap indicates that a write operation has been submitted but is not yet finished. Upon submitting a write request (to update a data block or a parity block) to the storage facility, a controller sets the corresponding data unit “dirty” in the bitmap. After receiving an acknowledgement from the storage facility indicating that the operation has been completed, the controller clears the corresponding data unit. If a controller fails during a write operation, another controller can use the bitmap to re-establish data consistency.Type: GrantFiled: October 30, 2007Date of Patent: November 2, 2010Assignee: Network Appliance, Inc.Inventor: Thomas Rudolf Wenzel
-
Patent number: 7818504Abstract: A storage system may include storage, a main pipeline to carry data for the storage, and a store pipeline to carry data for the storage. The storage system may also include a controller to prioritize data storage requests for the storage based upon available interleaves and which pipeline is associated with the data storage requests.Type: GrantFiled: September 26, 2007Date of Patent: October 19, 2010Assignee: International Business Machines CorporationInventors: Derrin M. Berger, Michael A. Blake, Garrett M Drapala, Pak-kin Mak
-
Patent number: 7779205Abstract: A multi processor system 1 comprises a plurality of processors 21 to 25, a system bus 30 and a main system memory 40. Each processor 21 to 25 is connected to a respective cache memory 41 to 45, with each cache memory 41 to 45 in turn being connected to the system bus 30. The cache memories 41 to 45 store copies of data or instructions that are used frequently by the respective processors 21 to 25, thereby eliminating the need for the processors 21 to 25 to access the main system memory 40 during each read or write operation. Processor 25 is connected to a local memory 50 having a plurality of data blocks (not shown). According to the invention, the local memory 50 has a first port 51 for connection to its respective processor 25. In addition, the local memory 50 has a second port 52 connected to the system bus 30, thereby allowing one or more of the other processors 21 to 24 to access the local memory 50.Type: GrantFiled: November 8, 2005Date of Patent: August 17, 2010Assignee: Koninklijke Philips Electronics N.V.Inventor: Jan Hoogerbrugge
-
Patent number: 7747828Abstract: A static random access memory (SRAM) includes an input read register (IRR) for monitoring the state of external binary devices and an output drive register (ODR) for controlling the state of external binary devices. The SRAM can be a multi-port device for access by multiple processors or controllers. Each bit of the IRR can mirror the state of a connected external binary device, and can be read to a connected processor using a standard read instruction. Each bit of the ODR can manipulate the state of a connected external binary device by providing the device with a path to the SRAM supply voltage. Each bit of the ODR can also be read without changing the state, or interrupting the operation of, the connected external binary device. When set to the proper mode, the addresses used for the IRR and ODR can be used with the SRAM main memory array for standard memory operations.Type: GrantFiled: November 17, 2004Date of Patent: June 29, 2010Assignee: Integrated Device Technology, Inc.Inventors: Yunsheng Wang, Casey Springer, Tak Kwong Wong, Bill Beane
-
Patent number: 7711900Abstract: A method, system and program product for equitable sharing of a CAM (Content Addressable Memory) table among multiple users of a switch. The method includes reserving buffers in the table to be shared, the remaining buffers being allocated to each user. The method further includes establishing whether or not an address contained in a packet from a user is listed in a buffer in the table, if the address is listed, updating a time-to-live value for the buffer for forwarding the packet and, if the address is not listed, determining whether or not the user has exceeded its allocated buffers and whether or not the reserved buffers have been exhausted, such that, if the user has exceeded its allocated buffers and the reserved buffers have been exhausted, the address is not added to the table and the user is precluded from using any additional buffers in the network switch.Type: GrantFiled: March 5, 2007Date of Patent: May 4, 2010Assignee: International Business Machines CorporationInventors: Bruce Booth, Mark E. Goodgion, Atef O. Zaghloul, John H. Zeiger
-
Patent number: 7710426Abstract: Buffers may be shared between components in a system. The components may be loosely coupled, allowing the components to be assembled into various different configurations, and yet buffers may still be shared. A buffer requirements negotiator of the system analyzes the buffer requirements of each of the components and determines, if possible, a set of requirements that satisfies all of the components. Accordingly, savings may be achieved in buffer memory, as well as in copying and converting between unshared buffers. Further, the individual components may operate as efficiently as possible because the buffer requirements of the components in the system are all met. One implementation accesses a first component's buffer requirements and a second component's buffer requirements, determines a reconciled set of buffer requirements that satisfies the buffer requirements of both components, and provides the reconciled set of buffer requirements to one or more components.Type: GrantFiled: April 25, 2005Date of Patent: May 4, 2010Assignee: Apple Inc.Inventor: John Samuel Bushell
-
Patent number: 7698506Abstract: A technique for partially offloading, from a main cache in a storage server, the storage of cache tags for data blocks in a victim cache of the storage server, is described. The technique includes storing, in the main cache, a first subset of the cache tag information for each of the data blocks, and storing, in a victim cache of the storage server, a second subset of the cache tag information for each of the data blocks. This technique avoids the need to store the second subset of the cache tag information in the main cache.Type: GrantFiled: April 26, 2007Date of Patent: April 13, 2010Assignee: Network Appliance, Inc.Inventors: Robert L. Fair, William P. McGovern, Thomas C. Holland, Jason Sylvain
-
Patent number: 7694077Abstract: A multi-port instruction/data integrated cache which is provided between a parallel processor and a main memory and stores therein a part of instructions and data stored in the main memory has a plurality of banks, and a plurality of ports including an instruction port unit consisting of at least one instruction port used to access an instruction from the parallel processor and a data port unit consisting of at least one data port used to access data from the parallel processor. Further, a data width which can be specified to the bank from the instruction port is set larger than a data width which can be specified to the bank from the data port.Type: GrantFiled: February 20, 2008Date of Patent: April 6, 2010Assignee: Semiconductor Technology Academic Research CenterInventors: Tetsuo Hironaka, Hans Jürgen Mattausch, Tetsushi Koide, Tai Hirakawa, Koh Johguchi
-
Patent number: 7680988Abstract: A shared memory is usable by concurrent threads in a multithreaded processor, with any addressable storage location in the shared memory being readable and writeable by any of the threads. Processing engines that execute the threads are coupled to the shared memory via an interconnect that transfers data in only one direction (e.g., from the shared memory to the processing engines); the same interconnect supports both read and write operations. The interconnect advantageously supports multiple parallel read or write operations.Type: GrantFiled: October 30, 2006Date of Patent: March 16, 2010Assignee: NVIDIA CorporationInventors: John R. Nickolls, Brett W. Coon, Ming Y. Siu, Stuart F. Oberman, Samuel Liu
-
Publication number: 20100011167Abstract: A multi-core processor providing heterogeneous processor cores and a shared cache is presented.Type: ApplicationFiled: July 6, 2009Publication date: January 14, 2010Inventors: Frank T. Hady, Mason B. Cabot, John Beck, Mark B. Rosenbluth