Patents by Inventor Jin Chin Wang
Jin Chin Wang has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20150032931Abstract: A system includes a first processing component, a second processing component, and an adapted bus linking the first and second processing components. The adapted bus may account for a circuit characteristic or on-chip variation in the system. For example, the bus may be adapted to include a wider data width because of an effect of the on-chip variation that limits performance of the bus at a lower data width. The bus may include a widened data width for a portion of the bus. In that regard, the bus may include a bus expander and a bus narrower for adjusting the data width of the bus.Type: ApplicationFiled: September 27, 2013Publication date: January 29, 2015Inventors: David Alan Baer, Brian Schoner, Jin-Chin Wang
-
Patent number: 7711906Abstract: Systems and methods that cache are provided. In one example, a system may include a spatial cache system coupled to a processing unit and to a memory. The spatial cache system may be adapted to reduce the memory latency of the processing unit. The spatial cache system may be adapted to store prefetched blocks, each stored prefetched block including a plurality of cache lines. If a cache line requested by the processing unit resides in one of the stored prefetched blocks and does not reside in the processing unit, then the spatial cache system may be adapted to provide the processing unit with the requested cache line.Type: GrantFiled: May 13, 2005Date of Patent: May 4, 2010Assignee: Broadcom CorporationInventors: Kimming So, Jin Chin Wang
-
Patent number: 7627720Abstract: Systems and methods that provide directional prefetching are provided. In one embodiment, a method may include one or more of the following: storing a first block and a second block in a prefetch buffer; associating a first block access with a backward prefetch scheme; associating a second block access with a forward prefetch scheme; and, if the first block is accessed before the second block, then performing a backward prefetch with respect to the first block.Type: GrantFiled: April 8, 2005Date of Patent: December 1, 2009Assignee: Broadcom CorporationInventors: Kimming So, Jin Chin Wang
-
Patent number: 7167954Abstract: Systems and methods that cache are provided. In one example, a system may include a spatial cache system coupled to a processing unit and to a memory. The spatial cache system may be adapted to reduce the memory latency of the processing unit. The spatial cache system may be adapted to store prefetched blocks, each stored prefetched block including a plurality of cache lines. If a cache line requested by the processing unit resides in one of the stored prefetched blocks and does not reside in the processing unit, then the spatial cache system may be adapted to provide the processing unit with the requested cache line.Type: GrantFiled: November 14, 2002Date of Patent: January 23, 2007Assignee: Broadcom CorporationInventors: Kimming So, Jin Chin Wang
-
Patent number: 6931494Abstract: Systems and methods that provide directional prefetching are provided. In one embodiment, a method may include one or more of the following: storing a first block and a second block in a prefetch buffer; associating a first block access with a backward prefetch scheme; associating a second block access with a forward prefetch scheme; and, if the first block is accessed before the second block, then performing a backward prefetch with respect to the first block.Type: GrantFiled: November 14, 2002Date of Patent: August 16, 2005Assignee: Broadcom CorporationInventors: Kimming So, Jin Chin Wang
-
Patent number: 6715040Abstract: Described is a data processing system including a processor, a plurality of caches, and main memory, the secondary caches being implemented as being non-inclusive, i.e., the lower order caches not storing a superset of the data stored in the next higher order cache. The non-inclusive cache structure provides increased flexibility in the storage of data. The operation of a write request operation when the target data line is not found in the primary cache. By using the dirty bit associated with each data line, the interaction between the processor and the primary cache can be reduced. By using the invalidity bit associated with each data line, the interaction between the processor and the primary cache can be reduced.Type: GrantFiled: January 7, 2002Date of Patent: March 30, 2004Assignee: NEC Electronics, Inc.Inventors: Jin Chin Wang, Maciek P. Kozyrczak
-
Publication number: 20040049639Abstract: Systems and methods that cache are provided. In one example, a system may include a spatial cache system coupled to a processing unit and to a memory. The spatial cache system may be adapted to reduce the memory latency of the processing unit. The spatial cache system may be adapted to store prefetched blocks, each stored prefetched block including a plurality of cache lines. If a cache line requested by the processing unit resides in one of the stored prefetched blocks and does not reside in the processing unit, then the spatial cache system may be adapted to provide the processing unit with the requested cache line.Type: ApplicationFiled: November 14, 2002Publication date: March 11, 2004Inventors: Kimming So, Jin Chin Wang
-
Publication number: 20040049640Abstract: Systems and methods that provide directional prefetching are provided. In one embodiment, a method may include one or more of the following: storing a first block and a second block in a prefetch buffer; associating a first block access with a backward prefetch scheme; associating a second block access with a forward prefetch scheme; and, if the first block is accessed before the second block, then performing a backward prefetch with respect to the first block.Type: ApplicationFiled: November 14, 2002Publication date: March 11, 2004Inventors: Kimming So, Jin Chin Wang
-
Publication number: 20020174304Abstract: Described is a data processing system including a processor, a plurality of caches, and main memory, the secondary caches being implemented as being non-inclusive, i.e., the lower order caches not storing a superset of the data stored in the next higher order cache. The non-inclusive cache structure provides increased flexibility in the storage of data. The operation of a write request operation when the target data line is not found in the primary cache. By using the dirty bit associated with each data line, the interaction between the processor and the primary cache can be reduced. By using the invalidity bit associated with each data line, the interaction between the processor and the primary cache can be reduced.Type: ApplicationFiled: January 7, 2002Publication date: November 21, 2002Inventors: Jin Chin Wang, Maciek P. Kozyrczak
-
Patent number: 5784394Abstract: A method and apparatus in a data processing system having a plurality of node controllers and a memory unit for each of the node controllers. Each one of the node controllers including at least one processor having a cache. Each memory unit including a plurality of entries each having an exclusive bit, an address tag, and an inclusion field. Each bit of the inclusion field representing one of the node controllers. The method and apparatus allow error recovery for errors occurring within the entries without using the ECC implementation. Specifically, two parity bits are used for detecting errors within an entry and logic for flushing any cache lines represented by the entry in error. The method and apparatus also includes means for detecting persistent errors and for indicating whether the error is generated by either hardware or software.Type: GrantFiled: November 15, 1996Date of Patent: July 21, 1998Assignee: International Business Machines CorporationInventors: Manuel Joseph Alvarez, II, Gary Dale Carpenter, Kai Cheng, Jeffrey Holland Gruger, Jin Chin Wang
-
Patent number: 5680577Abstract: A method and system for processing multiple requests for data residing at the same memory address. The multiple requests are associated with an individual duplicate bit flag that indicates whether the request can be processed. Thus, manipulation of the duplicate bit flag controls the order of processing for each of the received requests, thereby maintaining data coherency and integrity.Type: GrantFiled: April 27, 1995Date of Patent: October 21, 1997Assignee: International Business Machines CorporationInventors: Steven George Aden, Kai Cheng, Jin Chin Wang, Ramanathan Raghavan
-
Patent number: 5655103Abstract: A system and method for identifying which incoming write data is valid and for insuring that stale data does not overwrite valid data within system memory within a symmetrical multiprocessor data processing system. Upon receipt of a Load Miss request from a processor, a stale bit is established and set equal to zero. A determination is then made of which other processor has ownership of the requested cache line. The requested cache line is then transferred in a cache-to-cache transfer from the second processor to the first processor. If the first processor further modifies the cache line and writes back the cache line to system memory before the original owner of the cache line writes back the stale data with an acknowledgment of the cache-to-cache transfer, the stale bit is set to one. Upon receipt from the acknowledgment from the original owner of the cache line, the stale data is dropped when it is determined that the stale bit has been set.Type: GrantFiled: February 13, 1995Date of Patent: August 5, 1997Assignee: International Business Machines CorporationInventors: Kai Cheng, Kimming So, Jin Chin Wang