Patents by Inventor Zhigang Hu

Zhigang Hu has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20080244186
    Abstract: A write filter cache system for protecting a microprocessor core from soft errors and method thereof are provided. In one aspect, data coming from a processor core to be written in primary cache memory, for instance, L1 cache memory system, is buffered in a write filter cache placed between the primary cache memory and the processor core. The data from the write filter is move to the main cache memory only if it is verified that main thread's data is soft error free, for instance, by comparing the main thread's data with that of its redundant thread. The main cache memory only keeps clean data associated with accepted checkpoints.
    Type: Application
    Filed: June 12, 2008
    Publication date: October 2, 2008
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Pradip Bose, Zhigang Hu, Xiaodong Li, Jude A. Rivers
  • Publication number: 20080234477
    Abstract: This invention involves a new method for preparing chitosan nano-particles. This method uses the nano-cavity technology so as to obtain pure chitosan nano-particles and pure chitosan nano-particle latex in a simple way.
    Type: Application
    Filed: May 27, 2008
    Publication date: September 25, 2008
    Applicant: The Hong Kong Polytechnic University
    Inventors: Yau Shan Szeto, Zhigang Hu
  • Publication number: 20080229068
    Abstract: A multithreaded processor, fetch control for a multithreaded processor and a method of fetching in the multithreaded processor. Processor event and use (EU) signals are monitored for downstream pipeline conditions indicating pipeline execution thread states. Instruction cache fetches are skipped for any thread that is incapable of receiving fetched cache contents, e.g., because the thread is full or stalled. Also, consecutive fetches may be selected for the same thread, e.g., on a branch mis-predict. Thus, the processor avoids wasting power on unnecessary or place keeper fetches.
    Type: Application
    Filed: April 21, 2008
    Publication date: September 18, 2008
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPERATION
    Inventors: PRADIP BOSE, ALPER BUYUKTOSUNOGLU, RICHARD J. EICKEMEYER, LEE E. EISEN, PHILIP G. EMMA, JOHN B. GRISWELL, ZHIGANG HU, HUNG Q. LE, DOUGLAS R. LOGAN, BALARAM SINHAROY
  • Patent number: 7392366
    Abstract: A multithreaded processor, fetch control for a multithreaded processor and a method of fetching in the multithreaded processor. Processor event and use (EU) signs are monitored for downstream pipeline conditions indicating pipeline execution thread states. Instruction cache fetches are skipped for any thread that is incapable of receiving fetched cache contents, e.g., because the thread is full or stalled. Also, consecutive fetches may be selected for the same thread, e.g., on a branch mis-predict. Thus, the processor avoids wasting power on unnecessary or place keeper fetches.
    Type: Grant
    Filed: September 16, 2005
    Date of Patent: June 24, 2008
    Assignee: International Business Machines Corp.
    Inventors: Pradip Bose, Alper Buyuktosunoglu, Richard J. Eickemeyer, Lee E. Eisen, Philip G. Emma, John B. Griswell, Zhigang Hu, Hung Q. Le, Douglas R. Logan, Balaram Sinharoy
  • Publication number: 20080133886
    Abstract: A multithreaded processor, fetch control for a multithreaded processor and a method of fetching in the multithreaded processor. Processor event and use (EU) signals are monitored for downstream pipeline conditions indicating pipeline execution thread states. Instruction cache fetches are skipped for any thread that is incapable of receiving fetched cache contents, e.g., because the thread is full or stalled. Also, consecutive fetches may be selected for the same thread, e.g., on a branch mis-predict. Thus, the processor avoids wasting power on unnecessary or place keeper fetches.
    Type: Application
    Filed: October 30, 2007
    Publication date: June 5, 2008
    Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: PRADIP BOSE, ALPER BUYUKTOSUNOGLU, RICHARD J. EICKEMEYER, LEE E. EISEN, PHILIP G. EMMA, JOHN B. GRISWELL, ZHIGANG HU, HUNG Q. LE, DOUGLAS R. LOGAN, BALARAM SINHAROY
  • Publication number: 20080126771
    Abstract: An instruction cache (I-Cache) for a processor is configured to include a Branch Target Extension associated with each Instruction Sector. When an Instruction Sector is fetched, the Branch Target Extension is simultaneously fetched. If the Instruction Sector has a branch instruction that is predicted taken, then the branch target address in the branch extension is used to access the next Instruction Sector. In other embodiments, each Instruction Sector has a plurality of Branch Target Extensions each corresponding to a potential branch instruction in an Instruction Sector. In this case, the Branch Target Extensions are partitioned into an instruction index field for locating branch instruction in the Instruction Sector, a local predictor field for predicted taken status and a target address field for the branch target address. The least significant bits of the instruction fetch address are compared to the instruction indexes to determine a particular Branch Target Extension to use.
    Type: Application
    Filed: July 25, 2006
    Publication date: May 29, 2008
    Inventors: Lei Chen, Zhigang Hu, Lixin Zhang
  • Publication number: 20080016393
    Abstract: A write filter cache system for protecting a microprocessor core from soft errors and method thereof are provided. In one aspect, data coming from a processor core to be written in primary cache memory, for instance, L1 cache memory system, is buffered in a write filter cache placed between the primary cache memory and the processor core. The data from the write filter is move to the main cache memory only if it is verified that main thread's data is soft error free, for instance, by comparing the main thread's data with that of its redundant thread. The main cache memory only keeps clean data associated with accepted checkpoints.
    Type: Application
    Filed: July 14, 2006
    Publication date: January 17, 2008
    Inventors: Pradip Bose, Zhigang Hu, Xiaodong Li, Jude A. Rivers
  • Patent number: 7284095
    Abstract: A method for replacing cache lines in a computer system having a non-uniform set associative cache memory is disclosed. The method incorporates access latency as an additional factor into the existing ranking guidelines for replacement of a line, the higher the rank of the line the sooner that it is likely to be evicted from the cache. Among a group of highest ranking cache lines in a cache set, the cache line chosen to be replaced is one that provides the lowest latency access to a requesting entity, such as a processor. The distance separating the requesting entity from the memory partition where the cache line is stored most affects access latency.
    Type: Grant
    Filed: August 18, 2004
    Date of Patent: October 16, 2007
    Assignee: International Business Machines Corporation
    Inventors: Zhigang Hu, William Robert Reohr
  • Patent number: 7228388
    Abstract: Arrangements and method for enabling and disabling cache bypass in a computer system with a cache hierarchy. Cache bypass status is identified with respect to at least one cache line. A cache line identified as cache bypass enabled is transferred to one or more higher level caches of the cache hierarchy, whereby a next higher level cache in the cache hierarchy is bypassed, while a cache line identified as cache bypass disabled is transferred to one or more higher level caches of the cache hierarchy, whereby a next higher level cache in the cache hierarchy is not bypassed. Included is an arrangement for selectively enabling or disabling cache bypass with respect to at least one cache line based on historical cache access information.
    Type: Grant
    Filed: November 19, 2004
    Date of Patent: June 5, 2007
    Assignee: International Business Machines Corporation
    Inventors: Zhigang Hu, John T. Robinson, Xiaowei Shen, Balaram Sinharoy
  • Publication number: 20070043960
    Abstract: A device for controlling power parameters in a microprocessor includes a resource activation control unit for controlling the maximum power of the microprocessor and two or more resources. The resource activation control unit controls the activation of the resources such that the consumed and dissipated power of the microprocessor does not exceed a power bound which is configurable to a predetermined value below the maximum power.
    Type: Application
    Filed: August 19, 2005
    Publication date: February 22, 2007
    Inventors: Pradip Bose, Alper Buyuktosunoglu, Zhigang Hu, Hans Jacobson, Vijayalakshmi Srinivasan, Victor Zyuban
  • Publication number: 20060248287
    Abstract: Arrangements and methods for providing cache management. Preferably, a buffer arrangement is provided that is adapted to record incoming data into a first cache memory from a second cache memory, convey a data location in the first cache memory upon a prompt for corresponding data, in the event of a hit in the first cache memory, and refer to the second cache memory in the event of a miss in the first cache memory.
    Type: Application
    Filed: April 29, 2005
    Publication date: November 2, 2006
    Applicant: IBM Corporation
    Inventors: Alper Buyuktosunoglu, Zhigang Hu, Jude Rivers, John Robinson, Xiaowei Shen, Vijayalakshmi Srinivasan
  • Patent number: 7093075
    Abstract: A system and method for reducing latency in memory systems is provided. A copy way is established in a set of a set associative cache, which is physically closer to a requesting entity than other memory positions. Likely to be accessed data is copied to the copy way for subsequent access. In this way, subsequent accesses of the most likely data have their access time reduced due to the physical proximity of the data being close to the requesting entity. Methods herein further provide ranking and rearranging blocks in the cache based on coupled local and global least recently used (LRU) algorithms to reduce latency time.
    Type: Grant
    Filed: November 7, 2003
    Date of Patent: August 15, 2006
    Assignee: International Business Machines Corporation
    Inventors: William Robert Reohr, Zhigang Hu
  • Publication number: 20060155933
    Abstract: A hardware based method for determining when to migrate cache lines to the cache bank closest to the requesting processor to avoid remote access penalty for future requests. In a preferred embodiment, decay counters are enhanced and used in determining the cost of retaining a line as opposed to replacing it while not losing the data. In one embodiment, a minimization of off-chip communication is sought; this may be particularly useful in a CMP environment.
    Type: Application
    Filed: January 13, 2005
    Publication date: July 13, 2006
    Applicant: International Business Machines Corporation
    Inventors: Alper Buyuktosunoglu, Zhigang Hu, Jude Rivers, John Robinson, Xiaowei Shen, Vijayalakshmi Srinivasan
  • Publication number: 20060112233
    Abstract: Arrangements and method for enabling and disabling cache bypass in a computer system with a cache hierarchy. Cache bypass status is identified with respect to at least one cache line. A cache line identified as cache bypass enabled is transferred to one or more higher level caches of the cache hierarchy, whereby a next higher level cache in the cache hierarchy is bypassed, while a cache line identified as cache bypass disabled is transferred to one or more higher level caches of the cache hierarchy, whereby a next higher level cache in the cache hierarchy is not bypassed. Included is an arrangement for selectively enabling or disabling cache bypass with respect to at least one cache line based on historical cache access information.
    Type: Application
    Filed: November 19, 2004
    Publication date: May 25, 2006
    Applicant: IBM Corporation
    Inventors: Zhigang Hu, John Robinson, Xiaowei Shen, Balaram Sinharoy
  • Publication number: 20060101238
    Abstract: A multithreaded processor, fetch control for a multithreaded processor and a method of fetching in the multithreaded processor. Processor event and use (EU) signs are monitored for downstream pipeline conditions indicating pipeline execution thread states. Instruction cache fetches are skipped for any thread that is incapable of receiving fetched cache contents, e.g., because the thread is full or stalled. Also, consecutive fetches may be selected for the same thread, e.g., on a branch mis-predict. Thus, the processor avoids wasting power on unnecessary or place keeper fetches.
    Type: Application
    Filed: September 16, 2005
    Publication date: May 11, 2006
    Inventors: Pradip Bose, Alper Buyuktosunoglu, Richard Eickemeyer, Lee Eisen, Philip Emma, John Griswell, Zhigang Hu, Hung Le, Douglas Logan, Balaram Sinharoy
  • Publication number: 20060041769
    Abstract: An adaptive cache decay technique is disclosed that removes power from cache lines that have not been accessed for a variable time interval, referred to as the cache line decay interval, assuming that these cache lines are unlikely to be accessed in the future. The decay interval may be increased or decreased for each cache line to increase cache performance or save power, respectively. A default decay interval is initially established for the cache and the default decay interval may then be adjusted for a given cache line based on the performance of the cache line following a cache decay. The cache decay performance is evaluated by determining if a cache line was decayed too quickly. If a cache line is decayed and the same cache contents are again required, then the cache line was decayed too quickly and the cache line decay interval is increased. If a cache line is decayed and the cache line is then accessed to obtain a different cache content, the cache line decay interval can be decreased.
    Type: Application
    Filed: October 7, 2005
    Publication date: February 23, 2006
    Inventors: Zhigang Hu, Stefanos Kaxiras, Margaret Martonosi
  • Publication number: 20060041720
    Abstract: A method for replacing cache lines in a computer system having a non-uniform set associative cache memory is disclosed. The method incorporates access latency as an additional factor into the existing ranking guidelines for replacement of a line, the higher the rank of the line the sooner that it is likely to be evicted from the cache. Among a group of highest ranking cache lines in a cache set, the cache line chosen to be replaced is one that provides the lowest latency access to a requesting entity, such as a processor. The distance separating the requesting entity from the memory partition where the cache line is stored most affects access latency.
    Type: Application
    Filed: August 18, 2004
    Publication date: February 23, 2006
    Inventors: Zhigang Hu, William Reohr
  • Publication number: 20050102475
    Abstract: A system and method for reducing latency in memory systems is provided. A copy way is established in a set of a set associative cache, which is physically closer to a requesting entity than other memory positions. Likely to be accessed data is copied to the copy way for subsequent access. In this way, subsequent accesses of the most likely data have their access time reduced due to the physical proximity of the data being close to the requesting entity. Methods herein further provide ranking and rearranging blocks in the cache based on coupled local and global least recently used (LRU) algorithms to reduce latency time.
    Type: Application
    Filed: November 7, 2003
    Publication date: May 12, 2005
    Inventors: William Reohr, Zhigang Hu
  • Publication number: 20030145241
    Abstract: An adaptive cache decay technique is disclosed that removes power from cache lines that have not been accessed for a variable time interval, referred to as the cache line decay interval, assuming that these cache lines are unlikely to be accessed in the future. The decay interval may be increased or decreased for each cache line to increase cache performance or save power, respectively. A default decay interval is initially established for the cache and the default decay interval may then be adjusted for a given cache line based on the performance of the cache line following a cache decay. The cache decay performance is evaluated by determining if a cache line was decayed too quickly. If a cache line is decayed and the same cache contents are again required, then the cache line was decayed too quickly and the cache line decay interval is increased. If a cache line is decayed and the cache line is then accessed to obtain a different cache content, the cache line decay interval can be decreased.
    Type: Application
    Filed: January 30, 2002
    Publication date: July 31, 2003
    Inventors: Zhigang Hu, Stefanos Kaxiras, Margaret Martonosi