Patents by Inventor Robert G. Gelinas
Robert G. Gelinas has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 9424195Abstract: A method of managing cache memory includes accessing a cache memory at a primary index that corresponds to an address specified in an access request. A determination is made that accessing the cache memory at the primary index does not result in a cache hit on a cache line with an error-free status. In response to this determination, the primary index is mapped to a secondary index and data for the address is written to a cache line at the secondary index.Type: GrantFiled: April 15, 2014Date of Patent: August 23, 2016Assignee: ADVANCED MICRO DEVICES, INC.Inventors: John Kalamatianos, Johnsy Kanjirapallil John, Phillip E. Nevius, Robert G. Gelinas
-
Publication number: 20150293854Abstract: A method of managing cache memory includes accessing a cache memory at a primary index that corresponds to an address specified in an access request. A determination is made that accessing the cache memory at the primary index does not result in a cache hit on a cache line with an error-free status. In response to this determination, the primary index is mapped to a secondary index and data for the address is written to a cache line at the secondary index.Type: ApplicationFiled: April 15, 2014Publication date: October 15, 2015Applicant: Advanced Micro Devices, Inc.Inventors: John Kalamatianos, Johnsy Kanjirapallil John, Phillip E. Nevius, Robert G. Gelinas
-
Publication number: 20040034759Abstract: An apparatus and method for increasing throughput in a processor having a multi-threaded pipeline is provided. Throughput is increased by dynamically allocating hardware contexts to pipeline flows according to context issue rules. The context issue rules eliminate some hardware bypass paths allowing for a shorter clock period and minimize pipeline stalls. One context issue rule eliminates the need for an E-E bypass path by ensuring that no context is allowed to issue in two adjacent pipeline flows. Another context issue rule eliminates the need for an M-E bypass path by ensuring that data retrieved from memory in a pipeline flow for a context is available prior to a successive pipeline flow for the same context entering the execution stage. A beat issue rule looks for reduced utilization of the pipeline when no active context can issue an instruction due to the context issue rules.Type: ApplicationFiled: October 17, 2002Publication date: February 19, 2004Applicant: Lexra, Inc.Inventors: Solomon J. Katzman, Michael A. Cotsford, Robert G. Gelinas, W. Patrick Hays, Todd H. Snyder
-
Patent number: 6122756Abstract: A high availability computer system and methodology including a backplane, having at least one backplane communication bus and a diagnostic bus, a plurality of motherboards, each interfacing to the diagnostic bus. Each motherboard also includes a memory system including main memory distributed among the plurality of motherboards and a memory controller module for accessing said main memory interfacing to said motherboard communication bus. Each motherboard also includes at least one daughterboard, detachably connected to thereto. The motherboard further includes a backplane diagnostic bus interface mechanism interfacing each of the motherboards to the backplane diagnostic bus; a microcontroller for processing information and providing outputs and a test bus controller mechanism including registers therein.Type: GrantFiled: February 10, 1998Date of Patent: September 19, 2000Assignee: Data General CorporationInventors: William F. Baxter, Robert G. Gelinas, James M. Guyer, Dan R. Huck, Michael F. Hunt, David L. Keating, Jeff S. Kimmell, Phil J. Roux, Liz M. Truebenbach, Rob P. Valentine, Pat J. Weiler, Joseph Cox, Barry E. Gillott, Andrea Heyda, Rob J. Pike, Tom V. Radogna, Art A. Sherman, Micheal Sporer, Doug J. Tucker, Simon N. Yeung
-
Patent number: 6026461Abstract: A very fast, memory efficient, highly expandable, highly efficient CCNUMA processing system based on a hardware architecture that minimizes system bus contention, maximizes processing forward progress by maintaining strong ordering and avoiding retries, and implements a full-map directory structure cache coherency protocol. A Cache Coherent Non-Uniform Memory Access (CCNUMA) architecture is implemented in a system comprising a plurality of integrated modules each consisting of a motherboard and two daughterboards. The daughterboards, which plug into the motherboard, each contain two Job Processors (JPs), cache memory, and input/output (I/O) capabilities. Located directly on the motherboard are additional integrated I/O capabilities in the form of two Small Computer System Interfaces (SCSI) and one Local Area Network (LAN) interface. The motherboard includes main memory, a memory controller (MC) and directory DRAMs for cache coherency.Type: GrantFiled: December 9, 1998Date of Patent: February 15, 2000Assignee: Data General CorporationInventors: William F. Baxter, Robert G. Gelinas, James M. Guyer, Dan R. Huck, Michael F. Hunt, David L. Keating, Jeff S. Kimmell, Phil J. Roux, Liz M. Truebenbach, Rob P. Valentine, Pat J. Weiler, Joseph Cox, Barry E. Gillott, Andrea Heyda, Rob J. Pike, Tom V. Radogna, Art A. Sherman, Michael Sporer, Doug J. Tucker, Simon N. Yeung
-
Patent number: 5887146Abstract: A very fast, memory efficient, highly expandable, highly efficient CCNUMA processing system based on a hardware architecture that minimizes system bus contention, maximizes processing forward progress by maintaining strong ordering and avoiding retries, and implements a full-map directory structure cache coherency protocol. A Cache Coherent Non-Uniform Memory Access (CCNUMA) architecture is implemented in a system comprising a plurality of integrated modules each consisting of a motherboard and two daughterboards. The daughterboards, which plug into the motherboard, each contain two Job Processors (JPs), cache memory, and input/output (I/O) capabilities. Located directly on the motherboard are additional integrated I/O capabilities in the form of two Small Computer System Interfaces (SCSI) and one Local Area Network (LAN) interface. The motherboard includes main memory, a memory controller (MC) and directory DRAMs for cache coherency.Type: GrantFiled: August 12, 1996Date of Patent: March 23, 1999Assignee: Data General CorporationInventors: William F. Baxter, Robert G. Gelinas, James M. Guyer, Dan R. Huck, Michael F. Hunt, David L. Keating, Jeff S. Kimmell, Phil J. Roux, Liz M. Truebenbach, Rob P. Valentine, Pat J. Weiler, Joseph Cox, Barry E. Gillott, Andrea Heyda, Rob J. Pike, Tom V. Radogna, Art A. Sherman, Michael Sporer, Doug J. Tucker, Simon N. Yeung
-
Patent number: 4979141Abstract: A system for use in a floating point computation unit for providing a sign/magniture subtraction operation, which system uses propagate/generate logic responsive to the subtraction operands to produce intermediate and final propagate and generate outputs. First carry computation logic is responsive to the final propagate and generate outputs and to a carry-in bit to produce final carry outputs and a carry-out bit. The latter outputs are used to produce a first subtraction result. Second carry computation logic responds to the intermediate propagate and generate outputs to produce second final carry outputs. The latter outputs and selected intermediate propagate outputs are used to produce a second subtraction result. The carry-out bit then selects one of the two subtraction results as the final subtraction result.Type: GrantFiled: September 28, 1988Date of Patent: December 18, 1990Assignee: Data General CorporationInventors: Robert G. Gelinas, Thomas V. Radogna