Patents by Inventor Jean Jourdan

Jean Jourdan has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11947454
    Abstract: Apparatuses, systems, and methods for controlling cache allocations in a configurable combined private and shared cache in a processor-based system. The processor-based system is configured to receive a cache allocation request to allocate a line in a share cache structure, which may further include a client identification (ID). The cache allocation request and the client ID can be compared to a sub-non-uniform memory access (NUMA) (sub-NUMA) bit mask and a client allocation bit mask to generate a cache allocation vector. The sub-NUMA bit mask may have been programmed to indicate that processing cores associated with a sub-NUMA region are available, whereas processing cores associated with other sub-NUMA regions are not available, and the client allocation bit mask may have been programmed to indicate that processing cores are available.
    Type: Grant
    Filed: June 7, 2022
    Date of Patent: April 2, 2024
    Assignee: Ampere Computing LLC
    Inventors: Richard James Shannon, Stephan Jean Jourdan, Matthew Robert Erler, Jared Eric Bendt
  • Patent number: 11880306
    Abstract: Aspects disclosed in the detailed description include configuring a configurable combined private and shared cache in a processor. Related processor-based systems and methods are also disclosed. A combined private and shared cache structure is configurable to select a private cache portion and a shared cache portion.
    Type: Grant
    Filed: June 7, 2022
    Date of Patent: January 23, 2024
    Assignee: Ampere Computing LLC
    Inventors: Richard James Shannon, Stephan Jean Jourdan, Matthew Robert Erler, Jared Eric Bendt
  • Patent number: 11822487
    Abstract: A memory management unit (MMU) including a unified translation lookaside buffer (TLB) supporting a plurality of page sizes is disclosed. In one aspect, the MMU is further configured to store and dynamically update page size residency metadata associated with each of the plurality of page sizes. The page size residency metadata may include most recently used (MRU) page size data and/or a counter for each page size indicating how many pages of that page size are resident in the unified TLB. The unified TLB is configured to determine an order in which to perform a TLB lookup for at least a subset of page sizes of the plurality of page sizes based on the page size residency metadata.
    Type: Grant
    Filed: December 1, 2021
    Date of Patent: November 21, 2023
    Assignee: Ampere Computing LLC
    Inventors: George Van Horn Leming, III, John Gregory Favor, Stephan Jean Jourdan, Jonathan Christopher Perry, Bret Leslie Toll
  • Patent number: 11762660
    Abstract: A unified queue configured to perform decoupled prediction and fetch operations, and related apparatuses, systems, methods, and computer-readable media, is disclosed. The unified queue has a plurality of entries, where each entry is configured to store information associated with at least one instruction, and where the information comprises an identifier portion, a prediction information portion, and a tag information portion. The unified queue is configured to update the prediction information portion of each entry responsive to a prediction block, and to update the tag information portion of each entry responsive to a tag and TLB block. The prediction information may be updated more than once, and the unified queue is configured to take corrective action where a later prediction conflicts with an earlier prediction.
    Type: Grant
    Filed: June 23, 2020
    Date of Patent: September 19, 2023
    Assignee: Ampere Computing LLC
    Inventors: Brett Alan Ireland, Michael Stephen Chin, Stephan Jean Jourdan
  • Publication number: 20220398195
    Abstract: Aspects disclosed in the detailed description include configuring a configurable combined private and shared cache in a processor. Related processor-based systems and methods are also disclosed. A combined private and shared cache structure is configurable to select a private cache portion and a shared cache portion.
    Type: Application
    Filed: June 7, 2022
    Publication date: December 15, 2022
    Inventors: Richard James Shannon, Stephan Jean Jourdan, Matthew Robert Erler, Jared Eric Bendt
  • Publication number: 20220398196
    Abstract: Apparatuses, systems, and methods for configuring combined private and shared cache levels in a processor-based system. The processor-based system includes a processor that includes a plurality of processing cores each including execution circuits which are coupled to respective cache(s) and a configurable combined private and shared cache, and which may receive instructions and data on which to perform operations from the cache(s) and the combined private and shared cache. A shared cache portion of each configurable combined private and shared cache can be treated as an independently-assignable portion of the overall shared cache, which is effectively the shared cache portions of all of the processing cores. Each independently-assignable portion of the overall shared cache can be associated with a particular client running on the processor as an example. This approach can provide greater granularity of cache partitioning of a shared cache between particular clients running on a processor.
    Type: Application
    Filed: June 7, 2022
    Publication date: December 15, 2022
    Inventors: Richard James Shannon, Stephan Jean Jourdan, Matthew Robert Erler, Jared Eric Bendt
  • Publication number: 20220398193
    Abstract: Apparatuses, systems, and methods for controlling cache allocations in a configurable combined private and shared cache in a processor-based system. The processor-based system is configured to receive a cache allocation request to allocate a line in a share cache structure, which may further include a client identification (ID). The cache allocation request and the client ID can be compared to a sub-non-uniform memory access (NUMA) (sub-NUMA) bit mask and a client allocation bit mask to generate a cache allocation vector. The sub-NUMA bit mask may have been programmed to indicate that processing cores associated with a sub-NUMA region are available, whereas processing cores associated with other sub-NUMA regions are not available, and the client allocation bit mask may have been programmed to indicate that processing cores are available.
    Type: Application
    Filed: June 7, 2022
    Publication date: December 15, 2022
    Inventors: Richard James Shannon, Stephan Jean Jourdan, Matthew Robert Erler, Jared Eric Bendt
  • Patent number: 11386016
    Abstract: A memory management unit (MMU) including a unified translation lookaside buffer (TLB) supporting a plurality of page sizes is disclosed. In one aspect, the MMU is further configured to store and dynamically update page size residency metadata associated with each of the plurality of page sizes. The page size residency metadata may include most recently used (MRU) page size data and/or a counter for each page size indicating how many pages of that page size are resident in the unified TLB. The unified TLB is configured to determine an order in which to perform a TLB lookup for at least a subset of page sizes of the plurality of page sizes based on the page size residency metadata.
    Type: Grant
    Filed: December 20, 2019
    Date of Patent: July 12, 2022
    Assignee: Ampere Computing LLC
    Inventors: George Van Horn Leming, III, John Gregory Favor, Stephan Jean Jourdan, Jonathan Christopher Perry, Bret Leslie Toll
  • Publication number: 20220197807
    Abstract: An apparatus configured to provide latency-aware prefetching, and related systems, methods, and computer-readable media, are disclosed. The apparatus comprises a prefetch buffer comprising at least a first entry, and the first entry comprises a memory operation prefetch request portion storing a first previous memory operation prefetch request. The apparatus further comprises a prefetch buffer replacement circuit, which is configured to select an entry of the prefetch buffer storing a previous memory operation prefetch request for replacement with a subsequent memory operation prefetch request, and to replace the previous memory operation prefetch request in the selected entry with the subsequent memory operation prefetch request.
    Type: Application
    Filed: December 17, 2020
    Publication date: June 23, 2022
    Inventors: Jonathan Christopher Perry, Stephan Jean Jourdan, Mahesh Jagdish Madhav, Aarti Chandrashekhar
  • Publication number: 20220091997
    Abstract: A memory management unit (MMU) including a unified translation lookaside buffer (TLB) supporting a plurality of page sizes is disclosed. In one aspect, the MMU is further configured to store and dynamically update page size residency metadata associated with each of the plurality of page sizes. The page size residency metadata may include most recently used (MRU) page size data and/or a counter for each page size indicating how many pages of that page size are resident in the unified TLB.
    Type: Application
    Filed: December 1, 2021
    Publication date: March 24, 2022
    Inventors: George Van Horn Leming, III, John Gregory Favor, Stephan Jean Jourdan, Jonathan Christopher Perry, Bret Leslie Toll
  • Publication number: 20220004501
    Abstract: An apparatus configured to provide just-in-time synonym handling, and related systems, methods, and computer-readable media, are disclosed. The apparatus includes a first cache comprising a translation lookaside buffer (TLB) and a hit/miss block. The first cache is configured to form a miss request associated with an access to the first cache and provide the miss request to a second cache. The miss request comprises a physical address provided by the TLB and miss information provided by the hit/miss block. The first cache is further configured to receive, from the second cache, previously-stored metadata associated with an entry in the second cache. The entry in the second cache is associated with the miss request.
    Type: Application
    Filed: July 2, 2020
    Publication date: January 6, 2022
    Inventors: John Gregory Favor, Stephan Jean Jourdan, Jonathan Christopher Perry, Kjeld Svendsen, Bret Leslie Toll
  • Publication number: 20210397452
    Abstract: A unified queue configured to perform decoupled prediction and fetch operations, and related apparatuses, systems, methods, and computer-readable media, is disclosed. The unified queue has a plurality of entries, where each entry is configured to store information associated with at least one instruction, and where the information comprises an identifier portion, a prediction information portion, and a tag information portion. The unified queue is configured to update the prediction information portion of each entry responsive to a prediction block, and to update the tag information portion of each entry responsive to a tag and TLB block. The prediction information may be updated more than once, and the unified queue is configured to take corrective action where a later prediction conflicts with an earlier prediction.
    Type: Application
    Filed: June 23, 2020
    Publication date: December 23, 2021
    Inventors: Brett Alan Ireland, Michael Stephen Chin, Stephan Jean Jourdan
  • Publication number: 20210191877
    Abstract: A memory management unit (MMU) including a unified translation lookaside buffer (TLB) supporting a plurality of page sizes is disclosed. In one aspect, the MMU is further configured to store and dynamically update page size residency metadata associated with each of the plurality of page sizes. The page size residency metadata may include most recently used (MRU) page size data and/or a counter for each page size indicating how many pages of that page size are resident in the unified TLB.
    Type: Application
    Filed: December 20, 2019
    Publication date: June 24, 2021
    Inventors: George Van Horn Leming, III, John Gregory Favor, Stephan Jean Jourdan, Jonathan Christopher Perry, Bret Leslie Toll
  • Patent number: 8112743
    Abstract: The invention relates to a method of designing a system. The system includes an application having software components and an architecture having hardware components on which the application is run. The system has to satisfy at least one functional and one non-functional requirement. The functional analysis step (11) obtains a breakdown of the functional need relating to the application. A step defines the architecture (12). A step for designs hardware components (13) according to the architecture. A step design software components (14) based on the breakdown of the functional need. A step for integrates the software components in the hardware components (15). A step validates the functional requirements of the system (16). A step validates the non-functional requirement of the system (17). An upstream step (21) validates the non-functional requirement of the system, preceding the steps for designing hardware components (13) and software components (14).
    Type: Grant
    Filed: January 29, 2008
    Date of Patent: February 7, 2012
    Assignee: Thales
    Inventors: Martin Defour, Jean Jourdan, Franck Tailliez, Jean-Luc Voirin
  • Patent number: 8020174
    Abstract: The method of the invention consists, in a system of the type with three “tiers” , in inserting a fourth “tier” which is the business interaction and which makes it possible to obtain a user-system interaction component that is independent of the services of the application and of the media.
    Type: Grant
    Filed: September 29, 2003
    Date of Patent: September 13, 2011
    Assignee: THALES
    Inventors: Celestin Sedogbo, Pascal Bisson, Olivier Grisvard, Jean Jourdan, Thierry Poibeau
  • Publication number: 20080235655
    Abstract: The invention relates to a method of designing a system. The system includes an application having software components and an architecture having hardware components on which the application is run. The system has to satisfy at least one functional and one non-functional requirement. The functional analysis step (11) obtains a breakdown of the functional need relating to the application. A step defines the architecture (12). A step for designs hardware components (13) according to the architecture. A step design software components (14) based on the breakdown of the functional need. A step for integrates the software components in the hardware components (15). A step validates the functional requirements of the system (16). A step validates the non-functional requirement of the system (17). An upstream step (21) validates the non-functional requirement of the system, preceding the steps for designing hardware components (13) and software components (14).
    Type: Application
    Filed: January 29, 2008
    Publication date: September 25, 2008
    Applicant: THALES
    Inventors: Martin Defour, Jean Jourdan, Franck Tailliez, Jean-Luc Voirin
  • Publication number: 20050289560
    Abstract: The method of the invention consists, in a system of the type with three “tiers”, in inserting a fourth “tier” which is the business interaction and which makes it possible to obtain a user-system interaction component that is independent of the services of the application and of the media.
    Type: Application
    Filed: September 29, 2003
    Publication date: December 29, 2005
    Applicant: Thales
    Inventors: Celestin Sedogbo, Pascal Bisson, Olivier Grisvard, Jean Jourdan, Thiierry Poibeau