Patents by Inventor Sai Prashanth Muralidhara

Sai Prashanth Muralidhara has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11669454
    Abstract: A processor includes one or more cores having cache, a cache home agent (CHA), a near memory controller, to near memory, and a far memory controller, which is to: receive a first memory read operation from the CHA directed at a memory address; detect a miss for the first memory address at the near memory; issue a second memory read operation to the far memory controller to retrieve a cache line, having first data, from the memory address of far memory; receive the cache line from the far memory controller in response to the second memory read operation; and send the cache line to the CHA with a forced change to a directory state of the cache line at the CHA, the forced change to cause the CHA to snoop remote sockets to maintain data coherence for the cache line in an absence of directory state in the far memory.
    Type: Grant
    Filed: May 7, 2019
    Date of Patent: June 6, 2023
    Assignee: Intel Corporation
    Inventors: Vedaraman Geetha, Jeffrey Baxter, Sai Prashanth Muralidhara, Sharada Venkateswaran, Daniel Liu, Nishant Singh, Bahaa Fahim, Samuel D. Strom
  • Publication number: 20230093247
    Abstract: An embodiment of an integrated circuit may comprise local memory, a plurality of per-page counters located in a non-system-addressable region of the local memory, and circuitry coupled to the local memory, the circuitry to count accesses to pages of a system-addressable memory space with the plurality of per-page counters located in the non-system-addressable region of the local memory. Other embodiments are disclosed and claimed.
    Type: Application
    Filed: September 22, 2021
    Publication date: March 23, 2023
    Applicant: Intel Corporation
    Inventors: Sanjay Kumar, Bhargavi Narayanasetty, Andrew Anderson, Anupama Kurpad, Evgeny V. Voevodin, Patrick Ndouniama, Sai Prashanth Muralidhara, Rajat Agarwal, Mohamed Arafa
  • Publication number: 20230091205
    Abstract: Methods and apparatus relating to memory side prefetch architecture for improved memory bandwidth are described. In an embodiment, logic circuitry transmits a Direct Memory Controller Prefetch (DMCP) request to cause a prefetch of data from memory. A memory controller receives the DMCP request and issues a plurality of read operations to the memory in response to the DMCP request. Data read from the memory is stored in a storage structure in response to the plurality of read operations. Other embodiments are also disclosed and claimed.
    Type: Application
    Filed: September 20, 2021
    Publication date: March 23, 2023
    Applicant: Intel Corporation
    Inventors: Adrian Moga, Ugonna Echeruo, Eduard Roytman, Krishnakanth Sistla, Joseph Nuzman, Brinda Ganesh, Meenakshisundaram Chinthamani, Yen-Cheng Liu, Sai Prashanth Muralidhara, Vivek Kozhikkottu, Hanna Alam, Narasimha Sridhar Srirangam
  • Publication number: 20220222178
    Abstract: A system includes a multilevel memory such as a two level memory (2LM), where a first level memory acts as a cache for the second level memory. A memory controller or cache controller can detect a cache miss in the first level memory for a request for data. Instead of automatically performing a swap, the controller can determine whether to perform a swap based on a swap policy assigned to a memory region associated with the address of the requested data.
    Type: Application
    Filed: March 31, 2022
    Publication date: July 14, 2022
    Inventors: Rajat AGARWAL, Sai Prashanth MURALIDHARA, Wei P. CHEN, Nishant SINGH, Sharada VENKATESWARAN, Daniel W. LIU
  • Publication number: 20210216452
    Abstract: A two-level main memory in which both volatile memory and persistent memory are exposed to the operating system in a flat manner and data movement and management is performed in cache line granularity is provided. The operating system can allocate pages in the two-level main memory randomly across the first level main memory and the second level main memory in a memory-type agnostic manner, or, in a more intelligent manner by allocating predicted hot pages in first level main memory and predicted cold pages in second level main memory. The cache line granularity movement is performed in a “swap” manner, that is, a hot cache line in the second level main memory is swapped with a cold cache line in first level main memory because data is stored in either first level main memory or second level main memory not in both first level main memory and second level main memory.
    Type: Application
    Filed: March 27, 2021
    Publication date: July 15, 2021
    Inventors: Sai Prashanth MURALIDHARA, Alaa R. ALAMELDEEN, Rajat AGARWAL, Wei P. CHEN, Vivek KOZHIKKOTTU
  • Publication number: 20200356482
    Abstract: A processor includes one or more cores having cache, a cache home agent (CHA), a near memory controller, to near memory, and a far memory controller, which is to: receive a first memory read operation from the CHA directed at a memory address; detect a miss for the first memory address at the near memory; issue a second memory read operation to the far memory controller to retrieve a cache line, having first data, from the memory address of far memory; receive the cache line from the far memory controller in response to the second memory read operation; and send the cache line to the CHA with a forced change to a directory state of the cache line at the CHA, the forced change to cause the CHA to snoop remote sockets to maintain data coherence for the cache line in an absence of directory state in the far memory.
    Type: Application
    Filed: May 7, 2019
    Publication date: November 12, 2020
    Inventors: Vedaraman Geetha, Jeffrey Baxter, Sai Prashanth Muralidhara, Sharada Venkateswaran, Daniel Liu, Nishant Singh, Bahaa Fahim, Samuel D. Strom