Patents by Inventor Ramakrishna Huggahalli

Ramakrishna Huggahalli has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20230401160
    Abstract: In one example of the present technology, an input/output memory management unit (IOMMU) of a computing device is configured to: receive a prefetch message including a virtual address from a central processing unit (CPU) core of a processor of the computing device; perform a page walk on the virtual address through a page table stored in a main memory of the computing device to obtain a prefetched translation of the virtual address to a physical address; and store the prefetched translation of the virtual address to the physical address in a translation lookaside buffer (TLB) of the IOMMU.
    Type: Application
    Filed: June 9, 2022
    Publication date: December 14, 2023
    Inventors: Ramakrishna HUGGAHALLI, Shachar RAINDEL
  • Publication number: 20230305968
    Abstract: Embodiments of the present disclosure includes techniques for cache memory replacement in a processing unit. A first data production operation to store first data to a first cache line of the cache memory is detected at a first time. A retention status of the first cache line is updated to a first retention level as a result of the first data production operation. Protection against displacement of the first data in the first cache line is increased based on the first retention level. A first data consumption operation retrieving the first data from the first cache line is detected at a second time after the first time. The retention status of the first cache line is updated to a second retention level as a result of the first data consumption operation, the second retention level being a lower level of retention than the first retention level.
    Type: Application
    Filed: March 28, 2022
    Publication date: September 28, 2023
    Inventors: Ramakrishna HUGGAHALLI, Shachar RAINDEL
  • Publication number: 20230305739
    Abstract: Embodiments of the present disclosure includes techniques for partial memory updates in a computer system. A data structure template is received. A first write data of a first write operation is received from a first data source, the first write operation performed in connection with provisioning of a first data payload to memory communicatively coupled with a processing unit. A first merge operation is performed involving the first write data and the first data structure template to obtain a first data structure update. The first data structure update is written to the memory, thereby improving efficiency of updating a first data structure associated with the first data payload.
    Type: Application
    Filed: March 28, 2022
    Publication date: September 28, 2023
    Inventors: Ramakrishna HUGGAHALLI, Shachar RAINDEL
  • Patent number: 11641326
    Abstract: Examples are described herein that relate to a mesh in a switch fabric. The mesh can include one or more buses that permit operations (e.g., read, write, or responses) to continue in the same direction, drop off to a memory, drop off a bus to permit another operation to use the bus, or receive operations that are changing direction. A latency estimate can be determined at least for operations that drop off from a bus to permit another operation to use the bus or receive and channel operations that are changing direction. An operation with a highest latency estimate (e.g., time of traversing a mesh) can be permitted to use the bus, even causing another operation, that is not to change direction, to drop off the bus and re-enter later.
    Type: Grant
    Filed: August 23, 2019
    Date of Patent: May 2, 2023
    Assignee: Intel Corporation
    Inventors: Karl S. Papadantonakis, Robert Southworth, Arvind Srinivasan, Helia A. Naeimi, James E. McCormick, Jr., Jonathan Dama, Ramakrishna Huggahalli, Roberto Penaranda Cebrian
  • Patent number: 11575609
    Abstract: A switch or network interface can detect congestion caused by a flow of packets. The switch or network interface can generate a congestion hint packet and send the congestion hint packet directly to a source transmitter of the flow of packets that caused the congestion. The congestion hint packet can include information that the source transmitter can use to determine a remedial action to attempt to alleviate or stop congestion at the switch or network interface. For example, the transmitter can reduce a transmit rate of the flow of packets and/or select another route for the flow of packets. Some or all switches or network interfaces between the source transmitter and a destination endpoint can employ flow differentiation whereby a queue is selected to accommodate for a flow's sensitivity to latency.
    Type: Grant
    Filed: July 19, 2019
    Date of Patent: February 7, 2023
    Assignee: Intel Corporation
    Inventors: Arvind Srinivasan, Ramakrishna Huggahalli, Parthasarathy Sarangam, Sunil Ahluwalia, Mrittika Ganguli, Malek Musleh
  • Patent number: 11379342
    Abstract: There is disclosed in one example a computing apparatus, including: a processor; a multilevel cache including a plurality of cache levels; a peripheral device configured to write data directly to a selected cache level; and a cache monitoring circuit, including a cache counter to track cache lines evicted from the selected cache level without being processed; and logic to provide a direct write policy according to the cache counter.
    Type: Grant
    Filed: March 23, 2020
    Date of Patent: July 5, 2022
    Assignee: Intel Corporation
    Inventors: Ren Wang, Bin Li, Andrew J. Herdrich, Tsung-Yuan C. Tai, Ramakrishna Huggahalli
  • Publication number: 20200412666
    Abstract: Examples are described herein that relate to a mesh in a switch fabric. The mesh can include one or more buses that permit operations (e.g., read, write, or responses) to continue in the same direction, drop off to a memory, drop off a bus to permit another operation to use the bus, or receive operations that are changing direction. A latency estimate can be determined at least for operations that drop off from a bus to permit another operation to use the bus or receive and channel operations that are changing direction. An operation with a highest latency estimate (e.g., time of traversing a mesh) can be permitted to use the bus, even causing another operation, that is not to change direction, to drop off the bus and re-enter later.
    Type: Application
    Filed: August 23, 2019
    Publication date: December 31, 2020
    Inventors: Karl S. PAPADANTONAKIS, Robert SOUTHWORTH, Arvind SRINIVASAN, Helia A. NAEIMI, James E. McCORMICK, JR., Jonathan DAMA, Ramakrishna HUGGAHALLI, Roberto PENARANDA CEBRIAN
  • Publication number: 20200218631
    Abstract: There is disclosed in one example a computing apparatus, including: a processor; a multilevel cache including a plurality of cache levels; a peripheral device configured to write data directly to a selected cache level; and a cache monitoring circuit, including a cache counter to track cache lines evicted from the selected cache level without being processed; and logic to provide a direct write policy according to the cache counter.
    Type: Application
    Filed: March 23, 2020
    Publication date: July 9, 2020
    Applicant: Intel Corporation
    Inventors: Ren Wang, Bin Li, Andrew J. Herdrich, Tsung-Yuan C. Tai, Ramakrishna Huggahalli
  • Patent number: 10599548
    Abstract: There is disclosed in one example a computing apparatus, including: a processor; a multilevel cache including a plurality of cache levels; a peripheral device configured to write data directly to a directly writable cache; and a cache monitoring circuit, including cache counters La to be incremented when a cache line is allocated into the directly writable cache, Lp to be incremented when a cache line is processed by the processor and deallocated from the directly writable cache, and Le to be incremented when a cache line is evicted from the directly writable cache to the memory, wherein the cache monitoring circuit is to determine a direct write policy according to the cache counters.
    Type: Grant
    Filed: June 28, 2018
    Date of Patent: March 24, 2020
    Assignee: Intel Corporation
    Inventors: Ren Wang, Bin Li, Andrew J. Herdrich, Tsung-Yuan C. Tai, Ramakrishna Huggahalli
  • Publication number: 20190386924
    Abstract: A switch or network interface can detect congestion caused by a flow of packets. The switch or network interface can generate a congestion hint packet and send the congestion hint packet directly to a source transmitter of the flow of packets that caused the congestion. The congestion hint packet can include information that the source transmitter can use to determine a remedial action to attempt to alleviate or stop congestion at the switch or network interface. For example, the transmitter can reduce a transmit rate of the flow of packets and/or select another route for the flow of packets. Some or all switches or network interfaces between the source transmitter and a destination endpoint can employ flow differentiation whereby a queue is selected to accommodate for a flow's sensitivity to latency.
    Type: Application
    Filed: July 19, 2019
    Publication date: December 19, 2019
    Inventors: Arvind SRINIVASAN, Ramakrishna HUGGAHALLI, Parthasarathy SARANGAM, Sunil AHLUWALIA, Mrittika GANGULI, Malek MUSLEH
  • Publication number: 20190042388
    Abstract: There is disclosed in one example a computing apparatus, including: a processor; a multilevel cache including a plurality of cache levels; a peripheral device configured to write data directly to a directly writable cache; and a cache monitoring circuit, including cache counters La to be incremented when a cache line is allocated into the directly writable cache, Lp to be incremented when a cache line is processed by the processor and deallocated from the directly writable cache, and Le to be incremented when a cache line is evicted from the directly writable cache to the memory, wherein the cache monitoring circuit is to determine a direct write policy according to the cache counters.
    Type: Application
    Filed: June 28, 2018
    Publication date: February 7, 2019
    Applicant: Intel Corporation
    Inventors: Ren Wang, Bin Li, Andrew J. Herdrich, Tsung-Yuan C. Tai, Ramakrishna Huggahalli
  • Patent number: 9971391
    Abstract: A method of assessing energy efficiency of a High-performance computing (HPC) system, including: selecting a plurality of HPC workloads to run on a system under test (SUT) with one or more power constraints, wherein the SUT includes a plurality of HPC nodes in the HPC system, executing the plurality of HPC workloads on the SUT, and generating a benchmark metric for the SUT based on a baseline configuration for each selected HPC workload and a plurality of measured performance per power values for each executed workload at each selected power constraint is shown.
    Type: Grant
    Filed: December 23, 2015
    Date of Patent: May 15, 2018
    Assignee: Intel Corporation
    Inventors: Devadatta Bodas, Meenakshi Arunachalam, Ilya Sharapov, Charles R. Yount, Scott B. Huck, Ramakrishna Huggahalli, Justin J. Song, Brian J. Griffith, Muralidhar Rajappa, Lingdan (Linda) Zeng
  • Publication number: 20170185132
    Abstract: A method of assessing energy efficiency of a High-performance computing (HPC) system, including: selecting a plurality of HPC workloads to run on a system under test (SUT) with one or more power constraints, wherein the SUT includes a plurality of HPC nodes in the HPC system, executing the plurality of HPC workloads on the SUT, and generating a benchmark metric for the SUT based on a baseline configuration for each selected HPC workload and a plurality of measured performance per power values for each executed workload at each selected power constraint is shown.
    Type: Application
    Filed: December 23, 2015
    Publication date: June 29, 2017
    Inventors: Devadatta Bodas, Meenakshi Arunachalam, Ilya Sharapov, Charles R. Yount, Scott B. Huck, Ramakrishna Huggahalli, Justin J. Song, Brian J. Griffith, Muralidhar Rajappa, Lingdan (Linda) Zeng
  • Patent number: 9559953
    Abstract: In one embodiment this disclosure provides a network device that includes an input port configured to link to a first device to receive a packet from the first device, wherein the received packet having a first label encoded therein, the value of the first label is specific to the link between the network device and the first device; the input port having an input port identifier, the input port identifier and the first label form an input tuple; a plurality of output ports configured to link to respective ones of a plurality of second devices, each output port having a respective output port identifier; a forwarding table that includes at least one input tuple and a corresponding set of output tuples; wherein each output tuple comprises an output port identifier and a second label, the value of the second label is specific to the link between the network device and a respective one of the second plurality of devices; and routing circuitry configured to compare the input tuple of the received packet with at l
    Type: Grant
    Filed: November 14, 2011
    Date of Patent: January 31, 2017
    Assignee: Intel Corporation
    Inventors: Radia Perlman, Ramakrishna Huggahalli
  • Publication number: 20140185618
    Abstract: In one embodiment this disclosure provides a network device that includes an input port configured to link to a first device to receive a packet from the first device, wherein the received packet having a first label encoded therein, the value of the first label is specific to the link between the network device and the first device; the input port having an input port identifier, the input port identifier and the first label form an input tuple; a plurality of output ports configured to link to respective ones of a plurality of second devices, each output port having a respective output port identifier; a forwarding table that includes at least one input tuple and a corresponding set of output tuples; wherein each output tuple comprises an output port identifier and a second label, the value of the second label is specific to the link between the network device and a respective one of the second plurality of devices; and routing circuitry configured to compare the input tuple of the received packet with at l
    Type: Application
    Filed: November 14, 2011
    Publication date: July 3, 2014
    Inventors: Radia Perlman, Ramakrishna Huggahalli
  • Patent number: 8271748
    Abstract: In an embodiment, an apparatus is provided that may include circuitry to generate, at least in part, and/or receive, at least in part, at least one request to access at least one portion of data. The at least one request may indicate, at least in part, at least one subset of the at least one portion of the data that is of relatively higher importance than one or more other subsets of the at least one portion of the data that are of relatively lower importance. The at least one request may be to request, at least in part, that the at least one subset be accessed prior to the one or more other subsets are accessed. The at least one request may be comprised, at least in part, in at least one packet in accordance with a protocol that permits variable packet size.
    Type: Grant
    Filed: August 12, 2008
    Date of Patent: September 18, 2012
    Assignee: Intel Corporation
    Inventors: Steen K. Larsen, Ramakrishna Huggahalli
  • Patent number: 8001278
    Abstract: Methods and apparatus relating to network packet payload compression/decompression are described. In an embodiment, an uncompressed packet payload may be compressed before being transferred between various components of a computing system. For example, a packet payload may be compressed prior to transfer between network interface cards or controllers (NICs) and storage devices (e.g., including a main system memory and/or cache(s)), as well as between processors (or processor cores) and storage devices (e.g., including main system memory and/or caches). Other embodiments are also disclosed.
    Type: Grant
    Filed: September 28, 2007
    Date of Patent: August 16, 2011
    Assignee: Intel Corporation
    Inventors: Ramakrishna Huggahalli, Steen K. Larsen
  • Patent number: 7937534
    Abstract: Embodiments of an apparatus, method, and system for encoding direct cache access transactions based on a memory access data structure are disclosed. In one embodiment, an apparatus includes memory access logic and transaction logic. The memory access logic is to determine whether to allow a memory access based on a memory access data structure. The transaction logic is to assign direct cache access attributes to a transaction based on the memory access data structure.
    Type: Grant
    Filed: December 30, 2005
    Date of Patent: May 3, 2011
    Inventors: Rajesh Sankaran Madukkarumukumana, Sridhar Muthrasanallur, Ramakrishna Huggahalli, Rameshkumar G. Illikkal
  • Publication number: 20100042579
    Abstract: In an embodiment, an apparatus is provided that may include circuitry to generate, at least in part, and/or receive, at least in part, at least one request to access at least one portion of data. The at least one request may indicate, at least in part, at least one subset of the at least one portion of the data that is of relatively higher importance than one or more other subsets of the at least one portion of the data that are of relatively lower importance. The at least one request may be to request, at least in part, that the at least one subset be accessed prior to the one or more other subsets are accessed. The at least one request may be comprised, at least in part, in at least one packet in accordance with a protocol that permits variable packet size.
    Type: Application
    Filed: August 12, 2008
    Publication date: February 18, 2010
    Inventors: Steen K. Larsen, Ramakrishna Huggahalli
  • Publication number: 20090089454
    Abstract: Methods and apparatus relating to network packet payload compression/decompression are described. In an embodiment, an uncompressed packet payload may be compressed before being transferred between various components of a computing system. For example, a packet payload may be compressed prior to transfer between network interface cards or controllers (NICs) and storage devices (e.g., including a main system memory and/or cache(s)), as well as between processors (or processor cores) and storage devices (e.g., including main system memory and/or caches). Other embodiments are also disclosed.
    Type: Application
    Filed: September 28, 2007
    Publication date: April 2, 2009
    Inventors: Ramakrishna Huggahalli, Steen K. Larsen