Patents by Inventor Rochan Sankar
Rochan Sankar has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 12271337Abstract: A system for providing memory access is disclosed. In some embodiments, the system is configured to receive at a source server fabric adapter (SFA), from a server, a memory access request comprising a virtual memory address; using associative mapping, determining whether the virtual address corresponds to a source-local memory associated with the source SFA or to a remote memory. If the virtual address corresponds to the source-local memory, the virtual memory address is translated, at the source SFA, into a physical memory address of the source-local memory. If the virtual address corresponds to the remote memory, a request message is synthesized, and the synthesized request message is transmitted to the destination SFA using a network protocol.Type: GrantFiled: June 9, 2022Date of Patent: April 8, 2025Assignee: Enfabrica CorporationInventors: Thomas Norrie, Shrijeet Mukherjee, Rochan Sankar
-
Patent number: 12255826Abstract: An adaptive generic receive offload (A-GRO) system and method are disclosed. In some embodiments, the system comprises a host including a host protocol stack and a host memory, and a network interface card that is communicatively connectable to the host. The A-GRO system is configured to: receive a packet from a network, parse the packet to a header and a payload, classify and map the packet into a particular flow based on contexts associated with a plurality of flows and the header, and move the header and the payload to separate queues associated with the particular flow in the host memory, without holding and stalling the packet in hardware of the NIC. By maintain packet coherence information including header chains, the A-GRO allows the host to skip processing the packets between the first and last headers in a GRO aggregation. The A-GRO system also improves mis-ordering packet handling.Type: GrantFiled: January 12, 2023Date of Patent: March 18, 2025Assignee: Enfabrica CorporationInventors: Shrijeet Mukherjee, Carlo Contavalli, Shimon Muller, Ariel Hendel, Gurjeet Singh, Rochan Sankar
-
Patent number: 12244494Abstract: A server fabric adapter (SFA) communication system is disclosed. In some embodiments, the SFA communication system comprises an SFA communicatively coupled to a plurality of controlling hosts, a plurality of endpoints, and a plurality of network ports. The SFA is configured to receive a network packet from a network port of the plurality of network ports; separate the network packet into different portions, each portion including a header or a payload; map each portion of the network packet to: (i) a controlling host of the plurality controlling hosts, the controlling host being designated as a destination controlling host, or (ii) an endpoint of the plurality of endpoints, the endpoint being designated as a destination endpoint; and forward a respective portion of the network packet to the destination controlling host or the destination endpoint.Type: GrantFiled: July 19, 2024Date of Patent: March 4, 2025Assignee: Enfabrica CorporationInventors: Rochan Sankar, Shrijeet Mukherjee, Ariel Hendel, Carlo Contavalli, Shimon Muller
-
Publication number: 20250047621Abstract: In an aspect, a system for an optimally balanced networked system is disclosed. The system includes a fabric adapter communication system communicatively coupled to a plurality of network ports and a plurality of controlling hosts. The fabric adapter communication system is configured to receive a network packet from, or transmit a network packet to, a network port of the plurality of network ports. The fabric adapter communication system is configured to separate the network packet into different portions, each portion including a header or a payload. The fabric adapter communication system is configured to forward the headers of the different portions to one or more controlling hosts. The fabric adapter communication system is configured to forward multiple payloads of the different portions in parallel through a bundled interface to multiple memory buffers of a global memory pool based on one or more scatter gather lists (SGLs).Type: ApplicationFiled: June 18, 2024Publication date: February 6, 2025Inventors: Rochan Sankar, Shrijeet Mukherjee, Ariel Hendel, Carlo Contavalli, Shimon Muller
-
Publication number: 20240372805Abstract: A server fabric adapter (SFA) communication system is disclosed. In some embodiments, the SFA communication system comprises an SFA communicatively coupled to a plurality of controlling hosts, a plurality of endpoints, and a plurality of network ports. The SFA is configured to receive a network packet from a network port of the plurality of network ports; separate the network packet into different portions, each portion including a header or a payload; map each portion of the network packet to: (i) a controlling host of the plurality controlling hosts, the controlling host being designated as a destination controlling host, or (ii) an endpoint of the plurality of endpoints, the endpoint being designated as a destination endpoint; and forward a respective portion of the network packet to the destination controlling host or the destination endpoint.Type: ApplicationFiled: July 19, 2024Publication date: November 7, 2024Inventors: Rochan Sankar, Shrijeet Mukherjee, Ariel Hendel, Carlo Contavalli, Shimon Muller
-
Publication number: 20240345989Abstract: A system for providing memory access is disclosed. In some embodiments, the system is configured to receive at a source server fabric adapter (SFA), from a server, a memory access request comprising a virtual memory address; using associative mapping, determining whether the virtual address corresponds to a source-local memory associated with the source SFA or to a remote memory. If the virtual address corresponds to the source-local memory, the virtual memory address is translated, at the source SFA, into a physical memory address of the source-local memory. If the virtual address corresponds to the remote memory, a request message is synthesized, and the synthesized request message is transmitted to the destination SFA using a network protocol.Type: ApplicationFiled: June 26, 2024Publication date: October 17, 2024Inventors: Thomas Norrie, Shrijeet Mukherjee, Rochan Sankar
-
Patent number: 12120021Abstract: A server fabric adapter (SFA) communication system is disclosed. In some embodiments, the SFA communication system comprises an SFA communicatively coupled to a plurality of controlling hosts, a plurality of endpoints, and a plurality of network ports. The SFA is configured to receive a network packet from a network port of the plurality of network ports; separate the network packet into different portions, each portion including a header or a payload; map each portion of the network packet to: (i) a controlling host of the plurality controlling hosts, the controlling host being designated as a destination controlling host, or (ii) an endpoint of the plurality of endpoints, the endpoint being designated as a destination endpoint; and forward a respective portion of the network packet to the destination controlling host or the destination endpoint.Type: GrantFiled: January 6, 2022Date of Patent: October 15, 2024Assignee: Enfabrica CorporationInventors: Rochan Sankar, Shrijeet Mukherjee, Ariel Hendel, Carlo Contavalli, Shimon Muller
-
Publication number: 20240330221Abstract: A multi-plane, multi-protocol memory switch system is disclosed. In some embodiments, a memory switch includes a plurality of switch ports, the memory switch connectable to one or more root complex (RC) devices through one or more respective switch ports of the plurality of switch ports, and the memory switch connectable to a set of endpoints through a set of other switch ports of the plurality of switch ports, wherein the set includes zero or multiple endpoints; a cacheline exchange engine configured to provide a data-exchange path between two endpoints and to map an address space of one endpoint to an address space of another endpoint; and a bulk data transfer engine configured to facilitate data-exchange between two endpoints as a source-destination data stream, one endpoint being designated a source address and another endpoint being designated a destination address.Type: ApplicationFiled: June 13, 2024Publication date: October 3, 2024Inventors: Thomas Norrie, Shrijeet Mukherjee, John Greth, Rochan Sankar, Shimon Muller, Ariel Hendel, Gurjeet Singh
-
Publication number: 20240264964Abstract: A multi-plane, multi-protocol memory switch system is disclosed. In some embodiments, a memory switch includes a plurality of switch ports, the memory switch connectable to one or more root complex (RC) devices through one or more respective switch ports of the plurality of switch ports, and the memory switch connectable to a set of endpoints through a set of other switch ports of the plurality of switch ports, wherein the set includes zero or multiple endpoints; a cacheline exchange engine configured to provide a data-exchange path between two endpoints and to map an address space of one endpoint to an address space of another endpoint; and a bulk data transfer engine configured to facilitate data-exchange between two endpoints as a source-destination data stream, one endpoint being designated a source address and another endpoint being designated a destination address.Type: ApplicationFiled: April 18, 2024Publication date: August 8, 2024Inventors: Thomas Norrie, Shrijeet Mukherjee, John Greth, Rochan Sankar, Shimon Muller, Ariel Hendel, Gurjeet Singh
-
Publication number: 20240244005Abstract: An adaptive generic receive offload (A-GRO) system and method are disclosed. In some embodiments, the system comprises a host including a host protocol stack and a host memory, and a network interface card that is communicatively connectable to the host. The A-GRO system is configured to: receive a packet from a network, parse the packet to a header and a payload, classify and map the packet into a particular flow based on contexts associated with a plurality of flows and the header, and move the header and the payload to separate queues associated with the particular flow in the host memory, without holding and stalling the packet in hardware of the NIC. By maintain packet coherence information including header chains, the A-GRO allows the host to skip processing the packets between the first and last headers in a GRO aggregation. The A-GRO system also improves mis-ordering packet handling.Type: ApplicationFiled: January 12, 2023Publication date: July 18, 2024Inventors: Shrijeet Mukherjee, Carlo Contavalli, Shimon Muller, Ariel Hendel, Gurjeet Singh, Rochan Sankar
-
Publication number: 20240184732Abstract: A modular interconnection system is disclosed. In some embodiments, the modular interconnection system comprising a server fabric adapter (SFA) on a primary circuit board, the SFA configured to perform peripheral component interconnect express (PCIe) interconnection or compute express link (CXL) interconnection; a plurality of ports on one or more PCIe slots configured to connect the SFA to external resources; and a PCIe slot adaptation device configured to adapt a first lane count slot of the one or more PCIe slots to support a second lane count device.Type: ApplicationFiled: December 1, 2023Publication date: June 6, 2024Inventors: David Skirmont, Rochan Sankar, Shrijeet Mukherjee
-
Patent number: 11995017Abstract: A multi-plane, multi-protocol memory switch system is disclosed. In some embodiments, a memory switch includes a plurality of switch ports, the memory switch connectable to one or more root complex (RC) devices through one or more respective switch ports of the plurality of switch ports, and the memory switch connectable to a set of endpoints through a set of other switch ports of the plurality of switch ports, wherein the set includes zero or multiple endpoints; a cacheline exchange engine configured to provide a data-exchange path between two endpoints and to map an address space of one endpoint to an address space of another endpoint; and a bulk data transfer engine configured to facilitate data-exchange between two endpoints as a source-destination data stream, one endpoint being designated a source address and another endpoint being designated a destination address.Type: GrantFiled: June 8, 2022Date of Patent: May 28, 2024Assignee: Enfabrica CorporationInventors: Thomas Norrie, Shrijeet Mukherjee, John Greth, Rochan Sankar, Shimon Muller, Ariel Hendel, Gurjeet Singh
-
Publication number: 20230059755Abstract: A system for congestion control using a flow level transmit mechanism is disclosed. In some embodiments, the system comprises a source SFA and a receive SFA. The source SFA is configured to detect and classify a congestion notification packet (CNP) generated based on congestion in a network; select a receive block from a plurality of receive blocks based on the CNP; forward the CNP to a dedicated congestion notification queue of the receive block; identify a transmit queue from a plurality of transmit blocks based on processing the congestion notification queue, wherein the transmit queue originated a particular transmit flow causing the congestion; and stop the transmit queue.Type: ApplicationFiled: August 11, 2022Publication date: February 23, 2023Inventors: Shrijeet Mukherjee, Shimon Muller, Carlo Contavalli, Gurjeet Singh, Ariel Hendel, Rochan Sankar
-
Publication number: 20220398207Abstract: A multi-plane, multi-protocol memory switch system is disclosed. In some embodiments, a memory switch includes a plurality of switch ports, the memory switch connectable to one or more root complex (RC) devices through one or more respective switch ports of the plurality of switch ports, and the memory switch connectable to a set of endpoints through a set of other switch ports of the plurality of switch ports, wherein the set includes zero or multiple endpoints; a cacheline exchange engine configured to provide a data-exchange path between two endpoints and to map an address space of one endpoint to an address space of another endpoint; and a bulk data transfer engine configured to facilitate data-exchange between two endpoints as a source-destination data stream, one endpoint being designated a source address and another endpoint being designated a destination address.Type: ApplicationFiled: June 8, 2022Publication date: December 15, 2022Inventors: Thomas Norrie, Shrijeet Mukherjee, John Greth, Rochan Sankar, Shimon Muller, Ariel Hendel, Gurjeet Singh
-
Publication number: 20220398215Abstract: A system for providing memory access is disclosed. In some embodiments, the system is configured to receive at a source server fabric adapter (SFA), from a server, a memory access request comprising a virtual memory address; using associative mapping, determining whether the virtual address corresponds to a source-local memory associated with the source SFA or to a remote memory. If the virtual address corresponds to the source-local memory, the virtual memory address is translated, at the source SFA, into a physical memory address of the source-local memory. If the virtual address corresponds to the remote memory, a request message is synthesized, and the synthesized request message is transmitted to the destination SFA using a network protocol.Type: ApplicationFiled: June 9, 2022Publication date: December 15, 2022Inventors: Thomas Norrie, Shrijeet Mukherjee, Rochan Sankar
-
Publication number: 20220217085Abstract: A server fabric adapter (SFA) communication system is disclosed. In some embodiments, the SFA communication system comprises an SFA communicatively coupled to a plurality of controlling hosts, a plurality of endpoints, and a plurality of network ports. The SFA is configured to receive a network packet from a network port of the plurality of network ports; separate the network packet into different portions, each portion including a header or a payload; map each portion of the network packet to: (i) a controlling host of the plurality controlling hosts, the controlling host being designated as a destination controlling host, or (ii) an endpoint of the plurality of endpoints, the endpoint being designated as a destination endpoint; and forward a respective portion of the network packet to the destination controlling host or the destination endpoint.Type: ApplicationFiled: January 6, 2022Publication date: July 7, 2022Inventors: Rochan Sankar, Shrijeet Mukherjee, Ariel Hendel, Carlo Contavalli, Shimon Muller
-
Patent number: 10164796Abstract: A network switch for network communications includes an embedded programmable state machine to monitor data flows through the switch. The programmable state machine is configured to retain selectable states of selectable data packet fields. Programmable switch logic operative with the programmable state machine is configured to output one or more potential actions to be taken based on a selectable computation of detected selectable states. The programmable state machine can be implemented with either table lookups or flexible logic.Type: GrantFiled: May 16, 2016Date of Patent: December 25, 2018Assignee: AVAGO TECHNOLOGIES INTERNATIONAL SALES PTE. LIMITEDInventors: Mohan Kalkunte, Surendra Anubolu, Rochan Sankar
-
Publication number: 20170302477Abstract: A network switch for network communications includes an embedded programmable state machine to monitor data flows through the switch. The programmable state machine is configured to retain selectable states of selectable data packet fields. Programmable switch logic operative with the programmable state machine is configured to output one or more potential actions to be taken based on a selectable computation of detected selectable states. The programmable state machine can be implemented with either table lookups or flexible logic.Type: ApplicationFiled: May 16, 2016Publication date: October 19, 2017Applicant: BROADCOM CORPORATIONInventors: Mohan Kalkunte, Surendra Anubolu, Rochan Sankar
-
Patent number: 7254748Abstract: A CAM and method for operating a CAM are presented. Copies of a CAM database are duplicated and placed in a first set of CAM locations and a second set of CAM locations. An error detector is used to determine false matches in the case of soft errors within the entries producing those false matches. While the entries producing a match should have the same index location, errors might cause those match lines to have an offset. If so, the present CAM, through use of duplicative sets of CAM locations, will detect the offset and thereafter the values in each index location that produces a match, along with the corresponding parity or error detection encoding bit(s). If the parity or error detection encoding bit(s) indicate an error in a particular entry, then that error is located and the corresponding entry at the same index within the other, duplicative set of CAM locations is copied into the that erroneous entry.Type: GrantFiled: October 14, 2003Date of Patent: August 7, 2007Assignee: NetLogic Microsystems, Inc.Inventors: Andrew J. Wright, Eric H. Voelkel, Srinivasan Venkatachary, Rochan Sankar
-
Patent number: 7003545Abstract: A method for computing a sum or difference and a carry-out of numbers in product-term based programmable logic comprising the steps of: (A) generating (i) a portion of the sum or difference and (ii) a lookahead carry output in each of a plurality of logic blocks; (B) communicating the lookahead carry output of each of the logic blocks to a carry input of a next logic block; (C) presenting the lookahead carry output of a last logic block as the carry-out.Type: GrantFiled: September 11, 2001Date of Patent: February 21, 2006Assignee: Cypress Semiconductor Corp.Inventors: Haneef D. Mohammed, Rochan Sankar