Patents by Inventor Rishi Mehta

Rishi Mehta has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 10686716
    Abstract: Some embodiments provide a method for managing multiple queues of a network interface card (NIC) of a host computer that executes a data compute node (DCN). The method defines first, second, and third subsets of the queues. The first subset of queues is associated with a first feature for processing data messages received by the NIC, the second subset of queues is associated with a second feature, and the third subset is associated with both features. The method receives a request from the DCN to process data messages addressed to the DCN using both the first and second features. The method configures the NIC to direct data messages received for the DCN to a queue that is selected from the third subset of queues.
    Type: Grant
    Filed: July 23, 2018
    Date of Patent: June 16, 2020
    Assignee: VMWARE, INC.
    Inventors: Aditya G. Holla, Rishi Mehta, Boon Ang, Rajeev Nair, Wenyi Jiang
  • Publication number: 20200028792
    Abstract: Some embodiments provide a method for managing multiple queues of a network interface card (NIC) of a host computer that executes a data compute node (DCN). The method defines first, second, and third subsets of the queues. The first subset of queues is associated with a first feature for processing data messages received by the NIC, the second subset of queues is associated with a second feature, and the third subset is associated with both features. The method receives a request from the DCN to process data messages addressed to the DCN using both the first and second features. The method configures the NIC to direct data messages received for the DCN to a queue that is selected from the third subset of queues.
    Type: Application
    Filed: July 23, 2018
    Publication date: January 23, 2020
    Inventors: Aditya G. Holla, Rishi Mehta, Boon Ang, Rajeev Nair, Wenyi Jiang
  • Patent number: 10476801
    Abstract: An approach for dynamically distributing RSS engines to virtual machines based on flow data is disclosed. A method comprises receiving first absolute counts of data packets that belong to at least one data flow. Flow load indicator values are computed based on the first absolute counts, and stored in a lookup table. A sorted table is generated by sorting entries of the lookup table. A first count of filters that can be applied on RSS engines is determined. A second count of data flows in the sorted table and having corresponding flow load indicator values exceeding a threshold value is determined. If the second count exceeds the first count, then the first count of data flows is selected from the sorted table. The first count of filters that correspond to the data flows is determined, and the first count of the filters is assigned to at least one RSS engine.
    Type: Grant
    Filed: April 27, 2018
    Date of Patent: November 12, 2019
    Assignee: NICIRA, INC.
    Inventors: Aditya G. Holla, Shrikrishna Khare, Rajeev Nair, Aditya Sonthy, Wenyi Jiang, Rishi Mehta
  • Publication number: 20190334821
    Abstract: An approach for dynamically distributing RSS engines to virtual machines based on flow data is disclosed. A method comprises receiving first absolute counts of data packets that belong to at least one data flow. Flow load indicator values are computed based on the first absolute counts, and stored in a lookup table. A sorted table is generated by sorting entries of the lookup table. A first count of filters that can be applied on RSS engines is determined. A second count of data flows in the sorted table and having corresponding flow load indicator values exceeding a threshold value is determined. If the second count exceeds the first count, then the first count of data flows is selected from the sorted table. The first count of filters that correspond to the data flows is determined, and the first count of the filters is assigned to at least one RSS engine.
    Type: Application
    Filed: April 27, 2018
    Publication date: October 31, 2019
    Applicant: NICIRA, INC.
    Inventors: Aditya G. HOLLA, Shrikrishna KHARE, Rajeev NAIR, Aditya SONTHY, Wenyi JIANG, Rishi MEHTA
  • Patent number: 10338822
    Abstract: Systems and methods described herein align various types of hypervisor threads with a non-uniform memory access (NUMA) client of a virtual machine (VM) that is driving I/O transactions from an application so that no remote memory access is required and the I/O transactions can be completed with local accesses to CPUs, caches, and the I/O devices of a same NUMA node of a hardware NUMA system. First, hypervisor of the VM detects whether the VM runs on a single or multiple NUMA nodes. If the VM runs on multiple NUMA nodes, a NUMA client on which the application is executing the I/O transactions is identified and knowledge of resource sharing between the NUMA client and its related hypervisor threads is established. Such knowledge is then utilized to schedule the NUMA client and its related hypervisor threads to the same NUMA node of the NUMA system.
    Type: Grant
    Filed: June 15, 2016
    Date of Patent: July 2, 2019
    Assignee: VMware, Inc.
    Inventors: Amitabha Banerjee, Rishi Mehta, Xiaochuan Shen, Seongbeom Kim
  • Patent number: 10341245
    Abstract: In a computer-implemented method for reducing delay of bursty data transmission in a network employing a congestion control protocol, data is accessed that is to be periodically transmitted over a network employing a congestion control protocol. The data is to be periodically transmitted with a high burst rate followed by an idle period. The congestion control protocol progressively increases a data transmission rate during a data transmission rate increase period invoked immediately following a predetermined idle period. Prior to transmitting the data, priming data is transmitted during at least a portion of the idle period until the congestion control protocol progressively increases the data transmission rate to a desired transmission rate. The data is transmitted at the desired transmission rate.
    Type: Grant
    Filed: March 24, 2014
    Date of Patent: July 2, 2019
    Assignee: VMWare, Inc.
    Inventors: Kalyan Saladi, Rishi Mehta
  • Patent number: 10313926
    Abstract: Example methods are provided for a host to perform large receive offload (LRO) processing in a virtualized computing environment. The method may comprise receiving, via a physical network interface controller (NIC), incoming packets that are destined for the virtualized computing instance, and processing the incoming packets to generate at least one processed packet using a networking service pipeline that includes a packet aggregation service and multiple networking services. The packet aggregation service may be configured to aggregate the incoming packets into an aggregated packet and enabled at a service point along the networking service pipeline based on an LRO capability of at least one of the multiple networking services to process the aggregated packet. The method may also comprise forwarding the at least one processed packet generated by the networking service pipeline to the virtualized computing instance.
    Type: Grant
    Filed: May 31, 2017
    Date of Patent: June 4, 2019
    Assignee: NICIRA, INC.
    Inventors: Rishi Mehta, Boon Ang, Guolin Yang, Wenyi Jiang, Jayant Jain
  • Patent number: 10250450
    Abstract: Some embodiments provide a method for performing a multi-point capture of packets in a network. The method identifies multiple nodes for the multi-point capture in the network. The method configures each node of the multiple nodes to capture a set of packets. The method receives multiple captured packet sets from the multiple nodes. The method analyzes the multiple captured packet sets.
    Type: Grant
    Filed: December 20, 2016
    Date of Patent: April 2, 2019
    Assignee: NICIRA, INC.
    Inventors: Neelima Balakrishnan, Ninad Ghodke, Rishi Mehta, Banit Agrawal, Ramya Bolla, Siming Li
  • Publication number: 20190065265
    Abstract: Computer system and method for characterizing throughput performance of a datacenter utilize bandwidth information of physical network interfaces in the datacenter and results of benchmark testing on throughput on a single processor core to compute a plurality of throughput constraints that define a throughput capacity region for the datacenter to improve throughput performance of the datacenter.
    Type: Application
    Filed: August 25, 2018
    Publication date: February 28, 2019
    Inventors: Dexiang WANG, Bin QIAN, Jinqiang YANG, Naga S. S. Kishore KANKIPATI, Sanal PILLAI, Sujatha SUNDARARAMAN, Ganesan CHANDRASHEKHAR, Rishi MEHTA
  • Publication number: 20180352474
    Abstract: Example methods are provided for a host to perform large receive offload (LRO) processing in a virtualized computing environment. The method may comprise receiving, via a physical network interface controller (NIC), incoming packets that are destined for the virtualized computing instance, and processing the incoming packets to generate at least one processed packet using a networking service pipeline that includes a packet aggregation service and multiple networking services. The packet aggregation service may be configured to aggregate the incoming packets into an aggregated packet and enabled at a service point along the networking service pipeline based on an LRO capability of at least one of the multiple networking services to process the aggregated packet. The method may also comprise forwarding the at least one processed packet generated by the networking service pipeline to the virtualized computing instance.
    Type: Application
    Filed: May 31, 2017
    Publication date: December 6, 2018
    Applicant: Nicira, Inc.
    Inventors: Rishi MEHTA, Boon ANG, Guolin YANG, Wenyi JIANG, Jayant JAIN
  • Patent number: 9992113
    Abstract: Techniques disclosed herein provide an approach for using receive side scaling (RSS) offloads from a physical network interface controller (PNIC) to improve the performance of a virtual network interface controller (VNIC). In one embodiment, the PNIC is configured to write hash values it computes for RSS purposes to packets themselves. The VNIC then reads the hash values from the packets and places the packets into VNIC RSS queues, which are processed by respective CPUs, based on the hash values. CPU overhead is thereby reduced, as RSS processing by the VNIC no longer requires computing hash values. In another embodiment in which the number of PNIC RSS queues and VNIC RSS queues are identical, the VNIC may map packets from PNIC RSS queues to VNIC RSS queues using the PNIC RSS queue ID numbers, which also does not require the computing RSS hash values.
    Type: Grant
    Filed: June 30, 2015
    Date of Patent: June 5, 2018
    Assignee: VMware, Inc.
    Inventors: Rishi Mehta, Lenin Singaravelu
  • Publication number: 20180006926
    Abstract: Some embodiments provide a method for presenting packets captured in a network. The method identifies a first set of packets from a first packet group of multiple captured packet groups. The method identifies a second set of packets, that corresponds to the first set of packets, from a second packet group of the multiple captured packet groups. The method displays representations of the multiple captured packet groups. At least one of the first set of packets and the second set of packets are presented with a different appearance from other packets of their respective packet group.
    Type: Application
    Filed: December 20, 2016
    Publication date: January 4, 2018
    Inventors: Neelima Balakrishnan, Ninad Ghodke, Rishi Mehta, Banit Agrawal, Ramya Bolla, Siming Li
  • Publication number: 20180006908
    Abstract: Some embodiments provide a method for performing a multi-point capture of packets in a network. The method identifies multiple nodes for the multi-point capture in the network. The method configures each node of the multiple nodes to capture a set of packets. The method receives multiple captured packet sets from the multiple nodes. The method analyzes the multiple captured packet sets.
    Type: Application
    Filed: December 20, 2016
    Publication date: January 4, 2018
    Inventors: Neelima Balakrishnan, Ninad Ghodke, Rishi Mehta, Banit Agrawal, Ramya Bolla, Siming Li
  • Publication number: 20170364279
    Abstract: Systems and methods described herein align various types of hypervisor threads with a non-uniform memory access (NUMA) client of a virtual machine (VM) that is driving I/O transactions from an application so that no remote memory access is required and the I/O transactions can be completed with local accesses to CPUs, caches, and the I/O devices of a same NUMA node of a hardware NUMA system. First, hypervisor of the VM detects whether the VM runs on a single or multiple NUMA nodes. If the VM runs on multiple NUMA nodes, a NUMA client on which the application is executing the I/O transactions is identified and knowledge of resource sharing between the NUMA client and its related hypervisor threads is established. Such knowledge is then utilized to schedule the NUMA client and its related hypervisor threads to the same NUMA node of the NUMA system.
    Type: Application
    Filed: June 15, 2016
    Publication date: December 21, 2017
    Inventors: Amitabha Banerjee, Rishi Mehta, Xiaochuan Shen, Seongbeom Kim
  • Publication number: 20170180258
    Abstract: A method for load balancing IPTV channels is described. In one embodiment of the invention, a first Provider Edge (PE) network element of a label switched network, coupled with a second PE network element over multiple member links of an aggregate link, receives IPTV packets. For each IPTV packet received, the first PE network determines layer 3 information of the IPTV packet, and generates one or more channel load balancing keys based on the layer 3 information. The PE network element generates a hash value from the channel load balancing keys and determines which one of multiple member links to transmit the IPTV packet on based on the hash value, and transmits the IPTV packet to the second PE network element on the determined member link. Other methods and apparatuses are also described.
    Type: Application
    Filed: March 9, 2017
    Publication date: June 22, 2017
    Inventors: Pritam BARUAH, Rishi MEHTA
  • Patent number: 9628390
    Abstract: A method for load balancing IPTV channels is described. In one embodiment of the invention, a first Provider Edge (PE) network element of a label switched network, coupled with a second PE network element over multiple member links of an aggregate link, receives IPTV packets. For each IPTV packet received, the first PE network determines layer 3 information of the IPTV packet, and generates one or more channel load balancing keys based on the layer 3 information. The PE network element generates a hash value from the channel load balancing keys and determines which one of multiple member links to transmit the IPTV packet on based on the hash value, and transmits the IPTV packet to the second PE network element on the determined member link. Other methods and apparatuses are also described.
    Type: Grant
    Filed: July 17, 2013
    Date of Patent: April 18, 2017
    Inventors: Pritam Baruah, Rishi Mehta
  • Publication number: 20170005931
    Abstract: Techniques disclosed herein provide an approach for using receive side scaling (RSS) offloads from a physical network interface controller (PNIC) to improve the performance of a virtual network interface controller (VNIC). In one embodiment, the PNIC is configured to write hash values it computes for RSS purposes to packets themselves. The VNIC then reads the hash values from the packets and places the packets into VNIC RSS queues, which are processed by respective CPUs, based on the hash values. CPU overhead is thereby reduced, as RSS processing by the VNIC no longer requires computing hash values. In another embodiment in which the number of PNIC RSS queues and VNIC RSS queues are identical, the VNIC may map packets from PNIC RSS queues to VNIC RSS queues using the PNIC RSS queue ID numbers, which also does not require the computing RSS hash values.
    Type: Application
    Filed: June 30, 2015
    Publication date: January 5, 2017
    Inventors: Rishi MEHTA, Lenin SINGARAVELU
  • Patent number: 9495192
    Abstract: Systems and methods for preferentially assigning virtual machines (VMs) on a particular NUMA node with network queues on the same NUMA node are described. A load balancer process on a host assigns multiple VMs to network queues. The assignment of the VMs to a network queues is performed with a bias toward assigning VMs using a particular NUMA node to network queues on the same NUMA node. A scheduler on the host assigns VMs to NUMA nodes. The scheduler is biased toward assigning VMs to the same NUMA node as the PNIC and/or the same NUMA node as a network queue assigned to the VM.
    Type: Grant
    Filed: October 31, 2014
    Date of Patent: November 15, 2016
    Assignee: VMware, Inc.
    Inventors: Rishi Mehta, Xiaochuan Shen, Amitabha Banerjee, Ayyappan Veeraiyan
  • Patent number: 9459904
    Abstract: Systems and methods for preferentially assigning virtual machines (VMs) on a particular NUMA node with network queues on the same NUMA node are described. A load balancer process on a host assigns multiple VMs to network queues. The assignment of the VMs to a network queues is performed with a bias toward assigning VMs using a particular NUMA node to network queues on the same NUMA node. A scheduler on the host assigns VMs to NUMA nodes. The scheduler is biased toward assigning VMs to the same NUMA node as the PNIC and/or the same NUMA node as a network queue assigned to the VM.
    Type: Grant
    Filed: October 31, 2014
    Date of Patent: October 4, 2016
    Assignee: VMware, Inc.
    Inventors: Rishi Mehta, Xiaochuan Shen, Amitabha Banerjee, Ayyappan Veeraiyan
  • Publication number: 20160092259
    Abstract: Systems and methods for preferentially assigning virtual machines (VMs) on a particular NUMA node with network queues on the same NUMA node are described. A load balancer process on a host assigns multiple VMs to network queues. The assignment of the VMs to a network queues is performed with a bias toward assigning VMs using a particular NUMA node to network queues on the same NUMA node. A scheduler on the host assigns VMs to NUMA nodes. The scheduler is biased toward assigning VMs to the same NUMA node as the PNIC and/or the same NUMA node as a network queue assigned to the VM.
    Type: Application
    Filed: October 31, 2014
    Publication date: March 31, 2016
    Inventors: Rishi Mehta, Xiaochuan Shen, Amitabha Banerjee, Ayyappan Veeraiyan