Patents by Inventor Shilpi Agarwal

Shilpi Agarwal has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 10735341
    Abstract: An approach for a dynamic provisioning of multiple RSS engines is provided. In an embodiment, a method comprises monitoring a CPU usage of hardware queues implemented in a plurality of RSS pools, and determining whether a CPU usage of any hardware queue, implemented in a particular RSS pool of the plurality of RSS pools, has increased above a threshold value. In response to determining that a CPU usage of a particular hardware queue, implemented in the particular RSS pool, has increased above the threshold value, it is determined whether the particular RSS pool includes an unused hardware queue (a queue with light CPU usage). If such an unused hardware queue is presented, then an indirection table that is associated with the particular RSS pool is modified to remap one or more data flows from the particular hardware queue to the unused hardware queue.
    Type: Grant
    Filed: April 26, 2018
    Date of Patent: August 4, 2020
    Assignee: NICIRA, INC.
    Inventors: Aditya G. Holla, Rajeev Nair, Shilpi Agarwal, Subbarao Narahari, Zongyun Lai, Wenyi Jiang, Srikar Tati
  • Publication number: 20190334829
    Abstract: An approach for a dynamic provisioning of multiple RSS engines is provided. In an embodiment, a method comprises monitoring a CPU usage of hardware queues implemented in a plurality of RSS pools, and determining whether a CPU usage of any hardware queue, implemented in a particular RSS pool of the plurality of RSS pools, has increased above a threshold value. In response to determining that a CPU usage of a particular hardware queue, implemented in the particular RSS pool, has increased above the threshold value, it is determined whether the particular RSS pool includes an unused hardware queue (a queue with light CPU usage). If such an unused hardware queue is presented, then an indirection table that is associated with the particular RSS pool is modified to remap one or more data flows from the particular hardware queue to the unused hardware queue.
    Type: Application
    Filed: April 26, 2018
    Publication date: October 31, 2019
    Applicant: NICIRA, INC.
    Inventors: Aditya G. HOLLA, Rajeev NAIR, Shilpi AGARWAL, Subbarao NARAHARI, Zongyun LAI, Wenyi JIANG, Srikar TATI
  • Publication number: 20190284252
    Abstract: Engineered FGF1 and FGF2 polypeptides, polynucleotides encoding these polypeptides and DNA constructs, vectors and compositions including these engineered polypeptides are provided herein. The engineered FGF1 and FGF2 polypeptides are more stable than their wild-type counterparts and may be more effective at treating a variety of conditions that FGF1 and FGF2 are useful for treating such as wound healing.
    Type: Application
    Filed: March 18, 2019
    Publication date: September 19, 2019
    Applicant: THE BOARD OF TRUSTEES OF THE UNIVERSITY OF ARKANSAS
    Inventors: Suresh Kumar Thallapuranam, Shilpi Agarwal, Ravi Kumar Gindampati, Srinivas Jayanthi, Tengjiao Wang, Jake Jones, Olivia Kolenc, Ngoc Lam, Isabelle Niyonshuti, Kartik Balachandran, Kyle Quinn, Jingyi Chen
  • Publication number: 20190185790
    Abstract: The present disclosure relates to detergent compositions comprising xanthan lyase variants and methods for use of said compositions.
    Type: Application
    Filed: August 24, 2017
    Publication date: June 20, 2019
    Applicant: Henkel AG & Co. KGaA
    Inventors: NINA MUSSMANN, SUSANNE WIELAND, DANIELA HERBST, LARS ANDERSON, JENS ERIK NIELSEN, ALLAN SVENDSEN, RUNE NYGAARD MONRAD, RAJENDRA KULOTHUNGAN SAINATHAN, GEETHA HIREMATH MENDEZ, SOHEL DALAL, SHILPI AGARWAL, KASPER D. RAND, PERNILLE F. JENSEN
  • Publication number: 20190185841
    Abstract: The present invention relates to xanthan lyase variants and methods for obtaining xanthan lyase variants. The present invention also relates to polynucleotides encoding the variants; nucleic acid constructs, vectors, and host cells comprising the polynucleotides; and methods of using the variants.
    Type: Application
    Filed: August 24, 2017
    Publication date: June 20, 2019
    Applicant: NOVOZYMES A/S
    Inventors: Jens Erik Nielsen, Allan Svendsen, Lars Anderson, Rune Nygaard Monrad, Rajendra Kulothungan Sainathan, Pernille F. Jensen, Kasper D. Rand, Geetha Hiremath Mendez, Sohel Dalal, Shilpi Agarwal
  • Publication number: 20190140984
    Abstract: Some embodiments provide a queue management system that efficiently and dynamically manages multiple queues that process traffic to and from multiple virtual machines (VMs) executing on a host. This system manages the queues by (1) breaking up the queues into different priority pools with the higher priority pools reserved for particular types of traffic or VM (e.g., traffic for VMs that need low latency), (2) dynamically adjusting the number of queues in each pool (i.e., dynamically adjusting the size of the pools), (3) dynamically reassigning a VM to a new queue based on one or more optimization criteria (e.g., criteria relating to the underutilization or overutilization of the queue).
    Type: Application
    Filed: July 17, 2018
    Publication date: May 9, 2019
    Inventors: Shilpi Agarwal, Lenin Singaravelu, Ayyappan Veeraiyan, Chien-Chia Chen
  • Patent number: 10073711
    Abstract: A host computer has a virtualization software that supports execution of a plurality of virtual machines, where the virtualization software includes a virtual machine monitor for each of the virtual machines, and where each virtual machine monitor emulates a virtual central processing unit (CPU) for a corresponding virtual machine. A virtual machine monitor halts execution of a virtual CPU of a virtual machine by receiving a first halt instruction from a corresponding virtual machine and determining whether the virtual machine is latency sensitive. If the virtual machine is latency sensitive, then a second halt instruction is issued from the virtual machine monitor to halt a physical CPU on which the virtual CPU executes. If the virtual machine is not latency sensitive, then a system call to a kernel executing on the host computer is executed to indicate to the kernel that the virtual CPU is in an idle state.
    Type: Grant
    Filed: April 12, 2016
    Date of Patent: September 11, 2018
    Assignee: WMware, Inc.
    Inventors: Haoqiang Zheng, Lenin Singaravelu, Shilpi Agarwal, Daniel Michael Hecht, Garrett Smith
  • Patent number: 10061610
    Abstract: A host computer has one or more physical central processing units (CPUs) that support the execution of a plurality of containers, where the containers each include one or more processes. Each process of a container is assigned to execute exclusively on a corresponding physical CPU when the corresponding container is determined to be latency sensitive. The assignment of a process to execute exclusively on a corresponding physical CPU includes the migration of tasks from the corresponding physical CPU to one or more other physical CPUs of the host system, and the directing of task and interrupt processing to the one or more other physical CPUs. Tasks of of the process corresponding to the container are then executed on the corresponding physical CPU.
    Type: Grant
    Filed: May 11, 2017
    Date of Patent: August 28, 2018
    Assignee: VMware, Inc.
    Inventors: Haoqiang Zheng, Lenin Singaravelu, Shilpi Agarwal, Daniel Michael Hecht, Garrett Smith
  • Patent number: 10027605
    Abstract: Some embodiments provide a queue management system that efficiently and dynamically manages multiple queues that process traffic to and from multiple virtual machines (VMs) executing on a host. This system manages the queues by (1) breaking up the queues into different priority pools with the higher priority pools reserved for particular types of traffic or VM (e.g., traffic for VMs that need low latency), (2) dynamically adjusting the number of queues in each pool (i.e., dynamically adjusting the size of the pools), (3) dynamically reassigning a VM to a new queue based on one or more optimization criteria (e.g., criteria relating to the underutilization or overutilization of the queue).
    Type: Grant
    Filed: December 20, 2013
    Date of Patent: July 17, 2018
    Assignee: VMware, Inc.
    Inventors: Shilpi Agarwal, Lenin Singaravelu, Ayyappan Veeraiyan, Chien-Chia Chen
  • Publication number: 20180089053
    Abstract: In one embodiment, a method selects a percentage of a plurality hosts that are coupled together via a network fabric and calculates a number of workloads needed for the percentage of hosts based on a benchmark test to run. A plurality of data compute nodes are configured on one or more host pairs in the percentage of the plurality of hosts to send and receive the number of workloads through the network fabric to perform the benchmark test. A set of measurements is received for sending and receiving the workloads through the network fabric using the plurality of data compute nodes. The method increases the percentage of the plurality of hosts until the set of measurements fails a criteria or the percentage of the plurality of hosts is all of the plurality of hosts.
    Type: Application
    Filed: September 20, 2017
    Publication date: March 29, 2018
    Inventors: Chien-Chia Chen, Dongdong Deng, Jin Heo, Shilpi Agarwal
  • Patent number: 9843540
    Abstract: Some embodiments provide a queue management system that efficiently and dynamically manages multiple queues that process traffic to and from multiple virtual machines (VMs) executing on a host. This system manages the queues by (1) breaking up the queues into different priority pools with the higher priority pools reserved for particular types of traffic or VM (e.g., traffic for VMs that need low latency), (2) dynamically adjusting the number of queues in each pool (i.e., dynamically adjusting the size of the pools), (3) dynamically reassigning a VM to a new queue based on one or more optimization criteria (e.g., criteria relating to the underutilization or overutilization of the queue).
    Type: Grant
    Filed: December 20, 2013
    Date of Patent: December 12, 2017
    Assignee: VMware, Inc.
    Inventors: Shilpi Agarwal, Lenin Singaravelu, Ayyappan Veeraiyan, Chien-Chia Chen
  • Publication number: 20170308394
    Abstract: A host computer has a plurality of containers including a first container executing therein, where the host also includes a physical network interface controller (NIC). A packet handling interrupt is detected upon receipt of a first data packet associated with the first container. If the first virtual machine is latency sensitive, then the packet handling interrupt is processed. If the first virtual machine is not latency sensitive, then the first data packet is queued and and processing of the packet handling interrupt is delayed.
    Type: Application
    Filed: July 10, 2017
    Publication date: October 26, 2017
    Inventors: Haoqiang ZHENG, Lenin SINGARAVELU, Shilpi AGARWAL, Daniel Michael HECHT, Garrett SMITH
  • Publication number: 20170249186
    Abstract: A host computer has one or more physical central processing units (CPUs) that support the execution of a plurality of containers, where the containers each include one or more processes. Each process of a container is assigned to execute exclusively on a corresponding physical CPU when the corresponding container is determined to be latency sensitive. The assignment of a process to execute exclusively on a corresponding physical CPU includes the migration of tasks from the corresponding physical CPU to one or more other physical CPUs of the host system, and the directing of task and interrupt processing to the one or more other physical CPUs. Tasks of of the process corresponding to the container are then executed on the corresponding physical CPU.
    Type: Application
    Filed: May 11, 2017
    Publication date: August 31, 2017
    Inventors: Haoqiang ZHENG, Lenin SINGARAVELU, Shilpi AGARWAL, Daniel Michael HECHT, Garrett SMITH
  • Patent number: 9703589
    Abstract: A host computer has a plurality of containers including a first container executing therein, where the host also includes a physical network interface controller (NIC). A packet handling interrupt is detected upon receipt of a first data packet associated with the first container If the first virtual machine is latency sensitive, then the packet handling interrupt is processed. If the first virtual machine is not latency sensitive, then the first data packet is queued and processing of the packet handling interrupt is delayed.
    Type: Grant
    Filed: August 25, 2014
    Date of Patent: July 11, 2017
    Assignee: VMware, Inc.
    Inventors: Haoqiang Zheng, Lenin Singaravelu, Shilpi Agarwal, Daniel Michael Hecht, Garrett Smith
  • Patent number: 9652280
    Abstract: A host computer has one or more physical central processing units (CPUs) that support the execution of a plurality of containers, where the containers each include one or more processes. Each process of a container is assigned to execute exclusively on a corresponding physical CPU when the corresponding container is determined to be latency sensitive. The assignment of a process to execute exclusively on a corresponding physical CPU includes the migration of tasks from the corresponding physical CPU to one or more other physical CPUs of the host system, and the directing of task and interrupt processing to the one or more other physical CPUs. Tasks of the process corresponding to the container are then executed on the corresponding physical CPU.
    Type: Grant
    Filed: February 15, 2016
    Date of Patent: May 16, 2017
    Assignee: VMware, Inc.
    Inventors: Haoqiang Zheng, Lenin Singaravelu, Shilpi Agarwal, Daniel Michael Hecht, Garrett Smith
  • Patent number: 9571426
    Abstract: Some embodiments provide a queue management system that efficiently and dynamically manages multiple queues that process traffic to and from multiple virtual machines (VMs) executing on a host. This system manages the queues by (1) breaking up the queues into different priority pools with the higher priority pools reserved for particular types of traffic or VM (e.g., traffic for VMs that need low latency), (2) dynamically adjusting the number of queues in each pool (i.e., dynamically adjusting the size of the pools), (3) dynamically reassigning a VM to a new queue based on one or more optimization criteria (e.g., criteria relating to the underutilization or overutilization of the queue).
    Type: Grant
    Filed: December 20, 2013
    Date of Patent: February 14, 2017
    Assignee: VMware, Inc.
    Inventors: Shilpi Agarwal, Lenin Singaravelu, Ayyappan Veeraiyan, Chien-Chia Chen
  • Patent number: 9552216
    Abstract: A host computer has a plurality of virtual machines executing therein under the control of a hypervisor, where the host also includes a physical network interface controller (NIC). An interrupt controller detects an interrupt generated by the physical NIC, where the interrupt corresponds to a virtual machine. If the virtual machine has exclusive affinity to one or more physical central processing units (CPUs), then the interrupt is forwarded to the virtual machine. If the virtual machine does not have exclusive affinity, then a process in the hypervisor is invoked to forward the interrupt to the virtual machine.
    Type: Grant
    Filed: August 25, 2014
    Date of Patent: January 24, 2017
    Assignee: VMware, Inc.
    Inventors: Haoqiang Zheng, Lenin Singaravelu, Shilpi Agarwal, Daniel Michael Hecht, Garrett Smith
  • Publication number: 20160224370
    Abstract: A host computer has a virtualization software that supports execution of a plurality of virtual machines, where the virtualization software includes a virtual machine monitor for each of the virtual machines, and where each virtual machine monitor emulates a virtual central processing unit (CPU) for a corresponding virtual machine. A virtual machine monitor halts execution of a virtual CPU of a virtual machine by receiving a first halt instruction from a corresponding virtual machine and determining whether the virtual machine is latency sensitive. If the virtual machine is latency sensitive, then a second halt instruction is issued from the virtual machine monitor to halt a physical CPU on which the virtual CPU executes. If the virtual machine is not latency sensitive, then a system call to a kernel executing on the host computer is executed to indicate to the kernel that the virtual CPU is in an idle state.
    Type: Application
    Filed: April 12, 2016
    Publication date: August 4, 2016
    Inventors: Haoqiang ZHENG, Lenin SINGARAVELU, Shilpi AGARWAL, Daniel Michael HECHT, Garrett SMITH
  • Publication number: 20160162336
    Abstract: A host computer has one or more physical central processing units (CPUs) that support the execution of a plurality of containers, where the containers each include one or more processes. Each process of a container is assigned to execute exclusively on a corresponding physical CPU when the corresponding container is determined to be latency sensitive. The assignment of a process to execute exclusively on a corresponding physical CPU includes the migration of tasks from the corresponding physical CPU to one or more other physical CPUs of the host system, and the directing of task and interrupt processing to the one or more other physical CPUs. Tasks of the process corresponding to the container are then executed on the corresponding physical CPU.
    Type: Application
    Filed: February 15, 2016
    Publication date: June 9, 2016
    Inventors: Haoqiang ZHENG, Lenin SINGARAVELU, Shilpi AGARWAL, Daniel Michael HECHT, Garrett SMITH
  • Patent number: 9317318
    Abstract: A host computer has a virtualization software that supports execution of a plurality of virtual machines, where the virtualization software includes a virtual machine monitor for each of the virtual machines, and where each virtual machine monitor emulates a virtual central processing unit (CPU) for a corresponding virtual machine. A virtual machine monitor halts execution of a virtual CPU of a virtual machine by receiving a first halt instruction from a corresponding virtual machine and determining whether the virtual machine is latency sensitive. If the virtual machine is latency sensitive, then a second halt instruction is issued from the virtual machine monitor to halt a physical CPU on which the virtual CPU executes. If the virtual machine is not latency sensitive, then a system call to a kernel executing on the host computer is executed to indicate to the kernel that the virtual CPU is in an idle state.
    Type: Grant
    Filed: August 25, 2014
    Date of Patent: April 19, 2016
    Assignee: VMware, Inc.
    Inventors: Haoqiang Zheng, Lenin Singaravelu, Shilpi Agarwal, Daniel Michael Hecht, Garrett Smith