Patents by Inventor Lalit Chhabra

Lalit Chhabra has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11750522
    Abstract: Systems and methods of communicating in a network use rate limiting. Rate limiting units (either receive side or transmit side) can perform rate limiting in response to a) a maximum number of bytes that can be solicited over a first period of time is exceeded, b) a maximum number of bytes that are outstanding over a second period of time is exceeded; or c) a maximum number of commands that are outstanding over a period of time is exceeded as part of CMD_RXRL. The CMD_RXRL can have three components (a) max bytes, b) outstanding bytes, c) outstanding commands. TXRL contains the component of max bytes or maximum number of bytes that can be transmitted over a third period of time to match the speed of a receive link, or any node or link through the network/fabric.
    Type: Grant
    Filed: April 19, 2021
    Date of Patent: September 5, 2023
    Assignee: Avago Technologies International Sales Pte. Limited
    Inventors: Kenny Wu, James Winston Smart, Mark Karnowski, Ravi Shenoy, Gregorio Gervasio, Jr., Lalit Chhabra, Chakradhara Raj Yadav Aradhyula
  • Publication number: 20220334985
    Abstract: Systems and methods relate to a bus adapter for a storage network. The bus adaptor includes a context memory comprising a first storage for uncacheable exchange resource indicators (XRI) and a second storage for cacheable XRI. The bus adapter also includes a host backing store unit configured to provide access to the different tier memories present locally or externally in the host memory extension using several caching sub-units and with the capability of an optional pinning operation for the cacheable XRI based upon at least one of input/output phase, first in line up to a limit, a region of a virtual context address associated with the cacheable XRI indicators, a protocol associated with the cacheable XRI, a size of a transaction, or work queue information.
    Type: Application
    Filed: April 16, 2021
    Publication date: October 20, 2022
    Inventors: Marc Pegolotti, Kenny Wu, Ravi Shenoy, Gregorio Gervasio, JR., Lalit Chhabra, Mark Karnowski, James Winston Smart, Vuong Cao Nguyen
  • Publication number: 20220337524
    Abstract: Systems and methods of communicating in a network use rate limiting. Rate limiting units (either receive side or transmit side) can perform rate limiting in response to a) a maximum number of bytes that can be solicited over a first period of time is exceeded, b) a maximum number of bytes that are outstanding over a second period of time is exceeded; or c) a maximum number of commands that are outstanding over a period of time is exceeded as part of CMD_RXRL. The CMD_RXRL, can have three components (a) max bytes, b) outstanding bytes, c) outstanding commands. TXRL, contains the component of max bytes or maximum number of bytes that can be transmitted over a third period of time to match the speed of a receive link, or any node or link through the network/fabric.
    Type: Application
    Filed: April 19, 2021
    Publication date: October 20, 2022
    Inventors: Kenny Wu, James Winston Smart, Mark Karnowski, Ravi Shenoy, Gregorio Gervasio, JR., Lalit Chhabra, Chakradhara Raj Yadav Aradhyula
  • Patent number: 10996980
    Abstract: A number of command processing devices, architectures, and methods are described. One example of a command processing device is disclosed to include a classification engine configured to classify input commands, a sequencer in communication with the classification engine, one or more thread managers in communication with the sequencer, and one or more sub-processing engines in communication with each of the one or more thread managers. The sequencer may control staging of work across multiple threads and processing elements within threads. Each of the one or more thread managers are configured to delegate work to different sub-processing engines. Each of the one or more sub-processing engines are configured to perform sub-tasks in connection with completing processing of an input command received at the classification engine based on particular sub-tasks assigned to the one or more sub-processing engines by the one or more thread managers.
    Type: Grant
    Filed: April 23, 2018
    Date of Patent: May 4, 2021
    Assignee: AVAGO TECHNOLOGIES INTERNATIONAL SALES PTE. LIMITED
    Inventors: Lalit Chhabra, Gregorio Gervasio, Jr., Kenny Wu, Mark Karnowski
  • Patent number: 10735340
    Abstract: A networking adaptor and method of transferring data are depicted and described herein. One example of the networking adaptor is provided with a host interface and a network interface. The network interface may include a transmit portion and a receive portion. The transmit portion may include a first set of data paths and the receive portion may include a second set of data paths. Both the first set of data paths and second set of data paths are configurable to be aggregated or de-aggregated to support a single port operation that represents a combined bandwidth of the data paths in the first set of data paths or the second set of data paths.
    Type: Grant
    Filed: April 18, 2018
    Date of Patent: August 4, 2020
    Assignee: Avago Technologies International Sales Pte. Limited
    Inventors: Kenny Wu, Gregorio Gervasio, Jr., Lalit Chhabra, Ravi Shenoy
  • Patent number: 10664420
    Abstract: A system, method, and adaptor that facilitate data transmission are described. One example of the disclosed system facilitates the chip-to-chip transport of header descriptors and payloads. The system may include a source chip, a destination chip, and a set of queues describing buffer memory locations for staging header descriptors and payloads to be transferred from the source chip to the destination chip, where the set of queues are directly accessible to the source chip and to the destination chip.
    Type: Grant
    Filed: April 19, 2018
    Date of Patent: May 26, 2020
    Assignee: Avago Technologies International Sales Pte. Limited
    Inventors: Kenny Wu, Mark Karnowski, James Smart, Ravi Shenoy, Lalit Chhabra, Gregorio Gervasio, Jr., Tuong Le, Vuong Nguyen
  • Publication number: 20190327178
    Abstract: A networking adaptor and method of transferring data are depicted and described herein. One example of the networking adaptor is provided with a host interface and a network interface. The network interface may include a transmit portion and a receive portion. The transmit portion may include a first set of data paths and the receive portion may include a second set of data paths. Both the first set of data paths and second set of data paths are configurable to be aggregated or de-aggregated to support a single port operation that represents a combined bandwidth of the data paths in the first set of data paths or the second set of data paths.
    Type: Application
    Filed: April 18, 2018
    Publication date: October 24, 2019
    Inventors: Kenny Wu, Gregorio Gervasio, JR., Lalit Chhabra, Ravi Shenoy
  • Publication number: 20190324926
    Abstract: A system, method, and adaptor that facilitate data transmission are described. One example of the disclosed system facilitates the chip-to-chip transport of header descriptors and payloads. The system may include a source chip, a destination chip, and a set of queues describing buffer memory locations for staging header descriptors and payloads to be transferred from the source chip to the destination chip, where the set of queues are directly accessible to the source chip and to the destination chip.
    Type: Application
    Filed: April 19, 2018
    Publication date: October 24, 2019
    Inventors: Kenny Wu, Mark Karnowski, James Smart, Ravi Shenoy, Lalit Chhabra, Gregorio Gervasio, JR., Tuong Le, Vuong Nguyen
  • Publication number: 20190324800
    Abstract: A number of command processing devices, architectures, and methods are described. One example of a command processing device is disclosed to include a classification engine configured to classify input commands, a sequencer in communication with the classification engine, one or more thread managers in communication with the sequencer, and one or more sub-processing engines in communication with each of the one or more thread managers. The sequencer may control staging of work across multiple threads and processing elements within threads. Each of the one or more thread managers are configured to delegate work to different sub-processing engines. Each of the one or more sub-processing engines are configured to perform sub-tasks in connection with completing processing of an input command received at the classification engine based on particular sub-tasks assigned to the one or more sub-processing engines by the one or more thread managers.
    Type: Application
    Filed: April 23, 2018
    Publication date: October 24, 2019
    Inventors: Lalit Chhabra, Gregorio Gervasio, JR., Kenny Wu, Mark Karnowski