Patents by Inventor Debojyoti Dutta

Debojyoti Dutta has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 10733037
    Abstract: In one embodiment, a server in a network reports one or more symptoms of a monitored device that is malfunctioning to a user interface via a particular chatbot session. The server receives, via the particular chatbot session, a triage request to enter a triage mode regarding the one or more reported symptoms. The server predicts a corrective action using the one or more reported symptoms as input to a machine learning model. The machine learning model is trained using a history of observed symptoms in the network, a history of corrective actions initiated via chatbot sessions and associated with the observed symptoms, and a history of feedback regarding the corrective actions received via the chatbot sessions. The server provides the predicted corrective action to the user interface via the particular chatbot session as a suggested corrective action, in response to the received triage request.
    Type: Grant
    Filed: November 3, 2016
    Date of Patent: August 4, 2020
    Assignee: Cisco Technology, Inc.
    Inventors: Rahul Ramakrishna, Yathiraj B. Udupi, Debojyoti Dutta
  • Patent number: 10705882
    Abstract: Systems, methods, computer-readable media are disclosed for determining a point of delivery (POD) device or network component on a cloud for workload and resource placement in a multi-cloud environment. A method includes determining a first amount of data for transitioning from performing a first function on input data to performing a second function on a first outcome of the first function; determining a second amount of data for transitioning from performing the second function on the first outcome to performing a third function on a second outcome of the second function; determining a processing capacity for each of one or more network nodes on which the first function and the third function are implemented; and selecting the network node for implementing the second function based on the first amount of data, the second amount of data, and the processing capacity for each of the network nodes.
    Type: Grant
    Filed: December 21, 2017
    Date of Patent: July 7, 2020
    Assignee: CISCO TECHNOLOGY, INC.
    Inventors: Debojyoti Dutta, Xinyuan Huang
  • Patent number: 10691671
    Abstract: Systems, methods, and computer-readable media are provided for consistent data to be used for streaming and batch processing. The system includes one or more devices; a processor coupled to the one or more devices; and a non-volatile memory coupled to the processor and the one or more devices, wherein the non-volatile memory stores instructions that are configured to cause the processor to perform operations including receiving data from the one or more devices; validating the data to yield validated data; storing the validated data in a database on the non-volatile memory, the validated data being used for streaming processing and batch processing; and sending the validated data to a remote disk for batch processing.
    Type: Grant
    Filed: December 21, 2017
    Date of Patent: June 23, 2020
    Assignee: CISCO TECHNOLOGY, INC.
    Inventors: Johnu George, Amit Kumar Saha, Debojyoti Dutta, Madhu S. Kumar, Ralf Rantzau
  • Patent number: 10678444
    Abstract: Aspects of the technology provide improvements to a Serverless Computing (SLC) workflow by determining when and how to optimize SLC jobs for computing in a Distributed Computing Framework (DCF). DCF optimization can be performed by abstracting SLC tasks into different workflow configurations to determined optimal arrangements for execution in a DCF environment. A process of the technology can include steps for receiving an SLC job including one or more SLC tasks, executing one or more of the tasks to determine a latency metric and a throughput metric for the SLC tasks, and determining if the SLC tasks should be converted to a Distributed Computing Framework (DCF) format based on the latency metric and the throughput metric. Systems and machine-readable media are also provided.
    Type: Grant
    Filed: April 2, 2018
    Date of Patent: June 9, 2020
    Assignee: CISCO TECHNOLOGY, INC.
    Inventors: Xinyuan Huang, Johnu George, Marc Solanas Tarre, Komei Shimamura, Purushotham Kamath, Debojyoti Dutta
  • Patent number: 10671289
    Abstract: Embodiments include receiving an indication of a data storage module to be associated with a tenant of a distributed storage system, allocating a partition of a disk for data of the tenant, creating a first association between the data storage module and the disk partition, creating a second association between the data storage module and the tenant, and creating rules for the data storage module based on one or more policies configured for the tenant. Embodiments further include receiving an indication of a type of subscription model selected for the tenant, and selecting the disk partition to be allocated based, at least in part, on the subscription model selected for the tenant. More specific embodiments include generating a storage map indicating the first association between the data storage module and the disk partition and indicating the second association between the data storage module and the tenant.
    Type: Grant
    Filed: February 5, 2019
    Date of Patent: June 2, 2020
    Assignee: CISCO TECHNOLOGY, INC.
    Inventors: Johnu George, Kai Zhang, Yathiraj B. Udupi, Debojyoti Dutta
  • Patent number: 10664169
    Abstract: Approaches are disclosed for improving performance of logical disks. A logical disk can comprise several storage devices. In an object storage system (OSS), when a logical disk stores a file, fragments of the file are stored distributed across the storage devices. Each of the fragments of the file is asymmetrically stored in (write) and retrieved from (read) the storage devices. The performance of the logical disk is improved by reconfiguring one or more of the storage devices based on an influence that each of the storage devices has on performance of the logical disk and the asymmetric read and write operations of each of the storage devices. For example, latency of the logical disk can be reduced by reconfiguring one or more of the plurality of storage disks based on a proportion of the latency of the logical device that is attributable to each of the plurality of storage devices.
    Type: Grant
    Filed: June 24, 2016
    Date of Patent: May 26, 2020
    Assignee: CISCO TECHNOLOGY, INC.
    Inventors: Johnu George, Kai Zhang, Marc Solanas Tarre, Debojyoti Dutta, Manoj Sharma
  • Patent number: 10637756
    Abstract: In one embodiment, a service converts a stream of network telemetry data into sketches. The stream of network telemetry data comprises a plurality of characteristics of traffic observed in a network. The service forms a time series of the sketches. The service performs anomaly detection on the time series of the sketches in part by calculating a joint distribution of ranks and frequencies of a portion of the characteristics at different points in time of the time series. The service sends an anomaly detection alert, when an anomaly is detected from the time series of the sketches.
    Type: Grant
    Filed: November 13, 2017
    Date of Patent: April 28, 2020
    Assignee: Cisco Technology, Inc.
    Inventors: Ralf Rantzau, Xinyuan Huang, Purushotham Kamath, Debojyoti Dutta
  • Publication number: 20200089532
    Abstract: In one embodiment, a method for FPGA accelerated serverless computing comprises receiving, from a user, a definition of a serverless computing task comprising one or more functions to be executed. A task scheduler performs an initial placement of the serverless computing task to a first host determined to be a first optimal host for executing the serverless computing task. The task scheduler determines a supplemental placement of a first function to a second host determined to be a second optimal host for accelerating execution of the first function, wherein the first function is not able to accelerated by one or more FPGAs in the first host. The serverless computing task is executed on the first host and the second host according to the initial placement and the supplemental placement.
    Type: Application
    Filed: November 25, 2019
    Publication date: March 19, 2020
    Inventors: Komei Shimamura, Xinyuan Huang, Amit Kumar Saha, Debojyoti Dutta
  • Patent number: 10547524
    Abstract: In one embodiment, a server determines a trigger to diagnose a software as a service (SaaS) pipeline for a SaaS client, and sends a notification to a plurality of SaaS nodes in the pipeline that the client is in a diagnostic mode, the notification causing the plurality of SaaS nodes to establish taps to collect diagnostic information for the client. The server may then send client-specific diagnostic messages into the SaaS pipeline for the client, the client-specific diagnostic messages causing the taps on the plurality of SaaS nodes to collect client-specific diagnostic information and send the client-specific diagnostic information to the server. The server then receives the client-specific diagnostic information from the plurality of SaaS nodes, and creates a client-specific diagnostic report based on the client-specific diagnostic information.
    Type: Grant
    Filed: April 27, 2017
    Date of Patent: January 28, 2020
    Assignee: Cisco Technology, Inc.
    Inventors: Timothy Okwii, Amit Kumar Saha, Debojyoti Dutta
  • Publication number: 20200021663
    Abstract: Approaches are disclosed for distributing messages across multiple data centers where the data centers do not store messages using a same message queue protocol. In some embodiment, a network element translates messages from a message queue protocol (e.g., Kestrel, RABBITMQ, APACHE Kafka, and ACTIVEMQ) to an application layer messaging protocol (e.g., XMPP, MQTT, WebSocket protocol, or other application layer messaging protocols). In other embodiments, a network element translates messages from an application layer messaging protocol to a message queue protocol. Using the new approaches disclosed herein, data centers communicate using, at least in part, application layer messaging protocols to disconnect the message queue protocols used by the data centers and enable sharing messages between messages queues in the data centers.
    Type: Application
    Filed: September 24, 2019
    Publication date: January 16, 2020
    Inventors: Marc Solanas Tarre, Ralf Rantzau, Debojyoti Dutta, Manoj Sharma
  • Patent number: 10489195
    Abstract: In one embodiment, a method for FPGA accelerated serverless computing comprises receiving, from a user, a definition of a serverless computing task comprising one or more functions to be executed. A task scheduler performs an initial placement of the serverless computing task to a first host determined to be a first optimal host for executing the serverless computing task. The task scheduler determines a supplemental placement of a first function to a second host determined to be a second optimal host for accelerating execution of the first function, wherein the first function is not able to accelerated by one or more FPGAs in the first host. The serverless computing task is executed on the first host and the second host according to the initial placement and the supplemental placement.
    Type: Grant
    Filed: July 20, 2017
    Date of Patent: November 26, 2019
    Assignee: CISCO TECHNOLOGY, INC.
    Inventors: Komei Shimamura, Xinyuan Huang, Amit Kumar Saha, Debojyoti Dutta
  • Patent number: 10476982
    Abstract: Approaches are disclosed for distributing messages across multiple data centers where the data centers do not store messages using a same message queue protocol. In some embodiment, a network element translates messages from a message queue protocol (e.g., Kestrel, RABBITMQ, APACHE Kafka, and ACTIVEMQ) to an application layer messaging protocol (e.g., XMPP, MQTT, WebSocket protocol, or other application layer messaging protocols). In other embodiments, a network element translates messages from an application layer messaging protocol to a message queue protocol. Using the new approaches disclosed herein, data centers communicate using, at least in part, application layer messaging protocols to disconnect the message queue protocols used by the data centers and enable sharing messages between messages queues in the data centers.
    Type: Grant
    Filed: May 13, 2016
    Date of Patent: November 12, 2019
    Assignee: CISCO TECHNOLOGY, INC.
    Inventors: Marc Solanas Tarre, Ralf Rantzau, Debojyoti Dutta, Manoj Sharma
  • Publication number: 20190303018
    Abstract: Aspects of the technology provide improvements to a Serverless Computing (SLC) workflow by determining when and how to optimize SLC jobs for computing in a Distributed Computing Framework (DCF). DCF optimization can be performed by abstracting SLC tasks into different workflow configurations to determined optimal arrangements for execution in a DCF environment. A process of the technology can include steps for receiving an SLC job including one or more SLC tasks, executing one or more of the tasks to determine a latency metric and a throughput metric for the SLC tasks, and determining if the SLC tasks should be converted to a Distributed Computing Framework (DCF) format based on the latency metric and the throughput metric. Systems and machine-readable media are also provided.
    Type: Application
    Filed: April 2, 2018
    Publication date: October 3, 2019
    Inventors: Xinyuan Huang, Johnu George, Marc Solanas Tarre, Komei Shimamura, Purushotham Kamath, Debojyoti Dutta
  • Publication number: 20190294792
    Abstract: Systems, methods, computer-readable media, and devices are disclosed for creating a malware inference architecture. An instruction set is received at an endpoint in a network. At the endpoint, the instruction set is classified as potentially malicious or benign according to a first machine learning model based on a first parameter set. If the instruction set is determined by the first machine learning model to be potentially malicious, the instruction set is sent to a cloud system and is analyzed at the cloud system using a second machine learning model to determine if the instruction set comprises malicious code. The second machine learning model is configured to classify a type of security risk associated with the instruction set based on a second parameter set that is different from the first parameter set.
    Type: Application
    Filed: August 13, 2018
    Publication date: September 26, 2019
    Inventors: Abhishek Singh, Debojyoti Dutta
  • Publication number: 20190286945
    Abstract: In one embodiment, a device forms a neural network envelope cell that comprises a plurality of convolution-based filters in series or parallel. The device constructs a convolutional neural network by stacking copies of the envelope cell in series. The device trains, using a training dataset of images, the convolutional neural network to perform image classification by iteratively collecting variance metrics for each filter in each envelope cell, pruning filters with low variance metrics from the convolutional neural network, and appending a new copy of the envelope cell into the convolutional neural network.
    Type: Application
    Filed: November 1, 2018
    Publication date: September 19, 2019
    Inventors: Purushotham Kamath, Abhishek Singh, Debojyoti Dutta
  • Publication number: 20190286548
    Abstract: A controller can receive first and second metrics respectively indicating distributed computing system servers' CPU, memory, or disk utilization, throughput, or latency for a first time. The controller can receive third and fourth metrics for a second time. The controller can determine a first graph including vertices corresponding to the servers and edges indicating data flow between the servers, a second graph including edges indicating the first metrics satisfy a first threshold, a third graph including edges indicating the second metrics satisfy a second threshold, a fourth graph including edges indicating the third metrics fail to satisfy the first threshold, and a fifth graph including edges indicating the fourth metrics fail to satisfy the second threshold. The controller can display a sixth graph indicating at least one of first changes between the second graph and the fourth graph or second changes between the third graph and the fifth graph.
    Type: Application
    Filed: June 6, 2019
    Publication date: September 19, 2019
    Inventors: Mingye Chen, Xinyuan Huang, Debojyoti Dutta
  • Patent number: 10412021
    Abstract: Systems and methods are described for allocating resources in a cloud computing environment. The method includes receiving a computing request, the request for use of at least one virtual machine and a portion of memory. In response to the request, a plurality of hosts is identified and a cost function is formulated using at least a portion of those hosts. Based on the cost function, at least one host that is capable of hosting the virtual machine and memory is selected.
    Type: Grant
    Filed: August 21, 2017
    Date of Patent: September 10, 2019
    Assignee: CISCO TECHNOLOGY, INC.
    Inventors: Yathiraj B. Udupi, Debojyoti Dutta, Madhav V. Marathe, Raghunath O. Nambiar
  • Patent number: 10353800
    Abstract: Systems, methods, and computer-readable media are disclosed for graph based monitoring and management of network components of a distributed streaming system. In one aspect, a method includes generating, by a processor, a first metrics and a second metrics based on data collected on a system; generating, by the processor, a topology graph representing data flow within the system; generating, by the processor, at least one first metrics graph corresponding to the first metrics based in part on the topology graph; generating, by the processor, at least one second metrics graph corresponding to the second metrics based in part on the topology graph; identifying, by the processor, a malfunction within the system based on a change in at least one of the first metrics graph and the second metrics graph; and sending, by the processor, a feedback on the malfunction to an operational management component of the system.
    Type: Grant
    Filed: October 18, 2017
    Date of Patent: July 16, 2019
    Assignee: CISCO TECHNOLOGY, INC.
    Inventors: Mingye Chen, Xinyuan Huang, Debojyoti Dutta
  • Publication number: 20190208011
    Abstract: A method for accelerating data operations across a plurality of nodes of one or more clusters of a distributed computing environment. Rack awareness information characterizing the plurality of nodes is retrieved and a non-volatile memory (NVM) capability of each node is determined. A write operation is received at a management node of the plurality of nodes and one or more of the rack awareness information and the NVM capability of the plurality of nodes are analyzed to select one or more nodes to receive at least a portion of the write operation, wherein at least one of the selected nodes has an NVM capability. A multicast group for the write operation is then generated wherein the selected nodes are subscribers of the multicast group, and the multicast group is used to perform hardware accelerated read or write operations at one or more of the selected nodes.
    Type: Application
    Filed: December 28, 2017
    Publication date: July 4, 2019
    Inventors: Debojyoti Dutta, Amit Kumar Saha, Johnu George, Ramdoot Kumar Pydipaty, Marc Solanas Tarre
  • Publication number: 20190197146
    Abstract: Systems, methods, and computer-readable media are provided for consistent data to be used for streaming and batch processing. The system includes one or more devices; a processor coupled to the one or more devices; and a non-volatile memory coupled to the processor and the one or more devices, wherein the non-volatile memory stores instructions that are configured to cause the processor to perform operations including receiving data from the one or more devices; validating the data to yield validated data; storing the validated data in a database on the non-volatile memory, the validated data being used for streaming processing and batch processing; and sending the validated data to a remote disk for batch processing.
    Type: Application
    Filed: December 21, 2017
    Publication date: June 27, 2019
    Inventors: Johnu George, Amit Kumar Saha, Debojyoti Dutta, Madhu S. Kumar, Ralf Rantzau