Patents by Inventor Sudipta Sengupta

Sudipta Sengupta has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11475067
    Abstract: Techniques for generation of synthetic queries from customer data for training of document querying machine learning (ML) models as a service are described. A service may receive one or more documents from a user, generate a set of question and answer pairs from the one or more documents from the user using a machine learning model trained to predict a question from an answer, and store the set of question and answer pairs generated from the one or more documents from the user. The question and answer pairs may be used to train another machine learning model, for example, a document ranking model, a passage ranking model, a question/answer model, or a frequently asked question (FAQ) model.
    Type: Grant
    Filed: November 27, 2019
    Date of Patent: October 18, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Cicero Nogueira Dos Santos, Xiaofei Ma, Peng Xu, Ramesh M. Nallapati, Bing Xiang, Sudipta Sengupta, Zhiguo Wang, Patrick Ng
  • Patent number: 11467835
    Abstract: Techniques for partitioning data flow operations between execution on a compute instance and an attached accelerator instance are described. A set of operations supported by the accelerator is obtained. A set of operations associated with the data flow is obtained. An operation in the set of operations associated with the data flow is identified based on the set of operations supported by the accelerator. The accelerator executes the first operation.
    Type: Grant
    Filed: November 23, 2018
    Date of Patent: October 11, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Sudipta Sengupta, Poorna Chand Srinivas Perumalla, Jalaja Kurubarahalli, Samuel Oshin, Cory Pruce, Jun Wu, Eftiquar Shaikh, Pragya Agarwal, David Thomas, Karan Kothari, Daniel Evans, Umang Wadhwa, Mark Klunder, Rahul Sharma, Zdravko Pantic, Dominic Rajeev Divakaruni, Andrea Olgiati, Leo Dirac, Nafea Bshara, Bratin Saha, Matthew Wood, Swaminathan Sivasubramanian, Rajankumar Singh
  • Publication number: 20220308977
    Abstract: A technology landscape may be characterized using a performance characterization that includes scores assigned to performance metrics for the technology landscape and using at least one trained machine learning model. In response to a detected calibration trigger, a calibratable performance metric of the performance metrics may be determined. A relationship may be determined between conforming values of the calibratable performance metric during a conforming period for which the at least one trained machine learning model was trained, and non-conforming values of the calibratable performance metric occurring during a calibration period initiated by the calibration trigger. In this way, a score assigned to the calibratable performance metric may be calibrated, based on the relationship.
    Type: Application
    Filed: March 26, 2021
    Publication date: September 29, 2022
    Inventors: Nigel Slinger, Wenjie Zhu, Catherine Drummond, Sudipta Sengupta
  • Patent number: 11449796
    Abstract: Techniques for making machine learning inference calls for database query processing are described. In some embodiments, a method of making machine learning inference calls for database query processing may include generating a first batch of machine learning requests based at least on a query to be performed on data stored in a database service, wherein the query identifies a machine learning service, sending the first batch of machine learning requests to an input buffer of an asynchronous request handler, the asynchronous request handler to generate a second batch of machine learning requests based on the first batch of machine learning requests, and obtaining a plurality of machine learning responses from an output buffer of the asynchronous request handler, the machine learning responses generated by the machine learning service using a machine learning model in response to receiving the second batch of machine learning requests.
    Type: Grant
    Filed: September 20, 2019
    Date of Patent: September 20, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Sangil Song, Yongsik Yoon, Kamal Kant Gupta, Saileshwar Krishnamurthy, Stefano Stefani, Sudipta Sengupta, Jaeyun Noh
  • Patent number: 11422863
    Abstract: Implementations detailed herein include description of a computer-implemented method. In an implementation, the method at least includes provisioning an application instance and portions of at least one accelerator attached to the application instance to execute a machine learning model of an application of the application instance; loading the machine learning model onto the portions of the at least one accelerator; receiving scoring data in the application; and utilizing each of the portions of the attached at least one accelerator to perform inference on the scoring data in parallel and only using one response from the portions of the accelerator.
    Type: Grant
    Filed: June 27, 2018
    Date of Patent: August 23, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Sudipta Sengupta, Poorna Chand Srinivas Perumalla, Dominic Rajeev Divakaruni, Nafea Bshara, Leo Parker Dirac, Bratin Saha, Matthew James Wood, Andrea Olgiati, Swaminathan Sivasubramanian
  • Publication number: 20220237083
    Abstract: When a restart event is detected within a technology landscape, restart-impacted performance metrics and non-restart-impacted performance metrics may be identified. The non-restart-impacted performance metrics may continue to be included within a performance characterization of the technology landscape. The restart-impacted performance metrics may be monitored, while being excluded from the performance characterization. The restart-impacted performance metric of the restart-impacted performance metrics may be transitioned to a non-restart-impacted performance metric, based on a monitored value of the restart-impacted performance metric following the restart event.
    Type: Application
    Filed: January 22, 2021
    Publication date: July 28, 2022
    Inventors: Nigel Slinger, Wenjie Zhu, Catherine Drummond, Roxanne Kallman, Sudipta Sengupta, Jeremy Riegel, John Floumoy
  • Patent number: 11379437
    Abstract: Methods and systems enable a database reorganization to occur without a database outage. In one aspect, the method includes creating a shadow copy of a database, the shadow having at least one partition associated with a plurality of first data sets and the database having at least a first partition associated with a plurality of second data sets and a second partition associated with a plurality of third data sets. The method also includes reorganizing the at least one partition of the shadow, taking the first partition offline, replacing the plurality of second data sets with the plurality of first data sets in a schema, and restarting the first partition.
    Type: Grant
    Filed: May 10, 2019
    Date of Patent: July 5, 2022
    Assignee: BMC Software, Inc.
    Inventors: Bruce H. Mansur, Sudipta Sengupta, Gary L. Salazar
  • Patent number: 11366855
    Abstract: Techniques for searching documents are described. An exemplary method includes receiving a document search query; querying at least one index based upon the document search query to identify matching data; fetching the identified matched data; determining one or more of a top ranked passage and top ranked documents from the set of documents based upon one or more invocations of one or more machine learning models based at least on the fetched identified matched data and the document search query; and returning one or more of the top ranked passage and the proper subset of documents.
    Type: Grant
    Filed: November 27, 2019
    Date of Patent: June 21, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Jean-Pierre Dodel, Zhiheng Huang, Xiaofei Ma, Ramesh M. Nallapati, Krishnakumar Rajagopalan, Milan Saini, Sudipta Sengupta, Saurabh Kumar Singh, Dimitrios Soulios, Ankit Sultania, Dong Wang, Zhiguo Wang, Bing Xiang, Peng Xu, Yong Yuan
  • Patent number: 11210220
    Abstract: A data manager may include a data opaque interface configured to provide, to an arbitrarily selected page-oriented access method, interface access to page data storage that includes latch-free access to the page data storage. In another aspect, a swap operation may be initiated, of a portion of a first page in cache layer storage to a location in secondary storage, based on initiating a prepending of a partial swap delta record to a page state associated with the first page, the partial swap delta record including a main memory address indicating a storage location of a flush delta record that indicates a location in secondary storage of a missing part of the first page. In another aspect, a page manager may initiate a flush operation of a first page in cache layer storage to a location in secondary storage, based on atomic operations with flush delta records.
    Type: Grant
    Filed: December 19, 2018
    Date of Patent: December 28, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: David B. Lomet, Justin Levandoski, Sudipta Sengupta
  • Publication number: 20210304010
    Abstract: Methods and systems for training a neural network are provided. In one example, an apparatus comprises a memory that stores instructions; and a hardware processor configured to execute the instructions to: control a neural network processor to perform a loss gradient operation to generate data gradients; after the loss gradient operation completes, control the neural network processor to perform a forward propagation operation to generate intermediate outputs; control the neural network processor to perform a backward propagation operation based on the data gradients and the intermediate outputs to generate weight gradients; receive the weight gradients from the neural network processor; and update weights of a neural network based on the weight gradients.
    Type: Application
    Filed: March 31, 2020
    Publication date: September 30, 2021
    Inventors: Sudipta Sengupta, Randy Renfu Huang, Ron Diamant, Vignesh Vivekraja
  • Patent number: 11036799
    Abstract: Described is using flash memory (or other secondary storage), RAM-based data structures and mechanisms to access key-value pairs stored in the flash memory using only a low RAM space footprint. A mapping (e.g. hash) function maps key-value pairs to a slot in a RAM-based index. The slot includes a pointer that points to a bucket of records on flash memory that each had keys that mapped to the slot. The bucket of records is arranged as a linear-chained linked list, e.g., with pointers from the most-recently written record to the earliest written record. Also described are compacting non-contiguous records of a bucket onto a single flash page, and garbage collection. Still further described is load balancing to reduce variation in bucket sizes, using a bloom filter per slot to avoid unnecessary searching, and splitting a slot into sub-slots.
    Type: Grant
    Filed: February 8, 2020
    Date of Patent: June 15, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Sudipta Sengupta, Biplob Kumar Debnath, Jin Li
  • Publication number: 20210174238
    Abstract: Techniques for making machine learning inference calls for database query processing are described. In some embodiments, a method of making machine learning inference calls for database query processing may include generating a first batch of machine learning requests based at least on a query to be performed on data stored in a database service, wherein the query identifies a machine learning service, sending the first batch of machine learning requests to an input buffer of an asynchronous request handler, the asynchronous request handler to generate a second batch of machine learning requests based on the first batch of machine learning requests, and obtaining a plurality of machine learning responses from an output buffer of the asynchronous request handler, the machine learning responses generated by the machine learning service using a machine learning model in response to receiving the second batch of machine learning requests.
    Type: Application
    Filed: September 20, 2019
    Publication date: June 10, 2021
    Inventors: Sangil SONG, Yongsik YOON, Kamal Kant GUPTA, Saileshwar KRISHNAMURTHY, Stefano STEFANI, Sudipta SENGUPTA, Jaeyun NOH
  • Publication number: 20210157857
    Abstract: Techniques for generation of synthetic queries from customer data for training of document querying machine learning (ML) models as a service are described. A service may receive one or more documents from a user, generate a set of question and answer pairs from the one or more documents from the user using a machine learning model trained to predict a question from an answer, and store the set of question and answer pairs generated from the one or more documents from the user. The question and answer pairs may be used to train another machine learning model, for example, a document ranking model, a passage ranking model, a question/answer model, or a frequently asked question (FAQ) model.
    Type: Application
    Filed: November 27, 2019
    Publication date: May 27, 2021
    Inventors: Cicero NOGUEIRA DOS SANTOS, Xiaofei MA, Peng XU, Ramesh M. NALLAPATI, Bing XIANG, Sudipta SENGUPTA, Zhiguo WANG, Patrick NG
  • Publication number: 20210157845
    Abstract: Techniques for searching documents are described. An exemplary method includes receiving a document search query; querying at least one index based upon the document search query to identify matching data; fetching the identified matched data; determining one or more of a top ranked passage and top ranked documents from the set of documents based upon one or more invocations of one or more machine learning models based at least on the fetched identified matched data and the document search query; and returning one or more of the top ranked passage and the proper subset of documents.
    Type: Application
    Filed: November 27, 2019
    Publication date: May 27, 2021
    Inventors: Jean-Pierre DODEL, Zhiheng HUANG, Xiaofei MA, Ramesh M. NALLAPATI, Krishnakumar RAJAGOPALAN, Milan SAINI, Sudipta SENGUPTA, Saurabh Kumar SINGH, Dimitrios SOULIOS, Ankit SULTANIA, Dong WANG, Zhiguo WANG, Bing XIANG, Peng XU, Yong YUAN
  • Patent number: 10853129
    Abstract: Implementations detailed herein include description of a computer-implemented method to migrate a machine learning model from one accelerator portion (such as a portion of a graphical processor unit (GPU)) to a different accelerator portion. In some instances, a state of the first accelerator portion is persisted, the second accelerator portion is configured, the first accelerator portion is then detached from a client application instance, and at least a portion of an inference request is performed using the loaded at least a portion of the machine learning model on the second accelerator portion that had been configured.
    Type: Grant
    Filed: March 19, 2019
    Date of Patent: December 1, 2020
    Assignee: Amazon Technologies, Inc.
    Inventors: Sudipta Sengupta, Haifeng He, Pejus Manoj Das, Poorna Chand Srinivas Perumalla, Wei Xiao, Shirley Xue Yi Leung, Vladimir Mitrovic, Yongcong Luo, Jiacheng Guo, Stefano Stefani, Matthew Shawn Wilson
  • Publication number: 20200175070
    Abstract: Described is using flash memory (or other secondary storage), RAM-based data structures and mechanisms to access key-value pairs stored in the flash memory using only a low RAM space footprint. A mapping (e.g. hash) function maps key-value pairs to a slot in a RAM-based index. The slot includes a pointer that points to a bucket of records on flash memory that each had keys that mapped to the slot. The bucket of records is arranged as a linear-chained linked list, e.g., with pointers from the most-recently written record to the earliest written record. Also described are compacting non-contiguous records of a bucket onto a single flash page, and garbage collection. Still further described is load balancing to reduce variation in bucket sizes, using a bloom filter per slot to avoid unnecessary searching, and splitting a slot into sub-slots.
    Type: Application
    Filed: February 8, 2020
    Publication date: June 4, 2020
    Inventors: Sudipta SENGUPTA, Biplob Kumar DEBNATH, Jin LI
  • Patent number: 10558705
    Abstract: Described is using flash memory (or other secondary storage), RAM-based data structures and mechanisms to access key-value pairs stored in the flash memory using only a low RAM space footprint. A mapping (e.g. hash) function maps key-value pairs to a slot in a RAM-based index. The slot includes a pointer that points to a bucket of records on flash memory that each had keys that mapped to the slot. The bucket of records is arranged as a linear-chained linked list, e.g., with pointers from the most-recently written record to the earliest written record. Also described are compacting non-contiguous records of a bucket onto a single flash page, and garbage collection. Still further described is load balancing to reduce variation in bucket sizes, using a bloom filter per slot to avoid unnecessary searching, and splitting a slot into sub-slots.
    Type: Grant
    Filed: October 20, 2010
    Date of Patent: February 11, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Sudipta Sengupta, Biplob Kumar Debnath, Jin Li
  • Publication number: 20200004595
    Abstract: Implementations detailed herein include description of a computer-implemented method.
    Type: Application
    Filed: June 27, 2018
    Publication date: January 2, 2020
    Inventors: Sudipta SENGUPTA, Poorna Chand Srinivas PERUMALLA, Dominic Rajeev DIVAKARUNI, Nafea BSHARA, Leo Parker DIRAC, Bratin SAHA, Matthew James WOOD, Andrea OLGIATI, Swaminathan SIVASUBRAMANIAN
  • Publication number: 20200005124
    Abstract: Implementations detailed herein include description of a computer-implemented method.
    Type: Application
    Filed: June 27, 2018
    Publication date: January 2, 2020
    Inventors: Sudipta SENGUPTA, Poorna Chand Srinivas PERUMALLA, Dominic Rajeev DIVAKARUNI, Nafea BSHARA, Leo Parker DIRAC, Bratin SAHA, Matthew James WOOD, Andrea OLGIATI, Swaminathan SIVASUBRAMANIAN
  • Publication number: 20200004596
    Abstract: Implementations detailed herein include description of a computer-implemented method. In an implementation, the method at least includes receiving an application instance configuration, an application of the application instance to utilize a portion of an attached accelerator during execution of a machine learning model and the application instance configuration including: an indication of the central processing unit (CPU) capability to be used, an arithmetic precision of the machine learning model to be used, an indication of the accelerator capability to be used, a storage location of the application, and an indication of an amount of random access memory to use.
    Type: Application
    Filed: June 27, 2018
    Publication date: January 2, 2020
    Inventors: Sudipta SENGUPTA, Poorna Chand Srinivas PERUMALLA, Dominic Rajeev DIVAKARUNI, Nafea BSHARA, Leo Parker DIRAC, Bratin SAHA, Matthew James WOOD, Andrea OLGIATI, Swaminathan SIVASUBRAMANIAN