Patents by Inventor Nitish Shirish Keskar

Nitish Shirish Keskar has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11928600
    Abstract: A method for sequence-to-sequence prediction using a neural network model includes generating an encoded representation based on an input sequence using an encoder of the neural network model and predicting an output sequence based on the encoded representation using a decoder of the neural network model. The neural network model includes a plurality of model parameters learned according to a machine learning process. At least one of the encoder or the decoder includes a branched attention layer. Each branch of the branched attention layer includes an interdependent scaling node configured to scale an intermediate representation of the branch by a learned scaling parameter. The learned scaling parameter depends on one or more other learned scaling parameters of one or more other interdependent scaling nodes of one or more other branches of the branched attention layer.
    Type: Grant
    Filed: January 30, 2018
    Date of Patent: March 12, 2024
    Assignee: Salesforce, Inc.
    Inventors: Nitish Shirish Keskar, Karim Ahmed, Richard Socher
  • Patent number: 11829727
    Abstract: Approaches for cross-lingual regularization for multilingual generalization include a method for training a natural language processing (NLP) deep learning module. The method includes accessing a first dataset having a first training data entry, the first training data entry including one or more natural language input text strings in a first language; translating at least one of the one or more natural language input text strings of the first training data entry from the first language to a second language; creating a second training data entry by starting with the first training data entry and substituting the at least one of the natural language input text strings in the first language with the translation of the at least one of the natural language input text strings in the second language; adding the second training data entry to a second dataset; and training the deep learning module using the second dataset.
    Type: Grant
    Filed: April 23, 2021
    Date of Patent: November 28, 2023
    Assignee: salesforce.com, inc.
    Inventors: Jasdeep Singh, Nitish Shirish Keskar, Bryan McCann
  • Patent number: 11829721
    Abstract: Embodiments described herein provide dynamic blocking, a decoding algorithm which enables large-scale pretrained language models to generate high-quality paraphrases in an un-supervised setting. Specifically, in order to obtain an alternative surface form, when the language model emits a token that is present in the source sequence, the language model is prevented from generating the next token that is the same as the subsequent source token in the source sequence at the next time step. In this way, the language model is forced to generate a paraphrased sequence of the input source sequence, but with mostly different wording.
    Type: Grant
    Filed: January 28, 2021
    Date of Patent: November 28, 2023
    Assignee: salesforce.com, inc.
    Inventors: Tong Niu, Semih Yavuz, Yingbo Zhou, Nitish Shirish Keskar, Huan Wang, Caiming Xiong
  • Patent number: 11810298
    Abstract: An analytics system uses one or more machine-learned models to predict a hormone receptor status from a H&E stain image. The system partitions H&E stain images each into a plurality of non-overlapping image tiles. Bags of tiles are created through sampling of the image tiles. For each H&E stain image, the system generates a feature vector from a bag of tiles sampled from the partitioned image tiles. The analytics system trains one or more machine-learned models with training H&E stain images having a positive or negative receptor status. With the trained models, the analytics system predicts a hormone receptor status by applying a prediction model to the feature vector for a test H&E stain image.
    Type: Grant
    Filed: October 21, 2022
    Date of Patent: November 7, 2023
    Assignee: Salesforce, Inc.
    Inventors: Nikhil Naik, Ali Madani, Nitish Shirish Keskar
  • Patent number: 11769013
    Abstract: A multi-tenant system performs custom configuration of a tenant-specific chatbot to process and act upon natural language requests. The multi-tenant system configures the tenant-specific chatbots without requiring tenant-specific training. The multi-tenant system providing a user interface for configuring a tenant-specific set of permitted actions. The multi-tenant system determines a set of example phrases for each of the selected permitted actions. The multi-tenant system receives a natural language request from a user and identifies the action that the user wants to perform. The multi-tenant system uses a neural network to compare the natural language request with example phrases to identify an example phrase that matches the natural language request. The multi-tenant system performs the action corresponding to the matching example phrase.
    Type: Grant
    Filed: November 11, 2019
    Date of Patent: September 26, 2023
    Assignee: Salesforce, Inc.
    Inventors: Michael Machado, James Douglas Harrison, Caiming Xiong, Xinyi Yang, Thomas Archie Cook, Roojuta Lalani, Jean-Marc Soumet, Karl Ryszard Skucha, Juan Rodriguez, Manju Vijayakumar, Vishal Motwani, Tian Xie, Bryan McCann, Nitish Shirish Keskar, Zhihao Zou, Chitra Gulabrani, Minal Khodani, Adarsha Badarinath, Rohiniben Thakar, Srikanth Kollu, Kevin Schoen, Qiong Liu, Amit Hetawal, Kevin Zhang, Kevin Zhang, Johnson Liu, Rafael Amsili
  • Publication number: 20230252345
    Abstract: Embodiments described herein provide methods and systems for training a sequential recommendation model. A system receives a plurality of user behavior sequences, and encodes those sequences into a plurality of user interest representations. The system predicts a next item using a sequential recommendation model, producing a probability distribution over a set of items. The next interacted item in a sequence is selected as a positive sample, and a negative sample is selected based on the generated probability distribution. The positive and negative samples are used to compute a contrastive loss and update the sequential recommendation model.
    Type: Application
    Filed: May 27, 2022
    Publication date: August 10, 2023
    Inventors: Yongjun Chen, Jia LI, Nitish Shirish Keskar, Caiming Xiong
  • Patent number: 11657233
    Abstract: Systems and methods for unifying question answering and text classification via span extraction include a preprocessor for preparing a source text and an auxiliary text based on a task type of a natural language processing task, an encoder for receiving the source text and the auxiliary text from the preprocessor and generating an encoded representation of a combination of the source text and the auxiliary text, and a span-extractive decoder for receiving the encoded representation and identifying a span of text within the source text that is a result of the NLP task. The task type is one of entailment, classification, or regression. In some embodiments, the source text includes one or more of text received as input when the task type is entailment, a list of classifications when the task type is entailment or classification, or a list of similarity options when the task type is regression.
    Type: Grant
    Filed: February 16, 2022
    Date of Patent: May 23, 2023
    Assignee: salesforce.com, inc.
    Inventors: Nitish Shirish Keskar, Bryan McCann, Richard Socher, Caiming Xiong
  • Patent number: 11615249
    Abstract: Approaches for multitask learning as question answering include an input layer for encoding a context and a question, a self-attention based transformer including an encoder and a decoder, a first bi-directional long-term short-term memory (biLSTM) for further encoding an output of the encoder, a long-term short-term memory (LSTM) for generating a context-adjusted hidden state from the output of the decoder and a hidden state, an attention network for generating first attention weights based on an output of the first biLSTM and an output of the LSTM, a vocabulary layer for generating a distribution over a vocabulary, a context layer for generating a distribution over the context, and a switch for generating a weighting between the distributions over the vocabulary and the context, generating a composite distribution based on the weighting, and selecting a word of an answer using the composite distribution.
    Type: Grant
    Filed: August 18, 2020
    Date of Patent: March 28, 2023
    Assignee: salesforce.com, inc.
    Inventors: Bryan McCann, Nitish Shirish Keskar, Caiming Xiong, Richard Socher
  • Patent number: 11600194
    Abstract: Approaches for natural language processing include a multi-layer encoder for encoding words from a context and words from a question in parallel, a multi-layer decoder for decoding the encoded context and the encoded question, a pointer generator for generating distributions over the words from the context, the words from the question, and words in a vocabulary based on an output from the decoder, and a switch. The switch generates a weighting of the distributions over the words from the context, the words from the question, and the words in the vocabulary, generates a composite distribution based on the weighting of the distribution over the first words from the context, the distribution over the second words from the question, and the distribution over the words in the vocabulary, and selects words for inclusion in an answer using the composite distribution.
    Type: Grant
    Filed: June 12, 2018
    Date of Patent: March 7, 2023
    Assignee: Salesforce.com, Inc.
    Inventors: Bryan McCann, Nitish Shirish Keskar, Caiming Xiong, Richard Socher
  • Publication number: 20230042318
    Abstract: An analytics system uses one or more machine-learned models to predict a hormone receptor status from a H&E stain image. The system partitions H&E stain images each into a plurality of non-overlapping image tiles. Bags of tiles are created through sampling of the image tiles. For each H&E stain image, the system generates a feature vector from a bag of tiles sampled from the partitioned image tiles. The analytics system trains one or more machine-learned models with training H&E stain images having a positive or negative receptor status. With the trained models, the analytics system predicts a hormone receptor status by applying a prediction model to the feature vector for a test H&E stain image.
    Type: Application
    Filed: October 21, 2022
    Publication date: February 9, 2023
    Inventors: Nikhil Naik, Ali Madani, Nitish Shirish Keskar
  • Publication number: 20220391640
    Abstract: Embodiments are directed to pre-training a transformer model using more parameters for sophisticated patterns (PSP++). The transformer model is divided into a held-out model and a main model. A forward pass and a backward pass are performed on the held-out model, where the forward pass determines self-attention hidden states of the held-out model and the backward pass determines loss of the held-out model. A forward pass on the main model is performed to determine a self-attention hidden states of the main model. The self-attention hidden states of the main model are concatenated with the self-attention hidden states of the held-out model. A backward pass is performed on the main model to determine a loss of the main model. The parameters of the held-out model are updated to reflect the loss of the held-out model and parameters of the main model are updated to reflect the loss of the main model.
    Type: Application
    Filed: November 22, 2021
    Publication date: December 8, 2022
    Inventors: Chen Xing, Wenhao Liu, Chu Hong Hoi, Nitish Shirish Keskar, Caiming Xiong
  • Patent number: 11508481
    Abstract: An analytics system uses one or more machine-learned models to predict a hormone receptor status from a H&E stain image. The system partitions H&E stain images each into a plurality of image tiles. Bags of tiles are created through sampling of the image tiles. The analytics system trains one or more machine-learned models with training H&E stain images having a positive or negative receptor status. The analytics system generates, via a tile featurization model, a tile feature vector for each image tile a test bag for a test H&E stain image. The analytics system generates, via an attention model, an aggregate feature vector for the test bag by aggregating the tile feature vectors of the test bag, wherein an attention weight is determined for each tile feature vector. The analytics system predicts a hormone receptor status by applying a prediction model to the aggregate feature vector for the test bag.
    Type: Grant
    Filed: June 8, 2020
    Date of Patent: November 22, 2022
    Assignee: Salesforce, Inc.
    Inventors: Nikhil Naik, Ali Madani, Nitish Shirish Keskar
  • Patent number: 11501076
    Abstract: Approaches for multitask learning as question answering include a method for training that includes receiving a plurality of training samples including training samples from a plurality of task types, presenting the training samples to a neural model to generate an answer, determining an error between the generated answer and the natural language ground truth answer for each training sample presented, and adjusting parameters of the neural model based on the error. Each of the training samples includes a natural language context, question, and ground truth answer. An order in which the training samples are presented to the neural model includes initially selecting the training samples according to a first training strategy and switching to selecting the training samples according to a second training strategy. In some embodiments the first training strategy is a sequential training strategy and the second training strategy is a joint training strategy.
    Type: Grant
    Filed: May 8, 2018
    Date of Patent: November 15, 2022
    Assignee: SALESFORCE.COM, INC.
    Inventors: Nitish Shirish Keskar, Bryan McCann, Caiming Xiong, Richard Socher
  • Publication number: 20220171943
    Abstract: Systems and methods for unifying question answering and text classification via span extraction include a preprocessor for preparing a source text and an auxiliary text based on a task type of a natural language processing task, an encoder for receiving the source text and the auxiliary text from the preprocessor and generating an encoded representation of a combination of the source text and the auxiliary text, and a span-extractive decoder for receiving the encoded representation and identifying a span of text within the source text that is a result of the NLP task. The task type is one of entailment, classification, or regression. In some embodiments, the source text includes one or more of text received as input when the task type is entailment, a list of classifications when the task type is entailment or classification, or a list of similarity options when the task type is regression.
    Type: Application
    Filed: February 16, 2022
    Publication date: June 2, 2022
    Inventors: Nitish Shirish Keskar, Bryan McCann, Richard Socher, Caiming Xiong
  • Publication number: 20220129629
    Abstract: Embodiments described herein provide dynamic blocking, a decoding algorithm which enables large-scale pretrained language models to generate high-quality paraphrases in an un-supervised setting. Specifically, in order to obtain an alternative surface form, when the language model emits a token that is present in the source sequence, the language model is prevented from generating the next token that is the same as the subsequent source token in the source sequence at the next time step. In this way, the language model is forced to generate a paraphrased sequence of the input source sequence, but with mostly different wording.
    Type: Application
    Filed: January 28, 2021
    Publication date: April 28, 2022
    Inventors: Tong Niu, Semih Yavuz, Yingbo Zhou, Nitish Shirish Keskar, Huan Wang, Caiming Xiong
  • Patent number: 11281863
    Abstract: Systems and methods for unifying question answering and text classification via span extraction include a preprocessor for preparing a source text and an auxiliary text based on a task type of a natural language processing task, an encoder for receiving the source text and the auxiliary text from the preprocessor and generating an encoded representation of a combination of the source text and the auxiliary text, and a span-extractive decoder for receiving the encoded representation and identifying a span of text within the source text that is a result of the NLP task. The task type is one of entailment, classification, or regression. In some embodiments, the source text includes one or more of text received as input when the task type is entailment, a list of classifications when the task type is entailment or classification, or a list of similarity options when the task type is regression.
    Type: Grant
    Filed: July 22, 2019
    Date of Patent: March 22, 2022
    Assignee: salesforce.com, inc.
    Inventors: Nitish Shirish Keskar, Bryan McCann, Richard Socher, Caiming Xiong
  • Patent number: 11276002
    Abstract: Hybrid training of deep networks includes a multi-layer neural network. The training includes setting a current learning algorithm for the multi-layer neural network to a first learning algorithm. The training further includes iteratively applying training data to the neural network, determining a gradient for parameters of the neural network based on the applying of the training data, updating the parameters based on the current learning algorithm, and determining whether the current learning algorithm should be switched to a second learning algorithm based on the updating. The training further includes, in response to the determining that the current learning algorithm should be switched to a second learning algorithm, changing the current learning algorithm to the second learning algorithm and initializing a learning rate of the second learning algorithm based on the gradient and a step used by the first learning algorithm to update the parameters of the neural network.
    Type: Grant
    Filed: March 20, 2018
    Date of Patent: March 15, 2022
    Assignee: salesforce.com, inc.
    Inventors: Nitish Shirish Keskar, Richard Socher
  • Publication number: 20210375269
    Abstract: Embodiments described herein utilize pre-trained masked language models as the backbone for dialogue act tagging and provide cross-domain generalization of the resulting dialogue acting taggers. For example, a pre-trained MASK token of BERT model may be used as a controllable mechanism for augmenting text input, e.g., generating tags for an input of unlabeled dialogue history. The pre-trained MASK model can be trained with semi-supervised learning, e.g., using multiple objectives from supervised tagging loss, masked tagging loss, masked language model loss, and/or a disagreement loss.
    Type: Application
    Filed: August 21, 2020
    Publication date: December 2, 2021
    Inventors: Semih Yavuz, Kazuma Hashimoto, Wenhao Liu, Nitish Shirish Keskar, Richard Socher, Caiming Xiong
  • Publication number: 20210280311
    Abstract: An analytics system uses one or more machine-learned models to predict a hormone receptor status from a H&E stain image. The system partitions H&E stain images each into a plurality of image tiles. Bags of tiles are created through sampling of the image tiles. The analytics system trains one or more machine-learned models with training H&E stain images having a positive or negative receptor status. The analytics system generates, via a tile featurization model, a tile feature vector for each image tile a test bag for a test H&E stain image. The analytics system generates, via an attention model, an aggregate feature vector for the test bag by aggregating the tile feature vectors of the test bag, wherein an attention weight is determined for each tile feature vector. The analytics system predicts a hormone receptor status by applying a prediction model to the aggregate feature vector for the test bag.
    Type: Application
    Filed: June 8, 2020
    Publication date: September 9, 2021
    Inventors: Nikhil Naik, Ali Madani, Nitish Shirish Keskar
  • Publication number: 20210240943
    Abstract: Approaches for cross-lingual regularization for multilingual generalization include a method for training a natural language processing (NLP) deep learning module. The method includes accessing a first dataset having a first training data entry, the first training data entry including one or more natural language input text strings in a first language; translating at least one of the one or more natural language input text strings of the first training data entry from the first language to a second language; creating a second training data entry by starting with the first training data entry and substituting the at least one of the natural language input text strings in the first language with the translation of the at least one of the natural language input text strings in the second language; adding the second training data entry to a second dataset; and training the deep learning module using the second dataset.
    Type: Application
    Filed: April 23, 2021
    Publication date: August 5, 2021
    Inventors: Jasdeep SINGH, Nitish Shirish KESKAR, Bryan MCCANN