Patents by Inventor Shiva Kumar Pentyala

Shiva Kumar Pentyala has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240411992
    Abstract: Embodiments described herein provide a training framework for generative NLP models. Specifically, the training input, e.g., in the form of a sequence of tokens representing a user-agent dialogue, may be randomly masked for a few spans, which can be one or more tokens, one or more words, one or more sentences, or one or more paragraphs. These masked spans are replaced with their embeddings generated from pre-trained large language models are then used for training the NLP model.
    Type: Application
    Filed: June 15, 2023
    Publication date: December 12, 2024
    Inventors: Shiva Kumar Pentyala, Prafulla Kumar Choubey, Shashank Harinath, Sitaram Asur, Chien-Sheng Jason Wu, Zachary Alexander, Caiming Xiong
  • Publication number: 20240411991
    Abstract: Embodiments described herein provide a training framework for generative NLP models that operate on previously learnt knowledge from pretrained large language models. Specifically, to train an NLP model to generate a response to a user utterance (e.g., “resolve login issue”), document embeddings of support IT documents encoded by a pretrained LLM are fed to an NLP decoder together with a training dialogue (e.g., a dialogue between the chat agent on how to “resolve login issue”). The NLP decoder can thus be trained by a causal language modeling loss computed based on the predicted next token and the ground-truth token from the training dialogue.
    Type: Application
    Filed: June 6, 2023
    Publication date: December 12, 2024
    Inventors: Shiva Kumar Pentyala, Prafulla Kumar Choubey, Shashank Harinath, Sitaram Asur, Chien-Sheng Jason Wu, Zachary Alexander, Caiming Xiong
  • Publication number: 20240412059
    Abstract: Embodiments described herein provide A method for training a neural network based model. The methods include receiving a training dataset with a plurality of training samples, and those samples are encoded into representations in feature space. A positive sample is determined from the raining dataset based on a relationship between the given query and the positive sample in feature space. For a given query, a positive sample from the training dataset is selected based on a relationship between the given query and the positive sample in a feature space. One or more negative samples from the training dataset that are within a reconfigurable distance to the positive sample in the feature space are selected, and a loss is computed based on the positive sample and the one or more negative samples. The neural network is trained based on the loss.
    Type: Application
    Filed: June 7, 2023
    Publication date: December 12, 2024
    Inventors: Regunathan Radhakrishnan, Zachary Alexander, Sitaram Asur, Shashank Harinath, Na Cheng, Shiva Kumar Pentyala
  • Publication number: 20240241820
    Abstract: Embodiments described herein provide an automated testing pipeline for providing a testing dataset for testing a trained neural network model trained using a first training dataset. A first testing dataset for the trained neural network including a first plurality of user queries is received. A dependency parser is used to filter the first plurality of user queries based on one or more action verbs. A pretrained language model is used to rank the remaining user queries based on respective relationships with queries in the first training dataset. Further, user queries that are classified as keyword matches with the queries in the first training dataset using a bag of words classifier are removed. A second testing dataset is generated using the ranked remaining user queries. Testing outputs are generated, by the trained neural network model, using the second testing dataset.
    Type: Application
    Filed: January 18, 2023
    Publication date: July 18, 2024
    Inventors: Shiva Kumar PENTYALA, Shashank HARINATH, Sitaram ASUR, Zachary ALEXANDER
  • Patent number: 12001798
    Abstract: A system performs named entity recognition for performing natural language processing, for example, for conversation engines. The system uses context information in named entity recognition. The system includes the context of a sentence during model training and execution. The system generates high quality contextual data for training NER models. The system utilizes labeled and unlabeled contextual data for training NER models. The system provides NER models for execution in production environments. The system uses heuristics to determine whether to use a context-based NER model or a simple NER model that does not use context information. This allows the system to use simple NER models when the likelihood of improving the accuracy of prediction based on context is low.
    Type: Grant
    Filed: March 15, 2021
    Date of Patent: June 4, 2024
    Assignee: Salesforce, Inc.
    Inventors: Jingyuan Liu, Abhishek Sharma, Suhail Sanjiv Barot, Gurkirat Singh, Mridul Gupta, Shiva Kumar Pentyala, Ankit Chadha
  • Patent number: 11880659
    Abstract: Methods and systems for hierarchical natural language understanding are described. A representation of an utterance is inputted to a first machine learning model to obtain information on the first utterance. According to the information on the utterance a determination that the representation of the utterance is to be inputted to a second machine learning model that performs a dedicated natural language task is performed. In response to determining that the representation of the utterance is to be inputted to a second machine learning model, the utterance is inputted to the second machine learning model to obtain an output of the dedicated natural language task.
    Type: Grant
    Filed: January 29, 2021
    Date of Patent: January 23, 2024
    Assignee: Salesforce, Inc.
    Inventors: Shiva Kumar Pentyala, Jean-Marc Soumet, Shashank Harinath, Shilpa Bhagavath, Johnson Liu, Ankit Chadha
  • Patent number: 11599721
    Abstract: A natural language processing system that trains task models for particular natural language tasks programmatically generates additional utterances for inclusion in the training set, based on the existing utterances in the training set and the existing state of a task model as generated from the original (non-augmented) training set. More specifically, the training augmentation module 220 identifies specific textual units of utterances and generates variants of the utterances based on those identified units. The identification is based on determined importances of the textual units to the output of the task model, as well as on task rules that correspond to the natural language task for which the task model is being generated. The generation of the additional utterances improves the quality of the task model without the expense of manual labeling of utterances for training set inclusion.
    Type: Grant
    Filed: August 25, 2020
    Date of Patent: March 7, 2023
    Assignee: Salesforce, Inc.
    Inventors: Shiva Kumar Pentyala, Mridul Gupta, Ankit Chadha, Indira Iyer, Richard Socher
  • Publication number: 20220245349
    Abstract: Methods and systems for hierarchical natural language understanding are described. A representation of an utterance is inputted to a first machine learning model to obtain information on the first utterance. According to the information on the utterance a determination that the representation of the utterance is to be inputted to a second machine learning model that performs a dedicated natural language task is performed. In response to determining that the representation of the utterance is to be inputted to a second machine learning model, the utterance is inputted to the second machine learning model to obtain an output of the dedicated natural language task.
    Type: Application
    Filed: January 29, 2021
    Publication date: August 4, 2022
    Inventors: Shiva Kumar Pentyala, Jean-Marc Soumet, Shashank Harinath, Shilpa Bhagavath, Johnson Liu, Ankit Chadha
  • Publication number: 20220222441
    Abstract: A system performs named entity recognition for performing natural language processing, for example, for conversation engines. The system uses context information in named entity recognition. The system includes the context of a sentence during model training and execution. The system generates high quality contextual data for training NER models. The system utilizes labeled and unlabeled contextual data for training NER models. The system provides NER models for execution in production environments. The system uses heuristics to determine whether to use a context-based NER model or a simple NER model that does not use context information. This allows the system to use simple NER models when the likelihood of improving the accuracy of prediction based on context is low.
    Type: Application
    Filed: March 15, 2021
    Publication date: July 14, 2022
    Inventors: Jingyuan Liu, Abhishek Sharma, Suhail Sanjiv Barot, Gurkirat Singh, Mridul Gupta, Shiva Kumar Pentyala, Ankit Chadha
  • Publication number: 20220222489
    Abstract: A system performs named entity recognition for performing natural language processing, for example, for conversation engines. The system uses context information in named entity recognition. The system includes the context of a sentence during model training and execution. The system generates high quality contextual data for training NER models. The system utilizes labeled and unlabeled contextual data for training NER models. The system provides NER models for execution in production environments. The system uses heuristics to determine whether to use a context-based NER model or a simple NER model that does not use context information. This allows the system to use simple NER models when the likelihood of improving the accuracy of prediction based on context is low.
    Type: Application
    Filed: March 15, 2021
    Publication date: July 14, 2022
    Inventors: Jingyuan Liu, Abhishek Sharma, Suhail Sanjiv Barot, Gurkirat Singh, Mridul Gupta, Shiva Kumar Pentyala, Ankit Chadha
  • Publication number: 20220067277
    Abstract: A natural language processing system that trains task models for particular natural language tasks programmatically generates additional utterances for inclusion in the training set, based on the existing utterances in the training set and the existing state of a task model as generated from the original (non-augmented) training set. More specifically, the training augmentation module 220 identifies specific textual units of utterances and generates variants of the utterances based on those identified units. The identification is based on determined importances of the textual units to the output of the task model, as well as on task rules that correspond to the natural language task for which the task model is being generated. The generation of the additional utterances improves the quality of the task model without the expense of manual labeling of utterances for training set inclusion.
    Type: Application
    Filed: August 25, 2020
    Publication date: March 3, 2022
    Inventors: Shiva Kumar Pentyala, Mridul Gupta, Ankit Chadha, Indira Iyer, Richard Socher