Patents by Inventor Shafiq Rayhan Joty

Shafiq Rayhan Joty has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240020102
    Abstract: Embodiments described herein a code generation and understanding model that builds on a Transformer-based encoder-decoder framework. The code generation and understanding model is configured to derive generic representations for programming language (PL) and natural language (NL) in code domain via pre-training on unlabeled code corpus, and then to benefit many code-related downstream tasks with fine-tuning. Apart from the denoising sequence-to-sequence objectives widely adopted for pre-training on natural language, identifier tagging and prediction pre-training objective is adopted to enable the model to better leverage the crucial token type information from PL, which specifically are the identifiers assigned by developers.
    Type: Application
    Filed: September 26, 2023
    Publication date: January 18, 2024
    Inventors: Yue Wang, Weishi Wang, Shafiq Rayhan Joty, Chu Hong Hoi
  • Publication number: 20230419037
    Abstract: Embodiments described herein provide label modular prompts for a text classification task. A label modular prompt generator may determine a set of class labels of interest from a set of possible class labels associated with an input text sequence. The label modular prompt generator may generate a plurality of label prompts based on the set of class labels of interest. A first class label and a sequence of soft tokens that are generated based on representations associated with the first class label are concatenated into a first label prompt. The soft tokens are tunable using a plurality of parameters of the label modular prompt generator. The label modular prompt generator may provide an input of the input text sequence prepended with the plurality of label prompts to a pretrained language model. The pretrained language model may generate a task output in response to the input text sequence.
    Type: Application
    Filed: November 28, 2022
    Publication date: December 28, 2023
    Inventors: Hailin CHEN, Amrita SAHA, Shafiq Rayhan JOTY, Chu Hong HOI
  • Publication number: 20230419049
    Abstract: Embodiments described herein provide training a prompt generator for text classification. A first training dataset associated with a first plurality of class labels is received for a first training process. For a first instance of the first training dataset, a set of labels of interest is generated by sampling from a set of possible class labels including the first plurality of class labels. The prompt generator generates a first prompt based on the set of labels of interest. A pretrained language model generates a task output in response to an input of the first instance prepended with the first prompt. A loss objective is generated based on the task output and the set of labels of interest. Parameters of the prompt generator are updated based on the computed loss function via backpropagation while the pretrained language model is frozen.
    Type: Application
    Filed: November 28, 2022
    Publication date: December 28, 2023
    Inventors: Hailin CHEN, Amrita SAHA, Shafiq Rayhan JOTY, Chu Hong HOI
  • Publication number: 20230376401
    Abstract: Systems and methods for automatic program repair using neural network models are described. After a first buggy code patch is received, a first representation of the first buggy code patch is generated using a retriever encoder of a patch retriever. The patch retriever retrieves, based on the first representation, a first bug-fix code pair from a first plurality of bug-fix code pairs. A first augmented buggy code patch is generated based on the first buggy code patch and the first bug-fix code pair. A patch generator generates a fixed code patch based on the first augmented buggy code patch.
    Type: Application
    Filed: August 26, 2022
    Publication date: November 23, 2023
    Inventors: Yue Wang, Weishi Wang, Shafiq Rayhan Joty, Chu Hong Hoi
  • Patent number: 11782686
    Abstract: Embodiments described herein a code generation and understanding model that builds on a Transformer-based encoder-decoder framework. The code generation and understanding model is configured to derive generic representations for programming language (PL) and natural language (NL) in code domain via pre-training on unlabeled code corpus, and then to benefit many code-related downstream tasks with fine-tuning. Apart from the denoising sequence-to-sequence objectives widely adopted for pre-training on natural language, identifier tagging and prediction pre-training objective is adopted to enable the model to better leverage the crucial token type information from PL, which specifically are the identifiers assigned by developers.
    Type: Grant
    Filed: August 27, 2021
    Date of Patent: October 10, 2023
    Assignee: SALESFORCE.COM, INC.
    Inventors: Yue Wang, Weishi Wang, Shafiq Rayhan Joty, Chu Hong Hoi
  • Patent number: 11755847
    Abstract: Embodiments described herein provide adversarial attacks targeting the cross-lingual generalization ability of massive multilingual representations, demonstrating their effectiveness on multilingual models for natural language inference and question answering. An efficient adversarial training scheme can thus be implemented with the adversarial attacks, which takes the same number of steps as standard supervised training and show that it encourages language-invariance in representations, thereby improving both clean and robust accuracy.
    Type: Grant
    Filed: January 15, 2021
    Date of Patent: September 12, 2023
    Assignee: Salesforce, Inc.
    Inventors: Samson Min Rong Tan, Shafiq Rayhan Joty
  • Publication number: 20230237275
    Abstract: Embodiments provide a software framework for evaluating and troubleshooting real-world task-oriented bot systems. Specifically, the evaluation framework includes a generator that infers dialog acts and entities from bot definitions and generates test cases for the system via model-based paraphrasing. The framework may also include a simulator for task-oriented dialog user simulation that supports both regression testing and end-to-end evaluation. The framework may also include a remediator to analyze and visualize the simulation results, remedy some of the identified issues, and provide actionable suggestions for improving the task-oriented dialog system.
    Type: Application
    Filed: June 2, 2022
    Publication date: July 27, 2023
    Inventors: Guangsen Wang, Samson Min Rong Tan, Shafiq Rayhan Joty, Gang Wu, Chu Hong Hoi, Ka Chun Au
  • Patent number: 11640505
    Abstract: Embodiments described herein provide systems and methods for an Explicit Memory Tracker (EMT) that tracks each rule sentence to perform decision making and to generate follow-up clarifying questions. Specifically, the EMT first segments the regulation text into several rule sentences and allocates the segmented rule sentences into memory modules, and then feeds information regarding the user scenario and dialogue history into the EMT sequentially to update each memory module separately. At each dialogue turn, the EMT makes a decision among based on current memory status of the memory modules whether further clarification is needed to come up with an answer to a user question. The EMT determines that further clarification is needed by identifying an underspecified rule sentence span by modulating token-level span distributions with sentence-level selection scores. The EMT extracts the underspecified rule sentence span and rephrases the underspecified rule sentence span to generate a follow-up question.
    Type: Grant
    Filed: April 30, 2020
    Date of Patent: May 2, 2023
    Assignee: Salesforce.com, Inc.
    Inventors: Yifan Gao, Chu Hong Hoi, Shafiq Rayhan Joty, Chien-Sheng Wu
  • Patent number: 11615240
    Abstract: Embodiments described herein provide an attention-based tree encoding mechanism. Specifically, the attention layer receives as input the pre-parsed constituency tree of a sentence and the lower-layer representations of all nodes. The attention layer then performs upward accumulation to encode the tree structure from leaves to the root in a bottom-up fashion. Afterwards, weighted aggregation is used to compute the final representations of non-terminal nodes.
    Type: Grant
    Filed: September 24, 2019
    Date of Patent: March 28, 2023
    Assignee: Salesforce.com, Inc
    Inventors: Xuan Phi Nguyen, Shafiq Rayhan Joty, Chu Hong Hoi
  • Patent number: 11580975
    Abstract: Embodiments described herein provide a dynamic topic tracking mechanism that tracks how the conversation topics change from one utterance to another and use the tracking information to rank candidate responses. A pre-trained language model may be used for response selection in the multi-party conversations, which consists of two steps: (1) a topic-based pre-training to embed topic information into the language model with self-supervised learning, and (2) a multi-task learning on the pretrained model by jointly training response selection and dynamic topic prediction and disentanglement tasks.
    Type: Grant
    Filed: September 8, 2020
    Date of Patent: February 14, 2023
    Assignee: salesforce.com, inc.
    Inventors: Weishi Wang, Shafiq Rayhan Joty, Chu Hong Hoi
  • Patent number: 11562147
    Abstract: A visual dialogue model receives image input and text input that includes a dialogue history between the model and a current utterance by a human user. The model generates a unified contextualized representation using a transformer encoder network, in which the unified contextualized representation includes a token level encoding of the image input and text input. The model generates an encoded visual dialogue input from the unified contextualized representation using visual dialogue encoding layers. The encoded visual dialogue input includes a position level encoding and a segment type encoding. The model generates an answer prediction from the encoded visual dialogue input using a first self-attention mask associated with discriminative settings or a second self-attention mask associated with generative settings. Dense annotation fine tuning may be performed to increase accuracy of the answer prediction. The model provides the answer prediction as a response to the current utterance of the human user.
    Type: Grant
    Filed: July 15, 2020
    Date of Patent: January 24, 2023
    Assignee: Salesforce.com, Inc.
    Inventors: Yue Wang, Chu Hong Hoi, Shafiq Rayhan Joty
  • Publication number: 20220382527
    Abstract: Embodiments described herein a code generation and understanding model that builds on a Transformer-based encoder-decoder framework. The code generation and understanding model is configured to derive generic representations for programming language (PL) and natural language (NL) in code domain via pre-training on unlabeled code corpus, and then to benefit many code-related downstream tasks with fine-tuning. Apart from the denoising sequence-to-sequence objectives widely adopted for pre-training on natural language, identifier tagging and prediction pre-training objective is adopted to enable the model to better leverage the crucial token type information from PL, which specifically are the identifiers assigned by developers.
    Type: Application
    Filed: August 27, 2021
    Publication date: December 1, 2022
    Inventors: Yue Wang, Weishi Wang, Shafiq Rayhan Joty, Chu Hong Hoi
  • Publication number: 20220374595
    Abstract: Embodiments described herein provides a contrastive learning framework that leverages hard negative examples, that are mined globally from the entire training corpus for a given query to improve the quality of code and natural language representations. Specifically, similar examples from the training corpus are extracted and used as hard negatives in an online manner during training while keeping the minibatch construction random.
    Type: Application
    Filed: November 19, 2021
    Publication date: November 24, 2022
    Inventors: Akhilesh Deepak Gotmare, Junnan Li, Shafiq Rayhan Joty, Chu Hong Hoi
  • Publication number: 20220164547
    Abstract: Embodiments described herein provide adversarial attacks targeting the cross-lingual generalization ability of massive multilingual representations, demonstrating their effectiveness on multilingual models for natural language inference and question answering. An efficient adversarial training scheme can thus be implemented with the adversarial attacks, which takes the same number of steps as standard supervised training and show that it encourages language-invariance in representations, thereby improving both clean and robust accuracy.
    Type: Application
    Filed: January 15, 2021
    Publication date: May 26, 2022
    Inventors: Samson Min Rong Tan, Shafiq Rayhan Joty
  • Publication number: 20220108169
    Abstract: Embodiments described herein provide systems and methods for a partially supervised training model for questioning answering tasks. Specifically, the partially supervised training model may include two modules—a query parsing module and a program execution module. The query parsing module parses queries into a grogram, and the program execution module execute the program to reach an answer through explicit reasoning and partial supervision. In this way, the partially supervised training model can be trained with answers as supervision, obviating the need for supervision by gold program operations and gold query-span attention at each step of the program.
    Type: Application
    Filed: January 29, 2021
    Publication date: April 7, 2022
    Inventors: Amrita Saha, Shafiq Rayhan Joty, Chu Hong Hoi
  • Patent number: 11256754
    Abstract: Embodiments described herein provide systems and methods for generating an adversarial sample with inflectional perturbations for training a natural language processing (NLP) system. A natural language sentence is received at an inflection perturbation module. Tokens are generated from the natural language sentence. For each token that has a part of speech that is a verb, adjective, or an adverb, an inflected form is determined. An adversarial sample of the natural language sentence is generated by detokenizing inflected forms of the tokens. The NLP system is trained using the adversarial sample.
    Type: Grant
    Filed: May 8, 2020
    Date of Patent: February 22, 2022
    Assignee: salesforce.com, inc.
    Inventors: Samson Min Rong Tan, Shafiq Rayhan Joty
  • Publication number: 20210375441
    Abstract: A method can be implemented at one or more computing machines. The method can include receiving, using a server, time-series data corresponding to monitoring instrumentation in a medical care facility. The time-series data corresponds to a selected care recipient. The time-series data is stored in one or more data storage units. The time-series data includes data correlated with a plurality of regular time intervals. The method includes receiving, using a server, aperiodic data corresponding to clinical notes collected in the medical care facility and corresponding to the selected care recipient. The aperiodic data is stored in one or more data storage units. The aperiodic data includes a time stamp. The method includes generating, using a deep neural network and the time-series data and using a convolutional neural network (CNN) and the aperiodic data, a plurality of computer-generated data corresponding to management of the medical care facility or medical condition of the care recipient.
    Type: Application
    Filed: May 26, 2021
    Publication date: December 2, 2021
    Inventors: Karan Aggarwal, Swaraj Khadanga, Shafiq Rayhan Joty, Jaideep Srivastava
  • Publication number: 20210375280
    Abstract: Embodiments described herein provide a dynamic topic tracking mechanism that tracks how the conversation topics change from one utterance to another and use the tracking information to rank candidate responses. A pre-trained language model may be used for response selection in the multi-party conversations, which consists of two steps: (1) a topic-based pre-training to embed topic information into the language model with self-supervised learning, and (2) a multi-task learning on the pretrained model by jointly training response selection and dynamic topic prediction and disentanglement tasks.
    Type: Application
    Filed: September 8, 2020
    Publication date: December 2, 2021
    Inventors: Weishi Wang, Shafiq Rayhan Joty, Chu Hong Hoi
  • Publication number: 20210232773
    Abstract: A visual dialogue model receives image input and text input that includes a dialogue history between the model and a current utterance by a human user. The model generates a unified contextualized representation using a transformer encoder network, in which the unified contextualized representation includes a token level encoding of the image input and text input. The model generates an encoded visual dialogue input from the unified contextualized representation using visual dialogue encoding layers. The encoded visual dialogue input includes a position level encoding and a segment type encoding. The model generates an answer prediction from the encoded visual dialogue input using a first self-attention mask associated with discriminative settings or a second self-attention mask associated with generative settings. Dense annotation fine tuning may be performed to increase accuracy of the answer prediction. The model provides the answer prediction as a response to the current utterance of the human user.
    Type: Application
    Filed: July 15, 2020
    Publication date: July 29, 2021
    Inventors: Yue WANG, Chu Hong HOI, Shafiq Rayhan JOTY
  • Publication number: 20210174023
    Abstract: Embodiments described herein provide systems and methods for an Explicit Memory Tracker (EMT) that tracks each rule sentence to perform decision making and to generate follow-up clarifying questions. Specifically, the EMT first segments the regulation text into several rule sentences and allocates the segmented rule sentences into memory modules, and then feeds information regarding the user scenario and dialogue history into the EMT sequentially to update each memory module separately. At each dialogue turn, the EMT makes a decision among based on current memory status of the memory modules whether further clarification is needed to come up with an answer to a user question. The EMT determines that further clarification is needed by identifying an underspecified rule sentence span by modulating token-level span distributions with sentence-level selection scores. The EMT extracts the underspecified rule sentence span and rephrases the underspecified rule sentence span to generate a follow-up question.
    Type: Application
    Filed: April 30, 2020
    Publication date: June 10, 2021
    Inventors: Yifan GAO, Chu Hong HOI, Shafiq Rayhan JOTY, Chien-Sheng WU