Patents by Inventor Shafiq Rayhan Joty
Shafiq Rayhan Joty has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240020102Abstract: Embodiments described herein a code generation and understanding model that builds on a Transformer-based encoder-decoder framework. The code generation and understanding model is configured to derive generic representations for programming language (PL) and natural language (NL) in code domain via pre-training on unlabeled code corpus, and then to benefit many code-related downstream tasks with fine-tuning. Apart from the denoising sequence-to-sequence objectives widely adopted for pre-training on natural language, identifier tagging and prediction pre-training objective is adopted to enable the model to better leverage the crucial token type information from PL, which specifically are the identifiers assigned by developers.Type: ApplicationFiled: September 26, 2023Publication date: January 18, 2024Inventors: Yue Wang, Weishi Wang, Shafiq Rayhan Joty, Chu Hong Hoi
-
Publication number: 20230419037Abstract: Embodiments described herein provide label modular prompts for a text classification task. A label modular prompt generator may determine a set of class labels of interest from a set of possible class labels associated with an input text sequence. The label modular prompt generator may generate a plurality of label prompts based on the set of class labels of interest. A first class label and a sequence of soft tokens that are generated based on representations associated with the first class label are concatenated into a first label prompt. The soft tokens are tunable using a plurality of parameters of the label modular prompt generator. The label modular prompt generator may provide an input of the input text sequence prepended with the plurality of label prompts to a pretrained language model. The pretrained language model may generate a task output in response to the input text sequence.Type: ApplicationFiled: November 28, 2022Publication date: December 28, 2023Inventors: Hailin CHEN, Amrita SAHA, Shafiq Rayhan JOTY, Chu Hong HOI
-
Publication number: 20230419049Abstract: Embodiments described herein provide training a prompt generator for text classification. A first training dataset associated with a first plurality of class labels is received for a first training process. For a first instance of the first training dataset, a set of labels of interest is generated by sampling from a set of possible class labels including the first plurality of class labels. The prompt generator generates a first prompt based on the set of labels of interest. A pretrained language model generates a task output in response to an input of the first instance prepended with the first prompt. A loss objective is generated based on the task output and the set of labels of interest. Parameters of the prompt generator are updated based on the computed loss function via backpropagation while the pretrained language model is frozen.Type: ApplicationFiled: November 28, 2022Publication date: December 28, 2023Inventors: Hailin CHEN, Amrita SAHA, Shafiq Rayhan JOTY, Chu Hong HOI
-
Publication number: 20230376401Abstract: Systems and methods for automatic program repair using neural network models are described. After a first buggy code patch is received, a first representation of the first buggy code patch is generated using a retriever encoder of a patch retriever. The patch retriever retrieves, based on the first representation, a first bug-fix code pair from a first plurality of bug-fix code pairs. A first augmented buggy code patch is generated based on the first buggy code patch and the first bug-fix code pair. A patch generator generates a fixed code patch based on the first augmented buggy code patch.Type: ApplicationFiled: August 26, 2022Publication date: November 23, 2023Inventors: Yue Wang, Weishi Wang, Shafiq Rayhan Joty, Chu Hong Hoi
-
Patent number: 11782686Abstract: Embodiments described herein a code generation and understanding model that builds on a Transformer-based encoder-decoder framework. The code generation and understanding model is configured to derive generic representations for programming language (PL) and natural language (NL) in code domain via pre-training on unlabeled code corpus, and then to benefit many code-related downstream tasks with fine-tuning. Apart from the denoising sequence-to-sequence objectives widely adopted for pre-training on natural language, identifier tagging and prediction pre-training objective is adopted to enable the model to better leverage the crucial token type information from PL, which specifically are the identifiers assigned by developers.Type: GrantFiled: August 27, 2021Date of Patent: October 10, 2023Assignee: SALESFORCE.COM, INC.Inventors: Yue Wang, Weishi Wang, Shafiq Rayhan Joty, Chu Hong Hoi
-
Patent number: 11755847Abstract: Embodiments described herein provide adversarial attacks targeting the cross-lingual generalization ability of massive multilingual representations, demonstrating their effectiveness on multilingual models for natural language inference and question answering. An efficient adversarial training scheme can thus be implemented with the adversarial attacks, which takes the same number of steps as standard supervised training and show that it encourages language-invariance in representations, thereby improving both clean and robust accuracy.Type: GrantFiled: January 15, 2021Date of Patent: September 12, 2023Assignee: Salesforce, Inc.Inventors: Samson Min Rong Tan, Shafiq Rayhan Joty
-
Publication number: 20230237275Abstract: Embodiments provide a software framework for evaluating and troubleshooting real-world task-oriented bot systems. Specifically, the evaluation framework includes a generator that infers dialog acts and entities from bot definitions and generates test cases for the system via model-based paraphrasing. The framework may also include a simulator for task-oriented dialog user simulation that supports both regression testing and end-to-end evaluation. The framework may also include a remediator to analyze and visualize the simulation results, remedy some of the identified issues, and provide actionable suggestions for improving the task-oriented dialog system.Type: ApplicationFiled: June 2, 2022Publication date: July 27, 2023Inventors: Guangsen Wang, Samson Min Rong Tan, Shafiq Rayhan Joty, Gang Wu, Chu Hong Hoi, Ka Chun Au
-
Patent number: 11640505Abstract: Embodiments described herein provide systems and methods for an Explicit Memory Tracker (EMT) that tracks each rule sentence to perform decision making and to generate follow-up clarifying questions. Specifically, the EMT first segments the regulation text into several rule sentences and allocates the segmented rule sentences into memory modules, and then feeds information regarding the user scenario and dialogue history into the EMT sequentially to update each memory module separately. At each dialogue turn, the EMT makes a decision among based on current memory status of the memory modules whether further clarification is needed to come up with an answer to a user question. The EMT determines that further clarification is needed by identifying an underspecified rule sentence span by modulating token-level span distributions with sentence-level selection scores. The EMT extracts the underspecified rule sentence span and rephrases the underspecified rule sentence span to generate a follow-up question.Type: GrantFiled: April 30, 2020Date of Patent: May 2, 2023Assignee: Salesforce.com, Inc.Inventors: Yifan Gao, Chu Hong Hoi, Shafiq Rayhan Joty, Chien-Sheng Wu
-
Patent number: 11615240Abstract: Embodiments described herein provide an attention-based tree encoding mechanism. Specifically, the attention layer receives as input the pre-parsed constituency tree of a sentence and the lower-layer representations of all nodes. The attention layer then performs upward accumulation to encode the tree structure from leaves to the root in a bottom-up fashion. Afterwards, weighted aggregation is used to compute the final representations of non-terminal nodes.Type: GrantFiled: September 24, 2019Date of Patent: March 28, 2023Assignee: Salesforce.com, IncInventors: Xuan Phi Nguyen, Shafiq Rayhan Joty, Chu Hong Hoi
-
Patent number: 11580975Abstract: Embodiments described herein provide a dynamic topic tracking mechanism that tracks how the conversation topics change from one utterance to another and use the tracking information to rank candidate responses. A pre-trained language model may be used for response selection in the multi-party conversations, which consists of two steps: (1) a topic-based pre-training to embed topic information into the language model with self-supervised learning, and (2) a multi-task learning on the pretrained model by jointly training response selection and dynamic topic prediction and disentanglement tasks.Type: GrantFiled: September 8, 2020Date of Patent: February 14, 2023Assignee: salesforce.com, inc.Inventors: Weishi Wang, Shafiq Rayhan Joty, Chu Hong Hoi
-
Patent number: 11562147Abstract: A visual dialogue model receives image input and text input that includes a dialogue history between the model and a current utterance by a human user. The model generates a unified contextualized representation using a transformer encoder network, in which the unified contextualized representation includes a token level encoding of the image input and text input. The model generates an encoded visual dialogue input from the unified contextualized representation using visual dialogue encoding layers. The encoded visual dialogue input includes a position level encoding and a segment type encoding. The model generates an answer prediction from the encoded visual dialogue input using a first self-attention mask associated with discriminative settings or a second self-attention mask associated with generative settings. Dense annotation fine tuning may be performed to increase accuracy of the answer prediction. The model provides the answer prediction as a response to the current utterance of the human user.Type: GrantFiled: July 15, 2020Date of Patent: January 24, 2023Assignee: Salesforce.com, Inc.Inventors: Yue Wang, Chu Hong Hoi, Shafiq Rayhan Joty
-
Publication number: 20220382527Abstract: Embodiments described herein a code generation and understanding model that builds on a Transformer-based encoder-decoder framework. The code generation and understanding model is configured to derive generic representations for programming language (PL) and natural language (NL) in code domain via pre-training on unlabeled code corpus, and then to benefit many code-related downstream tasks with fine-tuning. Apart from the denoising sequence-to-sequence objectives widely adopted for pre-training on natural language, identifier tagging and prediction pre-training objective is adopted to enable the model to better leverage the crucial token type information from PL, which specifically are the identifiers assigned by developers.Type: ApplicationFiled: August 27, 2021Publication date: December 1, 2022Inventors: Yue Wang, Weishi Wang, Shafiq Rayhan Joty, Chu Hong Hoi
-
Publication number: 20220374595Abstract: Embodiments described herein provides a contrastive learning framework that leverages hard negative examples, that are mined globally from the entire training corpus for a given query to improve the quality of code and natural language representations. Specifically, similar examples from the training corpus are extracted and used as hard negatives in an online manner during training while keeping the minibatch construction random.Type: ApplicationFiled: November 19, 2021Publication date: November 24, 2022Inventors: Akhilesh Deepak Gotmare, Junnan Li, Shafiq Rayhan Joty, Chu Hong Hoi
-
Publication number: 20220164547Abstract: Embodiments described herein provide adversarial attacks targeting the cross-lingual generalization ability of massive multilingual representations, demonstrating their effectiveness on multilingual models for natural language inference and question answering. An efficient adversarial training scheme can thus be implemented with the adversarial attacks, which takes the same number of steps as standard supervised training and show that it encourages language-invariance in representations, thereby improving both clean and robust accuracy.Type: ApplicationFiled: January 15, 2021Publication date: May 26, 2022Inventors: Samson Min Rong Tan, Shafiq Rayhan Joty
-
Publication number: 20220108169Abstract: Embodiments described herein provide systems and methods for a partially supervised training model for questioning answering tasks. Specifically, the partially supervised training model may include two modules—a query parsing module and a program execution module. The query parsing module parses queries into a grogram, and the program execution module execute the program to reach an answer through explicit reasoning and partial supervision. In this way, the partially supervised training model can be trained with answers as supervision, obviating the need for supervision by gold program operations and gold query-span attention at each step of the program.Type: ApplicationFiled: January 29, 2021Publication date: April 7, 2022Inventors: Amrita Saha, Shafiq Rayhan Joty, Chu Hong Hoi
-
Patent number: 11256754Abstract: Embodiments described herein provide systems and methods for generating an adversarial sample with inflectional perturbations for training a natural language processing (NLP) system. A natural language sentence is received at an inflection perturbation module. Tokens are generated from the natural language sentence. For each token that has a part of speech that is a verb, adjective, or an adverb, an inflected form is determined. An adversarial sample of the natural language sentence is generated by detokenizing inflected forms of the tokens. The NLP system is trained using the adversarial sample.Type: GrantFiled: May 8, 2020Date of Patent: February 22, 2022Assignee: salesforce.com, inc.Inventors: Samson Min Rong Tan, Shafiq Rayhan Joty
-
Publication number: 20210375441Abstract: A method can be implemented at one or more computing machines. The method can include receiving, using a server, time-series data corresponding to monitoring instrumentation in a medical care facility. The time-series data corresponds to a selected care recipient. The time-series data is stored in one or more data storage units. The time-series data includes data correlated with a plurality of regular time intervals. The method includes receiving, using a server, aperiodic data corresponding to clinical notes collected in the medical care facility and corresponding to the selected care recipient. The aperiodic data is stored in one or more data storage units. The aperiodic data includes a time stamp. The method includes generating, using a deep neural network and the time-series data and using a convolutional neural network (CNN) and the aperiodic data, a plurality of computer-generated data corresponding to management of the medical care facility or medical condition of the care recipient.Type: ApplicationFiled: May 26, 2021Publication date: December 2, 2021Inventors: Karan Aggarwal, Swaraj Khadanga, Shafiq Rayhan Joty, Jaideep Srivastava
-
Publication number: 20210375280Abstract: Embodiments described herein provide a dynamic topic tracking mechanism that tracks how the conversation topics change from one utterance to another and use the tracking information to rank candidate responses. A pre-trained language model may be used for response selection in the multi-party conversations, which consists of two steps: (1) a topic-based pre-training to embed topic information into the language model with self-supervised learning, and (2) a multi-task learning on the pretrained model by jointly training response selection and dynamic topic prediction and disentanglement tasks.Type: ApplicationFiled: September 8, 2020Publication date: December 2, 2021Inventors: Weishi Wang, Shafiq Rayhan Joty, Chu Hong Hoi
-
Publication number: 20210232773Abstract: A visual dialogue model receives image input and text input that includes a dialogue history between the model and a current utterance by a human user. The model generates a unified contextualized representation using a transformer encoder network, in which the unified contextualized representation includes a token level encoding of the image input and text input. The model generates an encoded visual dialogue input from the unified contextualized representation using visual dialogue encoding layers. The encoded visual dialogue input includes a position level encoding and a segment type encoding. The model generates an answer prediction from the encoded visual dialogue input using a first self-attention mask associated with discriminative settings or a second self-attention mask associated with generative settings. Dense annotation fine tuning may be performed to increase accuracy of the answer prediction. The model provides the answer prediction as a response to the current utterance of the human user.Type: ApplicationFiled: July 15, 2020Publication date: July 29, 2021Inventors: Yue WANG, Chu Hong HOI, Shafiq Rayhan JOTY
-
Publication number: 20210174023Abstract: Embodiments described herein provide systems and methods for an Explicit Memory Tracker (EMT) that tracks each rule sentence to perform decision making and to generate follow-up clarifying questions. Specifically, the EMT first segments the regulation text into several rule sentences and allocates the segmented rule sentences into memory modules, and then feeds information regarding the user scenario and dialogue history into the EMT sequentially to update each memory module separately. At each dialogue turn, the EMT makes a decision among based on current memory status of the memory modules whether further clarification is needed to come up with an answer to a user question. The EMT determines that further clarification is needed by identifying an underspecified rule sentence span by modulating token-level span distributions with sentence-level selection scores. The EMT extracts the underspecified rule sentence span and rephrases the underspecified rule sentence span to generate a follow-up question.Type: ApplicationFiled: April 30, 2020Publication date: June 10, 2021Inventors: Yifan GAO, Chu Hong HOI, Shafiq Rayhan JOTY, Chien-Sheng WU