Patents by Inventor Weishi Wang

Weishi Wang has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240020102
    Abstract: Embodiments described herein a code generation and understanding model that builds on a Transformer-based encoder-decoder framework. The code generation and understanding model is configured to derive generic representations for programming language (PL) and natural language (NL) in code domain via pre-training on unlabeled code corpus, and then to benefit many code-related downstream tasks with fine-tuning. Apart from the denoising sequence-to-sequence objectives widely adopted for pre-training on natural language, identifier tagging and prediction pre-training objective is adopted to enable the model to better leverage the crucial token type information from PL, which specifically are the identifiers assigned by developers.
    Type: Application
    Filed: September 26, 2023
    Publication date: January 18, 2024
    Inventors: Yue Wang, Weishi Wang, Shafiq Rayhan Joty, Chu Hong Hoi
  • Publication number: 20230376401
    Abstract: Systems and methods for automatic program repair using neural network models are described. After a first buggy code patch is received, a first representation of the first buggy code patch is generated using a retriever encoder of a patch retriever. The patch retriever retrieves, based on the first representation, a first bug-fix code pair from a first plurality of bug-fix code pairs. A first augmented buggy code patch is generated based on the first buggy code patch and the first bug-fix code pair. A patch generator generates a fixed code patch based on the first augmented buggy code patch.
    Type: Application
    Filed: August 26, 2022
    Publication date: November 23, 2023
    Inventors: Yue Wang, Weishi Wang, Shafiq Rayhan Joty, Chu Hong Hoi
  • Patent number: 11782686
    Abstract: Embodiments described herein a code generation and understanding model that builds on a Transformer-based encoder-decoder framework. The code generation and understanding model is configured to derive generic representations for programming language (PL) and natural language (NL) in code domain via pre-training on unlabeled code corpus, and then to benefit many code-related downstream tasks with fine-tuning. Apart from the denoising sequence-to-sequence objectives widely adopted for pre-training on natural language, identifier tagging and prediction pre-training objective is adopted to enable the model to better leverage the crucial token type information from PL, which specifically are the identifiers assigned by developers.
    Type: Grant
    Filed: August 27, 2021
    Date of Patent: October 10, 2023
    Assignee: SALESFORCE.COM, INC.
    Inventors: Yue Wang, Weishi Wang, Shafiq Rayhan Joty, Chu Hong Hoi
  • Patent number: 11580975
    Abstract: Embodiments described herein provide a dynamic topic tracking mechanism that tracks how the conversation topics change from one utterance to another and use the tracking information to rank candidate responses. A pre-trained language model may be used for response selection in the multi-party conversations, which consists of two steps: (1) a topic-based pre-training to embed topic information into the language model with self-supervised learning, and (2) a multi-task learning on the pretrained model by jointly training response selection and dynamic topic prediction and disentanglement tasks.
    Type: Grant
    Filed: September 8, 2020
    Date of Patent: February 14, 2023
    Assignee: salesforce.com, inc.
    Inventors: Weishi Wang, Shafiq Rayhan Joty, Chu Hong Hoi
  • Publication number: 20220382527
    Abstract: Embodiments described herein a code generation and understanding model that builds on a Transformer-based encoder-decoder framework. The code generation and understanding model is configured to derive generic representations for programming language (PL) and natural language (NL) in code domain via pre-training on unlabeled code corpus, and then to benefit many code-related downstream tasks with fine-tuning. Apart from the denoising sequence-to-sequence objectives widely adopted for pre-training on natural language, identifier tagging and prediction pre-training objective is adopted to enable the model to better leverage the crucial token type information from PL, which specifically are the identifiers assigned by developers.
    Type: Application
    Filed: August 27, 2021
    Publication date: December 1, 2022
    Inventors: Yue Wang, Weishi Wang, Shafiq Rayhan Joty, Chu Hong Hoi
  • Publication number: 20210375280
    Abstract: Embodiments described herein provide a dynamic topic tracking mechanism that tracks how the conversation topics change from one utterance to another and use the tracking information to rank candidate responses. A pre-trained language model may be used for response selection in the multi-party conversations, which consists of two steps: (1) a topic-based pre-training to embed topic information into the language model with self-supervised learning, and (2) a multi-task learning on the pretrained model by jointly training response selection and dynamic topic prediction and disentanglement tasks.
    Type: Application
    Filed: September 8, 2020
    Publication date: December 2, 2021
    Inventors: Weishi Wang, Shafiq Rayhan Joty, Chu Hong Hoi