Patents by Inventor Shaogang REN

Shaogang REN has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11947908
    Abstract: Described herein are system and method embodiments to improve word representation learning. Embodiments of a probabilistic prior may seamlessly integrate statistical disentanglement with word embedding. Different from previous deterministic methods, word embedding may be taken as a probabilistic generative model, and it enables imposing a prior that may identify independent factors generating word representation vectors. The probabilistic prior not only enhances the representation of word embedding, but also improves the model's robustness and stability. Furthermore, embodiments of the disclosed method may be flexibly plugged in various word embedding models. Extensive experimental results show that embodiments of the presented method may improve word representation on different tasks.
    Type: Grant
    Filed: April 7, 2021
    Date of Patent: April 2, 2024
    Assignee: Baidu USA LLC
    Inventors: Shaogang Ren, Ping Li
  • Patent number: 11816533
    Abstract: Learning disentangled representations is an important topic in machine learning for a wide range of applications. Disentangled latent variables represent interpretable semantic information and reflect separate factors of variation in data. Although generative models may learn latent representations and generate data samples as well, existing models may ignore the structural information among latent representations. Described in the present disclosure are embodiments to learn disentangled latent structural representations from data using decomposable variational auto-encoders, which simultaneously learn component representations and encode component relationships. Embodiments of a novel structural prior for latent representations are disclosed to capture interactions among different data components. Embodiments are applied to data segmentation and latent relation discovery among different data components. Experiments on several datasets demonstrate the utility of the present model embodiments.
    Type: Grant
    Filed: November 18, 2020
    Date of Patent: November 14, 2023
    Assignee: Baidu USA LLC
    Inventors: Shaogang Ren, Hongliang Fei, Dingcheng Li, Ping Li
  • Patent number: 11783198
    Abstract: The thriving of deep models and generative models provides approaches to model high dimensional distributions. Generative adversarial networks (GANs) can approximate data distributions and generate data samples from the learned data manifolds as well. Presented herein are embodiments to estimate the implicit likelihoods of GAN models. In one or more embodiments, a stable inverse function of the generator is learned with the help of a variance network of the generator. The local variance of the sample distribution may be approximated by the normalized distance in the latent space. Simulation studies and likelihood testing on data sets validate embodiments, which outperformed several baseline methods in these tasks. An embodiment was also applied to anomaly detection. Experiments show that the embodiments herein can achieve state-of-the-art anomaly detection performance.
    Type: Grant
    Filed: April 3, 2020
    Date of Patent: October 10, 2023
    Assignee: Baidu USA LLC
    Inventors: Dingcheng Li, Shaogang Ren, Zhixin Zhou, Ping Li
  • Patent number: 11748567
    Abstract: Described herein are embodiments of a framework named as total correlation variational autoencoder (TC_VAE) to disentangle syntax and semantics by making use of total correlation penalties of KL divergences. One or more Kullback-Leibler (KL) divergence terms in a loss for a variational autoencoder are discomposed so that generated hidden variables may be separated. Embodiments of the TC_VAE framework were examined on semantic similarity tasks and syntactic similarity tasks. Experimental results show that better disentanglement between syntactic and semantic representations have been achieved compared with state-of-the-art (SOTA) results on the same data sets in similar settings.
    Type: Grant
    Filed: July 10, 2020
    Date of Patent: September 5, 2023
    Assignee: Baidu USA LLC
    Inventors: Dingcheng Li, Shaogang Ren, Ping Li
  • Publication number: 20220335216
    Abstract: Described herein are system and method embodiments to improve word representation learning. Embodiments of a probabilistic prior may seamlessly integrate statistical disentanglement with word embedding. Different from previous deterministic methods, word embedding may be taken as a probabilistic generative model, and it enables imposing a prior that may identify independent factors generating word representation vectors. The probabilistic prior not only enhances the representation of word embedding, but also improves the model's robustness and stability. Furthermore, embodiments of the disclosed method may be flexibly plugged in various word embedding models. Extensive experimental results show that embodiments of the presented method may improve word representation on different tasks.
    Type: Application
    Filed: April 7, 2021
    Publication date: October 20, 2022
    Applicant: Baidu USA LLC
    Inventors: Shaogang REN, Ping LI
  • Publication number: 20220156612
    Abstract: Learning disentangled representations is an important topic in machine learning for a wide range of applications. Disentangled latent variables represent interpretable semantic information and reflect separate factors of variation in data. Although generative models may learn latent representations and generate data samples as well, existing models may ignore the structural information among latent representations. Described in the present disclosure are embodiments to learn disentangled latent structural representations from data using decomposable variational auto-encoders, which simultaneously learn component representation and encodes component relationships. Embodiments of a novel structural prior for latent representations are disclosed to capture interactions among different data components. Embodiments are applied to data segmentation and latent relation discovery among different data components. Experiments on several datasets demonstrate the utility of the present model embodiments.
    Type: Application
    Filed: November 18, 2020
    Publication date: May 19, 2022
    Applicant: Baidu USA LLC
    Inventors: Shaogang REN, Hongliang FEI, Dingcheng LI, Ping LI
  • Publication number: 20220043975
    Abstract: Described herein are embodiments of a framework named decomposable variational autoencoder (DecVAE) to disentangle syntax and semantics by using total correlation penalties of Kullback-Leibler (KL) divergences. KL divergence term of the original VAE are decomposed such that the hidden variables generated may be separated in a clear-cut and interpretable way. Embodiments of DecVAE models are evaluated on various semantic similarity and syntactic similarity datasets. Experimental results show that embodiments of DecVAE models achieve state-of-the-art (SOTA) performance in disentanglement between syntactic and semantic representations.
    Type: Application
    Filed: August 5, 2020
    Publication date: February 10, 2022
    Applicant: Baidu USA LLC
    Inventors: Dingcheng LI, Shaogang REN, Ping LI
  • Publication number: 20220012425
    Abstract: Described herein are embodiments of a framework named as total correlation variational autoencoder (TC_VAE) to disentangle syntax and semantics by making use of total correlation penalties of KL divergences. One or more Kullback-Leibler (KL) divergence terms in a loss for a variational autoencoder are discomposed so that generated hidden variables may be separated. Embodiments of the TC_VAE framework were examined on semantic similarity tasks and syntactic similarity tasks. Experimental results show that better disentanglement between syntactic and semantic representations have been achieved compared with state-of-the-art (SOTA) results on the same data sets in similar settings.
    Type: Application
    Filed: July 10, 2020
    Publication date: January 13, 2022
    Applicant: Baidu USA LLC
    Inventors: Dingcheng LI, Shaogang REN, Ping LI
  • Publication number: 20210319302
    Abstract: The thriving of deep models and generative models provides approaches to model high dimensional distributions. Generative adversarial networks (GANs) can approximate data distributions and generate data samples from the learned data manifolds as well. Presented herein are embodiments to estimate the implicit likelihoods of GAN models. In one or more embodiments, a stable inverse function of the generator is learned with the help of a variance network of the generator. The local variance of the sample distribution may be approximated by the normalized distance in the latent space. Simulation studies and likelihood testing on data sets validate embodiments, which outperformed several baseline methods in these tasks. An embodiment was also applied to anomaly detection. Experiments show that the embodiments herein can achieve state-of-the-art anomaly detection performance.
    Type: Application
    Filed: April 3, 2020
    Publication date: October 14, 2021
    Applicant: Baidu USA LLC
    Inventors: Dingcheng LI, Shaogang REN, Zhixin ZHOU, Ping LI