Patents by Inventor Shaogang REN
Shaogang REN has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11947908Abstract: Described herein are system and method embodiments to improve word representation learning. Embodiments of a probabilistic prior may seamlessly integrate statistical disentanglement with word embedding. Different from previous deterministic methods, word embedding may be taken as a probabilistic generative model, and it enables imposing a prior that may identify independent factors generating word representation vectors. The probabilistic prior not only enhances the representation of word embedding, but also improves the model's robustness and stability. Furthermore, embodiments of the disclosed method may be flexibly plugged in various word embedding models. Extensive experimental results show that embodiments of the presented method may improve word representation on different tasks.Type: GrantFiled: April 7, 2021Date of Patent: April 2, 2024Assignee: Baidu USA LLCInventors: Shaogang Ren, Ping Li
-
Patent number: 11816533Abstract: Learning disentangled representations is an important topic in machine learning for a wide range of applications. Disentangled latent variables represent interpretable semantic information and reflect separate factors of variation in data. Although generative models may learn latent representations and generate data samples as well, existing models may ignore the structural information among latent representations. Described in the present disclosure are embodiments to learn disentangled latent structural representations from data using decomposable variational auto-encoders, which simultaneously learn component representations and encode component relationships. Embodiments of a novel structural prior for latent representations are disclosed to capture interactions among different data components. Embodiments are applied to data segmentation and latent relation discovery among different data components. Experiments on several datasets demonstrate the utility of the present model embodiments.Type: GrantFiled: November 18, 2020Date of Patent: November 14, 2023Assignee: Baidu USA LLCInventors: Shaogang Ren, Hongliang Fei, Dingcheng Li, Ping Li
-
Patent number: 11783198Abstract: The thriving of deep models and generative models provides approaches to model high dimensional distributions. Generative adversarial networks (GANs) can approximate data distributions and generate data samples from the learned data manifolds as well. Presented herein are embodiments to estimate the implicit likelihoods of GAN models. In one or more embodiments, a stable inverse function of the generator is learned with the help of a variance network of the generator. The local variance of the sample distribution may be approximated by the normalized distance in the latent space. Simulation studies and likelihood testing on data sets validate embodiments, which outperformed several baseline methods in these tasks. An embodiment was also applied to anomaly detection. Experiments show that the embodiments herein can achieve state-of-the-art anomaly detection performance.Type: GrantFiled: April 3, 2020Date of Patent: October 10, 2023Assignee: Baidu USA LLCInventors: Dingcheng Li, Shaogang Ren, Zhixin Zhou, Ping Li
-
Patent number: 11748567Abstract: Described herein are embodiments of a framework named as total correlation variational autoencoder (TC_VAE) to disentangle syntax and semantics by making use of total correlation penalties of KL divergences. One or more Kullback-Leibler (KL) divergence terms in a loss for a variational autoencoder are discomposed so that generated hidden variables may be separated. Embodiments of the TC_VAE framework were examined on semantic similarity tasks and syntactic similarity tasks. Experimental results show that better disentanglement between syntactic and semantic representations have been achieved compared with state-of-the-art (SOTA) results on the same data sets in similar settings.Type: GrantFiled: July 10, 2020Date of Patent: September 5, 2023Assignee: Baidu USA LLCInventors: Dingcheng Li, Shaogang Ren, Ping Li
-
Publication number: 20220335216Abstract: Described herein are system and method embodiments to improve word representation learning. Embodiments of a probabilistic prior may seamlessly integrate statistical disentanglement with word embedding. Different from previous deterministic methods, word embedding may be taken as a probabilistic generative model, and it enables imposing a prior that may identify independent factors generating word representation vectors. The probabilistic prior not only enhances the representation of word embedding, but also improves the model's robustness and stability. Furthermore, embodiments of the disclosed method may be flexibly plugged in various word embedding models. Extensive experimental results show that embodiments of the presented method may improve word representation on different tasks.Type: ApplicationFiled: April 7, 2021Publication date: October 20, 2022Applicant: Baidu USA LLCInventors: Shaogang REN, Ping LI
-
Publication number: 20220156612Abstract: Learning disentangled representations is an important topic in machine learning for a wide range of applications. Disentangled latent variables represent interpretable semantic information and reflect separate factors of variation in data. Although generative models may learn latent representations and generate data samples as well, existing models may ignore the structural information among latent representations. Described in the present disclosure are embodiments to learn disentangled latent structural representations from data using decomposable variational auto-encoders, which simultaneously learn component representation and encodes component relationships. Embodiments of a novel structural prior for latent representations are disclosed to capture interactions among different data components. Embodiments are applied to data segmentation and latent relation discovery among different data components. Experiments on several datasets demonstrate the utility of the present model embodiments.Type: ApplicationFiled: November 18, 2020Publication date: May 19, 2022Applicant: Baidu USA LLCInventors: Shaogang REN, Hongliang FEI, Dingcheng LI, Ping LI
-
Publication number: 20220043975Abstract: Described herein are embodiments of a framework named decomposable variational autoencoder (DecVAE) to disentangle syntax and semantics by using total correlation penalties of Kullback-Leibler (KL) divergences. KL divergence term of the original VAE are decomposed such that the hidden variables generated may be separated in a clear-cut and interpretable way. Embodiments of DecVAE models are evaluated on various semantic similarity and syntactic similarity datasets. Experimental results show that embodiments of DecVAE models achieve state-of-the-art (SOTA) performance in disentanglement between syntactic and semantic representations.Type: ApplicationFiled: August 5, 2020Publication date: February 10, 2022Applicant: Baidu USA LLCInventors: Dingcheng LI, Shaogang REN, Ping LI
-
Publication number: 20220012425Abstract: Described herein are embodiments of a framework named as total correlation variational autoencoder (TC_VAE) to disentangle syntax and semantics by making use of total correlation penalties of KL divergences. One or more Kullback-Leibler (KL) divergence terms in a loss for a variational autoencoder are discomposed so that generated hidden variables may be separated. Embodiments of the TC_VAE framework were examined on semantic similarity tasks and syntactic similarity tasks. Experimental results show that better disentanglement between syntactic and semantic representations have been achieved compared with state-of-the-art (SOTA) results on the same data sets in similar settings.Type: ApplicationFiled: July 10, 2020Publication date: January 13, 2022Applicant: Baidu USA LLCInventors: Dingcheng LI, Shaogang REN, Ping LI
-
Publication number: 20210319302Abstract: The thriving of deep models and generative models provides approaches to model high dimensional distributions. Generative adversarial networks (GANs) can approximate data distributions and generate data samples from the learned data manifolds as well. Presented herein are embodiments to estimate the implicit likelihoods of GAN models. In one or more embodiments, a stable inverse function of the generator is learned with the help of a variance network of the generator. The local variance of the sample distribution may be approximated by the normalized distance in the latent space. Simulation studies and likelihood testing on data sets validate embodiments, which outperformed several baseline methods in these tasks. An embodiment was also applied to anomaly detection. Experiments show that the embodiments herein can achieve state-of-the-art anomaly detection performance.Type: ApplicationFiled: April 3, 2020Publication date: October 14, 2021Applicant: Baidu USA LLCInventors: Dingcheng LI, Shaogang REN, Zhixin ZHOU, Ping LI