Patents by Inventor Shuangfei Zhai

Shuangfei Zhai has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240161014
    Abstract: A method of modelling data, comprising training an objective function of a linear classifier, based on a set of labeled data, to derive a set of classifier weights; defining a posterior probability distribution on the set of classifier weights of the linear classifier; approximating a marginalized loss function for an autoencoder as a Bregman divergence, based on the posterior probability distribution on the set of classifier weights learned from the linear classifier; and classifying unlabeled data using the autoencoder according to the marginalized loss function.
    Type: Application
    Filed: January 7, 2024
    Publication date: May 16, 2024
    Inventors: Zhongfei Zhang, Shuangfei Zhai
  • Patent number: 11868862
    Abstract: A method of modelling data, comprising: training an objective function of a linear classifier, based on a set of labeled data, to derive a set of classifier weights; defining a posterior probability distribution on the set of classifier weights of the linear classifier; approximating a marginalized loss function for an autoencoder as a Bregman divergence, based on the posterior probability distribution on the set of classifier weights learned from the linear classifier; and classifying unlabeled data using the autoencoder according to the marginalized loss function.
    Type: Grant
    Filed: December 19, 2021
    Date of Patent: January 9, 2024
    Assignee: The Research Foundation for The State University of New York
    Inventors: Zhongfei Zhang, Shuangfei Zhai
  • Patent number: 11496769
    Abstract: Techniques for coding sets of images with neural networks include transforming a first image of a set of images into coefficients with an encoder neural network, encoding a group of the coefficients as an integer patch index into coding table of table entries each having vectors of coefficients, and storing a collection of patch indices as a first coded image. The encoder neural network may be configured with encoder weights determined by jointly with corresponding decoder weights of a decoder neural network on the set of images.
    Type: Grant
    Filed: March 31, 2020
    Date of Patent: November 8, 2022
    Assignee: APPLE INC.
    Inventors: Shuangfei Zhai, Joshua M. Susskind
  • Publication number: 20220114405
    Abstract: A method of modelling data, comprising: training an objective function of a linear classifier, based on a set of labeled data, to derive a set of classifier weights; defining a posterior probability distribution on the set of classifier weights of the linear classifier; approximating a marginalized loss function for an autoencoder as a Bregman divergence, based on the posterior probability distribution on the set of classifier weights learned from the linear classifier; and classifying unlabeled data using the autoencoder according to the marginalized loss function.
    Type: Application
    Filed: December 19, 2021
    Publication date: April 14, 2022
    Inventors: Zhongfei Zhang, Shuangfei Zhai
  • Publication number: 20220108212
    Abstract: Attention-free transformers are disclosed. Various implementations of attention-free transformers include a gating and pooling operation that allows the attention-free transformers to provide comparable or better results to those of a standard attention-based transformer, with improved efficiency and reduced computational complexity with respect to space and time.
    Type: Application
    Filed: May 4, 2021
    Publication date: April 7, 2022
    Inventors: Shuangfei ZHAI, Walter A. TALBOTT, Nitish SRIVASTAVA, Chen HUANG, Hanlin GOH, Joshua M. SUSSKIND
  • Patent number: 11205103
    Abstract: A method of modelling data, comprising: training an objective function of a linear classifier, based on a set of labeled data, to derive a set of classifier weights; defining a posterior probability distribution on the set of classifier weights of the linear classifier; approximating a marginalized loss function for an autoencoder as a Bregman divergence, based on the posterior probability distribution on the set of classifier weights learned from the linear classifier; and classifying unlabeled data using the autoencoder according to the marginalized loss function.
    Type: Grant
    Filed: December 11, 2017
    Date of Patent: December 21, 2021
    Assignee: The Research Foundation for The State University
    Inventors: Zhongfei Zhang, Shuangfei Zhai
  • Publication number: 20210099731
    Abstract: Techniques for coding sets of images with neural networks include transforming a first image of a set of images into coefficients with an encoder neural network, encoding a group of the coefficients as an integer patch index into coding table of table entries each having vectors of coefficients, and storing a collection of patch indices as a first coded image. The encoder neural network may be configured with encoder weights determined by jointly with corresponding decoder weights of a decoder neural network on the set of images.
    Type: Application
    Filed: March 31, 2020
    Publication date: April 1, 2021
    Inventors: Shuangfei ZHAI, Joshua M. SUSSKIND
  • Patent number: 10354182
    Abstract: A computer-implemented technique is described herein for identifying one or more content items that are relevant to an input linguistic item (e.g., an input query) using a deep-structured neural network, trained based on a corpus of click-through data. The input linguistic item has a collection of input tokens. The deep-structured neural network includes a first part that produces word embeddings associated with the respective input tokens, a second part that generates state vectors that capture context information associated with the input tokens, and a third part which distinguishes important parts of the input linguistic item from less important parts. The second part of the deep-structured neural network can be implemented as a recurrent neural network, such as a bi-directional neural network. The third part of the deep-structured neural network can generate a concept vector by forming a weighted sum of the state vectors.
    Type: Grant
    Filed: October 29, 2015
    Date of Patent: July 16, 2019
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Keng-hao Chang, Ruofei Zhang, Shuangfei Zhai
  • Publication number: 20180165554
    Abstract: A method of modelling data, comprising: training an objective function of a linear classifier, based on a set of labeled data, to derive a set of classifier weights; defining a posterior probability distribution on the set of classifier weights of the linear classifier; approximating a marginalized loss function for an autoencoder as a Bregman divergence, based on the posterior probability distribution on the set of classifier weights learned from the linear classifier; and classifying unlabeled data using the autoencoder according to the marginalized loss function.
    Type: Application
    Filed: December 11, 2017
    Publication date: June 14, 2018
    Inventors: Zhongfei Zhang, Shuangfei Zhai
  • Publication number: 20170124447
    Abstract: A computer-implemented technique is described herein for identifying one or more content items that are relevant to an input linguistic item (e.g., an input query) using a deep-structured neural network, trained based on a corpus of click-through data. The input linguistic item has a collection of input tokens. The deep-structured neural network includes a first part that produces word embeddings associated with the respective input tokens, a second part that generates state vectors that capture context information associated with the input tokens, and a third part which distinguishes important parts of the input linguistic item from less important parts. The second part of the deep-structured neural network can be implemented as a recurrent neural network, such as a bi-directional neural network. The third part of the deep-structured neural network can generate a concept vector by forming a weighted sum of the state vectors.
    Type: Application
    Filed: October 29, 2015
    Publication date: May 4, 2017
    Inventors: Keng-hao Chang, Ruofei Zhang, Shuangfei Zhai