Patents by Inventor Xihui LIU

Xihui LIU has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11809822
    Abstract: Certain embodiments involve a method for generating a search result. The method includes processing devices performing operations including receiving a query having a text input by a joint embedding model trained to generate an image result. Training the joint embedding model includes accessing a set of images and textual information. Training further includes encoding the images into image feature vectors based on spatial features. Further, training includes encoding the textual information into textual feature vectors based on semantic information. Training further includes generating a set of image-text pairs based on matches between image feature vectors and textual feature vectors. Further, training includes generating a visual grounding dataset based on spatial information. Training further includes generating a set of visual-semantic joint embeddings by grounding the image-text pairs with the visual grounding dataset.
    Type: Grant
    Filed: February 27, 2020
    Date of Patent: November 7, 2023
    Assignee: Adobe Inc.
    Inventors: Zhe Lin, Xihui Liu, Quan Tran, Jianming Zhang, Handong Zhao
  • Publication number: 20230274472
    Abstract: Apparatuses, systems, and techniques are presented to generate one or more images. In at least one embodiment, one or more neural networks are used to generate one or more images of one or more objects based, at least in part, on a model of the one or more objects and texture information.
    Type: Application
    Filed: April 27, 2023
    Publication date: August 31, 2023
    Inventors: Xihui LIU, Ming-Yu LIU, Ting-Chun WANG
  • Patent number: 11574142
    Abstract: The technology described herein is directed to a reinforcement learning based framework for training a natural media agent to learn a rendering policy without human supervision or labeled datasets. The reinforcement learning based framework feeds the natural media agent a training dataset to implicitly learn the rendering policy by exploring a canvas and minimizing a loss function. Once trained, the natural media agent can be applied to any reference image to generate a series (or sequence) of continuous-valued primitive graphic actions, e.g., sequence of painting strokes, that when rendered by a synthetic rendering environment on a canvas, reproduce an identical or transformed version of the reference image subject to limitations of an action space and the learned rendering policy.
    Type: Grant
    Filed: July 30, 2020
    Date of Patent: February 7, 2023
    Assignee: Adobe Inc.
    Inventors: Zhe Lin, Xihui Liu, Quan Hung Tran, Jianming Zhang, Handong Zhao
  • Patent number: 11455788
    Abstract: A method and apparatus for positioning a description statement in an image includes: analyzing a to-be-analyzed description statement and a to-be-analyzed image to obtain a plurality of statement attention weights of the to-be-analyzed description statement and a plurality of image attention weights of the to-be-analyzed image; obtaining a plurality of first matching scores based on the plurality of statement attention weights and a subject feature, a location feature and a relationship feature of the to-be-analyzed image; obtaining a second matching score between the to-be-analyzed description statement and the to-be-analyzed image based on the plurality of first matching scores and the plurality of image attention weights; and determining a positioning result of the to-be-analyzed description statement in the to-be-analyzed image based on the second matching score.
    Type: Grant
    Filed: March 24, 2020
    Date of Patent: September 27, 2022
    Assignee: BEIJING SENSETIME TECHNOLOGY DEVELOPMENT CO., LTD.
    Inventors: Xihui Liu, Jing Shao, Zihao Wang, Hongsheng Li, Xiaogang Wang
  • Publication number: 20220207770
    Abstract: Apparatuses, systems, and techniques to produce an image of a first subject positioned in a pose demonstrated by an image of a second subject. In at least one embodiment, an image of a first subject can be generated from a variety of points of view.
    Type: Application
    Filed: February 2, 2021
    Publication date: June 30, 2022
    Inventors: Ming-Yu Liu, Ting-Chun Wang, Xihui Liu
  • Publication number: 20220036127
    Abstract: The technology described herein is directed to a reinforcement learning based framework for training a natural media agent to learn a rendering policy without human supervision or labeled datasets. The reinforcement learning based framework feeds the natural media agent a training dataset to implicitly learn the rendering policy by exploring a canvas and minimizing a loss function. Once trained, the natural media agent can be applied to any reference image to generate a series (or sequence) of continuous-valued primitive graphic actions, e.g., sequence of painting strokes, that when rendered by a synthetic rendering environment on a canvas, reproduce an identical or transformed version of the reference image subject to limitations of an action space and the learned rendering policy.
    Type: Application
    Filed: July 30, 2020
    Publication date: February 3, 2022
    Inventors: Zhe Lin, Xihui Liu, Quan Hung Tran, Jianming Zhang, Handong Zhao
  • Publication number: 20210295115
    Abstract: A method and device for cross-modal information retrieval, and a storage medium are provided. The method includes: acquiring first modal information and second modal information; performing feature fusion on a modal feature of the first modal information and a modal feature of the second modal information, and determining a first fused feature corresponding to the first modal information and a second fused feature corresponding to the second modal information; and determining the degree of similarity between the first modal information and the second modal information on the basis of the first fused feature and the second fused feature.
    Type: Application
    Filed: June 3, 2021
    Publication date: September 23, 2021
    Inventors: Zihao WANG, Xihui Liu, Jing Shao, Hongsheng Li, Lv Sheng, Junjie Yan, Xiaogang Wang
  • Publication number: 20210271707
    Abstract: Certain embodiments involve a method for generating a search result. The method includes processing devices performing operations including receiving a query having a text input by a joint embedding model trained to generate an image result. Training the joint embedding model includes accessing a set of images and textual information. Training further includes encoding the images into image feature vectors based on spatial features. Further, training includes encoding the textual information into textual feature vectors based on semantic information. Training further includes generating a set of image-text pairs based on matches between image feature vectors and textual feature vectors. Further, training includes generating a visual grounding dataset based on spatial information. Training further includes generating a set of visual-semantic joint embeddings by grounding the image-text pairs with the visual grounding dataset.
    Type: Application
    Filed: February 27, 2020
    Publication date: September 2, 2021
    Inventors: Zhe Lin, Xihui Liu, Quan Tran, Jianming Zhang, Handong Zhao
  • Publication number: 20200226410
    Abstract: A method and apparatus for positioning a description statement in an image includes: analyzing a to-be-analyzed description statement and a to-be-analyzed image to obtain a plurality of statement attention weights of the to-be-analyzed description statement and a plurality of image attention weights of the to-be-analyzed image; obtaining a plurality of first matching scores based on the plurality of statement attention weights and a subject feature, a location feature and a relationship feature of the to-be-analyzed image; obtaining a second matching score between the to-be-analyzed description statement and the to-be-analyzed image based on the plurality of first matching scores and the plurality of image attention weights; and determining a positioning result of the to-be-analyzed description statement in the to-be-analyzed image based on the second matching score.
    Type: Application
    Filed: March 24, 2020
    Publication date: July 16, 2020
    Inventors: Xihui LIU, Jing SHAO, Zihao WANG, Hongsheng LI, Xiaogang WANG