Patents by Inventor Sanqiang Zhao

Sanqiang Zhao has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20250246186
    Abstract: Synthetic conference transcripts are generated and used to train a natural processing engine to derive intelligence from conference recordings or conference transcripts. A server generates, using a natural language processing engine, synthetic conference transcripts. The server compares the synthetic conference transcripts with conference data to identify artifacts in the synthetic conference transcripts. The server provides additional training to the natural language processing engine using online learning based on the identified artifacts. The server outputs a portion of the synthetic conference transcripts selected based on the identified artifacts.
    Type: Application
    Filed: January 25, 2024
    Publication date: July 31, 2025
    Inventors: Yuanling Geng, Liwei Wu, Bing Zhao, Sanqiang Zhao
  • Publication number: 20250157463
    Abstract: Techniques for rendering visual content, in response to one or more utterances, are described. A device receives one or more utterances that define a parameter(s) for desired output content. A system (or the device) identifies natural language data corresponding to the desired content, and uses natural language generation processes to update the natural language data based on the parameter(s). The system (or the device) then generates an image based on the updated natural language data. The system (or the device) also generates video data of an avatar. The device displays the image and the avatar, and synchronizes movements of the avatar with output of synthesized speech of the updated natural language data. The device may also display subtitles of the updated natural language data, and cause a word of the subtitles to be emphasized when synthesized speech of the word is being output.
    Type: Application
    Filed: January 13, 2025
    Publication date: May 15, 2025
    Inventors: Taehwan Kim, Sanqiang Zhao, Robinson Piramuthu, Seokhwan Kim, Yang Liu, Gokhan Tur, Eshan Bhatnagar
  • Patent number: 12260340
    Abstract: Provided is a knowledge distillation technique for training a student language model that, relative to a larger teacher language model, has a significantly smaller vocabulary, lower embedding dimensions, and/or hidden state dimensions. Specifically, aspects of the present disclosure are directed to a dual-training mechanism that trains the teacher and student language models simultaneously to obtain optimal word embeddings for the student vocabulary. In some implementations, this approach can be combined with learning shared projection matrices that transfer layer-wise knowledge from the teacher language model to the student language model. Example experimental results have also demonstrated higher compression efficiency and accuracy when compared with other state-of-the-art compression techniques, including the ability to compress the BERTBASE model by more than 60×, with only a minor drop in downstream task metrics, resulting in a language model with a footprint of under 7 MB.
    Type: Grant
    Filed: September 21, 2023
    Date of Patent: March 25, 2025
    Assignee: GOOGLE LLC
    Inventors: Yang Song, Raghav Gupta, Dengyong Zhou, Sanqiang Zhao
  • Patent number: 12205577
    Abstract: Techniques for rendering visual content, in response to one or more utterances, are described. A device receives one or more utterances that define a parameter(s) for desired output content. A system (or the device) identifies natural language data corresponding to the desired content, and uses natural language generation processes to update the natural language data based on the parameter(s). The system (or the device) then generates an image based on the updated natural language data. The system (or the device) also generates video data of an avatar. The device displays the image and the avatar, and synchronizes movements of the avatar with output of synthesized speech of the updated natural language data. The device may also display subtitles of the updated natural language data, and cause a word of the subtitles to be emphasized when synthesized speech of the word is being output.
    Type: Grant
    Filed: March 30, 2021
    Date of Patent: January 21, 2025
    Assignee: Amazon Technologies, Inc.
    Inventors: Taehwan Kim, Sanqiang Zhao, Robinson Piramuthu, Seokhwan Kim, Yang Liu, Gokhan Tur, Eshan Bhatnagar
  • Publication number: 20240095987
    Abstract: Techniques for generating content associated with a user input/system generated response are described. Natural language data associated with a user input may be generated. For each portion of the natural language data, ambiguous references to entities in the portion may be replaced with the corresponding entity. Entities included in the portion may be extracted, and image data representing the entity may be determined. Background image data associated with the entities and the portion may be determined, and attributes which modify the entities in the natural language sentence may be extracted. Spatial relationships between two or more of the entities may further be extracted. Image data representing the natural language data may be generated based on the background image data, the entities, the attributes, and the spatial relationships. Video data may be generated based on the image data, where the video data includes animations of the entities moving.
    Type: Application
    Filed: December 14, 2022
    Publication date: March 21, 2024
    Inventors: Robinson Piramuthu, Sanqiang Zhao, Yadunandana Rao, Zhiyuan Fang
  • Publication number: 20240013059
    Abstract: Provided is a knowledge distillation technique for training a student language model that, relative to a larger teacher language model, has a significantly smaller vocabulary, lower embedding dimensions, and/or hidden state dimensions. Specifically, aspects of the present disclosure are directed to a dual-training mechanism that trains the teacher and student language models simultaneously to obtain optimal word embeddings for the student vocabulary. In some implementations, this approach can be combined with learning shared projection matrices that transfer layer-wise knowledge from the teacher language model to the student language model. Example experimental results have also demonstrated higher compression efficiency and accuracy when compared with other state-of-the-art compression techniques, including the ability to compress the BERTBASE model by more than 60×, with only a minor drop in downstream task metrics, resulting in a language model with a footprint of under 7 MB.
    Type: Application
    Filed: September 21, 2023
    Publication date: January 11, 2024
    Inventors: Yang Song, Raghav Gupta, Dengyong Zhou, Sanqiang Zhao
  • Patent number: 11797862
    Abstract: Provided is a knowledge distillation technique for training a student language model that, relative to a larger teacher language model, has a significantly smaller vocabulary, lower embedding dimensions, and/or hidden state dimensions. Specifically, aspects of the present disclosure are directed to a dual-training mechanism that trains the teacher and student language models simultaneously to obtain optimal word embeddings for the student vocabulary. In some implementations, this approach can be combined with learning shared projection matrices that transfer layer-wise knowledge from the teacher language model to the student language model. Example experimental results have also demonstrated higher compression efficiency and accuracy when compared with other state-of-the-art compression techniques, including the ability to compress the BERTBASE model by more than 60×, with only a minor drop in downstream task metrics, resulting in a language model with a footprint of under 7 MB.
    Type: Grant
    Filed: January 22, 2020
    Date of Patent: October 24, 2023
    Assignee: GOOGLE LLC
    Inventors: Yang Song, Raghav Gupta, Dengyong Zhou, Sanqiang Zhao
  • Publication number: 20210224660
    Abstract: Provided is a knowledge distillation technique for training a student language model that, relative to a larger teacher language model, has a significantly smaller vocabulary, lower embedding dimensions, and/or hidden state dimensions. Specifically, aspects of the present disclosure are directed to a dual-training mechanism that trains the teacher and student language models simultaneously to obtain optimal word embeddings for the student vocabulary. In some implementations, this approach can be combined with learning shared projection matrices that transfer layer-wise knowledge from the teacher language model to the student language model. Example experimental results have also demonstrated higher compression efficiency and accuracy when compared with other state-of-the-art compression techniques, including the ability to compress the BERTBAsE model by more than 60×, with only a minor drop in downstream task metrics, resulting in a language model with a footprint of under 7 MB.
    Type: Application
    Filed: January 22, 2020
    Publication date: July 22, 2021
    Inventors: Yang Song, Raghav Gupta, Dengyong Zhou, Sanqiang Zhao