Patents by Inventor Mingbo MA

Mingbo MA has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20230169281
    Abstract: Representation learning for text and speech has improved many language-related tasks. However, existing methods only learn from one input modality, while a unified representation for both speech and text is needed for tasks such as end-to-end speech translation. Consequently, these methods cannot exploit various large-scale text and speech data and their performance is limited by the scarcity of parallel speech translation data. To address these problems, embodiments of a fused acoustic and text masked language model (FAT-MLM) are disclosed. FAT-MLM embodiments jointly learn a unified representation for both acoustic and text input from various types of corpora including parallel data for speech recognition and machine translation, and pure speech and text data. Within this cross-modal representation learning framework, an end-to-end model is further presented for fused acoustic and text speech translation.
    Type: Application
    Filed: November 23, 2021
    Publication date: June 1, 2023
    Applicant: Baidu USA LLC
    Inventors: Renjie ZHENG, Junkun CHEN, Mingbo MA, Liang HUANG
  • Patent number: 11126800
    Abstract: Presented herein are embodiments of a prefix-to-prefix framework for simultaneous translation that implicitly learns to anticipates in a single translation. Within these frameworks are effective “wait-k” policy model embodiments that may be trained to generate a target sentence concurrently with a source sentence but lag behind by a predefined number of words. Embodiments of the prefix-to-prefix framework achieve low latency and better quality when compared to full-sentence translation in four directions: Chinese?English and German?English. Also presented herein is a novel latency metric that addresses deficiencies of previous latency metrics.
    Type: Grant
    Filed: May 10, 2019
    Date of Patent: September 21, 2021
    Assignee: Baidu USA LLC.
    Inventors: Mingbo Ma, Liang Huang, Hao Xiong, Kaibo Liu, Chuanqiang Zhang, Renjie Zheng, Zhongjun He, Hairong Liu, Xing Li, Hua Wu, Haifeng Wang, Baigong Zheng
  • Publication number: 20200104371
    Abstract: Presented herein are embodiments of a prefix-to-prefix framework for simultaneous translation that implicitly learns to anticipates in a single translation. Within these frameworks are effective “wait-k” policy model embodiments that may be trained to generate a target sentence concurrently with a source sentence but lag behind by a predefined number of words. Embodiments of the prefix-to-prefix framework achieve low latency and better quality when compared to full-sentence translation in four directions: Chinese?English and German?English. Also presented herein is a novel latency metric that addresses deficiencies of previous latency metrics.
    Type: Application
    Filed: May 10, 2019
    Publication date: April 2, 2020
    Applicant: Baidu USA LLC
    Inventors: Mingbo MA, Liang HUANG, Hao XIONG, Kaibo LIU, Chuanqiang ZHANG, Renjie ZHENG, Zhongjun HE, Hairong LIU, Xing LI, Hua Wu, Haifeng WANG, Baigong ZHENG