Patents by Inventor MINGFA ZHU

MINGFA ZHU has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11902346
    Abstract: The present disclosure provides a method and apparatus for processing a streaming media service, an electronic device, and a storage medium, and relates to the technical field of computers, particularly to technical fields such as industrial vision, deep learning, streaming media, and information flow. A specific implementation solution involves: acquiring registration information of an input source, the registration information including process information of a streaming media service process of the input source and streaming media address information of the input source; enabling the streaming media service process according to the process information; and controlling, by using the streaming media address information, the streaming media service process to process streaming media data of the input source.
    Type: Grant
    Filed: October 28, 2022
    Date of Patent: February 13, 2024
    Assignee: BEIJING BAIDU NETCOM SCIENCE TECHNOLOGY CO., LTD.
    Inventors: Shuo Li, Xuechao Wei, Yonggao Fu, Jiabing Leng, Yawen Liu, Mingfa Zhu, Feng Huang
  • Publication number: 20230319124
    Abstract: The present disclosure provides a method and apparatus for processing a streaming media service, an electronic device, and a storage medium, and relates to the technical field of computers, particularly to technical fields such as industrial vision, deep learning, streaming media, and information flow. A specific implementation solution involves: acquiring registration information of an input source, the registration information including process information of a streaming media service process of the input source and streaming media address information of the input source; enabling the streaming media service process according to the process information; and controlling, by using the streaming media address information, the streaming media service process to process streaming media data of the input source.
    Type: Application
    Filed: October 28, 2022
    Publication date: October 5, 2023
    Applicant: BEIJING BAIDU NETCOM SCIENCE TECHNOLOGY CO., LTD.
    Inventors: Shuo LI, Xuechao WEI, Yonggao FU, Jiabing LENG, Yawen LIU, Mingfa ZHU, Feng HUANG
  • Patent number: 11282516
    Abstract: Embodiments of the present disclosure provide a human-machine interaction processing method, an apparatus thereof, a user terminal, a processing server and a system. On the user terminal side, the method includes: receiving an interaction request voice inputted from a user, and collecting video data of the user when inputting the interaction request voice; obtaining an interaction response voice corresponding to the interaction request voice, where the interaction response voice is obtained according to expression information of the user when inputting the interaction request voice and included in the video data; and outputting the interaction response voice to the user. The method imbues the interaction response voice with an emotional tone that matches the current emotion of the user, so that the human-machine interaction process is no longer monotonous, greatly enhancing the user experience.
    Type: Grant
    Filed: February 18, 2019
    Date of Patent: March 22, 2022
    Assignee: BEIJING BAIDU NETCOM SCIENCE TECHNOLOGY CO., LTD.
    Inventors: Shuangshuang Qiao, Kun Liu, Yang Liang, Xiangyue Lin, Chao Han, Mingfa Zhu, Jiangliang Guo, Xu Li, Jun Liu, Shuo Li, Shiming Yin
  • Patent number: 11138903
    Abstract: The present disclosure provides a method, an apparatus, a device and a system for sign language translation, where a server receives video information sent by a terminal device, and preprocesses the video information to obtain at least one sign language action; the at least one sign language action is input into a sign language model for classification and prediction to obtain a word corresponding to the at least one sign language action; each word is input into a language model to determine whether an intention expression is complete; and each word is sent to the terminal device when the intention expression is complete, so that the terminal device displays each word, thereby realizing the translation of the sign language action into text, enabling the ordinary persons to better understand intentions of the hearing impaired, thus improving efficiency of communications.
    Type: Grant
    Filed: February 18, 2019
    Date of Patent: October 5, 2021
    Assignee: BEIJING BAIDU NETCOM SCIENCE TECHNOLOGY CO., LTD.
    Inventors: Xiangyue Lin, Kun Liu, Shuangshuang Qiao, Yang Liang, Chao Han, Mingfa Zhu, Jiangliang Guo, Xu Li, Jun Liu, Shuo Li, Shiming Yin
  • Publication number: 20210012777
    Abstract: Embodiments of the present disclosure provide a context acquiring method based on voice interaction and a device, the method comprising: acquiring a scene image collected by an image collection device at a voice start point of a current conversation, and extracting a face feature of each user in the scene image; if it is determined that there is a second face feature matching a first face feature according to the face feature of each user and a face database, acquiring a first user identifier corresponding to the second face feature from the face database; if it is determined that a stored conversation corresponding to the first user identifier is stored in a voice database, determine a context of a voice interaction according to the current conversation and the stored conversation, and after the voice end point of the current conversation is obtained, storing the current conversation into the voice database.
    Type: Application
    Filed: July 23, 2020
    Publication date: January 14, 2021
    Applicant: Beijing Baidu Netcom Science Technology Co., Ltd.
    Inventors: Yang LIANG, Kun LIU, Shuangshuang QIAO, Xiangyue LIN, Chao HAN, Mingfa ZHU, Jiangliang GUO, Xu LI, Jun LIU, Shuo LI, Shiming YIN
  • Publication number: 20200005781
    Abstract: Embodiments of the present disclosure provide a human-machine interaction processing method, an apparatus thereof, a user terminal, a processing server and a system. On the user terminal side, the method includes: receiving an interaction request voice inputted from a user, and collecting video data of the user when inputting the interaction request voice; obtaining an interaction response voice corresponding to the interaction request voice, where the interaction response voice is obtained according to expression information of the user when inputting the interaction request voice and included in the video data; and outputting the interaction response voice to the user. The method imbues the interaction response voice with an emotional tone that matches the current emotion of the user, so that the human-machine interaction process is no longer monotonous, greatly enhancing the user experience.
    Type: Application
    Filed: February 18, 2019
    Publication date: January 2, 2020
    Inventors: SHUANGSHUANG QIAO, KUN LIU, YANG LIANG, XIANGYUE LIN, CHAO HAN, MINGFA ZHU, JIANGLIANG GUO, XU LI, JUN LIU, SHUO LI, SHIMING YIN
  • Publication number: 20200005673
    Abstract: The present disclosure provides a method, an apparatus, a device and a system for sign language translation, where a server receives video information sent by a terminal device, and preprocesses the video information to obtain at least one sign language action; the at least one sign language action is input into a sign language model for classification and prediction to obtain a word corresponding to the at least one sign language action; each word is input into a language model to determine whether an intention expression is complete; and each word is sent to the terminal device when the intention expression is complete, so that the terminal device displays each word, thereby realizing the translation of the sign language action into text, enabling the ordinary persons to better understand intentions of the hearing impaired, thus improving efficiency of communications.
    Type: Application
    Filed: February 18, 2019
    Publication date: January 2, 2020
    Inventors: XIANGYUE LIN, KUN LIU, SHUANGSHUANG QIAO, YANG LIANG, CHAO HAN, MINGFA ZHU, JIANGLIANG GUO, XU LI, JUN LIU, SHUO LI, SHIMING YIN