Patents by Inventor Weibao GONG

Weibao GONG has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20230206080
    Abstract: A model training system includes at least one first cluster and a second cluster communicating with the at least first cluster. The at least one first cluster is configured to acquire a sample data set, generate training data according to the sample data set, and send the training data to the second cluster; and the second cluster is configured to train a pre-trained model according to the training data sent by the at least one first cluster.
    Type: Application
    Filed: March 7, 2023
    Publication date: June 29, 2023
    Inventors: Shuohuan WANG, Weibao GONG, Zhihua WU, Yu SUN, Siyu DING, Yaqian HAN, Yanbin ZHAO, Yuang LIU, Dianhai YU
  • Patent number: 11574146
    Abstract: A method for updating a parameter of a model, a distributed training system, and an electric device are related to a field of deep learning technologies. The method includes: obtaining a batch training period of batch training data to be trained for a model; increasing priorities of tasks ranked at a bottom in a sequence of gradient communication tasks for parameters of the model when the batch training period is greater than or equal to a preset period threshold; and performing a communication of gradients of the parameters and updating the parameters based on priorities of the gradient communication tasks for the parameters in the model.
    Type: Grant
    Filed: November 25, 2020
    Date of Patent: February 7, 2023
    Assignee: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY CO., LTD.
    Inventors: Long Li, Haifeng Wang, Weibao Gong
  • Publication number: 20220374713
    Abstract: The present disclosure provides a method and apparatus for performing distributed training on a deep learning model. The method may include: generating a distributed computation view based on data information of a to-be-trained deep learning model; generating a cluster resource view based on property information of a cluster hardware resource corresponding to the to-be-trained deep learning model; determining a target segmentation strategy of a distributed training task based on the distributed computation view and the cluster resource view; and performing distributed training on the to-be-trained deep learning model based on the target segmentation strategy.
    Type: Application
    Filed: August 3, 2022
    Publication date: November 24, 2022
    Inventors: Zhihua WU, Dianhai YU, Yulong AO, Weibao GONG
  • Publication number: 20210406767
    Abstract: The present application discloses a distributed training method and system, a device and a storage medium, and relates to technical fields of deep learning and cloud computing. The method includes: sending, by a task information server, a first training request and information of an available first computing server to at least a first data server; sending, by the first data server, a first batch of training data to the first computing server, according to the first training request; performing, by the first computing server, model training according to the first batch of training data, sending model parameters to the first data server so as to be stored after the training is completed, and sending identification information of the first batch of training data to the task information server so as to be recorded; wherein the model parameters are not stored at any one of the computing servers.
    Type: Application
    Filed: January 6, 2021
    Publication date: December 30, 2021
    Applicant: Beijing Baidu Netcom Science and Technology Co., Ltd.
    Inventors: Daxiang DONG, Weibao GONG, Yi LIU, Dianhai YU, Yanjun MA, Haifeng WANG
  • Publication number: 20210287044
    Abstract: A method for updating a parameter of a model, a distributed training system, and an electric device are related to a field of deep learning technologies. The method includes: obtaining a batch training period of batch training data to be trained for a model; increasing priorities of tasks ranked at the bottom in a sequence of gradient communication tasks for parameters of the model when the batch training period is greater than or equal to a preset period threshold; and performing a communication of gradients of the parameters and updating the parameters based on priorities of the gradient communication tasks for the parameters in the model.
    Type: Application
    Filed: November 25, 2020
    Publication date: September 16, 2021
    Inventors: Long LI, Haifeng WANG, Weibao GONG