Patents by Inventor Zhiyu Cheng

Zhiyu Cheng has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11769327
    Abstract: Presented herein are systems, methods, and datasets for automatically and precisely generating highlight or summary videos of content. For example, in one or more embodiments, videos of sporting events may be digested or condensed into highlights, which will dramatically benefit sports media, broadcasters, video creators or commentators, or other short video creators, in terms of cost reduction, fast, and mass production, and saving tedious engineering hours. Embodiment of the framework may also be used or adapted for use to better promote sports teams, players, and/or games, and produce stories to glorify the spirit of sports or its players. While presented in the context of sports, it shall be noted that the methodologies may be used for videos comprising other content and events.
    Type: Grant
    Filed: August 3, 2021
    Date of Patent: September 26, 2023
    Assignee: Baidu USA LLC
    Inventors: Zhiyu Cheng, Le Kang, Xin Zhou, Hao Tian, Xing Li
  • Patent number: 11681920
    Abstract: Embodiments of the present disclosure disclose a method and apparatus for compressing a deep learning model. An embodiment of the method includes: acquiring a to-be-compressed deep learning model; pruning each layer of weights of the to-be-compressed deep learning model in units of channels to obtain a compressed deep learning model; and sending the compressed deep learning model to a terminal device, so that the terminal device stores the compressed deep learning model. By pruning each layer of weights of the deep learning model in units of channels, the parameter redundancy of the deep learning model is effectively reduced, thereby improving the computational speed of the deep learning model and maintaining the model accuracy.
    Type: Grant
    Filed: September 27, 2019
    Date of Patent: June 20, 2023
    Assignee: BAIDU USA LLC
    Inventors: Zhiyu Cheng, Yingze Bao
  • Patent number: 11640528
    Abstract: A method for information processing for accelerating neural network training. The method includes: acquiring a neural network corresponding to a deep learning task; and performing iterations of iterative training on the neural network based on a training data set. The training data set includes task data corresponding to the deep learning task. The iterative training includes: processing the task data in the training data set using a current neural network, and determining, based on a processing result of the neural network on the task data in a current iterative training, prediction loss of the current iterative training; determining a learning rate and a momentum in the current iterative training; and updating weight parameters of the current neural network by gradient descent based on a preset weight decay, and the learning rate, the momentum, and the prediction loss in the current iterative training. This method achieves efficient and low-cost deep learning-based neural network training.
    Type: Grant
    Filed: October 22, 2019
    Date of Patent: May 2, 2023
    Assignee: Baidu USA LLC
    Inventors: Zhiyu Cheng, Baopu Li, Yingze Bao
  • Publication number: 20230114661
    Abstract: The present disclosure provides a multifunctional climbing operation platform and an operation method thereof. The multifunctional climbing operation platform includes a main machine, clamping devices, a bolt retightening device, and an auxiliary sleeve replacement box, wherein the clamping devices each include: a fixed seat; and a pair of claw bars swingably arranged on the fixed seat and each including a rear bar body and a front bar body hinged together, wherein when the pair of claw bars are opened, an inner angle between the rear bar body and the front bar body is greater than 180°, the main machine and the clamping devices can climb along an object to be climbed, the bolt retightening device can replace different sleeves from the auxiliary sleeve replacement box, and can screw a bolt on the object. The present disclosure has the advantages of a large avoidance range and diverse functions.
    Type: Application
    Filed: December 13, 2022
    Publication date: April 13, 2023
    Inventors: Jinfeng Zhang, Jun Liu, Zhiyu Cheng, Lei Sun, Zhiwen Gong, Bingyu Sun, Tianzhong Zhang, Shenghe Wang, Yihua Luo, Jie Huang, Daping Liu, Yong Liu, Chengzhi Liu, Daojing Wang, Lanbo Yao, Xingyuan Guo
  • Publication number: 20230055636
    Abstract: With rapidly evolving technologies and emerging tools, sports-related videos generated online are rapidly increasing. To automate the sports video editing/highlight generation process, a key task is to precisely recognize and locate events-of-interest in videos. Embodiments herein comprise a two-stage paradigm to detect categories of events and when these events happen in videos. In one or more embodiments, multiple action recognition models extract high-level semantic features, and a transformer-based temporal detection module locates target events. These novel approaches achieved state-of-the-art performance in both action spotting and replay grounding. While presented in the context of sports, it shall be noted that the systems and methods herein may be used for videos comprising other content and events.
    Type: Application
    Filed: January 10, 2022
    Publication date: February 23, 2023
    Applicant: Baidu USA LLC
    Inventors: Zhiyu CHENG, Le KANG, Xin ZHOU, Hao TIAN, Xing LI, Bo HE, Jingyu XIN
  • Patent number: 11443173
    Abstract: Embodiments disclose an artificial intelligence chip and a convolutional neural network applied to the artificial intelligence chip comprising a processor, at least one parallel computing unit, and a pooling computation unit. The method comprises: dividing a convolution task into convolution subtasks and corresponding pooling subtasks; executing convolution subtasks at different parallel computing units, and performing convolution, batch normalization, and non-linear computing operation in a same parallel computing unit; sending an execution result of each parallel computing unit from executing the convolution subtask to the pooling computation unit for executing the corresponding pooling subtask; merging executing results of the pooling computation unit from performing pooling operations on the executing results outputted by respective convolution subtasks to obtain an execution result of the convolution task.
    Type: Grant
    Filed: April 24, 2019
    Date of Patent: September 13, 2022
    Assignee: BAIDU USA LLC
    Inventors: Zhiyu Cheng, Haofeng Kou, Yingze Bao
  • Publication number: 20220189173
    Abstract: Presented herein are systems, methods, and datasets for automatically and precisely generating highlight or summary videos of content. In one or more embodiments, the inputs comprise a text (e.g., an article) of the key event(s) (e.g., a goal, a player action, etc.) in an activity (e.g., a game, a concert, etc.) and a video or videos of the activity. In one or more embodiments, the output is a short video of an event or events in the text, in which the video may include commentary about the highlighted events and/or other audio (e.g., music), which may also be automatically synthesized.
    Type: Application
    Filed: November 23, 2021
    Publication date: June 16, 2022
    Applicant: Baidu USA LLC
    Inventors: Xin ZHOU, Le KANG, Zhiyu CHENG, Hao TIAN, Daming LU, Dapeng LI, Jingya XUN, Jeff WANG, Xi CHEN, Xing LI
  • Publication number: 20220188550
    Abstract: Presented herein are systems, methods, and datasets for automatically and precisely generating highlight or summary videos of content. For example, in one or more embodiments, videos of sporting events may be digested or condensed into highlights, which will dramatically benefit sports media, broadcasters, video creators or commentators, or other short video creators, in terms of cost reduction, fast, and mass production, and saving tedious engineering hours. Embodiment of the framework may also be used or adapted for use to better promote sports teams, players, and/or games, and produce stories to glorify the spirit of sports or its players. While presented in the context of sports, it shall be noted that the methodology may be used for videos comprising other content and events.
    Type: Application
    Filed: August 3, 2021
    Publication date: June 16, 2022
    Applicant: Baidu USA LLC
    Inventors: Zhiyu CHENG, Le KANG, Xin ZHOU, Hao TIAN, Xing LI
  • Publication number: 20220121926
    Abstract: Various embodiments for tensor decomposition to compress neural network models are presented. In one or more embodiments, one or more neural network layers from a neural network model are compressed using tensor ring (TR) decomposition. In one or more embodiments, a TR-decomposed neural network utilizes fewer resources and can be more readily deployed on resource-constraint devices. With a smaller model size, during inference time, the compressed model runs faster than its original model and consumes less power. Furthermore, the TR-compressed model can achieve better performance in terms of stability and model accuracy, compared to other tensor decomposition methods.
    Type: Application
    Filed: October 21, 2020
    Publication date: April 21, 2022
    Applicant: Baidu USA LLC
    Inventor: Zhiyu CHENG
  • Patent number: 11144790
    Abstract: Presented herein are embodiments of a training deep learning models. In one or more embodiments, a compact deep learning model comprises fewer layers, which require fewer floating-point operations (FLOPs). Presented herein are also embodiments of a new learning rate function, which can adaptively change the learning rate between two linear functions. In one or more embodiments, combinations of half-precision floating point format training together with larger batch size in the training process may also be employed to aid the training process.
    Type: Grant
    Filed: October 11, 2019
    Date of Patent: October 12, 2021
    Assignee: Baidu USA LLC
    Inventors: Baopu Li, Zhiyu Cheng, Yingze Bao
  • Publication number: 20210241094
    Abstract: Tensor decomposition can be advantageous for compressing deep neural networks (DNNs). In many applications of DNNs, reducing the number of parameters and computation workload is helpful to accelerate inference speed in deployment. Modern DNNs comprise multiple layers with multi-array weights where tensor decomposition is a natural way to perform compression—in which the weight tensors in convolutional layers or fully-connected layers are decomposed with specified tensor ranks (e.g., canonical ranks, tensor train ranks). Conventional tensor decomposition with DNNs involves selecting ranks manually, which requires tedious human efforts to finetune the performance. Accordingly, presented herein are rank selection embodiments, which are inspired by reinforcement learning, to automatically select ranks in tensor decomposition.
    Type: Application
    Filed: November 26, 2019
    Publication date: August 5, 2021
    Applicants: Baidu USA LLC, Baidu.com Times Technology (Beijing) Co., Ltd.
    Inventors: Zhiyu CHENG, Baopu LI, Yanwen FAN, Yingze BAO
  • Publication number: 20210232890
    Abstract: Deep neural networks (DNN) model quantization may be used to reduce storage and computation burdens by decreasing the bit width. Presented herein are novel cursor-based adaptive quantization embodiments. In embodiments, a multiple bits quantization mechanism is formulated as a differentiable architecture search (DAS) process with a continuous cursor that represents a possible quantization bit. In embodiments, the cursor-based DAS adaptively searches for a quantization bit for each layer. The DAS process may be accelerated via an alternative approximate optimization process, which is designed for mixed quantization scheme of a DNN model. In embodiments, a new loss function is used in the search process to simultaneously optimize accuracy and parameter size of the model. In a quantization step, the closest two integers to the cursor may be adopted as the bits to quantize the DNN together to reduce the quantization noise and avoid the local convergence problem.
    Type: Application
    Filed: September 24, 2019
    Publication date: July 29, 2021
    Applicants: Baidu USA LLC, Baidu.com Times Technology (Beijing) Co., Ltd.
    Inventors: Baopu LI, Yanwen FAN, Zhiyu CHENG, Yingze BAO
  • Publication number: 20210117776
    Abstract: A method for information processing for accelerating neural network training. The method includes: acquiring a neural network corresponding to a deep learning task; and performing iterations of iterative training on the neural network based on a training data set. The training data set includes task data corresponding to the deep learning task. The iterative training includes: processing the task data in the training data set using a current neural network, and determining, based on a processing result of the neural network on the task data in a current iterative training, prediction loss of the current iterative training; determining a learning rate and a momentum in the current iterative training; and updating weight parameters of the current neural network by gradient descent based on a preset weight decay, and the learning rate, the momentum, and the prediction loss in the current iterative training. This method achieves efficient and low-cost deep learning-based neural network training.
    Type: Application
    Filed: October 22, 2019
    Publication date: April 22, 2021
    Inventors: Zhiyu Cheng, Baopu Li, Yingze Bao
  • Publication number: 20210110213
    Abstract: Presented herein are embodiments of a training deep learning models. In one or more embodiments, a compact deep learning model comprises fewer layers, which require fewer floating-point operations (FLOPs). Presented herein are also embodiments of a new learning rate function, which can adaptively change the learning rate between two linear functions. In one or more embodiments, combinations of half-precision floating point format training together with larger batch size in the training process may also be employed to aid the training process.
    Type: Application
    Filed: October 11, 2019
    Publication date: April 15, 2021
    Applicant: Baidu USA LLC
    Inventors: Baopu LI, Zhiyu CHENG, Yingze BAO
  • Publication number: 20210097394
    Abstract: Embodiments of the present disclosure disclose a method and apparatus for compressing a deep learning model. An embodiment of the method includes: acquiring a to-be-compressed deep learning model; pruning each layer of weights of the to-be-compressed deep learning model in units of channels to obtain a compressed deep learning model; and sending the compressed deep learning model to a terminal device, so that the terminal device stores the compressed deep learning model. By pruning each layer of weights of the deep learning model in units of channels, the parameter redundancy of the deep learning model is effectively reduced, thereby improving the computational speed of the deep learning model and maintaining the model accuracy.
    Type: Application
    Filed: September 27, 2019
    Publication date: April 1, 2021
    Inventors: Zhiyu CHENG, Yingze BAO
  • Publication number: 20200342292
    Abstract: Embodiments disclose an artificial intelligence chip and a convolutional neural network applied to the artificial intelligence chip comprising a processor, at least one parallel computing unit, and a pooling computation unit. The method comprises: dividing a convolution task into convolution subtasks and corresponding pooling subtasks; executing convolution subtasks at different parallel computing units, and performing convolution, batch normalization, and non-linear computing operation in a same parallel computing unit; sending an execution result of each parallel computing unit from executing the convolution subtask to the pooling computation unit for executing the corresponding pooling subtask; merging executing results of the pooling computation unit from performing pooling operations on the executing results outputted by respective convolution subtasks to obtain an execution result of the convolution task.
    Type: Application
    Filed: April 24, 2019
    Publication date: October 29, 2020
    Inventors: Zhiyu CHENG, Haofeng KOU, Yingze BAO
  • Patent number: 10038463
    Abstract: Digital pre-distortion is performed on a received signal using a set of pre-distortion coefficients to produce a digital pre-distorted signal. The digital pre-distorted signal is converted to an analog signal, which is amplified to produce a transmission output signal. The transmission output signal is converted to a digital feedback signal. A plurality of fractional delay filters is applied to the digital feedback signal to obtain a plurality of fractional delay compensated (FDC) candidates, and gain compensation is applied to each of the plurality of FDC candidates to obtain a plurality of gain and fractional delay compensated (XFT) candidates. The digital pre-distorted signal is used as a reference signal, and the XFT candidates and the reference signal are used to select a selected XFT candidate of the plurality of XFT candidates. The selected XFT candidate is used to generate the set of pre-distortion coefficients.
    Type: Grant
    Filed: November 14, 2017
    Date of Patent: July 31, 2018
    Assignee: NXP USA, Inc.
    Inventors: Jayakrishnan Cheriyath Mundarath, Zhiyu Cheng, Leo Dehner