Patents by Inventor Zhibiao Zhao

Zhibiao Zhao has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11822964
    Abstract: Embodiments of the disclosure discloses a method and system for a virtualization environment for a data processing (DP) accelerator. In one embodiment, a data processing (DP) accelerator includes one or more statically partitioned resources and one or more virtual functions (VFs) each associated with one of the one or more statically partitioned resources. A virtual machine (VM) of a host is assigned one of the one or more VFs to access the statically partitioned resources associated with the assigned VF. The VM has no access to the rest of the one or more statically partitioned resources of the DP accelerator.
    Type: Grant
    Filed: June 3, 2020
    Date of Patent: November 21, 2023
    Assignees: BAIDU USA LLC, KUNLUNXIN TECHNOLOGY (BEIJING) COMPANY LIMITED
    Inventors: Yueqiang Cheng, Zhibiao Zhao
  • Patent number: 11615295
    Abstract: A data processing system includes a central processing unit (CPU) and accelerator cards coupled to the CPU over a bus, each of the accelerator cards having a plurality of data processing (DP) accelerators to receive DP tasks from the CPU and to perform the received DP tasks. At least two of the accelerator cards are coupled to each other via an inter-card connection, and at least two of the DP accelerators are coupled to each other via an inter-chip connection. Each of the inter-card connection and the inter-chip connection is capable of being dynamically activated or deactivated, such that in response to a request received from the CPU, any one of the accelerator cards or any one of the DP accelerators within any one of the accelerator cards can be enabled or disabled to process any one of the DP tasks received from the CPU.
    Type: Grant
    Filed: November 15, 2019
    Date of Patent: March 28, 2023
    Assignees: BAIDU USA LLC, BAIDU.COM TIMES TECHNOLOGY (BEIJING) CO., LTD., KUNLUNXIN TECHNOLOGY (BEIJING) COMPANY LIMITED
    Inventors: Hefei Zhu, Jian Ouyang, Zhibiao Zhao, Xiaozhang Gong, Qingshu Chen
  • Patent number: 11595588
    Abstract: An image capturing method and a terminal device are provided. The method includes entering a camera application to start a lens and display a viewfinder interface, converting an original image captured by the lens into a red-green-blue (RGB) image, and decreasing luminance of the RGB image to be less than first luminance or increasing the luminance of the RGB image to be greater than second luminance, to obtain a first image; converting the RGB image into N frames of high-dynamic-range (HDR) images, and fusing color information of pixels in any same location on the first image and the N frames of HDR images to obtain a final image.
    Type: Grant
    Filed: October 10, 2019
    Date of Patent: February 28, 2023
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Liyao Zhang, Jing Ma, Zhibiao Zhao
  • Patent number: 11544067
    Abstract: According to various embodiments, methods and systems are provided to accelerate artificial intelligence (AI) model training with advanced interconnect communication technologies and systematic zero-value compression over a distributed training system. According to an exemplary method, during each iteration of a Scatter-Reduce process performed on a cluster of processors arranged in a logical ring to train a neural network model, a processor receives a compressed data block from a prior processor in the logical ring, performs an operation on the received compressed data block and a compressed data block generated on the processor to obtain a calculated data block, and sends the calculated data block to a following processor in the logical ring. A compressed data block calculated from corresponding data blocks from the processors can be identified on each processor and distributed to each other processor and decompressed therein for use in the AI model training.
    Type: Grant
    Filed: October 12, 2019
    Date of Patent: January 3, 2023
    Assignees: BAIDU USA LLC, BAIDU.COM TIMES TECHNOLOGY (BEIJING) CO., LTD., KUNLUNXIN TECHNOLOGY (BEIJING) COMPANY LIMITED
    Inventors: Zhibiao Zhao, Jian Ouyang, Hefei Zhu, Qingshu Chen, Wei Qi
  • Patent number: 11485376
    Abstract: An automatic processing system, a system on chip and a method for monitoring a processing module are described herein. The automatic driving processing system comprises: an automatic driving processing module, configured for receiving an input data stream and processing the input data stream based on a deep learning model so as to generate a processing result; a fault detection module, configured for generating a control signal and a fault detection stimulating data stream, and receiving the processing result from the automatic driving processing module; and a multi-way selection module, configured for receiving an automatic driving data stream as well as the control signal and the fault detection stimulating data stream, and selectively outputting the automatic driving data stream or the fault detection stimulating data stream to the automatic driving processing module based on the control signal, as an input data stream.
    Type: Grant
    Filed: December 11, 2019
    Date of Patent: November 1, 2022
    Assignees: Beijing Baidu Netcom Science And Technology Co., Ltd., Kunlunxin Technology (Beijing) Company Limited
    Inventors: Chongqin Wang, Zhibiao Zhao, Hefei Zhu, Ningyi Xu, Jian Ouyang
  • Publication number: 20220214903
    Abstract: Systems and methods are disclosed for migrating a virtual machine (VM) having a virtual function that maps resources of an artificial intelligence (AI) accelerator to the VM. A driver for the AI accelerator can generate a checkpoint of VM processes that make calls to the AI accelerator, and can the checkpoint can include a list and configuration of resources mapped to the AI accelerator by the virtual function. The driver can also access the code, data, and memory of the AI accelerator to generate a checkpoint of the AI accelerator status. When the VM is migrated to a new host, then either, or both, of these checkpoint frames can be used to ensure that resuming the VM on a new host having appropriate AI accelerator resources, can be successful resumed on the new host. One or both checkpoint frames can be captured based upon an event, in anticipation of the need to migrate the VM.
    Type: Application
    Filed: January 6, 2021
    Publication date: July 7, 2022
    Inventors: Zhibiao ZHAO, Yueqiang CHENG
  • Publication number: 20220214902
    Abstract: Systems and methods are disclosed for migrating a virtual machine (VM) having a virtual function that maps resources of an artificial intelligence (AI) accelerator to the VM. A driver for the AI accelerator can generate a checkpoint of VM processes that make calls to the AI accelerator, and can the checkpoint can include a list and configuration of resources mapped to the AI accelerator by the virtual function. The driver can also access the code, data, and memory of the AI accelerator to generate a checkpoint of the AI accelerator status. When the VM is migrated to a new host, then either, or both, of these checkpoint frames can be used to ensure that resuming the VM on a new host having appropriate AI accelerator resources, can be successful resumed on the new host. One or both checkpoint frames can be captured based upon an event, in anticipation of the need to migrate the VM.
    Type: Application
    Filed: January 6, 2021
    Publication date: July 7, 2022
    Inventors: Zhibiao ZHAO, Yueqiang CHENG
  • Patent number: 11301255
    Abstract: Methods, apparatuses, devices, and storage media for performing a processing task are provided. A portion of portions of the processing task can include a group of operations that are to be performed at a processing unit of processing units. The group of operations can include operations of a first type and operations of a second type. In the method, a first queue for performing the operations of the first type and a second queue for performing the operations of the second type can be built, respectively. Based on a definition of the processing task, a dependency relationship between a group of operations to be performed at the processing unit and a group of operations to be performed at other processing units in the plurality of processing units can be obtained. Operations in the first queue and operations in the second queue can be performed respectively based on the dependency relationship.
    Type: Grant
    Filed: December 30, 2019
    Date of Patent: April 12, 2022
    Assignee: Kunlunxin Technology (Beijing) Company Limited
    Inventors: Qingshu Chen, Zhibiao Zhao, Hefei Zhu, Xiaozhang Gong, Yong Wang, Jian Ouyang
  • Publication number: 20210382737
    Abstract: Embodiments of the disclosure discloses a method and system of a virtualization environment for a data processing (DP) accelerator. In one embodiment, a data processing (DP) accelerator includes a resource management unit and one or more dynamically isolated resources managed by the resource management unit. The DP accelerator includes one or more virtual functions (VFs) each associated with one of the one or more dynamically isolated resources, where a virtual machine (VM) of a host is assigned one of the one or more VFs to access the dynamically isolated resources associated with the assigned VF, and where the VM has no access to the rest of the one or more dynamically isolated resources.
    Type: Application
    Filed: June 3, 2020
    Publication date: December 9, 2021
    Inventors: YUEQIANG CHENG, ZHIBIAO ZHAO
  • Publication number: 20210385368
    Abstract: An image capturing method and a terminal device are provided. The method includes entering a camera application to start a lens and display a viewfinder interface, converting an original image captured by the lens into a red-green-blue (RGB) image, and decreasing luminance of the RGB image to be less than first luminance or increasing the luminance of the RGB image to be greater than second luminance, to obtain a first image; converting the RGB image into N frames of high-dynamic-range (HDR) images, and fusing color information of pixels in any same location on the first image and the N frames of HDR images to obtain a final image.
    Type: Application
    Filed: October 10, 2019
    Publication date: December 9, 2021
    Inventors: Liyao Zhang, Jing Ma, Zhibiao Zhao
  • Publication number: 20210382756
    Abstract: Embodiments of the disclosure discloses a method and system for a virtualization environment for a data processing (DP) accelerator. In one embodiment, a data processing (DP) accelerator includes one or more statically partitioned resources and one or more virtual functions (VFs) each associated with one of the one or more statically partitioned resources. A virtual machine (VM) of a host is assigned one of the one or more VFs to access the statically partitioned resources associated with the assigned VF. The VM has no access to the rest of the one or more statically partitioned resources of the DP accelerator.
    Type: Application
    Filed: June 3, 2020
    Publication date: December 9, 2021
    Inventors: YUEQIANG CHENG, ZHIBIAO ZHAO
  • Patent number: 11163714
    Abstract: Embodiments of the present disclosure relate to a method, an apparatus, an electronic device and a computer readable storage medium for determining connection relationships among a plurality of chips. The method includes determining identity information of a plurality of chips managed by a host, the plurality of chips being connected by respective inter-chip communication interfaces for inter-chip communication. The method further includes allowing one or more of the plurality of chips to acquire identity information of other chips connected to the inter-chip communication interface of the one or more chips. The method further includes reading identity information of the other chips by means of a management interface of the one or more chips with regard to communicating with the host, so as to determine connection relationships among the plurality of chips.
    Type: Grant
    Filed: December 11, 2019
    Date of Patent: November 2, 2021
    Inventors: Xianglun Leng, Hefei Zhu, Qingshu Chen, Zhibiao Zhao, Xiaozhang Gong
  • Publication number: 20210318878
    Abstract: According to various embodiments, methods and systems are provided to accelerate artificial intelligence (AI) model training with advanced interconnect communication technologies and systematic zero-value compression over a distributed training system. According to an exemplary method, during each iteration of a Scatter-Reduce process performed on a cluster of processors arranged in a logical ring to train a neural network model, a processor receives a compressed data block from a prior processor in the logical ring, performs an operation on the received compressed data block and a compressed data block generated on the processor to obtain a calculated data block, and sends the calculated data block to a following processor in the logical ring. A compressed data block calculated from corresponding data blocks from the processors can be identified on each processor and distributed to each other processor and decompressed therein for use in the AI model training.
    Type: Application
    Filed: October 12, 2019
    Publication date: October 14, 2021
    Inventors: Zhibiao ZHAO, Jian OUYANG, Hefei ZHU, Qingshu CHEN, Wei QI
  • Publication number: 20210174174
    Abstract: A data processing system includes a central processing unit (CPU) and accelerator cards coupled to the CPU over a bus, each of the accelerator cards having a plurality of data processing (DP) accelerators to receive DP tasks from the CPU and to perform the received DP tasks. At least two of the accelerator cards are coupled to each other via an inter-card connection, and at least two of the DP accelerators are coupled to each other via an inter-chip connection. Each of the inter-card connection and the inter-chip connection is capable of being dynamically activated or deactivated, such that in response to a request received from the CPU, any one of the accelerator cards or any one of the DP accelerators within any one of the accelerator cards can be enabled or disabled to process any one of the DP tasks received from the CPU.
    Type: Application
    Filed: November 15, 2019
    Publication date: June 10, 2021
    Inventors: Hefei ZHU, Jian OUYANG, Zhibiao ZHAO, Xiaozhang GONG, Qingshu CHEN
  • Publication number: 20210072996
    Abstract: Methods, apparatuses, devices, and storage media for performing a processing task are provided. A portion of portions of the processing task can include a group of operations that are to be performed at a processing unit of processing units. The group of operations can include operations of a first type and operations of a second type. In the method, a first queue for performing the operations of the first type and a second queue for performing the operations of the second type can be built, respectively. Based on a definition of the processing task, a dependency relationship between a group of operations to be performed at the processing unit and a group of operations to be performed at other processing units in the plurality of processing units can be obtained. Operations in the first queue and operations in the second queue can be performed respectively based on the dependency relationship.
    Type: Application
    Filed: December 30, 2019
    Publication date: March 11, 2021
    Inventors: Qingshu CHEN, Zhibiao ZHAO, Hefei ZHU, Xiaozhang GONG, Yong WANG, Jian OUYANG
  • Publication number: 20210049045
    Abstract: Embodiments of the present disclosure relate to a method and apparatus for resource management, an electronic device, and a computer-readable storage medium. The method may include: determining a plurality of virtual functions to be supported, where each of the plurality of virtual functions corresponds to a virtual machine running on a computing device. The method may further include: dividing a physical resource set into a plurality of physical resource subsets according to a predetermined ratio, a number of the physical resource subsets being identical to a number of the virtual functions. The method may further include: allocating the plurality of physical resource subsets to the plurality of virtual functions respectively.
    Type: Application
    Filed: March 4, 2020
    Publication date: February 18, 2021
    Inventors: Xianglun Leng, Zhibiao Zhao, Jinchen Han, Jian Ouyang, Wei Qi, Yong Wang
  • Publication number: 20210004679
    Abstract: Presented herein are embodiments of an improved asymmetric quantization, which may generally be referred to as improved asymmetric quantization (IAQ) embodiments. IAQ embodiments combine the benefits of conventional asymmetric quantization and symmetric quantization but also provide additional computation efficiencies. Embodiments of IAQ adopt an asymmetric range of the weights of a neural network layer, so they circumvent the limitation of symmetric range of symmetric quantization. Moreover, the inference process of a neural network quantized by an IAQ embodiment is much faster than that of the neural network quantized by conventional asymmetric quantization by quantizing an offset value of each layer.
    Type: Application
    Filed: May 19, 2020
    Publication date: January 7, 2021
    Applicant: Baidu USA LLC
    Inventors: Yingzhen YANG, Zhibiao ZHAO, Baoxin ZHAO, Jun HUAN, Jian OUYANG, Yong WANG, Jiaxin SHI
  • Publication number: 20200409895
    Abstract: Embodiments of the present disclosure relate to a method, an apparatus, an electronic device and a computer readable storage medium for determining connection relationships among a plurality of chips. The method includes determining identity information of a plurality of chips managed by a host, the plurality of chips being connected by respective inter-chip communication interfaces for inter-chip communication. The method further includes allowing one or more of the plurality of chips to acquire identity information of other chips connected to the inter-chip communication interface of the one or more chips. The method further includes reading identity information of the other chips by means of a management interface of the one or more chips with regard to communicating with the host, so as to determine connection relationships among the plurality of chips.
    Type: Application
    Filed: December 11, 2019
    Publication date: December 31, 2020
    Inventors: Xianglun Leng, Hefei Zhu, Qingshu Chen, Zhibiao Zhao, Xiaozhang Gong
  • Publication number: 20200409603
    Abstract: Embodiments of the present disclosure provide a data accessing method and apparatus, and a computer-readable storage medium, and relate to a computer field. The data accessing method includes: obtaining an identification of a virtual function corresponding to a virtual machine of a computing device and an address related to data in a memory to be accessed by the virtual machine, in which the identification of the virtual function and the address are determined based on an access request received from the virtual machine of the computing device; determining a range of storage resource in the memory corresponding to the virtual machine based on the identification; determining whether the address is within the range; and in response to determining that the address is within the range, accessing the data related to the address.
    Type: Application
    Filed: February 13, 2020
    Publication date: December 31, 2020
    Inventors: Xianglun LENG, Zhibiao ZHAO, Jincheng HAN, Wei QI
  • Publication number: 20200353941
    Abstract: An automatic processing system, a system on chip and a method for monitoring a processing module are described herein. The automatic driving processing system comprises: an automatic driving processing module, configured for receiving an input data stream and processing the input data stream based on a deep learning model so as to generate a processing result; a fault detection module, configured for generating a control signal and a fault detection stimulating data stream, and receiving the processing result from the automatic driving processing module; and a multi-way selection module, configured for receiving an automatic driving data stream as well as the control signal and the fault detection stimulating data stream, and selectively outputting the automatic driving data stream or the fault detection stimulating data stream to the automatic driving processing module based on the control signal, as an input data stream.
    Type: Application
    Filed: December 11, 2019
    Publication date: November 12, 2020
    Inventors: Chonggin Wang, Zhibiao Zhao, Hefei Zhu, Ningyi Xu, Jian Ouyang