Patents by Inventor Zidong Du

Zidong Du has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20200097827
    Abstract: The present disclosure provides a processing device including: a coarse-grained pruning unit configured to perform coarse-grained pruning on a weight of a neural network to obtain a pruned weight, an operation unit configured to train the neural network according to the pruned weight. The coarse-grained pruning unit is specifically configured to select M weights from the weights of the neural network through a sliding window, and when the M weights meet a preset condition, all or part of the M weights may be set to 0. The processing device can reduce the memory access while reducing the amount of computation, thereby obtaining an acceleration ratio and reducing energy consumption.
    Type: Application
    Filed: November 28, 2019
    Publication date: March 26, 2020
    Inventors: Zai Wang, Xuda Zhou, Zidong Du, Tianshi Chen
  • Publication number: 20200097793
    Abstract: The present disclosure relates to a fused vector multiplier for computing an inner product between vectors, where vectors to be computed are a multiplier number vector {right arrow over (A)}{AN . . . A2A1A0} and a multiplicand number {right arrow over (B)} {BN . . . B2B1B0}, {right arrow over (A)} and {right arrow over (B)} have the same dimension which is N+1.
    Type: Application
    Filed: November 27, 2019
    Publication date: March 26, 2020
    Inventors: Tianshi Chen, Shengyuan Zhou, Zidong Du, Qi Guo
  • Publication number: 20200097826
    Abstract: The present disclosure provides a processing device including: a coarse-grained pruning unit configured to perform coarse-grained pruning on a weight of a neural network to obtain a pruned weight, an operation unit configured to train the neural network according to the pruned weight. The coarse-grained pruning unit is specifically configured to select M weights from the weights of the neural network through a sliding window, and when the M weights meet a preset condition, all or part of the M weights may be set to 0. The processing device can reduce the memory access while reducing the amount of computation, thereby obtaining an acceleration ratio and reducing energy consumption.
    Type: Application
    Filed: November 28, 2019
    Publication date: March 26, 2020
    Inventors: Zidong Du, Xuda Zhou, Shaoli Liu, Tianshi Chen
  • Publication number: 20200089535
    Abstract: The application provides a processor and processing method. The processor includes a task segmentation device configured to perform task segmentation according to a task segmentation granularity and a hardware resource division device configured to divide hardware resources of the processor according to a task segmentation result. The processor and processing method provided by the application improve the processing performance and reduce the overhead by performing task segmentation and configuring different hardware according to task segmentation.
    Type: Application
    Filed: November 25, 2019
    Publication date: March 19, 2020
    Inventors: Zidong DU, Shaoli LIU, Zai WANG, Tianshi CHEN, Shuai HU, Xuda ZHOU
  • Publication number: 20200050918
    Abstract: A processing device with dynamically configurable operation bit width, characterized by comprising: a memory for storing data, the data comprising data to be operated, intermediate operation result, final operation result, and data to be buffered in a neural network; a data width adjustment circuit for adjusting the width of the data to be operated, the intermediate operation result, the final operation result, and/or the data to be buffered; an operation circuit for operating the data to be operated, including performing operation on data to be operated of different bit widths by using an adder circuit and a multiplier; and a control circuit for controlling the memory, the data width adjustment circuit and the operation circuit. The device of the present disclosure can have the advantages of strong flexibility, high configurability, fast operation speed, low power consumption or the like.
    Type: Application
    Filed: April 17, 2018
    Publication date: February 13, 2020
    Inventors: Tianshi CHEN, Jie WEI, Tian ZHI, Zai WANG, Shaoli LIU, Yuzhe LUO, Qi GUO, Wei LI, Shengyuan ZHOU, Zidong DU
  • Publication number: 20190370642
    Abstract: The application provides an operation method and device. Quantized data is looked up to realize an operation, which simplifies the structure and reduces the computation energy consumption of the data, meanwhile, a plurality of operations are realized.
    Type: Application
    Filed: August 1, 2019
    Publication date: December 5, 2019
    Inventors: Shaoli LIU, Xuda ZHOU, Zidong DU, Daofu LIU
  • Patent number: 10496597
    Abstract: The present invention is directed to the storage technical field and discloses an on-chip data partitioning read-write method, the method comprises: a data partitioning step for storing on-chip data in different areas, and storing the on-chip data in an on-chip storage medium and an off-chip storage medium respectively, based on a data partitioning strategy; a pre-operation step for performing an operational processing of an on-chip address index of the on-chip storage data in advance when implementing data splicing; and a data splicing step, for splicing the on-chip storage data and the off-chip input data to obtain a representation of the original data based on a data splicing strategy. Also provided are a corresponding on-chip data partitioning read-write system and device. Thus, read and write of repeated data can be efficiently realized, reducing memory access bandwidth requirements while providing good flexibility, thus reducing on-chip storage overhead.
    Type: Grant
    Filed: August 9, 2016
    Date of Patent: December 3, 2019
    Assignee: INSTITUTE OF COMPUTING TECHNOLOGY, CHINESE ACADEMY OF SCIENCES
    Inventors: Tianshi Chen, Zidong Du, Qi Guo, Yunji Chen
  • Patent number: 10489704
    Abstract: Aspects for supporting operation data of different bit widths in neural networks are described herein. The aspects may include a processing module that includes one or more processors. The processor may be capable of processing data of one or more respective bit-widths. Further, the aspects may include a determiner module configured to receive one or more instructions that include one or more operands and one or more width fields. The operands may correspond to one or more operand types and each of the width fields may indicate an operand bit-width of one operand type. The determiner module may be further configured to identify at least one operand bit-widths that is greater than each of the bit-widths. In addition, the aspects may include a processor combiner configured to designate a combination of two or more of the processors to process the operands.
    Type: Grant
    Filed: February 5, 2019
    Date of Patent: November 26, 2019
    Assignee: CAMBRICON TECHNOLOGIES CORPORATION LIMITED
    Inventors: Tianshi Chen, Qi Guo, Zidong Du
  • Publication number: 20190347542
    Abstract: The present disclosure provides a neural network processor and neural network computation method that deploy a memory and a cache to perform a neural network computation, where the memory may be configured to store data and instructions of the neural network computation, the cache may be connected to the memory via a memory bus, thereby, the actual compute ability of hardware may be fully utilized, the cost and power consumption overhead may be reduced, parallelism of the network may be fully utilized, and the efficiency of the neural network computation may be improved.
    Type: Application
    Filed: July 23, 2019
    Publication date: November 14, 2019
    Inventors: Tianshi CHEN, Xiaobin CHEN, Tian ZHI, Zidong DU
  • Publication number: 20190311251
    Abstract: Aspects of reusing neural network instructions are described herein. The aspects may include a computing device configured to calculate a hash value of a neural network layer based on the layer information thereof. A determination unit may be configured to determine whether the hash value exists in a hash table. If the hash value is included in the hash table, one or more neural network instructions that correspond to the hash value may be reused.
    Type: Application
    Filed: May 29, 2019
    Publication date: October 10, 2019
    Inventors: Yunji CHEN, Yixuan REN, Zidong DU, Tianshi CHEN
  • Publication number: 20190250860
    Abstract: The present disclosure provides an integrated circuit chip device and related product thereof. The integrated circuit chip device includes an external interface and a processing circuit. The processing circuit is configured to quantize the first layer input data and the first layer weight group data to obtain a first layer quantized input data and a first layer quantized weight group data; query a first layer output data corresponding to the first layer quantized input data and the first layer quantized weight group data from a preset output result table, determine the first layer output data as a second layer input data, and input the second layer input data into n?1 layers to execute forward operations to obtain nth layer output data; the nth layer output data gradients is determined according to the nth layer output data and the nth layer back operations is obtained according to the training instructions.
    Type: Application
    Filed: February 11, 2019
    Publication date: August 15, 2019
    Inventors: Yukun TIAN, Zhou FANG, Zidong DU
  • Publication number: 20190251448
    Abstract: The present disclosure provides an integrated circuit chip device and related product thereof. The integrated circuit chip device includes an external interface and a processing circuit. The processing circuit is configured to quantize the first layer input data and the first layer weight group data to obtain a first layer quantized input data and a first layer quantized weight group data; query a first layer output data corresponding to the first layer quantized input data and the first layer quantized weight group data from a preset output result table, determine the first layer output data as a second layer input data, and input the second layer input data into n-1 layers to execute forward operations to obtain nth layer output data; the nth layer output data gradients is determined according to the nth layer output data and the nth layer back operations is obtained according to the training instructions.
    Type: Application
    Filed: February 11, 2019
    Publication date: August 15, 2019
    Inventors: Yukun TIAN, Zhou FANG, Zidong DU
  • Publication number: 20190236442
    Abstract: Aspects for supporting operation data of different bit widths in neural networks are described herein. The aspects may include a processing module that includes one or more processors. The processor may be capable of processing data of one or more respective bit-widths. Further, the aspects may include a determiner module configured to receive one or more instructions that include one or more operands and one or more width fields. The operands may correspond to one or more operand types and each of the width fields may indicate an operand bit-width of one operand type. The determiner module may be further configured to identify at least one operand bit-widths that is greater than each of the bit-widths. In addition, the aspects may include a processor combiner configured to designate a combination of two or more of the processors to process the operands.
    Type: Application
    Filed: February 5, 2019
    Publication date: August 1, 2019
    Inventors: Tianshi CHEN, Qi GUO, Zidong DU
  • Publication number: 20190205739
    Abstract: The present disclosure provides an operation device, comprising: an operation module for executing a neural network operation; and a power conversion module connected to the operation module, for converting input neuron data and/or output neuron data of the neural network operation into power neuron data. The present disclosure further provides an operation method. The operation device and method according to the present disclosure reduce the cost of storage resources and computing resources and increase the operation speed.
    Type: Application
    Filed: February 22, 2019
    Publication date: July 4, 2019
    Inventors: Shaoli LIU, Yimin ZHUANG, Qi GUO, Tianshi CHEN, Xiaobing CHEN, Tian ZHI, Zidong DU, Yifan HAO, Zai WANG, Wei LI, Lei ZHANG, Shuai HU
  • Publication number: 20190087716
    Abstract: The present disclosure provides a neural network processing system that comprises a multi-core processing module composed of a plurality of core processing modules and for executing vector multiplication and addition operations in a neural network operation, an on-chip storage medium, an on-chip address index module, and an ALU module for executing a non-linear operation not completable by the multi-core processing module according to input data acquired from the multi-core processing module or the on-chip storage medium, wherein the plurality of core processing modules share an on-chip storage medium and an ALU module, or the plurality of core processing modules have an independent on-chip storage medium and an ALU module. The present disclosure improves an operating speed of the neural network processing system, such that performance of the neural network processing system is higher and more efficient.
    Type: Application
    Filed: August 9, 2016
    Publication date: March 21, 2019
    Inventors: Zidong DU, Qi GUO, Tianshi CHEN, Yunji CHEN
  • Publication number: 20190026626
    Abstract: A neural network accelerator and an operation method thereof applicable in the field of neural network algorithms are disclosed. The neural network accelerator comprises an on-chip storage medium for storing data externally transmitted or for storing data generated during computing; an on-chip address index module for mapping to a correct storage address on the basis of an input index when an operation is performed; a core computing module for performing a neural network operation; and a multi-ALU device for obtaining input data from the core computing module or the on-chip storage medium to perform a nonlinear operation which cannot be completed by the core computing module. By introducing a multi -ALU design into the neural network accelerator, an operation speed of the nonlinear operation is increased, such that the neural network accelerator is more efficient.
    Type: Application
    Filed: August 9, 2016
    Publication date: January 24, 2019
    Inventors: Zidong Du, Qi Guo, Tianshi Chen, Yunji Chen
  • Publication number: 20190026246
    Abstract: The present invention is directed to the storage technical field and discloses an on-chip data partitioning read-write method, the method comprises: a data partitioning step for storing on-chip data in different areas, and storing the on-chip data in an on-chip storage medium and an off-chip storage medium respectively, based on a data partitioning strategy; a pre-operation step for performing an operational processing of an on-chip address index of the on-chip storage data in advance when implementing data splicing; and a data splicing step, for splicing the on-chip storage data and the off-chip input data to obtain a representation of the original data based on a data splicing strategy. Also provided are a corresponding on-chip data partitioning read-write system and device. Thus, read and write of repeated data can be efficiently realized, reducing memory access bandwidth requirements while providing good flexibility, thus reducing on-chip storage overhead.
    Type: Application
    Filed: August 9, 2016
    Publication date: January 24, 2019
    Inventors: Tianshi Chen, Zidong Du, Qi Guo, Yunji Chen