Patents by Inventor Jiachao Liu

Jiachao Liu has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 12236341
    Abstract: Embodiments disclose bank-balanced-sparse activation neural network models and methods to generate the bank-balanced-sparse activation neural network models. According to one embodiment, a neural network sparsification engine determines a first deep neural network (DNN) model having two or more hidden layers. The engine determines a bank size, a bank layout, and a target sparsity. The engine segments the activation feature maps into a plurality of banks based on the bank size and the bank layout. The engine generates a second DNN model by increasing a sparsity for each bank of activation feature map based on the target sparsity, wherein the second DNN model is used for inferencing.
    Type: Grant
    Filed: September 30, 2020
    Date of Patent: February 25, 2025
    Assignee: MOFFETT INTERNATIONAL CO., LIMITED
    Inventors: Enxu Yan, Dongkuan Xu, Jiachao Liu
  • Publication number: 20230111362
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for parallelizing convolution processing. An exemplary method comprises: segmenting an input tensor into a plurality of sub-tensors and a plurality of filters into a plurality of sub-filter groups; respectively assigning a plurality of combinations of the sub-tensors and the sub-filter groups to a plurality of processors; storing, by each of the plurality of processors, nonzero values of the sub-tensor and the sub-filter group in the assigned combination as index-value pairs; parallelly performing for a plurality of iterations, by the plurality of processors, multiply-and-accumulate (MAC) operations based on the index-value pairs to obtain a plurality of outputs, where the index-value pairs of the sub-filter groups are rotated among the plurality of processors across the plurality of iterations; and aggregating the plurality of outputs as an output tensor.
    Type: Application
    Filed: December 12, 2022
    Publication date: April 13, 2023
    Inventors: Enxu YAN, Yong LU, Wei WANG, Zhibin XIAO, Jiachao LIU, Hengchang XIONG
  • Patent number: 11379724
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for domain-specific pruning of neural networks are described. An exemplary method includes obtaining a first neural network trained based on a first training dataset; obtaining one or more second training datasets respectively from one or more domains; and training, based on the first neural network and the one or more second training datasets, a second neural network comprising the first neural network and one or more branches extended from the first neural network, wherein the second neural network is applicable for inferencing in the one or more domains, and the training comprises: training the one or more branches based respectively on the one or more second training datasets and an output of the first neural network.
    Type: Grant
    Filed: July 12, 2021
    Date of Patent: July 5, 2022
    Assignee: MOFFETT TECHNOLOGIES CO., LIMITED
    Inventors: Jiachao Liu, Enxu Yan
  • Publication number: 20220198272
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for domain-specific pruning of neural networks are described. An exemplary method includes obtaining a first neural network trained based on a first training dataset; obtaining one or more second training datasets respectively from one or more domains; and training, based on the first neural network and the one or more second training datasets, a second neural network comprising the first neural network and one or more branches extended from the first neural network, wherein the second neural network is applicable for inferencing in the one or more domains, and the training comprises: training the one or more branches based respectively on the one or more second training datasets and an output of the first neural network.
    Type: Application
    Filed: July 12, 2021
    Publication date: June 23, 2022
    Inventors: JIACHAO LIU, ENXU YAN
  • Publication number: 20220172059
    Abstract: Embodiments disclosed herein allowed neural networks to be pruned. The inputs and outputs generated by a reference neural network are used to prune the reference neural network. The pruned neural network may have a subset of the weights that are in the reference neural network.
    Type: Application
    Filed: November 30, 2020
    Publication date: June 2, 2022
    Inventors: ENXU YAN, DONGKUAN XU, JIACHAO LIU
  • Publication number: 20220101118
    Abstract: Embodiments disclose bank-balanced-sparse activation neural network models and methods to generate the bank-balanced-sparse activation neural network models. According to one embodiment, a neural network sparsification engine determines a first deep neural network (DNN) model having two or more hidden layers. The engine determines a bank size, a bank layout, and a target sparsity. The engine segments the activation feature maps into a plurality of banks based on the bank size and the bank layout. The engine generates a second DNN model by increasing a sparsity for each bank of activation feature map based on the target sparsity, wherein the second DNN model is used for inferencing.
    Type: Application
    Filed: September 30, 2020
    Publication date: March 31, 2022
    Inventors: ENXU YAN, DONGKUAN XU, JIACHAO LIU
  • Patent number: 11068786
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for domain-specific pruning of neural networks are described. An exemplary method includes obtaining a first neural network trained based on a first training dataset; obtaining one or more second training datasets respectively from one or more domains; training, based on the first neural network and the one or more second training datasets, a second neural network comprising the first neural network and one or more branches extended from the first neural network. The one or more branches respectively correspond to the one or more domains, and each comprises one or more layers trained based on one of the one or more second training datasets. The method may further include: pruning the second neural network by reducing a number of active neurons; and applying the pruned second neural network for inferencing in the one or more domains.
    Type: Grant
    Filed: December 17, 2020
    Date of Patent: July 20, 2021
    Assignee: MOFFETT TECHNOLOGIES CO., LIMITED
    Inventors: Jiachao Liu, Enxu Yan