Patents Assigned to Institute of Computing Technology, Chinese Academy
  • Patent number: 11977784
    Abstract: The present invention proposes a dynamic resources allocation method and system for guaranteeing tail latency SLO of latency-sensitive applications. A plurality of request queues is created in a storage server node of a distributed storage system with different types of requests located in different queues, and thread groups are allocated to the request queues according to logical thread resources of the service node and target tail latency requirements, and thread resources are dynamically allocated in real time, and the thread group of each request queue is bound to physical CPU resources of the storage server node. The client sends an application's requests to the storage server node; the storage server node stores the request in a request queue corresponding to its type, uses the thread group allocated for the current queue to process the application's requests, and sends responses to the client.
    Type: Grant
    Filed: July 6, 2020
    Date of Patent: May 7, 2024
    Assignee: Institute of Computing Technology, Chinese Academy of Sciences
    Inventors: Liuying Ma, Zhenqing Liu, Jin Xiong, Dejun Jiang
  • Patent number: 11841733
    Abstract: A method and system for realizing a FPGA server, wherein centralized monitoring and managing all SoC FPGA compute nodes within the server by a motherboard, the motherboard comprising: a plurality of self-defined management interfaces for connecting the SoC FPGA compute nodes to supply power and data switch to the SoC FPGA compute nodes; a management network switch module for interconnecting the SoC FPGA compute nodes and supplying management; and a core control unit for managing the SoC FPGA compute nodes through the self-defined management interfaces and a self-defined management interface protocol, and acquiring operating parameters of the SoC FPGA compute nodes to manage and monitor the SoC FPGA compute nodes based on the management interface protocol.
    Type: Grant
    Filed: January 8, 2020
    Date of Patent: December 12, 2023
    Assignee: Institute of Computing Technology, Chinese Academy of Sciences
    Inventors: Ke Zhang, Yazhou Wang, Mingyu Chen, Yisong Chang, Ran Zhao, Yungang Bao
  • Publication number: 20230128059
    Abstract: The present invention proposes a dynamic resources allocation method and system for guaranteeing tail latency SLO of latency-sensitive applications. A plurality of request queues is created in a storage server node of a distributed storage system with different types of requests located in different queues, and thread groups are allocated to the request queues according to logical thread resources of the service node and target tail latency requirements, and thread resources are dynamically allocated in real time, and the thread group of each request queue is bound to physical CPU resources of the storage server node. The client sends an application's requests to the storage server node; the storage server node stores the request in a request queue corresponding to its type, uses the thread group allocated for the current queue to process the application's requests, and sends responses to the client.
    Type: Application
    Filed: July 6, 2020
    Publication date: April 27, 2023
    Applicant: Institute of Computing Technology, Chinese Academy of Sciences
    Inventors: Liuying MA, Zhenqing LIU, Jin XIONG, Dejun JIANG
  • Publication number: 20230101208
    Abstract: A method and system for realizing a FPGA server, wherein centralized monitoring and managing all SoC FPGA compute nodes within the server by a motherboard, the motherboard comprising: a plurality of self-defined management interfaces for connecting the SoC FPGA compute nodes to supply power and data switch to the SoC FPGA compute nodes; a management network switch module for interconnecting the SoC FPGA compute nodes and supplying management; and a core control unit for managing the SoC FPGA compute nodes through the self-defined management interfaces and a self-defined management interface protocol, and acquiring operating parameters of the SoC FPGA compute nodes to manage and monitor the SoC FPGA compute nodes based on the management interface protocol.
    Type: Application
    Filed: January 8, 2020
    Publication date: March 30, 2023
    Applicant: Institute of Computing Technology, Chinese Academy of Sciences
    Inventors: Ke ZHANG, Yazhou WANG, Mingyu CHEN, Yisong CHANG, Ran ZHAO, Yungang BAO
  • Patent number: 11616662
    Abstract: The present invention provides a fractal tree structure-based data transmit device and method, a control device, and an intelligent chip. The device comprises: a central node that is as a communication data center of a network-on-chip and used for broadcasting or multicasting communication data to a plurality of leaf nodes; the plurality of leaf nodes that are as communication data nodes of the network-on-chip and for transmitting the communication data to a central leaf node; and forwarder modules for connecting the central node with the plurality of leaf nodes and forwarding the communication data; the central node, the forwarder modules and the plurality of leaf nodes are connected in the fractal tree network structure, and the central node is directly connected to M the forwarder modules and/or leaf nodes, any the forwarder module is directly connected to M the next level forwarder modules and/or leaf nodes.
    Type: Grant
    Filed: November 20, 2020
    Date of Patent: March 28, 2023
    Assignee: Institute of Computing Technology, Chinese Academy of Sciences
    Inventors: Jinhua Tao, Tao Luo, Shaoli Liu, Shijin Zhang, Yunji Chen
  • Patent number: 11580367
    Abstract: The present disclosure provides a neural network processing system that comprises a multi-core processing module composed of a plurality of core processing modules and for executing vector multiplication and addition operations in a neural network operation, an on-chip storage medium, an on-chip address index module, and an ALU module for executing a non-linear operation not completable by the multi-core processing module according to input data acquired from the multi-core processing module or the on-chip storage medium, wherein the plurality of core processing modules share an on-chip storage medium and an ALU module, or the plurality of core processing modules have an independent on-chip storage medium and an ALU module. The present disclosure improves an operating speed of the neural network processing system, such that performance of the neural network processing system is higher and more efficient.
    Type: Grant
    Filed: August 9, 2016
    Date of Patent: February 14, 2023
    Assignee: Institute of Computing Technology, Chinese Academy of Sciences
    Inventors: Zidong Du, Qi Guo, Tianshi Chen, Yunji Chen
  • Patent number: 11551068
    Abstract: The present invention provides a processing system for a binary weight convolutional neural network. The system comprises: at least one storage unit for storing data and instructions; at least one control unit for acquiring the instructions stored in the storage unit and sending out a control signal; and, at least one calculation unit for acquiring, from the storage unit, node values of a layer in a convolutional neural network and corresponding binary weight value data and obtaining node values of a next layer by performing addition and subtraction operations. With the system of the present invention, the data bit width during the calculation process of a convolutional neural network is reduced, the convolutional operation speed is improved, and the storage capacity and operational energy consumption are reduced.
    Type: Grant
    Filed: February 11, 2018
    Date of Patent: January 10, 2023
    Assignee: Institute of Computing Technology, Chinese Academy of Sciences
    Inventors: Yinhe Han, Haobo Xu, Ying Wang
  • Patent number: 11531889
    Abstract: Disclosed are a weight data storage method and a convolution computation method that may be implemented in a neural network. The weight data storage method comprises searching for effective weights in a weight convolution kernel matrix and acquiring an index of effective weights. The effective weights are non-zero weights, and the index of effective weights is used to mark the position of the effective weights in the weight convolution kernel matrix. The weight data storage method further comprises storing the effective weights and the index of effective weights. According to the weight data storage method and the convolution computation method of the present disclosure, storage space can be saved, and computation efficiency can be improved.
    Type: Grant
    Filed: February 28, 2018
    Date of Patent: December 20, 2022
    Assignee: INSTITUTE OF COMPUTING TECHNOLOGY, CHINESE ACADEMY OF SCIENCES
    Inventors: Yinhe Han, Feng Min, Haobo Xu, Ying Wang
  • Patent number: 11521048
    Abstract: The present invention relates to a weight management method and system for neural network processing. The method includes two stages, i.e., off-chip encryption stage and on-chip decryption stage: encrypting trained neural network weight data in advance, inputting the encrypted weight into a neural network processor chip, and decrypting the weight in real time by a decryption unit inside the neural network processor chip to perform related neural network calculation. The method and system realizes the protection of weight data without affecting the normal operation of a neural network processor.
    Type: Grant
    Filed: March 22, 2018
    Date of Patent: December 6, 2022
    Assignee: Institute of Computing Technology, Chinese Academy of Sciences
    Inventors: Yinhe Han, Haobo Xu, Ying Wang
  • Publication number: 20220374733
    Abstract: The disclosure provides a data packet classification method and system based on a convolutional neural network including merging each rule set in a training rule set to form a plurality of merging schemes, and determining an optimal merging scheme for each rule set in the training rule set on the basis of performance evaluation; converting a prefix combination distribution of each rule set in the training rule set and a target rule set into an image, and training a convolutional neural network model by taking the image and the corresponding optimal merging scheme as features; and classifying the target rule set on the basis of image similarity, and constructing a corresponding hash table for data packet classification.
    Type: Application
    Filed: December 27, 2019
    Publication date: November 24, 2022
    Applicant: INSTITUTE OF COMPUTING TECHNOLOGY, CHINESE ACADEMY OF SCIENCES
    Inventors: Gaogang XIE, Xinyi ZHANG, Penghao ZHANG
  • Patent number: 11468565
    Abstract: The invention relates to a TMB classification method and system and a TMB analysis device based on a pathological image, comprising: performing TMB classification and marking and pre-processing on a known pathological image to construct a training set; training a convolutional neural network by means of the training set to construct a classification model; pre-processing a target pathological image of a target case to obtain a plurality of target image blocks; classifying the target image blocks by means of the classification model to acquire an image block TMB classification result of the target case; and acquiring an image TMB classification result of the target case by means of classification voting using all the image block TMB classification results. The invention further relates to a TMB analysis device based on a pathological image. The TMB classification method of the invention has advantages of accuracy, a low cost and fast rapid.
    Type: Grant
    Filed: October 28, 2019
    Date of Patent: October 11, 2022
    Assignee: Institute of Computing Technology, Chinese Academy of Sciences
    Inventors: Fei Ren, Zhiyong Liu, Yudong Liu
  • Publication number: 20220207726
    Abstract: The invention relates to a TMB classification method and system and a TMB analysis device based on a pathological image, comprising: performing TMB classification and marking and pre-processing on a known pathological image to construct a training set; training a convolutional neural network by means of the training set to construct a classification model; pre-processing a target pathological image of a target case to obtain a plurality of target image blocks; classifying the target image blocks by means of the classification model to acquire an image block TMB classification result of the target case; and acquiring an image TMB classification result of the target case by means of classification voting using all the image block TMB classification results. The invention further relates to a TMB analysis device based on a pathological image. The TMB classification method of the invention has advantages of accuracy, a low cost and fast rapid.
    Type: Application
    Filed: October 28, 2019
    Publication date: June 30, 2022
    Applicant: Institute of Computing Technology, Chinese Academy of Sciences
    Inventors: Fei REN, Zhiyong LIU, Yudong LIU
  • Patent number: 11331794
    Abstract: An inverse kinematics solution system for use with a robot, which is used for obtaining a joint angle value corresponding to a target pose value on the basis of an inputted target pose value and degree of freedom of a robot and which comprises: a parameters initialization module, an inverse kinematics scheduler, a Jacobian calculating unit, a pose updating unit and a parameters selector. The system is implemented by means of hardware and may quickly obtain motion parameters, which are used for controlling a robot, while reducing power consumption.
    Type: Grant
    Filed: February 11, 2018
    Date of Patent: May 17, 2022
    Assignee: INSTITUTE OF COMPUTING TECHNOLOGY, CHINESE ACADEMY OF SCIENCES
    Inventors: Hang Xiao, Yinhe Han, Ying Wang, Shiqi Lian
  • Publication number: 20210357735
    Abstract: Disclosed embodiments relate to a split accumulator for a convolutional neural network accelerator, comprising: arranging original weights in a computation sequence and aligning by bit to obtain a weight matrix, removing slack bits in the weight matrix, allowing essential bits in each column of the weight matrix to fill the vacancies according to the computation sequence to obtain an intermediate matrix, removing null rows in the intermediate matrix, obtain a kneading matrix, wherein each row of the kneading matrix serves as a kneading weight; obtaining positional information of the activation corresponding to each bit of the kneading weight; divides the kneading weight by bit into multiple weight segments, processing summation of the weight segments and the corresponding activations according to the positional information, and sending a processing result to an adder tree to obtain an output feature map by means of executing shift-and-add on the processing result.
    Type: Application
    Filed: May 21, 2019
    Publication date: November 18, 2021
    Applicant: Institute of Computing Technology, Chinese Academy of Sciences
    Inventors: Xiaowei LI, Xin WEI, Hang LU
  • Publication number: 20210350214
    Abstract: Disclosed embodiments relate to a convolutional neural network computing method and system based on weight kneading, comprising: arranging original weights in a computation sequence and aligning by bit to obtain a weight matrix, removing slack bits in the weight matrix, allowing essential bits in each column of the weight matrix to fill the vacancies according to the computation sequence to obtain an intermediate matrix, removing null rows in the intermediate matrix, obtain a kneading matrix, wherein each row of the kneading matrix serves as a kneading weight; obtaining positional information of the activation corresponding to each bit of the kneading weight; divides the kneading weight by bit into multiple weight segments, processing summation of the weight segments and the corresponding activations according to the positional information, and sending a processing result to an adder tree to obtain an output feature map by means of executing shift-and-add on the processing result.
    Type: Application
    Filed: May 21, 2019
    Publication date: November 11, 2021
    Applicant: Institute of Computing Technology, Chinese Academy of Sciences
    Inventors: Xiaowei LI, Xin WEI, Hang LU
  • Publication number: 20210350204
    Abstract: Disclosed embodiments relate to a convolutional neural network accelerator, comprising: arranging original weights in a computation sequence and aligning by bit to obtain a weight matrix, removing slack bits in the weight matrix, allowing essential bits in each column of the weight matrix to fill the vacancies according to the computation sequence to obtain an intermediate matrix, removing null rows in the intermediate matrix, obtain a kneading matrix, wherein each row of the kneading matrix serves as a kneading weight; obtaining positional information of the activation corresponding to each bit of the kneading weight; divides the kneading weight by bit into multiple weight segments, processing summation of the weight segments and the corresponding activations according to the positional information, and sending a processing result to an adder tree to obtain an output feature map by means of executing shift-and-add on the processing result.
    Type: Application
    Filed: May 21, 2019
    Publication date: November 11, 2021
    Applicant: Institute of Computing Technology, Chinese Academy of Sciences
    Inventors: Xiaowei LI, Xin WEI, Hang LU
  • Publication number: 20210182666
    Abstract: Disclosed are a weight data storage method and a convolution computation method that may be implemented in a neural network. The weight data storage method comprises searching for effective weights in a weight convolution kernel matrix and acquiring an index of effective weights. The effective weights are non-zero weights, and the index of effective weights is used to mark the position of the effective weights in the weight convolution kernel matrix. The weight data storage method further comprises storing the effective weights and the index of effective weights. According to the weight data storage method and the convolution computation method of the present disclosure, storage space can be saved, and computation efficiency can be improved.
    Type: Application
    Filed: February 28, 2018
    Publication date: June 17, 2021
    Applicant: INSTITUTE OF COMPUTING TECHNOLOGY, CHINESE ACADEMY OF SCIENCES
    Inventors: Yinhe HAN, Feng MIN, Haobo XU, Ying WANG
  • Patent number: 10904034
    Abstract: One example of a device comprises: a central node that is as a communication data center of a network-on-chip; a plurality of leaf nodes that are as communication data nodes of the network-on-chip and for transmitting the communication data to a central leaf node; forwarder modules for connecting the central node with the plurality of leaf nodes and forwarding the communication data, wherein the plurality of leaf nodes are divided into N groups, each group having the same number of leaf nodes, the central node is individually in communication connection with each group of leaf nodes by means of the forwarder module, a communication structure constituted by each group of leaf nodes has self-similarity, and the plurality of leaf nodes are in communication connection with the central node in a complete multi-way tree approach by means of the forwarder modules of multiple levels.
    Type: Grant
    Filed: June 17, 2016
    Date of Patent: January 26, 2021
    Assignee: Institute of Computing Technology, Chinese Academy of Sciences
    Inventors: Jinhua Tao, Tao Luo, Shaoli Liu, Shijin Zhang, Yunji Chen
  • Patent number: 10866924
    Abstract: An example device comprises a central node for receiving vector data returned by leaf nodes, a plurality of leaf nodes for calculating and shifting the vector data, and forwarder modules comprising a local cache structure and a data processing component, wherein the plurality of leaf nodes are divided into N groups, each group having the same number of leaf nodes; the central node is individually in communication connection with each group of leaf nodes by means of the forwarder modules; a communication structure constituted by each group of leaf nodes has self-similarity; the plurality of leaf nodes are in communication connection with the central node in a complete M-way tree approach by means of the forwarder modules of multiple levels; each of the leaf nodes comprises a setting bit.
    Type: Grant
    Filed: June 17, 2016
    Date of Patent: December 15, 2020
    Assignee: Institute of Computing Technology, Chinese Academy of Sciences
    Inventors: Dong Han, Tao Luo, Shaoli Liu, Shijin Zhang, Yunji Chen
  • Patent number: 10805233
    Abstract: A communication structure comprises: a central node that is a communication data center of a network-on-chip and used for broadcasting or multicasting communication data to a plurality of leaf nodes; a plurality of leaf nodes that are communication data nodes of the network-on-chip and used for transmitting the communication data to the central node; and forwarder modules for connecting the central node with the plurality of leaf nodes and forwarding the communication data, wherein the plurality of leaf nodes are divided into N groups, each group having the same number of leaf nodes, the central node is individually in communication connection with each group of leaf nodes by means of the forwarder modules, the communication structure is a fractal-tree structure, the communication structure constituted by each group of leaf nodes has self-similarity, and the forwarder modules comprises a central forwarder module, leaf forwarder modules, and intermediate forwarder modules.
    Type: Grant
    Filed: June 17, 2016
    Date of Patent: October 13, 2020
    Assignee: INSTITUTE OF COMPUTING TECHNOLOGY, CHINESE ACADEMY OF SCIENCE
    Inventors: Huiying Lan, Tao Luo, Shaoli Liu, Shijin Zhang, Yunji Chen