Patents by Inventor Chao-Yang Kao

Chao-Yang Kao has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11836214
    Abstract: A matrix calculation device including a storing unit, a multiply accumulate (MAC) circuit, a pre-fetch circuit, and a control circuit, and an operation method thereof are provided. The storing unit stores a first and second matrixes. The MAC circuit is configured to execute MAC calculation. The pre-fetch circuit pre-fetches at least one column of the first matrix from the storing unit to act as pre-fetch data, pre-fetches at least one row of the second matrix from the storing unit to act as the pre-fetch data, or pre-fetches at least one column of the first matrix and at least one row of the second matrix from the storing unit to act as the pre-fetch data. The control circuit decides whether to perform the MAC calculation on a current column of the first matrix and a current row of the second matrix through the MAC circuit according to the pre-fetch data.
    Type: Grant
    Filed: September 28, 2020
    Date of Patent: December 5, 2023
    Assignee: NEUCHIPS CORPORATION
    Inventors: Chiung-Liang Lin, Chao-Yang Kao
  • Patent number: 11782839
    Abstract: A feature map caching method of a convolutional neural network includes a connection analyzing step and a plurality of layer operation steps. The connection analyzing step is for analyzing a network to establish a convolutional neural network connection list. The convolutional neural network connection list includes a plurality of tensors and a plurality of layer operation coefficients. Each of the layer operation coefficients includes a step index, at least one input operand label and an output operand label. The step index as a processing order for the layer operation step. At least one of the layer operation steps is for flushing at least one of the tensors in a cache according to a distance between the at least one of the layer operation steps and a future layer operation step of the layer operation steps. The distance is calculated according to the convolutional neural network connection list.
    Type: Grant
    Filed: August 19, 2019
    Date of Patent: October 10, 2023
    Assignee: NEUCHIPS CORPORATION
    Inventors: Ping Chao, Chao-Yang Kao, Youn-Long Lin
  • Patent number: 11615286
    Abstract: A computing system and a compressing method for neural network parameters are provided. In the method, multiple neural network parameters are obtained. The neural network parameters are used for a neural network algorithm. Every at least two neural network parameters are grouped into an encoding combination. The number of neural network parameters in each encoding combination is the same. The encoding combinations are compressed with the same compression target bit number. Each encoding combination is compressed independently. The compression target bit number is not larger than a bit number of each encoding combination. Thereby, the storage space can be saved and excessive power consumption for accessing the parameters can be prevented.
    Type: Grant
    Filed: July 18, 2019
    Date of Patent: March 28, 2023
    Assignee: NEUCHIPS CORPORATION
    Inventors: Youn-Long Lin, Chao-Yang Kao, Huang-Chih Kuo, Chiung-Liang Lin
  • Patent number: 11474937
    Abstract: A computing device and an operation method thereof are provided. The computing device includes multiple memories and an indexer circuit. The indexer circuit is separately coupled to the memories through multiple memory channels. The indexer circuit determines an arrangement of at least one lookup table to at least one of the memories according to a characteristic of the at least one lookup table and a transmission bandwidth of the memory channels, so as to balance a transmission load of the memory channels.
    Type: Grant
    Filed: November 20, 2020
    Date of Patent: October 18, 2022
    Assignee: NEUCHIPS CORPORATION
    Inventors: Chao-Yang Kao, Youn-Long Lin
  • Patent number: 11467968
    Abstract: A memory-adaptive processing method for a convolutional neural network includes a feature map counting step, a size relation counting step and a convolution calculating step. The feature map counting step is for counting a number of a plurality of input channels of a plurality of input feature maps, an input feature map tile size, a number of a plurality of output channels of a plurality of output feature maps and an output feature map tile size for a convolutional layer operation. The size relation counting step is for obtaining a cache free space size in a feature map cache and counting a size relation. The convolution calculating step is for performing the convolutional layer operation with the input feature maps to produce the output feature maps according to a memory-adaptive processing technique, and the memory-adaptive processing technique includes a dividing step and an output-group-first processing step.
    Type: Grant
    Filed: February 26, 2021
    Date of Patent: October 11, 2022
    Assignee: NEUCHIPS CORPORATION
    Inventors: Ping Chao, Chao-Yang Kao, Youn-Long Lin
  • Publication number: 20220237430
    Abstract: A harmonic densely connecting method includes an input step, a plurality of layer operation steps and an output step. The input step is for storing an original input tensor of the block into a memory. Each of the layer operation steps includes a layer-input tensor concatenating step and a convolution operation step. The layer-input tensor concatenating step is for selecting at least one layer-input element tensor of a layer-input set from the memory according to an input connection rule. When a number of the at least one layer-input element tensor is greater than 1, concatenating all of the layer-input element tensors and producing a layer-input tensor. The convolution operation step is for calculating a convolution operation to produce at least one result tensor and then storing the at least one result tensor into the memory. The output step is for outputting a block output.
    Type: Application
    Filed: April 13, 2022
    Publication date: July 28, 2022
    Inventors: Ping CHAO, Chao-Yang KAO, Youn-Long LIN
  • Publication number: 20220121565
    Abstract: A computing device and an operation method thereof are provided. The computing device includes multiple memories and an indexer circuit. The indexer circuit is separately coupled to the memories through multiple memory channels. The indexer circuit determines an arrangement of at least one lookup table to at least one of the memories according to a characteristic of the at least one lookup table and a transmission bandwidth of the memory channels, so as to balance a transmission load of the memory channels.
    Type: Application
    Filed: November 20, 2020
    Publication date: April 21, 2022
    Applicant: NEUCHIPS CORPORATION
    Inventors: Chao-Yang Kao, Youn-Long Lin
  • Patent number: 11307853
    Abstract: A matrix multiplication device and an operation method thereof are provided. The matrix multiplication device includes calculation circuits, a control circuit, a multiplication circuit, and a routing circuit. The calculation circuits produce multiply-accumulate values. The control circuit receives a plurality of first element values of a first matrix. The control circuit classifies the first element values into at least one classification value. The multiplication circuit multiplies the classification value by a second element value of a second matrix in a low power mode to obtain at least one product value. The routing circuit transmits each of the product values to at least one corresponding calculation circuit in the calculation circuits in the low power mode.
    Type: Grant
    Filed: October 29, 2019
    Date of Patent: April 19, 2022
    Assignee: NEUCHIPS CORPORATION
    Inventors: Chiung-Liang Lin, Chao-Yang Kao, Youn-Long Lin, Huang-Chih Kuo, Jian-Wen Chen
  • Publication number: 20220058238
    Abstract: A matrix calculation device including a storing unit, a multiply accumulate (MAC) circuit, a pre-fetch circuit, and a control circuit, and an operation method thereof are provided. The storing unit stores a first and second matrixes. The MAC circuit is configured to execute MAC calculation. The pre-fetch circuit pre-fetches at least one column of the first matrix from the storing unit to act as pre-fetch data, pre-fetches at least one row of the second matrix from the storing unit to act as the pre-fetch data, or pre-fetches at least one column of the first matrix and at least one row of the second matrix from the storing unit to act as the pre-fetch data. The control circuit decides whether to perform the MAC calculation on a current column of the first matrix and a current row of the second matrix through the MAC circuit according to the pre-fetch data.
    Type: Application
    Filed: September 28, 2020
    Publication date: February 24, 2022
    Applicant: NEUCHIPS CORPORATION
    Inventors: Chiung-Liang Lin, Chao-Yang Kao
  • Patent number: 11210215
    Abstract: A computing device and an operation method thereof are provided. The computing device includes a plurality of memories and a processing circuit. The processing circuit is coupled to the memories. The processing circuit dynamically determines which of the plurality of memories to store at least one lookup table according to characteristics of the at least one lookup table. The processing circuit may then execute at least one algorithm by using the at least one lookup table.
    Type: Grant
    Filed: February 18, 2020
    Date of Patent: December 28, 2021
    Assignee: NEUCHIPS CORPORATION
    Inventors: Youn-Long Lin, Chao-Yang Kao, Huang-Chih Kuo
  • Publication number: 20210182204
    Abstract: A memory-adaptive processing method for a convolutional neural network includes a feature map counting step, a size relation counting step and a convolution calculating step. The feature map counting step is for counting a number of a plurality of input channels of a plurality of input feature maps, an input feature map tile size, a number of a plurality of output channels of a plurality of output feature maps and an output feature map tile size for a convolutional layer operation. The size relation counting step is for obtaining a cache free space size in a feature map cache and counting a size relation. The convolution calculating step is for performing the convolutional layer operation with the input feature maps to produce the output feature maps according to a memory-adaptive processing technique, and the memory-adaptive processing technique includes a dividing step and an output-group-first processing step.
    Type: Application
    Filed: February 26, 2021
    Publication date: June 17, 2021
    Inventors: Ping CHAO, Chao-Yang KAO, Youn-Long LIN
  • Publication number: 20210096987
    Abstract: A computing device and an operation method thereof are provided. The computing device includes a plurality of memories and a processing circuit. The processing circuit is coupled to the memories. The processing circuit dynamically determines which of the plurality of memories to store at least one lookup table according to characteristics of the at least one lookup table. The processing circuit may then execute at least one algorithm by using the at least one lookup table.
    Type: Application
    Filed: February 18, 2020
    Publication date: April 1, 2021
    Applicant: NEUCHIPS CORPORATION
    Inventors: Youn-Long Lin, Chao-Yang Kao, Huang-Chih Kuo
  • Publication number: 20210097368
    Abstract: A processing system includes at least one signal processing unit and at least one neural network layer. A first signal processing unit of the at least one signal processing unit performs signal processing with at least one first parameter. A first neural network layer of the at least one neural network layer has at least one second parameter. The at least one first parameter and the at least one second parameter are trained together.
    Type: Application
    Filed: February 12, 2020
    Publication date: April 1, 2021
    Inventors: Youn-Long Lin, Chao-Yang Kao, Huang-Chih Kuo
  • Patent number: 10963390
    Abstract: A memory-adaptive processing method for a convolutional neural network includes a feature map counting step, a size relation counting step and a convolution calculating step. The feature map counting step is for counting a plurality of input channels of an input feature map tile and a plurality of output channels of an output feature map tile for a convolutional layer operation of the convolutional neural network. The size relation counting step is for obtaining a cache free space size in a feature map cache and counting a size relation among a total input size, a total output size and the cache free space size of the feature map cache. The convolution calculating step is for performing the convolutional layer operation according to a memory-adaptive processing technique.
    Type: Grant
    Filed: August 7, 2019
    Date of Patent: March 30, 2021
    Assignee: NEUCHIPS CORPORATION
    Inventors: Ping Chao, Chao-Yang Kao, Youn-Long Lin
  • Publication number: 20210064373
    Abstract: A matrix multiplication device and an operation method thereof are provided. The matrix multiplication device includes calculation circuits, a control circuit, a multiplication circuit, and a routing circuit. The calculation circuits produce multiply-accumulate values. The control circuit receives a plurality of first element values of a first matrix. The control circuit classifies the first element values into at least one classification value. The multiplication circuit multiplies the classification value by a second element value of a second matrix in a low power mode to obtain at least one product value. The routing circuit transmits each of the product values to at least one corresponding calculation circuit in the calculation circuits in the low power mode.
    Type: Application
    Filed: October 29, 2019
    Publication date: March 4, 2021
    Applicant: NEUCHIPS CORPORATION
    Inventors: Chiung-Liang Lin, Chao-Yang Kao, Youn-Long Lin, Huang-Chih Kuo, Jian-Wen Chen
  • Publication number: 20200410353
    Abstract: A harmonic densely connecting method includes an input step, a plurality of layer operation steps and an output step. The input step is for storing an original input tensor of the block into a memory. Each of the layer operation steps includes a layer-input tensor concatenating step and a convolution operation step. The layer-input tensor concatenating step is for selecting at least one layer-input element tensor of a layer-input set from the memory according to an input connection rule. When a number of the at least one layer-input element tensor is greater than 1, concatenating all of the layer-input element tensors and producing a layer-input tensor. The convolution operation step is for calculating a convolution operation to produce at least one result tensor and then storing the at least one result tensor into the memory. The output step is for outputting a block output.
    Type: Application
    Filed: June 25, 2019
    Publication date: December 31, 2020
    Inventors: Ping CHAO, Chao-Yang KAO, Youn-Long LIN
  • Publication number: 20200372320
    Abstract: A computing system and a compressing method for neural network parameters are provided. In the method, multiple neural network parameters are obtained. The neural network parameters are used for a neural network algorithm. Every at least two neural network parameters are grouped into an encoding combination. The number of neural network parameters in each encoding combination is the same. The encoding combinations are compressed with the same compression target bit number. Each encoding combination is compressed independently. The compression target bit number is not larger than a bit number of each encoding combination. Thereby, the storage space can be saved and excessive power consumption for accessing the parameters can be prevented.
    Type: Application
    Filed: July 18, 2019
    Publication date: November 26, 2020
    Applicant: NEUCHIPS CORPORATION
    Inventors: Youn-Long Lin, Chao-Yang Kao, Huang-Chih Kuo, Chiung-Liang Lin
  • Publication number: 20200065250
    Abstract: A feature map caching method of a convolutional neural network includes a connection analyzing step and a plurality of layer operation steps. The connection analyzing step is for analyzing a network to establish a convolutional neural network connection list. The convolutional neural network connection list includes a plurality of tensors and a plurality of layer operation coefficients. Each of the layer operation coefficients includes a step index, at least one input operand label and an output operand label. The step index as a processing order for the layer operation step. At least one of the layer operation steps is for flushing at least one of the tensors in a cache according to a distance between the at least one of the layer operation steps and a future layer operation step of the layer operation steps. The distance is calculated according to the convolutional neural network connection list.
    Type: Application
    Filed: August 19, 2019
    Publication date: February 27, 2020
    Inventors: Ping CHAO, Chao-Yang KAO, Youn-Long LIN
  • Publication number: 20200065251
    Abstract: A memory-adaptive processing method for a convolutional neural network includes a feature map counting step, a size relation counting step and a convolution calculating step. The feature map counting step is for counting a plurality of input channels of an input feature map tile and a plurality of output channels of an output feature map tile for a convolutional layer operation of the convolutional neural network. The size relation counting step is for obtaining a cache free space size in a feature reap cache and counting a size relation among a total input size, a total output size and the cache free space size of the feature map cache. The convolution calculating step is for performing the convolutional layer operation according to a memory-adaptive processing technique.
    Type: Application
    Filed: August 7, 2019
    Publication date: February 27, 2020
    Inventors: Ping CHAO, Chao-Yang KAO, Youn-Long LIN
  • Patent number: 10217394
    Abstract: A display driving apparatus including a pixel reorder circuit, an image processing circuit and a driver circuit is provided. The pixel reorder circuit is configured to reorder pixels of frame data. The frame data includes previous frame data. The image processing circuit is coupled to the pixel reorder circuit. The image processing circuit is configured to perform an image processing operation on the frame data that the pixels have been reordered. The driver circuit is coupled to the pixel reorder circuit. The driver circuit is configured to drive a display according to the previous frame data that pixels have been reordered and the current frame data. Each of the pixels of the frame data includes a first sub-pixel set and a second sub-pixel set. In addition, a display driving method is also provided.
    Type: Grant
    Filed: September 30, 2016
    Date of Patent: February 26, 2019
    Assignee: Novatek Microelectronics Corp.
    Inventors: Wenhui Yu, Chao-Yang Kao, Chiung-Liang Lin