Patents by Inventor Taro Sekiyama

Taro Sekiyama has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11880762
    Abstract: A computer-implemented method, a computer program product, and a computer processing system are provided for selecting from among multiple Graphics Processing Unit (GPU) execution modes for a Neural Network (NN) having a size greater than a threshold size. The multiple GPU execution modes include a normal memory mode, an Out-of-Core (OoC) execution mode, and a Unified Memory (UM) mode. The method includes starting an execution on the NN with the UM mode and measuring the memory usage for each of layers of the NN. The method further includes selecting an execution mode based on the memory usage of all of the layers.
    Type: Grant
    Filed: June 26, 2018
    Date of Patent: January 23, 2024
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Yasushi Negishi, Haruki Imai, Taro Sekiyama, Tung D. Le, Kiyokuni Kawachiya
  • Patent number: 11461637
    Abstract: A generated algorithm used by a neural network is captured during execution of an iteration of the neural network. A candidate algorithm is identified based on the generated algorithm. A determination is made that the candidate algorithm utilizes less memory than the generated algorithm. Based on the determination the neural network is updated by replacing the generated algorithm with the candidate algorithm.
    Type: Grant
    Filed: March 6, 2019
    Date of Patent: October 4, 2022
    Assignee: International Business Machines Corporation
    Inventors: Taro Sekiyama, Kiyokuni Kawachiya, Tung D. Le, Yasushi Negishi
  • Patent number: 11164079
    Abstract: A computer-implemented method, computer program product, and computer processing system are provided for accelerating neural network data parallel training in multiple graphics processing units (GPUs) using at least one central processing unit (CPU). The method includes forming a set of chunks. Each of the chunks includes a respective group of neural network layers other than a last layer. The method further includes performing one or more chunk-wise synchronization operations during a backward phase of the neural network data parallel training, by each of the multiple GPUs and the at least one CPU.
    Type: Grant
    Filed: December 15, 2017
    Date of Patent: November 2, 2021
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Tung D. Le, Haruki Imai, Taro Sekiyama, Yasushi Negishi
  • Patent number: 11106970
    Abstract: In an approach to localizing tree-based convolutional neural networks, a method includes creating a first tree-based convolution layer (TBCL) corresponding to a tree, where the tree includes a first plurality of nodes and a node that has been indicated to be a first pivotal node. The first TBCL includes a second plurality of nodes and a second pivotal node having a feature vector based on node data from the first pivotal node. The method also includes creating a second TBCL corresponding to the tree. The second TBCL may include a third plurality of nodes. The method further includes determining a feature vector a third pivotal node in the third plurality of nodes based on the feature vectors from: (i) the second pivotal node, (ii) a parent node of the second pivotal node, and (iii) a child node of the second pivotal node.
    Type: Grant
    Filed: November 17, 2017
    Date of Patent: August 31, 2021
    Assignee: International Business Machines Corporation
    Inventors: Tung D. Le, Taro Sekiyama
  • Patent number: 10915810
    Abstract: A cascading convolutional neural network (CCNN) comprising a plurality of convolutional neural networks (CNNs) that are trained by weighting training data based on loss values of each training datum between CNNs of the CCN. The CCNN can receiving an input image from plurality of images, classify the input image using the CCNN, and present a classification of the input image.
    Type: Grant
    Filed: November 12, 2019
    Date of Patent: February 9, 2021
    Assignee: International Business Machines Corporation
    Inventors: Taro Sekiyama, Masaharu Sakamoto, Hiroki Nakano, Kun Zhao
  • Patent number: 10782897
    Abstract: A method is provided for reducing consumption of a memory in a propagation process for a neural network (NN) having fixed structures for computation order and node data dependency. The memory includes memory segments for allocating to nodes. The method collects, in a NN training iteration, information for each node relating to an allocation, size, and lifetime thereof. The method chooses, responsive to the information, a first node having a maximum memory size relative to remaining nodes, and a second node non-overlapped with the first node lifetime. The method chooses another node non-overlapped with the first node lifetime, responsive to a sum of memory sizes of the second node and the other node not exceeding a first node memory size. The method reallocates a memory segment allocated to the first node to the second node and the other node to be reused by the second node and the other node.
    Type: Grant
    Filed: April 2, 2018
    Date of Patent: September 22, 2020
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Taro Sekiyama, Haruki Imai, Jun Doi, Yasushi Negishi
  • Publication number: 20200097800
    Abstract: A cascading convolutional neural network (CCNN) comprising a plurality of convolutional neural networks (CNNs) that are trained by weighting training data based on loss values of each training datum between CNNs of the CCN. The CCNN can receiving an input image from plurality of images, classify the input image using the CCNN, and present a classification of the input image.
    Type: Application
    Filed: November 12, 2019
    Publication date: March 26, 2020
    Inventors: Taro Sekiyama, Masaharu Sakamoto, Hiroki Nakano, Kun Zhao
  • Patent number: 10599978
    Abstract: A cascading convolutional neural network (CCNN) comprising a plurality of convolutional neural networks (CNNs) that are trained by weighting training data based on loss values of each training datum between CNNs of the CCN. The CCNN can receiving an input image from plurality of images, classify the input image using the CCNN, and present a classification of the input image.
    Type: Grant
    Filed: November 3, 2017
    Date of Patent: March 24, 2020
    Assignee: International Business Machines Corporation
    Inventors: Taro Sekiyama, Masaharu Sakamoto, Hiroki Nakano, Kun Zhao
  • Patent number: 10558914
    Abstract: A generated algorithm used by a neural network is captured during execution of an iteration of the neural network. A candidate algorithm is identified based on the generated algorithm. A determination is made that the candidate algorithm utilizes less memory than the generated algorithm. Based on the determination the neural network is updated by replacing the generated algorithm with the candidate algorithm.
    Type: Grant
    Filed: April 16, 2019
    Date of Patent: February 11, 2020
    Assignee: International Business Machines Corporation
    Inventors: Taro Sekiyama, Kiyokuni Kawachiya, Tung D. Le, Yasushi Negishi
  • Publication number: 20190392306
    Abstract: A computer-implemented method, a computer program product, and a computer processing system are provided for selecting from among multiple Graphics Processing Unit (GPU) execution modes for a Neural Network (NN) having a size greater than a threshold size. The multiple GPU execution modes include a normal memory mode, an Out-of-Core (OoC) execution mode, and a Unified Memory (UM) mode. The method includes starting an execution on the NN with the UM mode and measuring the memory usage for each of layers of the NN. The method further includes selecting an execution mode based on the memory usage of all of the layers.
    Type: Application
    Filed: June 26, 2018
    Publication date: December 26, 2019
    Inventors: Yasushi Negishi, Haruki Imai, Taro Sekiyama, Tung D. Le, Kiyokuni Kawachiya
  • Publication number: 20190303025
    Abstract: A method is provided for reducing consumption of a memory in a propagation process for a neural network (NN) having fixed structures for computation order and node data dependency. The memory includes memory segments for allocating to nodes. The method collects, in a NN training iteration, information for each node relating to an allocation, size, and lifetime thereof. The method chooses, responsive to the information, a first node having a maximum memory size relative to remaining nodes, and a second node non-overlapped with the first node lifetime. The method chooses another node non-overlapped with the first node lifetime, responsive to a sum of memory sizes of the second node and the other node not exceeding a first node memory size. The method reallocates a memory segment allocated to the first node to the second node and the other node to be reused by the second node and the other node.
    Type: Application
    Filed: April 2, 2018
    Publication date: October 3, 2019
    Inventors: Taro Sekiyama, Haruki Imai, Jun Doi, Yasushi Negishi
  • Publication number: 20190266488
    Abstract: A generated algorithm used by a neural network is captured during execution of an iteration of the neural network. A candidate algorithm is identified based on the generated algorithm. A determination is made that the candidate algorithm utilizes less memory than the generated algorithm. Based on the determination the neural network is updated by replacing the generated algorithm with the candidate algorithm.
    Type: Application
    Filed: April 16, 2019
    Publication date: August 29, 2019
    Inventors: Taro Sekiyama, Kiyokuni Kawachiya, Tung D. Le, Yasushi Negishi
  • Publication number: 20190205755
    Abstract: A generated algorithm used by a neural network is captured during execution of an iteration of the neural network. A candidate algorithm is identified based on the generated algorithm. A determination is made that the candidate algorithm utilizes less memory than the generated algorithm. Based on the determination the neural network is updated by replacing the generated algorithm with the candidate algorithm.
    Type: Application
    Filed: March 6, 2019
    Publication date: July 4, 2019
    Inventors: Taro Sekiyama, Kiyokuni Kawachiya, Tung D. Le, Yasushi Negishi
  • Publication number: 20190188560
    Abstract: A computer-implemented method, computer program product, and computer processing system are provided for accelerating neural network data parallel training in multiple graphics processing units (GPUs) using at least one central processing unit (CPU). The method includes forming a set of chunks. Each of the chunks includes a respective group of neural network layers other than a last layer. The method further includes performing one or more chunk-wise synchronization operations during a backward phase of the neural network data parallel training, by each of the multiple GPUs and the at least one CPU.
    Type: Application
    Filed: December 15, 2017
    Publication date: June 20, 2019
    Inventors: Tung D. Le, Haruki Imai, Taro Sekiyama, Yasushi Negishi
  • Publication number: 20190156184
    Abstract: In an approach to localizing tree-based convolutional neural networks, a method includes creating a first tree-based convolution layer (TBCL) corresponding to a tree, where the tree includes a first plurality of nodes and a node that has been indicated to be a first pivotal node. The first TBCL includes a second plurality of nodes and a second pivotal node having a feature vector based on node data from the first pivotal node. The method also includes creating a second TBCL corresponding to the tree. The second TBCL may include a third plurality of nodes. The method further includes determining a feature vector a third pivotal node in the third plurality of nodes based on the feature vectors from: (i) the second pivotal node, (ii) a parent node of the second pivotal node, and (ii) a child node of the second pivotal node.
    Type: Application
    Filed: November 17, 2017
    Publication date: May 23, 2019
    Inventors: Tung D. Le, Taro Sekiyama
  • Publication number: 20190138888
    Abstract: A cascading convolutional neural network (CCNN) comprising a plurality of convolutional neural networks (CNNs) that are trained by weighting training data based on loss values of each training datum between CNNs of the CCN. The CCNN can receiving an input image from plurality of images, classify the input image using the CCNN, and present a classification of the input image.
    Type: Application
    Filed: November 3, 2017
    Publication date: May 9, 2019
    Inventors: Taro Sekiyama, Masaharu Sakamoto, Hiroki Nakano, Kun Zhao
  • Patent number: 10268951
    Abstract: A generated algorithm used by a neural network is captured during execution of an iteration of the neural network. A candidate algorithm is identified based on the generated algorithm. A determination is made that the candidate algorithm utilizes less memory than the generated algorithm. Based on the determination the neural network is updated by replacing the generated algorithm with the candidate algorithm.
    Type: Grant
    Filed: June 14, 2017
    Date of Patent: April 23, 2019
    Assignee: International Business Machines Corporation
    Inventors: Taro Sekiyama, Kiyokuni Kawachiya, Tung D. Le, Yasushi Negishi
  • Patent number: 10169874
    Abstract: A target object may be identified by estimating a distribution of a plurality of orientations of a periphery of a target object, and identifying the target object based on the distribution.
    Type: Grant
    Filed: May 30, 2017
    Date of Patent: January 1, 2019
    Assignee: International Business Machines Corporation
    Inventors: Hiroki Nakano, Yasushi Negishi, Masaharu Sakamato, Taro Sekiyama, Kun Zhao
  • Publication number: 20180365558
    Abstract: A generated algorithm used by a neural network is captured during execution of an iteration of the neural network. A candidate algorithm is identified based on the generated algorithm. A determination is made that the candidate algorithm utilizes less memory than the generated algorithm. Based on the determination the neural network is updated by replacing the generated algorithm with the candidate algorithm.
    Type: Application
    Filed: June 14, 2017
    Publication date: December 20, 2018
    Inventors: Taro Sekiyama, Kiyokuni Kawachiya, Tung D. Le, Yasushi Negishi
  • Publication number: 20180357544
    Abstract: A computer-implemented method for optimizing neural networks for receiving plural input data having a form of a tree or a Directed Acyclic Graph (DAG). Finding a common node included in at least two of the input data in common. Reconstructing the plural input data by sharing the common node.
    Type: Application
    Filed: June 8, 2017
    Publication date: December 13, 2018
    Inventors: TUNG D. LE, TARO SEKIYAMA, KUN ZHAO