Patents by Inventor Kiyokuni Kawachiya
Kiyokuni Kawachiya has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11957984Abstract: A server computer is connected to a plurality of client computers through a network, and controls objects in a Metaverse accessed by the client computers. The server computer includes a storage unit for storing an object ID specifying an object accessible in the Metaverse by the plurality of client computers and authenticity information associated with the object ID. The authenticity information indicates that the object is genuine. The server computer also includes a communication unit for communicating with each of the client computers. The server computer also includes an enquiry unit for causing the communication unit to transmit the authenticity information corresponding to the object ID to at least one of the plurality of client computers upon receipt of an enquiry request to enquire about the object ID of the object from one of the plurality of client computers.Type: GrantFiled: March 16, 2021Date of Patent: April 16, 2024Assignee: Activision Publishing, Inc.Inventors: Kiyokuni Kawachiya, Michiaki Tatsubori
-
Patent number: 11880762Abstract: A computer-implemented method, a computer program product, and a computer processing system are provided for selecting from among multiple Graphics Processing Unit (GPU) execution modes for a Neural Network (NN) having a size greater than a threshold size. The multiple GPU execution modes include a normal memory mode, an Out-of-Core (OoC) execution mode, and a Unified Memory (UM) mode. The method includes starting an execution on the NN with the UM mode and measuring the memory usage for each of layers of the NN. The method further includes selecting an execution mode based on the memory usage of all of the layers.Type: GrantFiled: June 26, 2018Date of Patent: January 23, 2024Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Yasushi Negishi, Haruki Imai, Taro Sekiyama, Tung D. Le, Kiyokuni Kawachiya
-
Publication number: 20230153318Abstract: A method for converting a shape and a format of tensor data to meet a specific data format of a hardware accelerator is provided. The method receives input tensors L1 and L2, each being constants having a data format of < X x Y x Z >, and each further having an n-dimension input tensor shape as <Xn x Xn-1 x Xn-2 x ... x X1 >. The method stores input tensor shape. The method calculates an n-dimension modified shape of the input tensors by (a) setting a largest divisor of (Xn x Xn-1 x...x X1 ) ? L1 to S1, (b) setting a largest divisor of ((Xn x Xn-1 x...x X1 ) / S1) ? L2 to S2, (c) setting (((Xn x Xn-1 x... x X1 ) / (S1 x S2)) to S3, and (d) returning the n-dimension modified shape as < S3 x S2 x S1 >.Type: ApplicationFiled: November 16, 2021Publication date: May 18, 2023Inventors: YASUSHI NEGISHI, Tung D. Le, HARUKI IMAI, KIYOKUNI KAWACHIYA
-
Patent number: 11461637Abstract: A generated algorithm used by a neural network is captured during execution of an iteration of the neural network. A candidate algorithm is identified based on the generated algorithm. A determination is made that the candidate algorithm utilizes less memory than the generated algorithm. Based on the determination the neural network is updated by replacing the generated algorithm with the candidate algorithm.Type: GrantFiled: March 6, 2019Date of Patent: October 4, 2022Assignee: International Business Machines CorporationInventors: Taro Sekiyama, Kiyokuni Kawachiya, Tung D. Le, Yasushi Negishi
-
Patent number: 11403278Abstract: Verified snapshots are generated by obtaining, from one of a plurality of first nodes, a difference between a common data at a first time point and the common data at a second time point that is different from the first time point, generating a first snapshot of the common data at the first time point based on the difference, obtaining a hash of the common data at the first time point from one of the plurality of first nodes, and verifying the first snapshot at the first time point with the hash of the common data at the first time point.Type: GrantFiled: December 4, 2020Date of Patent: August 2, 2022Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Hiroshi Horii, Kohichi Kamijoh, Kiyokuni Kawachiya, Shohei Ohsawa, Hironori Takeuchi
-
Patent number: 11379316Abstract: An example operation may include one or more of recording, by a snapshot node, a plurality of snapshots of a key-value storage based on a plurality of delta offsets from an initial snapshot of the key-value storage, receiving, by the snapshot node, an audit request from an audit node that contains an audit time, generating, by the snapshot node, a current snapshot based on an aggregation of the snapshots from the plurality of the snapshots up to a time of a transaction closest to the audit time, and executing, by the snapshot node, a chaincode based on a delta offset of the current snapshot from the time of the transaction to the audit time to restore a snapshot at the audit time.Type: GrantFiled: June 4, 2019Date of Patent: July 5, 2022Assignee: International Business Machines CorporationInventors: Kohichi Kamijoh, Kiyokuni Kawachiya, Takaaki Tateishi, Ryo Kawahara
-
Patent number: 11362670Abstract: A method is presented for compressing data of a Rectified Linear Unit (ReLU) function on a graphical processing unit (GPU) employed in a learning process of a deep neural network. The method includes converting an initial data structure including nonzero data and zero data into a compressed data structure including only the nonzero data of the initial data structure as compressed data by generating a nonzero data bitmap region, generating a nonzero data number table region by employing a parallel reduction algorithm, calculating a nonzero data array index per block region of all blocks from the nonzero data number table region by employing a parallel prefix sum scan algorithm, allocating a buffer for the compressed data; and copying the nonzero data from the initial data structure into a nonzero data array region in a compressed data format in parallel.Type: GrantFiled: October 30, 2020Date of Patent: June 14, 2022Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Yasushi Negishi, Tung D. Le, Haruki Imai, Kiyokuni Kawachiya
-
Patent number: 11327751Abstract: In a distributed processing system having multiple processing nodes including alive nodes and dead nodes, a method is provided for collecting an object from the alive nodes. The method includes maintaining a separate count value for each of remote nodes at which the object is remotely-referenced. The method further includes collecting the object for garbage collection when the separate count value for all of the remotes nodes is zero. The method also includes adding at least one per node sending counter responsive to a remote reference of the object being sent from a first remote node to a particular one of the remote nodes. The at least one per node sending counter is added at the first remote node to count a number of remote-references of the object being sent to the particular one of the remote nodes.Type: GrantFiled: October 29, 2019Date of Patent: May 10, 2022Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventor: Kiyokuni Kawachiya
-
Publication number: 20220140841Abstract: A method is presented for compressing data of a Rectified Linear Unit (ReLU) function on a graphical processing unit (GPU) employed in a learning process of a deep neural network. The method includes converting an initial data structure including nonzero data and zero data into a compressed data structure including only the nonzero data of the initial data structure as compressed data by generating a nonzero data bitmap region, generating a nonzero data number table region by employing a parallel reduction algorithm, calculating a nonzero data array index per block region of all blocks from the nonzero data number table region by employing a parallel prefix sum scan algorithm, allocating a buffer for the compressed data; and copying the nonzero data from the initial data structure into a nonzero data array region in a compressed data format in parallel.Type: ApplicationFiled: October 30, 2020Publication date: May 5, 2022Inventors: Yasushi Negishi, Tung D. Le, Haruki Imai, Kiyokuni Kawachiya
-
Publication number: 20220138580Abstract: Methods and systems for training a neural network include identifying units within a neural network, including a first unit for memory swapping and a second unit for re-computation to balance memory efficiency with computational efficiency. Each unit includes at least one layer of the neural network. Each unit has a first layer that is a checkpoint operation. During a feed-forward training stage, feature maps are stored in a first memory. The feature maps are output by the at least one layer of the first unit. The feature maps are swapped from the first memory to a second memory. During a backpropagation stage, the feature maps for the first unit are swapped from the second memory to the first memory. Feature maps for the second unit are re-computed.Type: ApplicationFiled: November 4, 2020Publication date: May 5, 2022Inventors: Haruki Imai, Tung D. Le, Yasushi Negishi, Kiyokuni Kawachiya
-
Patent number: 11314640Abstract: A method for reducing the cost of stack scanning in garbage collection (GC) includes, in the GC of the first-generation heap area, registering, in a nursery object reference list prepared for each thread, one or more addresses, within each stack, which each refer to a nursery object, and updating a scanning unnecessary area starting pointer such that the addresses listed in the nursery object reference list are included in the area from the bottom of the stack to the address pointed to by the scanning unnecessary area starting pointer. The method further includes, in the next GC of the first-generation heap area, for the area from the bottom of the stack to the address pointed to by the scanning unnecessary area starting pointer, performing the GC processing on the addresses included in the nursery object reference list.Type: GrantFiled: December 11, 2014Date of Patent: April 26, 2022Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Kiyokuni Kawachiya, Tamiya Onodera
-
Publication number: 20210268389Abstract: A server computer is connected to a plurality of client computers through a network, and controls objects in a Metaverse accessed by the client computers. The server computer includes a storage unit for storing an object ID specifying an object accessible in the Metaverse by the plurality of client computers and authenticity information associated with the object ID. The authenticity information indicates that the object is genuine. The server computer also includes a communication unit for communicating with each of the client computers. The server computer also includes an enquiry unit for causing the communication unit to transmit the authenticity information corresponding to the object ID to at least one of the plurality of client computers upon receipt of an enquiry request to enquire about the object ID of the object from one of the plurality of client computers.Type: ApplicationFiled: March 16, 2021Publication date: September 2, 2021Inventors: Kiyokuni Kawachiya, Michiaki Tatsubori
-
Patent number: 10981069Abstract: A server computer is connected to a plurality of client computers through a network, and controls objects in a Metaverse accessed by the client computers. The server computer includes a storage unit for storing an object ID specifying an object accessible in the Metaverse by the plurality of client computers and authenticity information associated with the object ID. The authenticity information indicates that the object is genuine. The server computer also includes a communication unit for communicating with each of the client computers. The server computer also includes an enquiry unit for causing the communication unit to transmit the authenticity information corresponding to the object ID to at least one of the plurality of client computers upon receipt of an enquiry request to enquire about the object ID of the object from one of the plurality of client computers.Type: GrantFiled: October 5, 2017Date of Patent: April 20, 2021Assignee: Activision Publishing, Inc.Inventors: Kiyokuni Kawachiya, Michiaki Tatsubori
-
Patent number: 10984014Abstract: A method is provided for consistent data processing by first and second distributed processing systems having different data partitioning and routing mechanisms such that the first system is without states and the second system is with states. The method includes dividing data in each system into a same number of partitions based on a same key and a same hash function. The method includes mapping partitions between the systems in a one-to-one mapping. The mapping step includes calculating a partition ID based on the hash function and a total number of partitions, and dynamically mapping a partition in the first system to a partition in the second system, responsive to the partition in the first system being unmapped to the partition in the second system.Type: GrantFiled: February 4, 2020Date of Patent: April 20, 2021Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Kiyokuni Kawachiya, Yasushi Negishi, Mikio Takeuchi, Gaku Yamamoto
-
Publication number: 20210089511Abstract: Verified snapshots are generated by obtaining, from one of a plurality of first nodes, a difference between a common data at a first time point and the common data at a second time point that is different from the first time point, generating a first snapshot of the common data at the first time point based on the difference, obtaining a hash of the common data at the first time point from one of the plurality of first nodes, and verifying the first snapshot at the first time point with the hash of the common data at the first time point.Type: ApplicationFiled: December 4, 2020Publication date: March 25, 2021Inventors: Hiroshi Horii, Kohichi Kamijoh, Kiyokuni Kawachiya, Shohei Ohsawa, Hironori Takeuchi
-
Patent number: 10896165Abstract: Verified snapshots are generated by obtaining, from one of a plurality of first nodes, a difference between a common data at a first time point and the common data at a second time point that is different from the first time point, generating a first snapshot of the common data at the first time point based on the difference, obtaining a hash of the common data at the first time point from one of the plurality of first nodes, and verifying the first snapshot at the first time point with the hash of the common data at the first time point.Type: GrantFiled: May 3, 2017Date of Patent: January 19, 2021Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Hiroshi Horii, Kohichi Kamijoh, Kiyokuni Kawachiya, Shohei Ohsawa, Hironori Takeuchi
-
Patent number: 10896166Abstract: Verified snapshots are generated by obtaining, from one of a plurality of first nodes, a difference between a common data at a first time point and the common data at a second time point that is different from the first time point, generating a first snapshot of the common data at the first time point based on the difference, obtaining a hash of the common data at the first time point from one of the plurality of first nodes, and verifying the first snapshot at the first time point with the hash of the common data at the first time point.Type: GrantFiled: November 1, 2017Date of Patent: January 19, 2021Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Hiroshi Horii, Kohichi Kamijoh, Kiyokuni Kawachiya, Shohei Ohsawa, Hironori Takeuchi
-
Patent number: 10884755Abstract: A computer-implemented method is provided for managing GPU memory consumption by computational graph rewriting. The method includes constructing, by a hardware processor, a categorized topological ordering of a computational graph. The categorized topological ordering includes multiple computational nodes arranged in multiple levels. The method further includes estimating, by the hardware processor, the GPU memory consumption responsive to a level including two or more computational nodes from among the multiple computational nodes. The method also includes rewriting, by the hardware processor, the computational graph by linearizing the two or more computational nodes in the level to avoid overlapping of the GPU memory consumption by the two or more computational nodes responsive to the GPU memory consumption exceeding a threshold. The memory additionally includes managing the GPU memory consumption in accordance with the rewritten computational graph.Type: GrantFiled: July 31, 2019Date of Patent: January 5, 2021Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Tung D. Le, Haruki Imai, Yasushi Negishi, Kiyokuni Kawachiya
-
Publication number: 20200387432Abstract: An example operation may include one or more of recording, by a snapshot node, a plurality of snapshots of a key-value storage based on a plurality of delta offsets from an initial snapshot of the key-value storage, receiving, by the snapshot node, an audit request from an audit node that contains an audit time, generating, by the snapshot node, a current snapshot based on an aggregation of the snapshots from the plurality of the snapshots up to a time of a transaction closest to the audit time, and executing, by the snapshot node, a chaincode based on a delta offset of the current snapshot from the time of the transaction to the audit time to restore a snapshot at the audit time.Type: ApplicationFiled: June 4, 2019Publication date: December 10, 2020Inventors: Kohichi Kamijoh, KIYOKUNI KAWACHIYA, Takaaki Tateishi, Ryo Kawahara
-
Publication number: 20200174848Abstract: A method is provided for consistent data processing by first and second distributed processing systems having different data partitioning and routing mechanisms such that the first system is without states and the second system is with states. The method includes dividing data in each system into a same number of partitions based on a same key and a same hash function. The method includes mapping partitions between the systems in a one-to-one mapping. The mapping step includes calculating a partition ID based on the hash function and a total number of partitions, and dynamically mapping a partition in the first system to a partition in the second system, responsive to the partition in the first system being unmapped to the partition in the second system.Type: ApplicationFiled: February 4, 2020Publication date: June 4, 2020Inventors: Kiyokuni Kawachiya, Yasushi Negishi, Mikio Takeuchi, Gaku Yamamoto