Patents by Inventor Harumi Kuno
Harumi Kuno has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240406251Abstract: In some examples, a system having a plurality of computer nodes receives a command based on program code of a program being developed in an interactive programming session. The system distributes data items from a network-attached memory to a distributed data object having data in node memories of the plurality of computer nodes. A dataset manager performs an operation specified by the command on the distributed data object, the operation executed in parallel on the plurality of computer nodes. The dataset manager produces derived data generated by the operation on the distributed data object, the derived data accessible by a programmer in the interactive programming session.Type: ApplicationFiled: October 26, 2023Publication date: December 5, 2024Inventors: Harumi Kuno, John L. Byrne, Paolo Faraboschi, Sharad Singhal
-
Patent number: 11644882Abstract: One embodiment provides a system and method for predicting network power usage associated with workloads. During operation, the system configures a simulator to simulate operations of a plurality of network components, which comprises embedding one or more event counters in each simulated network component. A respective event counter is configured to count a number of network-power-related events. The system collects, based on values of the event counters, network-power-related performance data associated with one or more sample workloads applied to the simulator; and trains a machine-learning model with the collected network-power-related performance data and characteristics of the sample workloads as training data 1, thereby facilitating prediction of network-power-related performance associated with a to-be-evaluated workload.Type: GrantFiled: June 2, 2021Date of Patent: May 9, 2023Assignee: Hewlett Packard Enterprise Development LPInventors: Harumi Kuno, Alan Davis, Torsten Wilde, Daniel William Dauwe, Duncan Roweth, Ryan Dean Menhusen, Sergey Serebryakov, John L. Byrne, Vipin Kumar Kukkala, Sai Rahul Chalamalasetti
-
Patent number: 11556438Abstract: While scheduled checkpoints are being taken of a cluster of active compute nodes distributively executing an application in parallel, a likelihood of failure of the active compute nodes is periodically and independently predicted. Responsive to the likelihood of failure of a given active compute node exceeding a threshold, the given active compute node is proactively migrated to a spare compute node of the cluster at a next scheduled checkpoint. Another spare compute node of the cluster can perform prediction and migration. Prediction can be based on both hardware events and software events regarding the active compute nodes.Type: GrantFiled: August 17, 2020Date of Patent: January 17, 2023Assignee: Hewlett Packard Enterprise Development LPInventors: Cong Xu, Naveen Muralimanohar, Harumi Kuno
-
Publication number: 20220390999Abstract: One embodiment provides a system and method for predicting network power usage associated with workloads. During operation, the system configures a simulator to simulate operations of a plurality of network components, which comprises embedding one or more event counters in each simulated network component. A respective event counter is configured to count a number of network-power-related events. The system collects, based on values of the event counters, network-power-related performance data associated with one or more sample workloads applied to the simulator; and trains a machine-learning model with the collected network-power-related performance data and characteristics of the sample workloads as training data 1, thereby facilitating prediction of network-power-related performance associated with a to-be-evaluated workload.Type: ApplicationFiled: June 2, 2021Publication date: December 8, 2022Inventors: Harumi Kuno, Alan Davis, Torsten Wilde, Daniel William Dauwe, Duncan Roweth, Ryan Dean Menhusen, Sergey Serebryakov, John L. Byrne, Vipin Kumar Kukkala, Sai Rahul Chalamalasetti
-
Patent number: 11210089Abstract: Methods and systems for conducting vector send operations are provided. The processor of a sender node receives a request to perform a collective send operation (e.g., MPI_Broadcast) from a user application, requesting a copy of data in one or more send buffers by sent to each of a plurality of destinations in a destination vector. The processor invokes a vector send operation from a software communications library, placing a remote enqueue atomic send command for each destination node of the destination vector in an entry of a transmit data mover (XDM) command queue in a single call. The processor executes all of the commands in the XDM command queue and writes the data in the one or more send buffers into each receive queue of each destination identified in the destination vector.Type: GrantFiled: July 11, 2019Date of Patent: December 28, 2021Assignee: Hewlett Packard Enterprise Development LPInventors: John L. Byrne, Harumi Kuno, Jeffrey Drummond
-
Patent number: 10929399Abstract: Computer-implemented systems and associated operating methods take measurements and landmarks associated with robustness maps and perform tests evaluating the robustness of a database engine's operator implementations and/or query components. The illustrative computer-implemented system comprises logic that receives one or more robustness maps of measured database system performance acquired during database execution in a predetermined range of runtime conditions and uses information from the robustness map or maps to perform regression testing wherein landmarks in the robustness maps are operated upon as a robustness bugs describing conditions under which a predetermined implementation of a database operator or query component degrades in a manner different from a predetermined expected manner.Type: GrantFiled: February 2, 2009Date of Patent: February 23, 2021Assignee: Micro Focus LLCInventors: Goetz Graefe, Harumi Kuno, Janet L. Wiener
-
Patent number: 10929293Abstract: A system includes a plurality of processes, a network fabric, and a shared memory accessible by the plurality of processes over the network fabric, the shared memory to store a plurality of elements of a data structure. A first process is designated as being allowed to update a target variable stored in the shared memory, and a second process of the plurality of processes writes a request for an atomic operation to a first region in the shared memory. The first process is responsive to the request to perform the atomic operation that updates the target variable, and write a result including a value of the updated target variable to a second region in the shared memory, the second region readable by the second process, the request and the result being elements of the data structure.Type: GrantFiled: April 30, 2018Date of Patent: February 23, 2021Assignee: Hewlett Packard Enterprise Development LPInventors: John L. Byrne, Harumi Kuno, Khemraj Shukla, Wei Zhang
-
Publication number: 20210011720Abstract: Methods and systems for conducting vector send operations are provided. The processor of a sender node receives a request to perform a collective send operation (e.g., MPI_Broadcast) from a user application, requesting a copy of data in one or more send buffers by sent to each of a plurality of destinations in a destination vector. The processor invokes a vector send operation from a software communications library, placing a remote enqueue atomic send command for each destination node of the destination vector in an entry of a transmit data mover (XDM) command queue in a single call. The processor executes all of the commands in the XDM command queue and writes the data in the one or more send buffers into each receive queue of each destination identified in the destination vector.Type: ApplicationFiled: July 11, 2019Publication date: January 14, 2021Inventors: JOHN L. BYRNE, HARUMI KUNO, JEFFREY DRUMMOND
-
Publication number: 20200379858Abstract: While scheduled checkpoints are being taken of a cluster of active compute nodes distributively executing an application in parallel, a likelihood of failure of the active compute nodes is periodically and independently predicted. Responsive to the likelihood of failure of a given active compute node exceeding a threshold, the given active compute node is proactively migrated to a spare compute node of the cluster at a next scheduled checkpoint. Another spare compute node of the cluster can perform prediction and migration. Prediction can be based on both hardware events and software events regarding the active compute nodes.Type: ApplicationFiled: August 17, 2020Publication date: December 3, 2020Inventors: Cong Xu, Naveen Muralimanohar, Harumi Kuno
-
Patent number: 10776225Abstract: While scheduled checkpoints are being taken of a cluster of active compute nodes distributively executing an application in parallel, a likelihood of failure of the active compute nodes is periodically and independently predicted. Responsive to the likelihood of failure of a given active compute node exceeding a threshold, the given active compute node is proactively migrated to a spare compute node of the cluster at a next scheduled checkpoint. Another spare compute node of the cluster can perform prediction and migration. Prediction can be based on both hardware events and software events regarding the active compute nodes.Type: GrantFiled: June 29, 2018Date of Patent: September 15, 2020Assignee: Hewlett Packard Enterprise Development LPInventors: Cong Xu, Naveen Muralimanohar, Harumi Kuno
-
Patent number: 10613949Abstract: In some examples, a node of a computing system may include a failure identification engine and a failure response engine. The failure identification engine may identify a failure condition for a system function of the node and the failure response engine may store a failure indication in a shared memory to trigger takeover of the system function by a different node of the computing system.Type: GrantFiled: September 24, 2015Date of Patent: April 7, 2020Assignee: Hewlett Packard Enterprise Development LPInventors: Charles Johnson, Harumi Kuno, Al Davis
-
Patent number: 10565037Abstract: A high performance computing system that includes a shared fabric memory and a plurality of processors is disclosed. A first processor is coupled to a local storage and executes a first process that, in combination with other processes, causes the plurality of processors to perform certain actions including transferring, from the shared fabric memory to the local storage, a first value corresponding to a first cell of a first set of cells and a first sweep of a stencil code. The actions further include transferring, from a first logical partition in the shared fabric memory associated with the first cell to the local storage, a second value corresponding to a second cell related to the first cell and not in the first set of cells. Further, these actions include updating, by the first process, the first value based on at least the first value and the second value.Type: GrantFiled: December 19, 2017Date of Patent: February 18, 2020Assignee: Hewlett Packard Enterprise Development LPInventors: Charles Johnson, Mesut Kuscu, Onkar Patil, James Hyungsun Park, Harumi Kuno, Robert Schreiber
-
Patent number: 10540227Abstract: A high performance computing system including processing circuitry and a shared fabric memory is disclosed. The processing circuitry includes processors coupled to local storages. The shared fabric memory includes memory devices and is coupled to the processing circuitry. The shared fabric memory executes a first sweep of a stencil code by sequentially retrieving data stripes. Further, for each retrieved data stripe, a set of values of the retrieved data stripe are updated substantially simultaneously. For each retrieved data stripe, the updated set of values are stored in a free memory gap adjacent to the retrieved data stripe. For each retrieved data stripe, the free memory gap is advanced to an adjacent memory location. A sweep status indicator is incremented from the first sweep to a second sweep.Type: GrantFiled: January 3, 2018Date of Patent: January 21, 2020Assignee: Hewlett Packard Enterprise Development LPInventors: Charles Johnson, Onkar Patil, Mesut Kuscu, Tuan Tran, Joseph Tucek, Harumi Kuno, Milind Chabbi, William Scherer
-
Publication number: 20200004648Abstract: While scheduled checkpoints are being taken of a cluster of active compute nodes distributively executing an application in parallel, a likelihood of failure of the active compute nodes is periodically and independently predicted. Responsive to the likelihood of failure of a given active compute node exceeding a threshold, the given active compute node is proactively migrated to a spare compute node of the cluster at a next scheduled checkpoint. Another spare compute node of the cluster can perform prediction and migration. Prediction can be based on both hardware events and software events regarding the active compute nodes.Type: ApplicationFiled: June 29, 2018Publication date: January 2, 2020Inventors: Cong Xu, Naveen Muralimanohar, Harumi Kuno
-
Patent number: 10482013Abstract: Systems and methods associated with page modification are disclosed. One example method may be embodied on a non-transitory computer-readable medium storing computer-executable instructions. The instructions, when executed by a computer, may cause the computer to fetch a page to a buffer pool in a memory. The page may be fetched from at least one of a log and a backup using single page recovery. The instructions may also cause the computer to store a modification of the page to the log. The modification may be stored to the log as a log entry. The instructions may also cause the computer to evict the page from memory when the page is replaced in the buffer pool. Page writes associated with the eviction may be elided.Type: GrantFiled: September 30, 2014Date of Patent: November 19, 2019Assignee: HEWLETT PACKARD ENTERPRISE DEVELOPMENT LPInventors: Charles S. Johnson, Harumi Kuno, Goetz Graefe, Haris Volos, Mark Lillibridge, James Hyungsun Park, Wey Guy
-
Publication number: 20190332529Abstract: A system includes a plurality of processes, a network fabric, and a shared memory accessible by the plurality of processes over the network fabric, the shared memory to store a plurality of elements of a data structure. A first process is designated as being allowed to update a target variable stored in the shared memory, and a second process of the plurality of processes writes a request for an atomic operation to a first region in the shared memory. The first process is responsive to the request to perform the atomic operation that updates the target variable, and write a result including a value of the updated target variable to a second region in the shared memory, the second region readable by the second process, the request and the result being elements of the data structure.Type: ApplicationFiled: April 30, 2018Publication date: October 31, 2019Inventors: John L. Byrne, Harumi Kuno, Khemraj Shukla, Wei Zhang
-
Patent number: 10360206Abstract: Systems and methods associated with latch-free searching are disclosed. One example method includes receiving a key identifying data to be retrieved from a tree-based data structure. The method also includes performing a concurrent, latch-free search of the tree-based data structure until a leaf node is reached. The method also includes validating the leaf node. The method also includes retreading a portion of the search if the leaf node fails validation.Type: GrantFiled: January 16, 2014Date of Patent: July 23, 2019Assignee: HEWLETT PACKARD ENTERPRISE DEVELOPMENT LPInventors: Joseph Tucek, Mark Lillibridge, Harumi Kuno, Goetz Graefe
-
Publication number: 20190205205Abstract: A high performance computing system including processing circuitry and a shared fabric memory is disclosed. The processing circuitry includes processors coupled to local storages. The shared fabric memory includes memory devices and is coupled to the processing circuitry. The shared fabric memory executes a first sweep of a stencil code by sequentially retrieving data stripes. Further, for each retrieved data stripe, a set of values of the retrieved data stripe are updated substantially simultaneously. For each retrieved data stripe, the updated set of values are stored in a free memory gap adjacent to the retrieved data stripe. For each retrieved data stripe, the free memory gap is advanced to an adjacent memory location. A sweep status indicator is incremented from the first sweep to a second sweep.Type: ApplicationFiled: January 3, 2018Publication date: July 4, 2019Inventors: Charles Johnson, Onkar Patil, Mesut Kuscu, Tuan Tran, Joseph Tucek, Harumi Kuno, Milind Chabbi, William Scherer
-
Publication number: 20190187924Abstract: A high performance computing system that includes a shared fabric memory and a plurality of processors is disclosed. A first processor is coupled to a local storage and executes a first process that, in combination with other processes, causes the plurality of processors to perform certain actions including transferring, from the shared fabric memory to the local storage, a first value corresponding to a first cell of a first set of cells and a first sweep of a stencil code. The actions further include transferring, from a first logical partition in the shared fabric memory associated with the first cell to the local storage, a second value corresponding to a second cell related to the first cell and not in the first set of cells. Further, these actions include updating, by the first process, the first value based on at least the first value and the second value.Type: ApplicationFiled: December 19, 2017Publication date: June 20, 2019Inventors: Charles Johnson, Mesut Kuscu, Onkar Patil, James Hyungsun Park, Harumi Kuno, Robert Schreiber
-
Patent number: 10262035Abstract: Disclosed herein are a system, non transitory computer-readable medium, and method for estimating database performance. A request for an estimate of data is read. The estimate is calculated based at least partially on a node located in a data structure.Type: GrantFiled: November 14, 2013Date of Patent: April 16, 2019Assignee: Hewlett Packard Enterprise Development LPInventors: Harumi Kuno, Goetz Graefe