Abstract: An electronic apparatus includes a communication interface to communicate with external servers; and a processor to control the electronic apparatus by executing at least one instruction. The processor is configured to receive, from each of the external servers, values of bandwidths of a plurality of GPU pairs into which a plurality of GPUs included in the external servers are combined and information on whether the plurality of GPUs are used, based on a input job related to machine learning being received, identify a number of GPUs and a bandwidth value that are required for performing tasks included in the input job, and determine GPUs among the plurality of GPUs to perform the tasks based on the values of the bandwidths of the plurality of GPU pairs, the received information on whether the plurality of GPUs are used, and the number of GPUs and the bandwidth value.
Abstract: Techniques usable in optimization processing are described. A system includes an optimization processing unit (OPU). The OPU includes stochastic computing units and at least one programmable interconnect. Each of the stochastic computing units includes nodes and multiplication unit(s) configured to interconnect at least a portion of the nodes. The programmable interconnect(s) are configured to provide weights for and to selectably couple a portion of the stochastic computing units.
Abstract: Resegmenting chunks of data for load balancing is disclosed. A plurality of first chunks of data is received. The plurality of first chunks of data includes one or more entries that include raw data produced by a component of an information technology environment and that reflects activity in the information technology environment. The plurality of first chunks of data is resegmented into a plurality of second chunks of data based on a source type of the plurality of first chunks. A first subset of the plurality of second chunks of data is distributed to a first indexer of a set of indexers. An occurrence of a trigger event is determined, and in response to the trigger event, a second subset of the plurality of second chunks of data is distributed to a second indexer of the set of indexers.
Type:
Grant
Filed:
April 22, 2021
Date of Patent:
March 7, 2023
Assignee:
SPLUNK INC.
Inventors:
Jag Kerai, Anish Shrigondekar, Mitchell Blank, Jr., Hasan Alayli
Abstract: Embodiments of systems and methods for managing performance optimization of applications executed by an Information Handling System (IHS) are described. In an illustrative, non-limiting embodiment, an IHS may include computer-executable instructions to, for each of multiple resources used to execute a target application, receive one or more machine learning (ML)-based hints associated with each resource that have been generated by a ML-based optimization service, and generate one or more augmented hints for at least one of the resources using a ML-to-ML orchestration service. The ML-to-ML orchestration service then transmits the augmented hints to the ML-based optimization service that combines the augmented ML-based hints with the one or more internally generated hints to generate augmented profile recommendations that are, in turn, used to adjust one or more settings of the resource to optimize a performance of the target application executed by the resource.
Type:
Grant
Filed:
January 13, 2021
Date of Patent:
February 28, 2023
Assignee:
Dell Products, L.P.
Inventors:
Farzad Khosrowpour, Mitchell Anthony Markow
Abstract: In general, embodiments are disclosed herein for tracking and allocating graphics hardware resources. In one embodiment, a software and/or firmware process constructs a cross-application command queue utilization table based on one or more specified command queue quality of service (QoS) settings, in order to track the target and current utilization rates of each command queue on the graphics hardware over a given frame and to load work onto the graphics hardware in accordance with the utilization table. Based on the constructed utilization table for a given frame, any command queues that have exceed their respective target utilization value may be moved to an “inactive” status for the duration of the current frame. For any command queues that remain in an “active” status for the current frame, work from those command queues may be loaded on to slots of the appropriate data masters of the graphics hardware in any desired order.
Abstract: A computer-implemented method for efficient and scalable enclave protection for machine learning (ML) programs includes tailoring at least one ML program to generate at least one tailored ML program for execution within at least one enclave, and executing the at least one tailored ML program within the at least one enclave.
Type:
Grant
Filed:
March 12, 2020
Date of Patent:
February 7, 2023
Inventors:
Chung Hwan Kim, Junghwan Rhee, Xiao Yu, Luan Tang, Haifeng Chen, Kyungtae Kim
Abstract: A system and method for providing dynamic device virtualization is herein disclosed. According to one embodiment, the computer-implemented method includes providing a hypervisor and one or more guest virtual machines (VMs). Each guest VM is disposed to run a guest user process and the hypervisor is split into a device hypervisor and a compute hypervisor. The computer-implemented method further includes providing an interface between the device hypervisor and the compute hypervisor. The compute hypervisor manages an efficient use of CPU and memory of a host and the device hypervisor manages a device connected to the host by exploiting hardware acceleration of the device.
Abstract: In some embodiments, a method includes: displaying, on a first client device, a plurality of tasks; identifying, by the first client device, a task from the plurality of tasks, the task transferrable to a second client device in communication with the first client device; and sending, by the first client device, metadata for the task to the second client device in response to input received by the first client device, the task including metadata to allowing the second client device to display the task in the same manner as the task was displayed by the first client device.
Abstract: The present disclosure is directed to methods and apparatus for evaluating resources that would be used by machine learning model(s) for purposes of implementing the machine learning model(s) on resource-constrained devices. For example, in one aspect, a plurality of layers in a machine learning model may be identified. A plurality of respective output sizes corresponding to the plurality of layers may be calculated. Based on the plurality of output sizes, a maximum amount of volatile memory used for application of the machine learning model may be estimated and compared to a volatile memory constraint of a resource-constrained computing device. Output indicative of a result of the comparing may be provided at one or more output components.
Abstract: A specialized in-memory database health check process is utilized to resolve dependencies in a resource indicating requirements for an instance of an in-memory database. Specifically, when an instance of an in-memory database is created in response to a request, a list of one or more component handlers are obtained. These component handlers are modular functions, separate from each other but potentially dependent on one or more other component handlers, and act to validate various requirements listed in a resource for the request. Each of the component handlers are executed individually during execution of a Reconcile function. To the extent that the execution of any component handlers in the list is unsuccessful, the Reconcile function is rerun for another iteration. These iterations continue until all component handlers report back as successful. Instance creation is then considered successful and the instance of the in-memory database can be utilized by users.
Type:
Grant
Filed:
May 20, 2020
Date of Patent:
December 27, 2022
Assignee:
SAP SE
Inventors:
Jannick Stephan Fahlbusch, Bryon Hummel
Abstract: Systems, methods, and apparatuses for resource monitoring identification reuse are described. In an embodiment, a system comprising a hardware processor core to execute instructions storage for a resource monitoring identification (RMID) recycling instructions to be executed by a hardware processor core, a logical processor to execute on the hardware processor core, the logical processor including associated storage for a RMID and state, are described.
Type:
Grant
Filed:
October 22, 2020
Date of Patent:
December 20, 2022
Assignee:
Intel Corporation
Inventors:
Matthew Fleming, Edwin Verplanke, Andrew Herdrich, Ravishankar Iyer
Abstract: Systems, methods, and interfaces for the management of virtual machine instances and other programmatically controlled networks are provided. The hosted virtual networks are configured in a manner such that a virtual machine manager of the virtual network may monitor activity such as user requests, network traffic, and the status and execution of various virtual machine instances to determine possible security assessments. Aspects of the virtual network may be assessed for vulnerabilities at varying levels of granularity and sophistication when a suspicious event or triggering activity is detected. Illustrative embodiments of the systems and methods may be implemented on a virtual network overlaid on one or more intermediate physical networks that are used as a substrate network.
Type:
Grant
Filed:
December 17, 2014
Date of Patent:
December 6, 2022
Assignee:
Amazon Technologies, Inc.
Inventors:
Eric Jason Brandwine, Donald L. Bailey, Jr.
Abstract: A scaling manager manages deques that track groups of preinitialized instances used to scale respective groups of active compute instances. Various techniques for deque management include a technique where a total instance quantity is preconfigured for the total number of instances assigned to both the group and the deque of preinitialized instances. As the size of the group grows for scale-ups, the size of the deque may go down. For example, the deque is not replenished when the group scales, but does expand when the group scales down. The total instance quantity may be bounded, in some examples, and an additional “buffer amount” of preinitialized instances may be implemented to provide a safety margin for burst scaling, which can be further enhanced by transferring instances between data structures of different groups of instances in some cases.
Abstract: Methods, systems, and computer-readable media for resource usage restrictions in a time-series database are disclosed. Elements of a plurality of time series are stored into one or more storage tiers of a time-series database. The time series are associated with a plurality of clients of the time-series database. Execution of tasks is initiated using one or more resources of one or more hosts. The time-series elements represent inputs to the tasks. The tasks comprise a first task and a second task. A usage of the one or more resources by the first task is determined to violate one or more resource usage restrictions. Based at least in part on the usage, one or more actions are performed to modify the execution of the first task. The one or more actions increase an amount of the one or more resources available to the second task.
Abstract: Methods, apparatus, and processor-readable storage media for determining capacity in storage systems using machine learning techniques are provided herein. An example computer-implemented method includes obtaining capacity-related data from a storage system; forecasting, for a given temporal period, capacity of one or more storage objects of the storage system by applying machine learning techniques to at least a portion of the capacity-related data; aggregating the forecasted capacity for at least portions of the one or more storage objects; determining, based on the aggregated forecasted capacity of the storage objects, whether at least a portion of the storage system will run out of capacity in connection with the given temporal period; and performing one or more automated actions based at least in part on the determination as to whether the at least a portion of the at least one storage system will run out of capacity.
Abstract: Example embodiments relate generally to systems and methods for continuous data protection (CDP) and more specifically to an input and output (I/O) filtering framework and log management system to seek a near-zero recovery point objective (RPO).
Type:
Grant
Filed:
April 30, 2019
Date of Patent:
November 15, 2022
Assignee:
Rubrik, Inc.
Inventors:
Benjamin Travis Meadowcroft, Li Ding, Shaomin Chen, Hardik Vohra, Arijit Banerjee, Abhay Mitra, Kushaagra Goyal, Arnav Gautum Mishra, Samir Rishi Chaudhry, Suman Swaroop, Kunal Sean Munshani, Mudit Malpani
Abstract: A solution is proposed for resource management of a software application including a plurality of software components interacting with each other. A corresponding method includes monitoring present conditions of the software components and estimating a future consumption of one or more computing resources by each software component from the present conditions of the software components; an allocation of the computing resources to the software components is then controlled accordingly. A computer program and a computer program product for performing the method are also proposed. Moreover, a system for implementing the method is proposed.
Type:
Grant
Filed:
July 31, 2020
Date of Patent:
November 15, 2022
Assignee:
International Business Machines Corporation
Abstract: A method for scheduling jobs for the calculator includes measuring core utilization of the second-type processor, when the measured core utilization is less than a reference value, transmitting, by the first-type processor, a job suspension instruction to suspend a first job, which is currently being executed, to the second-type processor, in response to the job suspension instruction, copying data of a region occupied by the first job in a memory of the second-type processor to a main memory, copying data of a second job stored in the main memory to the memory of the second-type processor, and transmitting, by the first-type processor, an instruction to execute the second job to the second-type processor.
Type:
Grant
Filed:
October 25, 2019
Date of Patent:
November 8, 2022
Assignee:
SAMSUNG SDS CO., LTD.
Inventors:
Man Suk Suh, Hwan Kyun Roh, Gi Beom Pang
Abstract: Distributing computation workload among computing nodes of differing computing paradigms is provided. Compute gravity of each computing node in a cloud computing paradigm and each computing node in a client network computing paradigm within an Internet of Systems is calculated. Each component part of an algorithm is distributed to an appropriate computing node of the cloud computing paradigm and client network computing paradigm based on calculated compute gravity of each respective computing node within the Internet of Systems. Computation workload of each component part of the algorithm is distributed to a respective computing node of the cloud computing paradigm and the client network computing paradigm having a corresponding component part of the algorithm for processing.
Type:
Grant
Filed:
January 7, 2020
Date of Patent:
October 25, 2022
Assignee:
International Business Machines Corporation
Inventors:
Aaron K. Baughman, Stephen C. Hammer, Gray Cannon, Shikhar Kwatra
Abstract: This disclosure describes systems, devices, and techniques for live migrating virtualized resources between the main region and edge locations. Live migration enables virtualized resources to remain operational during migration. Edge locations are typically separated from secure data centers via the Internet, a direct connection, or some other intermediate network. Accordingly, to place virtualized resources within an edge location, the virtualized resources must be migrated over a secure communication tunnel that can protect virtualized resource data during transmission over the intermediate network. The secure communication tunnel may have limited data throughput. To efficiently utilize resources of the secure communication tunnel, virtualized resource data may be transferred over the tunnel in a two-stage process.
Type:
Grant
Filed:
November 21, 2019
Date of Patent:
October 4, 2022
Assignee:
Amazon Technologies, Inc.
Inventors:
Oleksii Tsai, Nikolay Krasilnikov, Anton Valter, Alexey Gadalin