Patents Examined by Jorge A Chu Joy
  • Patent number: 11609793
    Abstract: An electronic apparatus includes a communication interface to communicate with external servers; and a processor to control the electronic apparatus by executing at least one instruction. The processor is configured to receive, from each of the external servers, values of bandwidths of a plurality of GPU pairs into which a plurality of GPUs included in the external servers are combined and information on whether the plurality of GPUs are used, based on a input job related to machine learning being received, identify a number of GPUs and a bandwidth value that are required for performing tasks included in the input job, and determine GPUs among the plurality of GPUs to perform the tasks based on the values of the bandwidths of the plurality of GPU pairs, the received information on whether the plurality of GPUs are used, and the number of GPUs and the bandwidth value.
    Type: Grant
    Filed: August 13, 2020
    Date of Patent: March 21, 2023
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventor: Dongeun Suh
  • Patent number: 11604913
    Abstract: Techniques usable in optimization processing are described. A system includes an optimization processing unit (OPU). The OPU includes stochastic computing units and at least one programmable interconnect. Each of the stochastic computing units includes nodes and multiplication unit(s) configured to interconnect at least a portion of the nodes. The programmable interconnect(s) are configured to provide weights for and to selectably couple a portion of the stochastic computing units.
    Type: Grant
    Filed: April 13, 2020
    Date of Patent: March 14, 2023
    Assignee: Sync Computing Corp.
    Inventors: Jeffrey Chou, Suraj Bramhavar
  • Patent number: 11599396
    Abstract: Resegmenting chunks of data for load balancing is disclosed. A plurality of first chunks of data is received. The plurality of first chunks of data includes one or more entries that include raw data produced by a component of an information technology environment and that reflects activity in the information technology environment. The plurality of first chunks of data is resegmented into a plurality of second chunks of data based on a source type of the plurality of first chunks. A first subset of the plurality of second chunks of data is distributed to a first indexer of a set of indexers. An occurrence of a trigger event is determined, and in response to the trigger event, a second subset of the plurality of second chunks of data is distributed to a second indexer of the set of indexers.
    Type: Grant
    Filed: April 22, 2021
    Date of Patent: March 7, 2023
    Assignee: SPLUNK INC.
    Inventors: Jag Kerai, Anish Shrigondekar, Mitchell Blank, Jr., Hasan Alayli
  • Patent number: 11593178
    Abstract: Embodiments of systems and methods for managing performance optimization of applications executed by an Information Handling System (IHS) are described. In an illustrative, non-limiting embodiment, an IHS may include computer-executable instructions to, for each of multiple resources used to execute a target application, receive one or more machine learning (ML)-based hints associated with each resource that have been generated by a ML-based optimization service, and generate one or more augmented hints for at least one of the resources using a ML-to-ML orchestration service. The ML-to-ML orchestration service then transmits the augmented hints to the ML-based optimization service that combines the augmented ML-based hints with the one or more internally generated hints to generate augmented profile recommendations that are, in turn, used to adjust one or more settings of the resource to optimize a performance of the target application executed by the resource.
    Type: Grant
    Filed: January 13, 2021
    Date of Patent: February 28, 2023
    Assignee: Dell Products, L.P.
    Inventors: Farzad Khosrowpour, Mitchell Anthony Markow
  • Patent number: 11593175
    Abstract: In general, embodiments are disclosed herein for tracking and allocating graphics hardware resources. In one embodiment, a software and/or firmware process constructs a cross-application command queue utilization table based on one or more specified command queue quality of service (QoS) settings, in order to track the target and current utilization rates of each command queue on the graphics hardware over a given frame and to load work onto the graphics hardware in accordance with the utilization table. Based on the constructed utilization table for a given frame, any command queues that have exceed their respective target utilization value may be moved to an “inactive” status for the duration of the current frame. For any command queues that remain in an “active” status for the current frame, work from those command queues may be loaded on to slots of the appropriate data masters of the graphics hardware in any desired order.
    Type: Grant
    Filed: May 2, 2022
    Date of Patent: February 28, 2023
    Assignee: Apple Inc.
    Inventors: Kutty Banerjee, Michael Imbrogno
  • Patent number: 11573828
    Abstract: A computer-implemented method for efficient and scalable enclave protection for machine learning (ML) programs includes tailoring at least one ML program to generate at least one tailored ML program for execution within at least one enclave, and executing the at least one tailored ML program within the at least one enclave.
    Type: Grant
    Filed: March 12, 2020
    Date of Patent: February 7, 2023
    Inventors: Chung Hwan Kim, Junghwan Rhee, Xiao Yu, Luan Tang, Haifeng Chen, Kyungtae Kim
  • Patent number: 11573813
    Abstract: A system and method for providing dynamic device virtualization is herein disclosed. According to one embodiment, the computer-implemented method includes providing a hypervisor and one or more guest virtual machines (VMs). Each guest VM is disposed to run a guest user process and the hypervisor is split into a device hypervisor and a compute hypervisor. The computer-implemented method further includes providing an interface between the device hypervisor and the compute hypervisor. The compute hypervisor manages an efficient use of CPU and memory of a host and the device hypervisor manages a device connected to the host by exploiting hardware acceleration of the device.
    Type: Grant
    Filed: April 12, 2021
    Date of Patent: February 7, 2023
    Assignee: Dynavisor, Inc.
    Inventor: Sreekumar Ramakrishnan Nair
  • Patent number: 11561822
    Abstract: In some embodiments, a method includes: displaying, on a first client device, a plurality of tasks; identifying, by the first client device, a task from the plurality of tasks, the task transferrable to a second client device in communication with the first client device; and sending, by the first client device, metadata for the task to the second client device in response to input received by the first client device, the task including metadata to allowing the second client device to display the task in the same manner as the task was displayed by the first client device.
    Type: Grant
    Filed: June 1, 2020
    Date of Patent: January 24, 2023
    Inventors: Yuran Ou, Fenghua Jie
  • Patent number: 11551147
    Abstract: The present disclosure is directed to methods and apparatus for evaluating resources that would be used by machine learning model(s) for purposes of implementing the machine learning model(s) on resource-constrained devices. For example, in one aspect, a plurality of layers in a machine learning model may be identified. A plurality of respective output sizes corresponding to the plurality of layers may be calculated. Based on the plurality of output sizes, a maximum amount of volatile memory used for application of the machine learning model may be estimated and compared to a volatile memory constraint of a resource-constrained computing device. Output indicative of a result of the comparing may be provided at one or more output components.
    Type: Grant
    Filed: April 21, 2020
    Date of Patent: January 10, 2023
    Assignee: Koninklijke Philips N.V.
    Inventor: Maurice Leonardus Anna Stassen
  • Patent number: 11537437
    Abstract: A specialized in-memory database health check process is utilized to resolve dependencies in a resource indicating requirements for an instance of an in-memory database. Specifically, when an instance of an in-memory database is created in response to a request, a list of one or more component handlers are obtained. These component handlers are modular functions, separate from each other but potentially dependent on one or more other component handlers, and act to validate various requirements listed in a resource for the request. Each of the component handlers are executed individually during execution of a Reconcile function. To the extent that the execution of any component handlers in the list is unsuccessful, the Reconcile function is rerun for another iteration. These iterations continue until all component handlers report back as successful. Instance creation is then considered successful and the instance of the in-memory database can be utilized by users.
    Type: Grant
    Filed: May 20, 2020
    Date of Patent: December 27, 2022
    Assignee: SAP SE
    Inventors: Jannick Stephan Fahlbusch, Bryon Hummel
  • Patent number: 11531562
    Abstract: Systems, methods, and apparatuses for resource monitoring identification reuse are described. In an embodiment, a system comprising a hardware processor core to execute instructions storage for a resource monitoring identification (RMID) recycling instructions to be executed by a hardware processor core, a logical processor to execute on the hardware processor core, the logical processor including associated storage for a RMID and state, are described.
    Type: Grant
    Filed: October 22, 2020
    Date of Patent: December 20, 2022
    Assignee: Intel Corporation
    Inventors: Matthew Fleming, Edwin Verplanke, Andrew Herdrich, Ravishankar Iyer
  • Patent number: 11522896
    Abstract: Systems, methods, and interfaces for the management of virtual machine instances and other programmatically controlled networks are provided. The hosted virtual networks are configured in a manner such that a virtual machine manager of the virtual network may monitor activity such as user requests, network traffic, and the status and execution of various virtual machine instances to determine possible security assessments. Aspects of the virtual network may be assessed for vulnerabilities at varying levels of granularity and sophistication when a suspicious event or triggering activity is detected. Illustrative embodiments of the systems and methods may be implemented on a virtual network overlaid on one or more intermediate physical networks that are used as a substrate network.
    Type: Grant
    Filed: December 17, 2014
    Date of Patent: December 6, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Eric Jason Brandwine, Donald L. Bailey, Jr.
  • Patent number: 11520638
    Abstract: A scaling manager manages deques that track groups of preinitialized instances used to scale respective groups of active compute instances. Various techniques for deque management include a technique where a total instance quantity is preconfigured for the total number of instances assigned to both the group and the deque of preinitialized instances. As the size of the group grows for scale-ups, the size of the deque may go down. For example, the deque is not replenished when the group scales, but does expand when the group scales down. The total instance quantity may be bounded, in some examples, and an additional “buffer amount” of preinitialized instances may be implemented to provide a safety margin for burst scaling, which can be further enhanced by transferring instances between data structures of different groups of instances in some cases.
    Type: Grant
    Filed: June 25, 2020
    Date of Patent: December 6, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Nilesh Girish Telang, Abhishek Saha, Dhruven Nimesh Shah, Pratik Shilwant
  • Patent number: 11513854
    Abstract: Methods, systems, and computer-readable media for resource usage restrictions in a time-series database are disclosed. Elements of a plurality of time series are stored into one or more storage tiers of a time-series database. The time series are associated with a plurality of clients of the time-series database. Execution of tasks is initiated using one or more resources of one or more hosts. The time-series elements represent inputs to the tasks. The tasks comprise a first task and a second task. A usage of the one or more resources by the first task is determined to violate one or more resource usage restrictions. Based at least in part on the usage, one or more actions are performed to modify the execution of the first task. The one or more actions increase an amount of the one or more resources available to the second task.
    Type: Grant
    Filed: June 26, 2019
    Date of Patent: November 29, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Gaurav Saxena, Mustafa Ozan Ozen
  • Patent number: 11513938
    Abstract: Methods, apparatus, and processor-readable storage media for determining capacity in storage systems using machine learning techniques are provided herein. An example computer-implemented method includes obtaining capacity-related data from a storage system; forecasting, for a given temporal period, capacity of one or more storage objects of the storage system by applying machine learning techniques to at least a portion of the capacity-related data; aggregating the forecasted capacity for at least portions of the one or more storage objects; determining, based on the aggregated forecasted capacity of the storage objects, whether at least a portion of the storage system will run out of capacity in connection with the given temporal period; and performing one or more automated actions based at least in part on the determination as to whether the at least a portion of the at least one storage system will run out of capacity.
    Type: Grant
    Filed: December 30, 2019
    Date of Patent: November 29, 2022
    Assignee: EMC IP Holding Company LLC
    Inventors: Deepak Gowda, Bina K. Thakkar
  • Patent number: 11500664
    Abstract: Example embodiments relate generally to systems and methods for continuous data protection (CDP) and more specifically to an input and output (I/O) filtering framework and log management system to seek a near-zero recovery point objective (RPO).
    Type: Grant
    Filed: April 30, 2019
    Date of Patent: November 15, 2022
    Assignee: Rubrik, Inc.
    Inventors: Benjamin Travis Meadowcroft, Li Ding, Shaomin Chen, Hardik Vohra, Arijit Banerjee, Abhay Mitra, Kushaagra Goyal, Arnav Gautum Mishra, Samir Rishi Chaudhry, Suman Swaroop, Kunal Sean Munshani, Mudit Malpani
  • Patent number: 11500686
    Abstract: A solution is proposed for resource management of a software application including a plurality of software components interacting with each other. A corresponding method includes monitoring present conditions of the software components and estimating a future consumption of one or more computing resources by each software component from the present conditions of the software components; an allocation of the computing resources to the software components is then controlled accordingly. A computer program and a computer program product for performing the method are also proposed. Moreover, a system for implementing the method is proposed.
    Type: Grant
    Filed: July 31, 2020
    Date of Patent: November 15, 2022
    Assignee: International Business Machines Corporation
    Inventors: Gabriele de Capoa, Massimo Villani
  • Patent number: 11494228
    Abstract: A method for scheduling jobs for the calculator includes measuring core utilization of the second-type processor, when the measured core utilization is less than a reference value, transmitting, by the first-type processor, a job suspension instruction to suspend a first job, which is currently being executed, to the second-type processor, in response to the job suspension instruction, copying data of a region occupied by the first job in a memory of the second-type processor to a main memory, copying data of a second job stored in the main memory to the memory of the second-type processor, and transmitting, by the first-type processor, an instruction to execute the second job to the second-type processor.
    Type: Grant
    Filed: October 25, 2019
    Date of Patent: November 8, 2022
    Assignee: SAMSUNG SDS CO., LTD.
    Inventors: Man Suk Suh, Hwan Kyun Roh, Gi Beom Pang
  • Patent number: 11481259
    Abstract: Distributing computation workload among computing nodes of differing computing paradigms is provided. Compute gravity of each computing node in a cloud computing paradigm and each computing node in a client network computing paradigm within an Internet of Systems is calculated. Each component part of an algorithm is distributed to an appropriate computing node of the cloud computing paradigm and client network computing paradigm based on calculated compute gravity of each respective computing node within the Internet of Systems. Computation workload of each component part of the algorithm is distributed to a respective computing node of the cloud computing paradigm and the client network computing paradigm having a corresponding component part of the algorithm for processing.
    Type: Grant
    Filed: January 7, 2020
    Date of Patent: October 25, 2022
    Assignee: International Business Machines Corporation
    Inventors: Aaron K. Baughman, Stephen C. Hammer, Gray Cannon, Shikhar Kwatra
  • Patent number: 11461123
    Abstract: This disclosure describes systems, devices, and techniques for live migrating virtualized resources between the main region and edge locations. Live migration enables virtualized resources to remain operational during migration. Edge locations are typically separated from secure data centers via the Internet, a direct connection, or some other intermediate network. Accordingly, to place virtualized resources within an edge location, the virtualized resources must be migrated over a secure communication tunnel that can protect virtualized resource data during transmission over the intermediate network. The secure communication tunnel may have limited data throughput. To efficiently utilize resources of the secure communication tunnel, virtualized resource data may be transferred over the tunnel in a two-stage process.
    Type: Grant
    Filed: November 21, 2019
    Date of Patent: October 4, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Oleksii Tsai, Nikolay Krasilnikov, Anton Valter, Alexey Gadalin