Patents Examined by Gregory A Kessler
  • Patent number: 11886924
    Abstract: An example apparatus includes a dependency graph generator to generate a dependency graph based on a resource request file specifying a first resource and a second resource to deploy to a resource-based service, the dependency graph representative of the first resource being dependent on a second resource, a verification controller to generate a status indicator after a determination that a time-based ordering of a first request relative to a second request satisfies the dependency graph, and a resource controller to cause transmission of the first request and the second request to the resource-based service based on the dependency graph, and, after determining that the time-based ordering of the first request relative to the second request satisfies the dependency graph, cause transmission of the status indicator to a user device.
    Type: Grant
    Filed: September 9, 2022
    Date of Patent: January 30, 2024
    Assignee: VMware LLC
    Inventors: Sergio Sanchez, Georgi Muleshkov, Stanislav Asenov Hadjiiski, Miroslav Shipkovenski, Radostin Georgiev
  • Patent number: 11880708
    Abstract: Disclosed is a method for detecting a disappearance of a task in an environment including at least one recurring parent task that triggers, in nominal mode, on each occurrence, at least one child task, including the following steps: evaluating a parent duration elapsed between the last occurrence and the penultimate occurrence of the parent task, evaluating a child duration elapsed between the last occurrence and the penultimate occurrence of the child task, comparing the parent duration and the child duration, it being concluded that an occurrence of the child task has disappeared if the child duration, preferably with a margin, is longer than the parent duration.
    Type: Grant
    Filed: September 10, 2019
    Date of Patent: January 23, 2024
    Assignee: VITESCO TECHNOLOGIES GMBH
    Inventors: Stéphane Eloy, Nicolas Romea
  • Patent number: 11874817
    Abstract: Aspects of the disclosure relate to providing and maintaining efficient and effective processing of sets of work items in enterprise computing environments by optimizing distributed and parallelized batch data processing. A computing platform may initialize a monitoring process configured to monitor a pending workload in a work queue database. Subsequently, the computing platform may cause the monitoring process to query the work queue database and create one or more historical records indicative of a workload processing status associated with one or more processing workers. Then, the computing platform may identify one or more new parameter values for one or more processing parameters associated with the one or more processing workers based on the one or more historical records. Thereafter, the computing platform may configure the one or more processing workers based on the one or more new parameter values identified for the one or more processing parameters.
    Type: Grant
    Filed: March 31, 2021
    Date of Patent: January 16, 2024
    Assignee: Bank of America Corporation
    Inventor: Marcus Matos
  • Patent number: 11868809
    Abstract: A processor includes a task scheduling unit and a compute unit coupled to the task scheduling unit. The task scheduling unit performs a task dependency assessment of a task dependency graph and task data requirements that correspond to each task of the plurality of tasks. Based on the task dependency assessment, the task scheduling unit schedules a first task of the plurality of tasks and a second proxy object of a plurality of proxy objects specified by the task data requirements such that a memory transfer of the second proxy object of the plurality of proxy objects occurs while the first task is being executed.
    Type: Grant
    Filed: January 11, 2023
    Date of Patent: January 9, 2024
    Assignee: Advanced Micro Devices, Inc.
    Inventors: Muhammad Amber Hassaan, Anirudh Mohan Kaushik, Sooraj Puthoor, Gokul Subramanian Ravi, Bradford Beckmann, Ashwin Aji
  • Patent number: 11868814
    Abstract: An information processing apparatus that is one embodiment of the present invention: detects execution of software in any of a host environment, and one or more virtual environments; and acquires discrimination information indicating that a detected environment is a first environment, and first name information indicating a name of the software in a name space of the first environment. The information processing apparatus acquires, based on the discrimination information, second name information indicating a name of the first environment in a name space of a second environment. The information processing apparatus converts, based on the second name information, the first name information into third name information indicating a name of the software in the name space of the second environment. The information processing apparatus acquires, based on the third name information, information on the software from an accessible resource.
    Type: Grant
    Filed: February 26, 2021
    Date of Patent: January 9, 2024
    Assignee: Kabushiki Kaisha Toshiba
    Inventors: Toshiyuki Kantake, Jun Kanai, Tatsuya Uehara
  • Patent number: 11868821
    Abstract: A system and method for launching parallel processes on a server configured to process a number of parallel processes. A request is received from a parallel application to start a number of parallel processes. In response to this request a launcher creates a surrogate. The surrogate inherits communications channels from the launcher. The surrogate then executes activities related to the launch of the parallel processes, and then launches the parallel processes. The parallel processes are launched and the surrogate is terminated.
    Type: Grant
    Filed: January 17, 2023
    Date of Patent: January 9, 2024
    Assignee: International Business Machines Corporation
    Inventors: Joshua J. Hursey, David Solt, Austen William Lauria
  • Patent number: 11861408
    Abstract: The present disclosure includes systems, methods, and computer-readable mediums for discovering capabilities of a hardware (HW) accelerator card. A processor may communicate a request for a listing of acceleration services to a HW accelerator card connected to the processor via the communication interface. The HW accelerator card may retrieve the listing from memory and provide a response to the processor that includes a listing of the HW acceleration services provided by the HW accelerator card.
    Type: Grant
    Filed: June 18, 2021
    Date of Patent: January 2, 2024
    Assignee: Google LLC
    Inventors: Shrikant Kelkar, Lakshmi Sharma, Manoj Jayadevan, Gargi Adhav, Parveen Patel, Parthasarathy Ranganathan
  • Patent number: 11853748
    Abstract: The current document is directed to automated application-release-management facilities that, in a described implementation, coordinate continuous development and release of cloud-computing applications. The application-release-management process is specified, in the described implementation, by application-release-management pipelines, each pipeline comprising one or more stages, with each stage comprising one or more tasks. The currently described methods and systems allow resources to be shared among multiple, interdependent release pipelines and allow access to shared resources to be controlled.
    Type: Grant
    Filed: October 25, 2021
    Date of Patent: December 26, 2023
    Assignee: VMware, Inc.
    Inventors: Agila Govindaraju, Ravi Kasha, Mohammed Muneebuddin
  • Patent number: 11853811
    Abstract: Methods of arbitrating between requestors and a shared resource are described. The method comprises generating a vector with one bit per requestor, each initially set to one. Based on a plurality of select signals (one per decision node in a first layer of a binary decision tree, where each select signal is configured to be used by the corresponding decision node to select one of two child nodes), bits in the vector corresponding to non-selected requestors are set to zero. The method is repeated for each subsequent layer in the binary decision tree, based on the select signals for the decision nodes in those layers. The resulting vector is a priority vector in which only a single bit has a value of one. Access to the shared resource is granted, for a current processing cycle, to the requestor corresponding to the bit having a value of one.
    Type: Grant
    Filed: October 17, 2022
    Date of Patent: December 26, 2023
    Assignee: Imagination Technologies Limited
    Inventor: Casper Van Benthem
  • Patent number: 11853794
    Abstract: A pipeline task verification method and system is disclosed, and may use one or more processors. The method may comprise providing a data processing pipeline specification, wherein the data processing pipeline specification defines a plurality of data elements of a data processing pipeline. The method may further comprise identifying from the data processing pipeline specification one or more tasks defining a relationship between a first data element and a second data element. The method may further comprise receiving for a given task one or more data processing elements intended to receive the first data element and to produce the second data element. The method may further comprise verifying that the received one or more data processing elements receive the first data element and produce the second data element according to the defined relationship.
    Type: Grant
    Filed: September 30, 2022
    Date of Patent: December 26, 2023
    Assignee: Palantir Technologies Inc.
    Inventor: Kaan Tekelioglu
  • Patent number: 11853779
    Abstract: A host device and methods for efficient distributed security forensics. The method includes creating, at a host device configured to run a virtualization entity, an event index for the virtualization entity; encoding a plurality of events related to the virtualization entity, wherein each event includes a process having a process path; and updating the event index based on the encoded plurality of events.
    Type: Grant
    Filed: October 15, 2021
    Date of Patent: December 26, 2023
    Assignee: Twistlock, Ltd.
    Inventors: Liron Levin, Dima Stopel, Ami Bizamcher, Michael Kletselman, John Morello
  • Patent number: 11853789
    Abstract: In one embodiment, a system includes first host machines implementing a public-cloud computing environment, wherein at least one of the first host machines includes a resource manager that provides a public-cloud resource interface through which one or more public-cloud clients interact with one or more virtual machines, and second host machines implementing a private-cloud computing environment, wherein at least one of the second host machines includes one or more private-cloud virtual machines, wherein at least one of the first host machines further includes a private-cloud VM resource provider through which the resource manager interacts with the private-cloud virtual machines, wherein the VM resource provider translates requests to perform virtual machine operations from a public-cloud-resource interface to a private-cloud virtual machine interface, and the private-cloud virtual machines perform the requested virtual machine operations in response to receiving the translated requests from the VM resource
    Type: Grant
    Filed: November 23, 2022
    Date of Patent: December 26, 2023
    Assignee: Google LLC
    Inventors: Ilya Beyer, Manoj Sharma, Gururaj Pangal, Maurilio Cometto
  • Patent number: 11847498
    Abstract: A system and method for multi-region deployment of application jobs in a federated cloud computing infrastructure. A job is received for execution in two or more regions of the federated cloud computing infrastructure, each of the two or more regions comprising a collection of servers joined in a raft group for separate, regional execution of the job generating a copy of the job for each of the two or more regions. The job is then deployed to the two or more regions, the workload orchestrator deploying the job according to a deployment plan. A state indication is received from each of the two or more regions, the state indication representing a state of completion of the job by each respective region of the multi-cloud computing infrastructure.
    Type: Grant
    Filed: July 8, 2021
    Date of Patent: December 19, 2023
    Assignee: HashiCorp
    Inventor: Timothy Gross
  • Patent number: 11847497
    Abstract: Methods, apparatus, systems and articles of manufacture are disclosed that enable out-of-order pipelined execution of static mapping of a workload to one or more computational building blocks of an accelerator. An example apparatus includes an interface to load a first number of credits into memory; a comparator to compare the first number of credits to a threshold number of credits associated with memory availability in a buffer; and a dispatcher to, when the first number of credits meets the threshold number of credits, select a workload node of the workload to be executed at a first one of the one or more computational building blocks.
    Type: Grant
    Filed: December 23, 2021
    Date of Patent: December 19, 2023
    Assignee: Intel Corporation
    Inventors: Michael Behar, Moshe Maor, Ronen Gabbai, Roni Rosner, Zigi Walter, Oren Agam
  • Patent number: 11847533
    Abstract: A distributed computing network includes a quantum computation network and a processor. The quantum computation network includes one or more quantum processor units (QPUs) interconnected one with the other using quantum interconnects including each a quantum link and quantum network interface cards (QNICs), where each QPU is further connected to, using the QNIC, a quantum memory. The processor is configured to receive a quantum computation task, and, using a network interface card (NIC) (i) allocate the quantum computation task to the computation network, by activating any of the quantum interconnects between the QPUs according to the quantum computation task, and (ii) solve the quantum computation task using the quantum computation network.
    Type: Grant
    Filed: December 3, 2020
    Date of Patent: December 19, 2023
    Assignee: MELLANOX TECHNOLOGIES, LTD.
    Inventors: Elad Mentovich, Kyle Scheps, Juan Jose Vegas Olmos
  • Patent number: 11836642
    Abstract: A method, system, and computer program product for dynamically scheduling machine learning inference jobs receive or determine a plurality of performance profiles associated with a plurality of system resources, wherein each performance profile is associated with a machine learning model; receive a request for system resources for an inference job associated with the machine learning model; determine a system resource of the plurality of system resources for processing the inference job associated with the machine learning model based on the plurality of performance profiles and a quality of service requirement associated with the inference job; assign the system resource to the inference job for processing the inference job; receive result data associated with processing of the inference job with the system resource; and update based on the result data, a performance profile of the plurality of the performance profiles associated with the system resource and the machine learning model.
    Type: Grant
    Filed: December 23, 2022
    Date of Patent: December 5, 2023
    Assignee: Visa International Service Association
    Inventors: Yinhe Cheng, Yu Gu, Igor Karpenko, Peter Walker, Ranglin Lu, Subir Roy
  • Patent number: 11836536
    Abstract: Systems and methods for analyzing an event log for a plurality of instances of execution of a process to identify a bottleneck are provided. An event log for a plurality of instances of execution of a process is received and segments executed during one or more of the plurality of instances of execution are identified from the event log. The segments represent a pair of activities of the process. For each particular segment of the identified segments, a measure of performance is calculated for each of the one or more instances of execution of the particular segment based on the event log, each of the one or more instances of execution of the particular segment is classified based on the calculated measures of performance, and one or more metrics are computed for the particular segment based on the classified one or more instances of execution of the particular segment.
    Type: Grant
    Filed: March 14, 2022
    Date of Patent: December 5, 2023
    Assignee: UiPath, Inc.
    Inventors: Martijn Copier, Roeland Johannus Scheepens
  • Patent number: 11836524
    Abstract: Representative apparatus, method, and system embodiments are disclosed for configurable computing. A representative memory interface circuit comprises: a plurality of registers storing a plurality of tables, a state machine circuit, and a plurality of queues. The plurality of tables include a memory request table, a memory request identifier table, a memory response table, a memory data message table, and a memory response buffer. The state machine circuit is adapted to receive a load request, and in response, to obtain a first memory request identifier from the load request, to store the first memory request identifier in the memory request identifier table, to generate one or more memory load request data packets having the memory request identifier for transmission to the memory circuit, and to store load request information in the memory request table. The plurality of queues store one or more data packets for transmission.
    Type: Grant
    Filed: August 19, 2020
    Date of Patent: December 5, 2023
    Assignee: Micron Technology, Inc.
    Inventor: Tony M. Brewer
  • Patent number: 11836522
    Abstract: In one or more embodiments, one or more systems, one or more methods, and/or one or more processes may determine an identification of an application executing on an information handling system (IHS); determine a first performance profile based at least on a policy and based at least on the identification of the application; configure a processor to utilize power up to a first power level based at least on the first performance profile; determine that a user physically utilizes at least one human input device of the IHS within an amount of time transpiring; receive information indicating that the user is physically in contact with the IHS; determine a second performance profile based at least on the policy and based at least on the information; and configure the processor to utilize power up to a second power level based at least on the second performance profile.
    Type: Grant
    Filed: March 23, 2021
    Date of Patent: December 5, 2023
    Assignee: Dell Products L.P.
    Inventors: Travis Christian North, Daniel Lawrence Hamlin
  • Patent number: 11829805
    Abstract: A plurality of low-performance locks within a computing environment are monitored. It is identified that, during a time window, threads of one of the plurality of low-performance locks are in a lock queue for an average time that exceeds a time threshold. It is further identified that, during that same time window, the average queue depth of the one of the plurality of low-performance locks exceeds a depth threshold. The one of the plurality of low-performance locks is converted from a low-performance lock into a high-performance lock.
    Type: Grant
    Filed: May 27, 2021
    Date of Patent: November 28, 2023
    Assignee: International Business Machines Corporation
    Inventor: Louis A. Rasor