Patents Examined by Melissa Alfred
-
Patent number: 10169102Abstract: A load calculation program that causes a computer a process includes acquiring processor usage information including usage of a processor of an managed computer, which is a management object, and usage of the processor for each of a plurality of virtual machines generated by a hypervisor executed on the managed computer; acquiring data transmission information including a data transmission amount for each virtual network interface used by the plurality of virtual machines; detecting a first virtual network interface that performs data transmission without routing through the hypervisor among the plurality of virtual network interfaces, based on the processor usage information and the data transmission information; and calculating load information including processor usage for data transmission of each of the plurality of virtual machines, based on whether or not each of the virtual machines uses the first virtual network interface.Type: GrantFiled: January 5, 2016Date of Patent: January 1, 2019Assignee: FUJITSU LIMITEDInventors: Naoki Tateishi, Kiyoshi Kouge, Kensuke Tokoda, Ichiro Kotani, Satoshi Ohta, Takaaki Nakazawa, Satoshi Matsuda, Hui Li, Kazunori Kandani, Saeko Nakamura, Yukihisa Miyagawa, Yuki Torii, Yoko Yonemoto
-
Patent number: 10162657Abstract: An information processing device includes a processor that executes a process. The process includes: identifying a cause of a shift from non-privileged mode to privileged mode that has occurred in processing by a guest program in an upper level virtual machine in a nested virtualization environment in which a first level virtual machine monitor operates in privileged mode, and an upper level virtual machine monitor and the guest program operate in non-privileged mode; and when the identified cause is setting or updating a virtual translation table employed in a virtual translation mechanism provided to the guest program by virtualizing an address translation mechanism for hardware that uses a set translation table to translate addresses of DMA by an input/output device assigned to the upper level virtual machine, setting the translation table employed by the translation mechanism based on a correspondence relationship between guest memory space and host memory space.Type: GrantFiled: October 15, 2015Date of Patent: December 25, 2018Assignee: FUJITSU LIMITEDInventor: Hirotaka Fukushima
-
Patent number: 10140177Abstract: A method begins by a dispersed storage (DS) processing module determining that partial task processing resources of a first DST execution unit are projected to be available. The method continues with the DS processing module ascertaining that partial task processing resources of a second DST execution unit are projected to be overburdened. The method continues with the DS processing module receiving, from the second DST execution unit, a partial task assigned to the second DST execution unit in accordance with a partial task allocation transfer policy to produce an allocated partial task and executing the allocated partial task.Type: GrantFiled: December 6, 2012Date of Patent: November 27, 2018Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Gary W. Grube, Timothy W. Markison
-
Patent number: 9965321Abstract: One embodiment of the present invention sets forth a technique for error-checking a compute task. The technique involves receiving a pointer to a compute task, storing the pointer in a scheduling queue, determining that the compute task should be executed, retrieving the pointer from the scheduling queue, determining via an error-check procedure that the compute task is eligible for execution, and executing the compute task.Type: GrantFiled: December 9, 2011Date of Patent: May 8, 2018Assignee: NVIDIA CorporationInventors: Jerome F. Duluk, Jr., Timothy John Purcell, Jesse David Hall, Philip Alexander Cuadra
-
Patent number: 9946574Abstract: A system and method for adaptive determination of computing resources is presented. Sometimes, computing jobs are to be allocated among a set of computers. Historical information regarding input data size and output data size is retrieved. Capacity information about each machine is retrieved. A weighted function is used to calculate a desired allocation of computing resources based on historical input data size and output data size. Computing resources are allocated based on the calculated allocation. The computing job can be performed on the allocated computing resources. Other embodiments also are disclosed.Type: GrantFiled: January 4, 2016Date of Patent: April 17, 2018Assignee: WAL-MART STORES, INC.Inventors: Shenghua Wan, Prijith Chandran, Varun Srivastava
-
Patent number: 9921873Abstract: A technique for controlling the distribution of compute task processing in a multi-threaded system encodes each processing task as task metadata (TMD) stored in memory. The TMD includes work distribution parameters specifying how the processing task should be distributed for processing. Scheduling circuitry selects a task for execution when entries of a work queue for the task have been written. The work distribution parameters may define a number of work queue entries needed before a cooperative thread array” (“CTA”) may be launched to process the work queue entries according to the compute task. The work distribution parameters may define a number of CTAs that are launched to process the same work queue entries. Finally, the work distribution parameters may define a step size that is used to update pointers to the work queue entries.Type: GrantFiled: January 31, 2012Date of Patent: March 20, 2018Assignee: NVIDIA CorporationInventors: Lacky V. Shah, Karim M. Abdalla, Sean J. Treichler, Abraham B. de Waal
-
Patent number: 9898318Abstract: Novel tools and techniques might provide for implementing applications management, based at least in part on operations, administration, and management (“OAM”) information. A host computing system might comprise a dedicated OAM management agent. While normal application frame flow might be sent or received by VMs running on the host computing system, OAM frame flow might be sent or received by the OAM management agent, which might also serve as an OAM frame generator. Alternatively, or additionally, based on a determination that at least one OAM frame has changed (in response to a change in address of far-end and/or near-end OAM server functions), the OAM management agent might update a list associating the at least one OAM frame that has changed with corresponding at least one VM of the one or more VMs, without restarting any of the at least one VM, the OAM management agent, and/or the host computing system.Type: GrantFiled: December 29, 2014Date of Patent: February 20, 2018Assignee: CenturyLink Intellectual Property LLCInventor: Michael K. Bugenhagen
-
Patent number: 9898704Abstract: Systems and method for forecasting release times in a multisystem software development environment are disclosed. A forecast engine may analyze the system specific data for multiple systems in a particular transport chain of systems in the development environment to determine when object code should be released from a source system to reach a target system in time to be included in an event, such as a software build or quality assurance test. The system specific data may include event schedules for each system in the transport chain. By analyzing the system specific data and the transport chain, the forecast engine can generate a time by which object code needs to be released to traverse the transport chain so that it is available to the target system for a particular scheduled event.Type: GrantFiled: July 8, 2013Date of Patent: February 20, 2018Assignee: SAP SEInventors: Tobias Rasch, Andreas Meier
-
Patent number: 9898333Abstract: Provided is a method and apparatus of selecting a preemption technique for a computation unit included in a processor to execute a second task before the at least one computation unit finishes executing a first task. The method includes receiving a preemption request, predicting a cost of preemption techniques based on a progress of the first task until receipt of the preemption request, and selecting one of the preemption techniques based on the predicted cost.Type: GrantFiled: August 7, 2015Date of Patent: February 20, 2018Assignee: SAMSUNG ELECTRONICS CO., LTD.Inventors: Jason Jong Kyu Park, Scott Mahlke, Donghoon Yoo
-
Patent number: 9891935Abstract: A method for dynamically configuring multiple processors based on needs of applications includes receiving, from an application, an acceleration request message including a task to be accelerated. The method further includes determining a type of the task and searching a database of available accelerators to dynamically select a first accelerator based on the type of the task. The method further includes sending the acceleration request message to a first acceleration interface located at a configurable processing circuit. The first acceleration interface sends the acceleration request message to a first accelerator, and the first accelerator accelerates the task upon receipt of the acceleration request message.Type: GrantFiled: August 13, 2015Date of Patent: February 13, 2018Assignee: Altera CorporationInventors: Chee Hak Teh, Kenneth Chong Yin Tan
-
Patent number: 9886077Abstract: Various systems, processes, and products may be used to manage a processor. In particular implementations, managing a processor may include the ability to determine whether a thread is pausing for a short period of time and place a wait event for the thread in a queue based on a short thread pause occurring. Managing a processor may also include the ability to activate a delay thread that determines whether a wait time associated with the pause has expired and remove the wait event from the queue based on the wait time having expired.Type: GrantFiled: May 9, 2016Date of Patent: February 6, 2018Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Bernard A. King-Smith, Bret R. Olszewski, Stephen Rees, Basu Vaidyanathan
-
Patent number: 9886297Abstract: Generally described, aspects of the present disclosure relate to loading an updated virtual machine monitor on the physical computing device during a boot process. The updated virtual machine monitor may be loaded from an update manager external to the virtual machine monitor, such as the offload device or a server connected with the physical computing device over a network. In certain embodiments, the updated virtual machine monitor may be loaded in a tiered process by first loading a startup virtual machine monitor, which automatically updates by loading the updated virtual machine monitor. The startup virtual machine monitor may be a virtual machine monitor with less functionality than the updated machine manager, such as where the startup virtual machine monitor may be a “lite” or simple virtual machine monitor while the updated virtual machine monitor may be a fully functional virtual machine monitor of the most recent update or version.Type: GrantFiled: December 11, 2014Date of Patent: February 6, 2018Assignee: AMAZON TECHNOLOGIES, INC.Inventors: Anthony Nicholas Liguori, Matthew Shawn Wilson, Ian Paul Nowland
-
Patent number: 9886249Abstract: Described herein is a transcompiling framework. In accordance with one aspect, the framework generates a source abstract syntax representation corresponding to source code written in a source language. The framework may determine validity of constraints of a common denominator language by parsing the source abstract syntax representation, wherein the common denominator language is a set of intersects provided by the source language and reachable by a target language. In response to determining the constraints are valid, the source abstract syntax representation may be transformed to a target syntax representation associated with the target language. The target syntax representation is then printed as transcompiled source code in the target language.Type: GrantFiled: November 16, 2015Date of Patent: February 6, 2018Assignee: SAP SEInventors: Marco Pesarese, Stephan Dominik Matthee
-
Patent number: 9864621Abstract: A method, system and computer readable medium for joining multiple virtual machines (VMs). The method includes identifying a first VM executing a first operating system (OS) for joining with a second VM executing a second OS, wherein the first OS and the second OS are incompatible. A new VM having a first new disk and a second new disk is created. The method includes creating a first image of a first disk associated with the first VM and a second image of a second disk associated with the second VM, installing the first image into the first new disk and the second image into the second new disk, and configuring the new VM to include a dual boot function enabling boot up of the new VM with either the first OS or the second OS.Type: GrantFiled: October 3, 2013Date of Patent: January 9, 2018Assignee: Red Hat Israel, Ltd.Inventor: David Botzer
-
Patent number: 9858050Abstract: A computer network implemented system is provided that provides a development platform for streamlining the creation of graphical user interface having natural user interface (NUI) attributes (the “platform”). The platform is based on a novel and innovative development framework and workflow that enables the creation of executable GUI elements. A display utility provides a series of screens that provide access to functions that capture a designer's user interface design intent. A user interface is provided that allows a designer to capture the various elements and information indicative of their intent relative to the UI design, to enable creation of a series of UI assets that are encoded with sufficient information to enable their rendering with significantly reduced development effort than in prior art solutions. In one aspect, the assets referred to herein incorporate coding elements corresponding to relevant effects such as position, animation, and “look”.Type: GrantFiled: July 2, 2013Date of Patent: January 2, 2018Assignee: YOUI LABS INC.Inventor: Andrew T. Emmons
-
Patent number: 9858130Abstract: A method for distributed processing involves receiving a graph (G) of targets and of influencers, with each influencer related to at least one target, receiving an action graph of actions performed by one or more of the influencers, and key partitioning G across shards. The method further involves transposing the first graph (G) to obtain a first transposed graph (GT), valuing partitioning GT across the shards, storing the action graph on multiple shards, issuing, to a shard, a request specifying an influencer, to perform an intersection, receiving a response to the request of a set of influencers each of which is related to a target, and determining whether to send a recommendation to the target based on the response.Type: GrantFiled: September 26, 2014Date of Patent: January 2, 2018Assignee: Twitter, Inc.Inventors: Ajeet Grewal, Siva Gurumurthy, Venumadhav Satuluri, Pankaj Gupta, Brian Larson, Volodymyr Zhabuik, Aneesh Sharma, Ashish Goel
-
Patent number: 9851999Abstract: Virtualization of a physical telephone number mapping service is provided. An amount of memory needed for handling requests from users of a physical telephone number mapping service provided by a communication network is determined. A listing of different virtual machine flavors available for instantiation is searched. Each virtual machine flavor has a different respective memory capacity. A determination is made whether there is a virtual machine flavor available for instantiation that has a memory capacity adequate to handle the requests. Responsive to determining that there is a virtual machine flavor available for instantiation that has the memory capacity adequate to handle the requests, instantiation of the virtual machine flavor is initiated for handling the requests.Type: GrantFiled: July 30, 2015Date of Patent: December 26, 2017Assignee: AT&T Intellectual Property I, L.P.Inventors: Mehrad Yasrebi, James E. Jackson, Art Daddona
-
Patent number: 9851986Abstract: A method, system and program product for providing virtual configuration management for applications running in a virtual environment. A method is described that includes: receiving a request for application configuration information from a virtual machine for an application; retrieving a versioned configuration image for the application, wherein the versioned configuration image comprises formatted configuration information captured from a previously configured application; creating a configuration runtime instance from the versioned configuration image; and attaching the configuration runtime image to the virtual machine for the application.Type: GrantFiled: August 4, 2015Date of Patent: December 26, 2017Assignee: International Business Machines CorporationInventors: Li Long Chen, Holger Karn
-
Patent number: 9804959Abstract: A method for supporting in-flight packet processing is provided. Packet processing devices (microengines) can send a request for packet processing to a packet engine before a packet comes in. The request offers a twofold benefit. First, the microengines add themselves to a work queue to request for processing. Once the packet becomes available, the header portion is automatically provided to the corresponding microengine for packet processing. Only one bus transaction is involved in order for the microengines to start packet processing. Second, the microengines can process packets before the entire packet is written into the memory. This is especially useful for large sized packets because the packets do not have to be written into the memory completely when processed by the microengines.Type: GrantFiled: October 31, 2014Date of Patent: October 31, 2017Assignee: Netronome Systems, Inc.Inventors: Salma Mirza, Steven W. Zagorianakos, Gavin J. Stark
-
Patent number: 9798591Abstract: Multiple lock assemblies are distributed on a chip, each lock assembly manage a lock application message for applying for a lock and a lock release message for releasing a lock that are sent by one small core. Specifically, embodiments include receiving a lock message sent by a small core, where the lock message carries a memory address corresponding to a lock requested by a first thread in the small core; calculating, using the memory address of the requested lock, a code number of a lock assembly to which the requested lock belongs; and sending the lock message to the lock assembly corresponding to the code number, to request the lock assembly to process the lock message.Type: GrantFiled: October 1, 2015Date of Patent: October 24, 2017Assignee: Huawei Technologies Co., LtdInventors: Weizhi Xu, Zhimin Tang, Zhimin Zhang, Fenglong Song