Abstract: Automated techniques are disclosed for coordinating request or transaction processing in a data processing system. For example, a technique for handling compound requests, in a system comprising multiple nodes for executing requests in which an individual request is associated with a particular node, comprises the following steps. A compound request comprising at least two individual requests associated with a same node is received. It is determined if both of the at least two individual requests are executable. The compound request is executed if it is determined that all individual requests of the compound request can execute.
Type:
Grant
Filed:
July 30, 2007
Date of Patent:
February 17, 2015
Assignee:
International Business Machines Corporation
Inventors:
Paul M. Dantzig, Arun Kwangil Iyengar, Francis Nicholas Parr, Gong Su
Abstract: For optimizing resources for a parallel data integration job, a job request is received, which specifies a parallel data integration job to deploy in a grid. Grid resource utilizations are predicted for hypothetical runs of the specified job on respective hypothetical grid resource configurations. This includes automatically predicting grid resource utilizations by a resource optimizer module responsive to a model based on a plurality of actual runs of previous jobs. A grid resource configuration is selected for running the parallel data integration job, which includes the optimizer module automatically selecting a grid resource configuration responsive to the predicted grid resource utilizations and an optimization criterion.
Type:
Grant
Filed:
February 24, 2012
Date of Patent:
February 10, 2015
Assignee:
International Business Machines Corporation
Abstract: For optimizing resources for a parallel data integration job, a job request is received, which specifies a parallel data integration job to deploy in a grid. Grid resource utilizations are predicted for hypothetical runs of the specified job on respective hypothetical grid resource configurations. This includes automatically predicting grid resource utilizations by a resource optimizer module responsive to a model based on a plurality of actual runs of previous jobs. A grid resource configuration is selected for running the parallel data integration job, which includes the optimizer module automatically selecting a grid resource configuration responsive to the predicted grid resource utilizations and an optimization criterion.
Type:
Grant
Filed:
September 4, 2009
Date of Patent:
January 13, 2015
Assignee:
International Business Machines Corporation
Abstract: The present invention relates to a method of scheduling for multi-function radars. Specifically, the present invention relates to an efficient urgency-based scheduling method. The present invention provides a method of scheduling tasks in a radar apparatus including the steps of: receiving one or more tasks to schedule; calculating an urgency function for each said task; and storing the said tasks using said urgency function to order each said task relative to the other said tasks; wherein when a task is to be performed, the task having the highest value of urgency function is located.
Abstract: A method and apparatus for handling queues associated with web services of a business process. The method may include generating deployment descriptors for executing a business process as a web application, and determining a default queue for the business process using a business process management (BPM) configuration file. During execution of the business process, users are allowed to monitor the message load associated with the default queue. If a user decides to re-distribute the message load, the user is allowed to specify a new set of queues for the business process to improve performance of the business process at runtime.
Abstract: Mapping tasks to physical processors in parallel computing system may include partitioning tasks in the parallel computing system into groups of tasks, the tasks being grouped according to their communication characteristics (e.g., pattern and frequency); mapping, by a processor, the groups of tasks to groups of physical processors, respectively; and fine tuning, by the processor, the mapping within each of the groups.
Type:
Grant
Filed:
March 6, 2012
Date of Patent:
December 2, 2014
Assignee:
International Business Machines Corporation
Inventors:
I-Hsin Chung, David J. Klepacki, Che-Rung Lee, Hui-Fang Wen
Abstract: Automated techniques are disclosed for coordinating request or transaction processing in a data processing system. For example, a technique for handling requests in a data processing system comprises the following steps. A compound request comprising at least two individual requests of different types is received. An individual request r1 of the compound request is placed on a queue for requests of type t1, wherein t1 is the request type of r1. After r1 has reached the front of the queue, it is determined if at least one individual request of said compound request can execute. The compound request is executed if it is determined that all individual requests of the compound request can execute.
Type:
Grant
Filed:
July 30, 2007
Date of Patent:
November 25, 2014
Assignee:
International Business Machines Corporation
Inventors:
Paul M. Dantzig, Arun Kwangil Iyengar, Francis Nicholas Parr, Gong Su
Abstract: In a large-scale transaction such as the bulk loading of new records into an ordered, distributed database, a transaction limit such as an insert limit may be chosen, partitions on overfull storage servers may be designated to be moved to underfull storage servers, and the move assignments may be based, at least in part on the degree to which a storage server is underfull and the move and insertion costs of the partitions to be moved.
Type:
Grant
Filed:
April 11, 2008
Date of Patent:
November 18, 2014
Assignee:
Yahoo! Inc.
Inventors:
Raghu Ramakrishnan, Erik Vee, Ramana Yerneni, Utkarsh Srivastava, Brian Frank Cooper, Adam Silberstein
Abstract: A method for virtualizing software associated with external computer hardware devices. The method may include maintaining a virtualized device profile associated with an external hardware device and receiving a request to provide the virtualized device profile to a client system to which the external hardware device can be temporarily connected. The method may also include downloading the virtualized device profile to the client system and installing the virtualized device profile as one or more virtualization layers on the client system. The method may further include activating the one or more virtualization layers to enable the virtualized device profile to be used to interact with the external hardware device.
Abstract: A method, information processing system, and computer readable storage medium are provided for dynamically managing accelerator resources. A first set of hardware accelerator resources is initially assigned to a first information processing system, and a second set of hardware accelerator resources is initially assigned to a second information processing system. Jobs running on the first and second information processing systems are monitored. When one of the jobs fails to satisfy a goal, at least one hardware accelerator resource in the second set of hardware accelerator resources from the second information processing system are dynamically reassigned to the first information processing system.
Type:
Grant
Filed:
December 24, 2009
Date of Patent:
October 21, 2014
Assignee:
International Business Machines Corporation
Inventors:
Donna N. Dillenberger, Matthew S. Thoennes, Peter Yocom
Abstract: Systems and methods are provided for enabling communication between two systems using different commands, variables, protocols, methods, or instructions. In an embodiment, exit points in a software application may be used to insert additional code without altering the source code. In an embodiment, this additional code may interface with a backend abstraction layer, which may contain an itemization of functions to performed on external systems or data. In an embodiment, the abstraction layer may interface with a connectivity layer. In an embodiment, the connectivity layer may include instructions in a format recognized by the external system that correspond to the desired function in the backend abstraction layer that is to be executed at the external system.
Abstract: In accordance with the disclosed subject matter there is provided a method for segregating threads running in a computer system, and executing the threads according to this categorization.
Type:
Grant
Filed:
July 23, 2009
Date of Patent:
September 16, 2014
Assignee:
Empire Technology Development LLC
Inventors:
Gokhan Memik, Seda Ogrenci Memik, Bill Mangione-Smith
Abstract: A central manager receives tick subscription requests from subscribers, including a requested period and an allowable variance. The manager selects a group period for a group of requests, based on requested period(s) and allowable variance(s). In some cases, the group period is not a divisor of every requested period but nonetheless provides at least one tick within the allowable variance of each requested period. Ticks may be issued by invoking a callback function. Ticks may be issued in a priority order based on the subscriber's category, e.g., whether it is a user-interface process. An application platform may send a tick subscription request on behalf of an application process, e.g., a mobile device platform may submit subscription requests for processes which execute on a mobile computing device. Tick subscription requests may be sent during application execution, e.g., while the application's user interface is being built or modified.
Abstract: In a portable computing device having a node-based resource architecture, resource requests are batched or otherwise transactionized to help minimize inter-processing entity messaging or other messaging or provide other benefits. In a resource graph defining the architecture, each node or resource of the graph represents an encapsulation of functionality of one or more resources controlled by a processor or other processing entity, each edge represents a client request, and adjacent nodes of the graph represent resource dependencies. A single transaction of resource requests may be provided against two or more of the resources.
Abstract: A computing system for virtualization-based resource management includes a plurality of physical machines, a plurality of virtual machines and a management virtual machine. The virtual machines are configured by virtualizing each of the plurality of physical machines. The management virtual machine is located at any one of the plurality physical machines. The management virtual machine monitors amounts of network resources utilized by the plurality of physical machines and time costs of the plurality of virtual machines, and performs a resource reallocation and a resource reclamation.
Type:
Grant
Filed:
August 18, 2009
Date of Patent:
August 5, 2014
Assignee:
Electronics and Telecommunications Research Institute
Inventors:
Kwang Won Koh, Jin Mee Kim, Young Woo Jung, Young Choon Woo
Abstract: A representative method for coordinating and optimizing resources in the completion of a set of tasks includes providing multiple task types defined for a scenario and task priorities; combining multiple solutions of the respective multiple task types into multiple candidate decisions; determining candidate scores for the respective multiple candidate decisions based on the combined multiple solutions of the respective multiple task types; determining an optimal candidate score based on the candidate scores from the respective multiple candidate decisions and task priorities; and selecting an optimal response to a given scenario based on the candidate decision based on having the determined optimal candidate score.
Type:
Grant
Filed:
January 12, 2011
Date of Patent:
July 8, 2014
Assignee:
Lockheed Martin Corporation
Inventors:
Jeffrey H. Heisler, Michael J. DiMario, Janet Barbera
Abstract: Multi-threaded, transactional memory systems may allow concurrent execution of critical sections as speculative transactions. These transactions may abort due to contention among threads. Hardware feedback mechanisms may detect information about aborts and provide that information to software, hardware, or hybrid software/hardware contention management mechanisms. For example, they may detect occurrences of transactional aborts or conditions that may result in transactional aborts, and may update local readable registers or other storage entities (e.g., performance counters) with relevant contention information. This information may include identifying data (e.g., information outlining abort relationships between the processor and other specific physical or logical processors) and/or tallied data (e.g., values of event counters reflecting the number of aborted attempts by the current thread or the resources consumed by those attempts).
Type:
Grant
Filed:
November 26, 2008
Date of Patent:
July 8, 2014
Assignee:
Oracle America, Inc.
Inventors:
David Dice, Kevin E. Moore, Mark S. Moir
Abstract: A system management layer changes a current program with a program (door lock failure diagnosis judgment program, security judgment program, door lock judgment program, keyless entry judgment program or the like) to be executed by an application layer, in accordance with an operation mode of on-vehicle equipment. Priorities of programs are previously stored for each operation mode, and a priority judgment program contributes to judge the priority of operation request based on the operation mode. Thus, plural programs of each hierarchal layer are categorized into groups per operation mode, although complicating in the single hierarchal layer. Therefore, it is possible to prevent the priority judgment processing from complicating for the operation request output by each computer program.
Abstract: A device receives a request for an amount of a resource. It determines for each resource provider in a set of resource providers a current load, a requested load corresponding to the requested amount of the resource, and an additional load corresponding to an expected state of an application. It determines for each of the resource providers an expected total load on the basis of the current load, the requested load, and the additional load. It subsequently selects from the set of resource providers a preferred resource provider on the basis of the expected total loads. The resource may be one of the following: memory, processing time, data throughput, power, and usage of a device.
Type:
Grant
Filed:
November 24, 2008
Date of Patent:
June 10, 2014
Assignee:
Freescale Seimconductor, Inc.
Inventors:
Vladimir Litovtchenko, Florian Bogenberger
Abstract: One or more embodiments of the invention enable a system and method for ordering tasks with complex interrelationships. The present invention as described herein may be used to produce a linear ordering of tasks with complex interrelationships including dependencies and constraints. In one or more embodiments optional tasks may be permitted such that a given task may or may not be added to the execution queue depending on the scheduling of earlier tasks following evaluation of their dependencies—that is, the system of the invention supports the management of optional tasks in a task ordering operation where some or all of tasks have complex interdependencies.