Patents by Inventor Seetharami R. Seelam
Seetharami R. Seelam has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 9009227Abstract: Enabling meeting extensions using an electronic meeting scheduling system may include enabling a second user invited to a meeting by a first user via an electronic meeting scheduling system to invite one or more third users to the meeting; and applying one or more meeting attributes set by the second user to said one or more third users.Type: GrantFiled: April 4, 2011Date of Patent: April 14, 2015Assignee: International Business Machines CorporationInventors: Peter K. Malkin, Seetharami R. Seelam
-
Patent number: 8990790Abstract: A method for executing native code in a distributed Java Virtual Machine (JVM) is disclosed herein. The method may include receiving, in a first thread executing in a remote execution container, a first native code-generated call, such as a Java Native Interface (JNI) call, to a second thread, the first call including a first array write request. The first call may be stored in an instruction cache and bundled with a second native code-generated call and sent to the second thread. The calls are unbundled and executed in the second thread. An opaque handle to an array returned by the second call is bundled with corresponding array data and returned to the first thread. The array data of the bundle is stored in a data cache and retrieved in response to requests for the array data addressed to the second thread. A corresponding computer program product is also disclosed.Type: GrantFiled: March 15, 2013Date of Patent: March 24, 2015Assignee: International Business Machines CorporationInventors: Michael Hilton Dawson, Xavier Rene Guerin, Megumi Ito, Graeme Johnson, Seetharami R Seelam
-
Patent number: 8943485Abstract: Detecting localizable native methods may include statically analyzing a native binary file of a native method. For each function call invoked in the native binary, it is checked whether resources accessed through the function call is locally available or not. If all resources accessed though the native method is locally available, the method is annotated as localizable.Type: GrantFiled: September 13, 2012Date of Patent: January 27, 2015Assignee: International Business Machines CorporationInventors: Michael H. Dawson, Yuqing Gao, Megumi Ito, Graeme Johnson, Seetharami R. Seelam
-
Publication number: 20150026687Abstract: Various embodiments monitor system noise in a parallel computing system. In one embodiment, at least one set of system noise data is stored in a shared buffer during a first computation interval. The set of system noise data is detected during the first computation interval and is associated with at least one parallel thread in a plurality of parallel threads. Each thread in the plurality of parallel threads is a thread of a program. The set of system noise data is filtered during a second computation interval based on at least one filtering condition creating a filtered set of system noise data. The filtered set of system noise data is then stored.Type: ApplicationFiled: July 18, 2013Publication date: January 22, 2015Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Keun Soo YIM, Seetharami R. SEELAM, Liana L. FONG, Arun IYENGAR, John LEWARS
-
Patent number: 8881168Abstract: Methods and apparatus are provided for scheduling parallel jobs in a system of virtual containers. At least one parallel job is assigned to a plurality of containers competing for a total capacity of a larger container, wherein the at least one parallel job comprises a plurality of tasks. The assignment method comprises determining a current utilization and a potential free capacity for each of the plurality of competing containers; and assigning the tasks to one of the plurality of containers based on the potential free capacities and at least one predefined scheduling policy. The predefined scheduling policy may comprise, for example, one or more of load balancing, server consolidation, maximizing the current utilizations, minimizing a response time of the parallel job and satisfying quality of service requirements. The load balancing can be achieved, for example, by assigning a task to a container having a highest potential free capacity.Type: GrantFiled: September 11, 2012Date of Patent: November 4, 2014Assignee: International Business Machines CorporationInventors: Norman Bobroff, Liana Liyow Fong, Yanbin Liu, Seetharami R. Seelam
-
Publication number: 20140316701Abstract: Managing routes to meet one or more predetermined conditions, in one aspect, may comprise receiving user information associated with a user via a user's device. Based on the user information, at least a target location to where the user is traveling may be determined. Path information associated with one or more intermediary locations leading to the target location may be received. The path information may be received automatically from one or more sensors installed at the respective intermediary locations for detecting the path information. A route strategy that meets one or more conditions may be estimated by analyzing the user information and the path information. The user information may be obtained automatically from one or more of social network profile data associated with the user, electronic calendar data associated with the user, or historical data associated with the user stored in a user profile database.Type: ApplicationFiled: April 18, 2013Publication date: October 23, 2014Applicant: International Business Machines CorporationInventors: Carlos Henrique Cardonha, Dimitri Kanevsky, Peter K. Malkin, Seetharami R. Seelam
-
Publication number: 20140316702Abstract: Managing routes to meet one or more predetermined conditions, in one aspect, may comprise receiving user information associated with a user via a user's device. Based on the user information, at least a target location to where the user is traveling may be determined. Path information associated with one or more intermediary locations leading to the target location may be received. The path information may be received automatically from one or more sensors installed at the respective intermediary locations for detecting the path information. A route strategy that meets one or more conditions may be estimated by analyzing the user information and the path information. The user information may be obtained automatically from one or more of social network profile data associated with the user, electronic calendar data associated with the user, or historical data associated with the user stored in a user profile database.Type: ApplicationFiled: September 16, 2013Publication date: October 23, 2014Applicant: International Business Machines CorporationInventors: Carlos Henrique Cardonha, Dimitri Kanevsky, Peter K. Malkin, Seetharami R. Seelam
-
Patent number: 8869164Abstract: Methods and apparatus are provided for scheduling parallel jobs in a system of virtual containers. At least one parallel job is assigned to a plurality of containers competing for a total capacity of a larger container, wherein the at least one parallel job comprises a plurality of tasks. The assignment method comprises determining a current utilization and a potential free capacity for each of the plurality of competing containers; and assigning the tasks to one of the plurality of containers based on the potential free capacities and at least one predefined scheduling policy. The predefined scheduling policy may comprise, for example, one or more of load balancing, server consolidation, maximizing the current utilizations, minimizing a response time of the parallel job and satisfying quality of service requirements. The load balancing can be achieved, for example, by assigning a task to a container having a highest potential free capacity.Type: GrantFiled: September 2, 2010Date of Patent: October 21, 2014Assignee: International Business Machines CorporationInventors: Norman Bobroff, Liana Liyow Fong, Yanbin Liu, Seetharami R. Seelam
-
Patent number: 8869119Abstract: Affinity-based preferential call technique, in one aspect, may improve performance of distributed applications in a hybrid system having heterogeneous platforms. A segment of code in a program being executed on a processor may be intercepted or trapped in runtime. A platform is selected in the hybrid system for executing said segment of code, the platform determined to run the segment of code with best efficiency among a plurality of platforms in the hybrid system. The segment of code is dynamically executed on the selected platform determined to run the segment of code with best efficiency.Type: GrantFiled: September 14, 2012Date of Patent: October 21, 2014Assignee: International Business Machines CorporationInventors: Michael H. Dawson, Parijat Dube, Liana L. Fong, Yuqing Gao, Xavier R. Guerin, Michel H. T. Hack, Megumi Ito, Graeme Johnson, Nai K. Ling, Yanbin Liu, Xiaoqiao Meng, Pramod B. Nagaraja, Seetharami R. Seelam, Wei Tan, Li Zhang
-
Patent number: 8843894Abstract: Affinity-based preferential call technique, in one aspect, may improve performance of distributed applications in a hybrid system having heterogeneous platforms. A segment of code in a program being executed on a processor may be intercepted or trapped in runtime. A platform is selected in the hybrid system for executing said segment of code, the platform determined to run the segment of code with best efficiency among a plurality of platforms in the hybrid system. The segment of code is dynamically executed on the selected platform determined to run the segment of code with best efficiency.Type: GrantFiled: March 12, 2012Date of Patent: September 23, 2014Assignee: International Business Machines CorporationInventors: Michael H. Dawson, Parijat Dube, Liana L. Fong, Yuqing Gao, Xavier R. Guerin, Michel H. T. Hack, Megumi Ito, Graeme Johnson, Nai K. Ling, Yanbin Liu, Xiaoqiao Meng, Pramod B. Nagaraja, Seetharami R. Seelam, Wei Tan, Li Zhang
-
Publication number: 20140201302Abstract: An exemplary method in accordance with embodiments of this invention includes, at a virtual machine that forms a part of a cluster of virtual machines, computing a key for an instance of a memory page that is to be swapped out to a shared memory cache that is accessible by all virtual machines of the cluster of virtual machines; determining if the computed key is already present in a global hash map that is accessible by all virtual machines of the cluster of virtual machines; and only if it is determined that the computed key is not already present in the global hash map, storing the computed key in the global hash map and the instance of the memory page in the shared memory cache.Type: ApplicationFiled: January 16, 2013Publication date: July 17, 2014Applicant: International Business Machines CorporationInventors: Parijat Dube, Xavier R. Guerin, Seetharami R. Seelam
-
Publication number: 20140201451Abstract: A data processing system includes a plurality of virtual machines each having associated memory pages; a shared memory page cache that is accessible by each of the plurality of virtual machines; and a global hash map that is accessible by each of the plurality of virtual machines. The data processing system is configured such that, for a particular memory page stored in the shared memory page cache that is associated with two or more of the plurality of virtual machines, there is a single key stored in the global hash map that identifies at least a storage location in the shared memory page cache of the particular memory page. The system can be embodied at least partially in a cloud computing system.Type: ApplicationFiled: August 15, 2013Publication date: July 17, 2014Applicant: International Business Machines CorporationInventors: Parijat Dube, Xavier R. Guerin, Seetharami R. Seelam
-
Publication number: 20140189171Abstract: Managing buffers in a hybrid system, in one aspect, may comprise selecting a first buffer management method from a plurality of buffer management methods; capturing statistics associated with access to the buffer in the hybrid system running under the initial buffer management method; analyzing the captured statistics; identifying a second buffer management method based on the analyzed captured statistics; determining whether the second buffer management method is more optimal than the first buffer management method; in response to determining that the second buffer management method is more optimal than the first buffer management method, invoking the second buffer management method; and repeating the capturing, the analyzing, the identifying and the determining.Type: ApplicationFiled: August 28, 2013Publication date: July 3, 2014Applicant: International Business Machines CorporationInventors: Michael H. Dawson, Yuqing Gao, Megumi Ito, Graeme Johnson, Seetharami R. Seelam
-
Publication number: 20140189262Abstract: Managing buffers in a hybrid system, in one aspect, may comprise selecting a first buffer management method from a plurality of buffer management methods; capturing statistics associated with access to the buffer in the hybrid system running under the initial buffer management method; analyzing the captured statistics; identifying a second buffer management method based on the analyzed captured statistics; determining whether the second buffer management method is more optimal than the first buffer management method; in response to determining that the second buffer management method is more optimal than the first buffer management method, invoking the second buffer management method; and repeating the capturing, the analyzing, the identifying and the determining.Type: ApplicationFiled: January 2, 2013Publication date: July 3, 2014Applicant: International Business Machines CorporationInventors: Michael H. Dawson, Yuqing Gao, Megumi Ito, Graeme Johnson, Seetharami R. Seelam
-
Publication number: 20140068572Abstract: A method for executing native code in a distributed Java Virtual Machine (JVM) is disclosed herein. The method may include receiving, in a first thread executing in a remote execution container, a first native code-generated call, such as a Java Native Interface (JNI) call, to a second thread, the first call including a first array write request. The first call may be stored in an instruction cache and bundled with a second native code-generated call and sent to the second thread. The calls are unbundled and executed in the second thread. An opaque handle to an array returned by the second call is bundled with corresponding array data and returned to the first thread. The array data of the bundle is stored in a data cache and retrieved in response to requests for the array data addressed to the second thread. A corresponding computer program product is also disclosed.Type: ApplicationFiled: August 28, 2012Publication date: March 6, 2014Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Michael H. Dawson, Graeme Johnson, Seetharami R. Seelam, Megumi Ito, Xavier Rene Guerin
-
Patent number: 8627291Abstract: Detecting localizable native methods may include statically analyzing a native binary file of a native method. For each function call invoked in the native binary, it is checked whether resources accessed through the function call is locally available or not. If all resources accessed though the native method is locally available, the method is annotated as localizable.Type: GrantFiled: April 2, 2012Date of Patent: January 7, 2014Assignee: International Business Machines CorporationInventors: Michael H. Dawson, Yuqing Gao, Megumi Ito, Graeme Johnson, Seetharami R. Seelam
-
Publication number: 20130275988Abstract: A method, information processing system, and computer program product are provided for managing operating system interference on applications in a parallel processing system. A mapping of hardware multi-threading threads to at least one processing core is determined, and first and second sets of logical processors of the at least one processing core are determined. The first set includes at least one of the logical processors of the at least one processing core, and the second set includes at least one of a remainder of the logical processors of the at least one processing core. A processor schedules application tasks only on the logical processors of the first set of logical processors of the at least one processing core. Operating system interference events are scheduled only on the logical processors of the second set of logical processors of the at least one processing core.Type: ApplicationFiled: June 11, 2013Publication date: October 17, 2013Inventors: John DIVIRGILIO, Liana L. FONG, John LEWARS, Seetharami R. SEELAM, Brian F. VEALE
-
Publication number: 20130263097Abstract: Detecting localizable native methods may include statically analyzing a native binary file of a native method. For each function call invoked in the native binary, it is checked whether resources accessed through the function call is locally available or not. If all resources accessed though the native method is locally available, the method is annotated as localizable.Type: ApplicationFiled: April 2, 2012Publication date: October 3, 2013Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Michael H. Dawson, Yuqing Gao, Megumi Ito, Graeme Johnson, Seetharami R. Seelam
-
Publication number: 20130263101Abstract: Detecting localizable native methods may include statically analyzing a native binary file of a native method. For each function call invoked in the native binary, it is checked whether resources accessed through the function call is locally available or not. If all resources accessed though the native method is locally available, the method is annotated as localizable.Type: ApplicationFiled: September 13, 2012Publication date: October 3, 2013Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Michael H. Dawson, Yuqing Gao, Megumi Ito, Graeme Johnson, Seetharami R. Seelam
-
Publication number: 20130239128Abstract: Affinity-based preferential call technique, in one aspect, may improve performance of distributed applications in a hybrid system having heterogeneous platforms. A segment of code in a program being executed on a processor may be intercepted or trapped in runtime. A platform is selected in the hybrid system for executing said segment of code, the platform determined to run the segment of code with best efficiency among a plurality of platforms in the hybrid system. The segment of code is dynamically executed on the selected platform determined to run the segment of code with best efficiency.Type: ApplicationFiled: September 14, 2012Publication date: September 12, 2013Applicant: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Michael H. Dawson, Parijat Dube, Liana L. Fong, Yuqing Gao, Xavier R. Guerin, Michel H. T. Hack, Megumi Ito, Graeme Johnson, Nai K. Ling, Yanbin Liu, Xiaoqiao Meng, Pramod B. Nagaraja, Seetharami R. Seelam, Wei Tan, Li Zhang