Patents by Inventor Parijat Dube

Parijat Dube has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 9448845
    Abstract: Embodiments include methods, systems and computer program products for providing an extendable job structure for executing instructions on an accelerator. The method includes creating a number of data descriptor blocks, each memory location addresses and a pointer to a next of the number of the data descriptor block. The method further includes creating a last data descriptor block having memory location addresses and a last block indicator. Based on determining that additional memory is required for executing instructions on the accelerator, the method includes modifying the last data descriptor block to become a data extender block having a pointer to one of one or more new data descriptor blocks and creating a new last data descriptor block.
    Type: Grant
    Filed: April 7, 2016
    Date of Patent: September 20, 2016
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Sameh W. Asaad, Parijat Dube, Hong Min, Donald W. Schmidt, Bharat Sukhwani, Mathew S. Thoennes
  • Publication number: 20160224268
    Abstract: Embodiments include methods, systems and computer program products for providing an extendable job structure for executing instructions on an accelerator. The method includes creating a number of data descriptor blocks, each memory location addresses and a pointer to a next of the number of the data descriptor block. The method further includes creating a last data descriptor block having memory location addresses and a last block indicator. Based on determining that additional memory is required for executing instructions on the accelerator, the method includes modifying the last data descriptor block to become a data extender block having a pointer to one of one or more new data descriptor blocks and creating a new last data descriptor block.
    Type: Application
    Filed: April 7, 2016
    Publication date: August 4, 2016
    Inventors: Sameh W. Asaad, Parijat Dube, Hong Min, Donald W. Schmidt, Bharat Sukhwani, Mathew S. Thoennes
  • Publication number: 20160179669
    Abstract: A data processing system includes a plurality of virtual machines each having associated memory pages; a shared memory page cache that is accessible by each of the plurality of virtual machines; and a global hash map that is accessible by each of the plurality of virtual machines. The data processing system is configured such that, for a particular memory page stored in the shared memory page cache that is associated with two or more of the plurality of virtual machines, there is a single key stored in the global hash map that identifies at least a storage location in the shared memory page cache of the particular memory page. The system can be embodied at least partially in a cloud computing system.
    Type: Application
    Filed: March 1, 2016
    Publication date: June 23, 2016
    Inventors: Parijat Dube, Xavier R. Guerin, Seetharami R. Seelam
  • Publication number: 20160148155
    Abstract: Booking information including destination and origin and specifying a desired multi-modal freight shipment is obtained from a user; based on same and on route information from a carrier database, a plurality of feasible multi-modal routes for the desired freight shipment are generated with a route enumeration module. Based on cost information from the carrier database, cost for each of the feasible multi-modal routes is computed with a cost estimation sub-module of a metric computation module. Based on transit time information from the carrier database, transit time for each of the feasible multi-modal routes is computed with a transit time estimation sub-module of the metric computation module. Based on the cost for each of the feasible multi-modal routes and the transit time for each of the feasible multi-modal routes, multi-objective optimization under uncertainty is carried out with an optimization module, to obtain one or more preferred feasible multi-modal routes.
    Type: Application
    Filed: December 30, 2015
    Publication date: May 26, 2016
    Inventors: Francisco Barahona, Mark D. Bedeman, Parijat Dube, Joao P.M. Goncalves, Shilpa N. Mahatma, Milind R. Naphade
  • Publication number: 20160148153
    Abstract: Booking information including destination and origin and specifying a desired multi-modal freight shipment is obtained from a user; based on same and on route information from a carrier database, a plurality of feasible multi-modal routes for the desired freight shipment are generated with a route enumeration module. Based on cost information from the carrier database, cost for each of the feasible multi-modal routes is computed with a cost estimation sub-module of a metric computation module. Based on transit time information from the carrier database, transit time for each of the feasible multi-modal routes is computed with a transit time estimation sub-module of the metric computation module. Based on the cost for each of the feasible multi-modal routes and the transit time for each of the feasible multi-modal routes, multi-objective optimization under uncertainty is carried out with an optimization module, to obtain one or more preferred feasible multi-modal routes.
    Type: Application
    Filed: February 18, 2015
    Publication date: May 26, 2016
    Inventors: Francisco Barahona, Mark D. Bedeman, Parijat Dube, Joao P.M. Goncalves, Shilpa N. Mahatma, Milind R. Naphade
  • Patent number: 9348664
    Abstract: Predicting acceleration in a hybrid system may comprise determining a number of cross system calls in a first host-accelerator computer architecture running a workload. Host machine overhead and accelerator overhead in the first host-accelerator computer architecture associated with each of the cross system calls may be determined. Communication delay associated with each of the cross system calls in the first host-accelerator computer architecture running a workload may be determined. An application response time may be predicted for a candidate application to be run in a second host-accelerator computer architecture, based at least on the determined host machine overhead, the accelerator overhead, and the communication delay associated with each of the cross system calls in the first host-accelerator computer architecture running a workload.
    Type: Grant
    Filed: May 21, 2015
    Date of Patent: May 24, 2016
    Assignee: International Business Machines Corporation
    Inventors: Parijat Dube, Xiaoqiao Meng, Li Zhang
  • Publication number: 20160139885
    Abstract: A method for scaling a cloud infrastructure, comprises receiving at least one of resource-level metrics and application-level metrics, estimating parameters of at least one application based on the received metrics, automatically and dynamically determining directives for scaling application deployment based on the estimated parameters, and providing the directives to a cloud service provider to execute the scaling.
    Type: Application
    Filed: January 28, 2016
    Publication date: May 19, 2016
    Inventors: Parijat Dube, Anshul Gandhi, Alexei Karve, Andrzej Kochut, Li Zhang
  • Publication number: 20160142265
    Abstract: A method for scaling a cloud infrastructure, comprises receiving at least one of resource-level metrics and application-level metrics, estimating parameters of at least one application based on the received metrics, automatically and dynamically determining directives for scaling application deployment based on the estimated parameters, and providing the directives to a cloud service provider to execute the scaling.
    Type: Application
    Filed: January 28, 2016
    Publication date: May 19, 2016
    Inventors: Parijat Dube, Anshul Gandhi, Alexei Karve, Andrzej Kochut, Li Zhang
  • Patent number: 9336056
    Abstract: Embodiments include methods, systems and computer program products for providing an extendable job structure for executing instructions on an accelerator. The method includes creating a number of data descriptor blocks, each having a fixed number of memory location addresses and a pointer to a next of the number of the data descriptor block. The method further includes creating a last data descriptor block having the fixed number of memory location addresses and a last block indicator. Based on determining that additional memory is required for executing instructions on the accelerator, the method includes modifying the last data descriptor block to become a data extender block having a pointer to one of one or more new data descriptor blocks and creating a new last data descriptor block.
    Type: Grant
    Filed: December 31, 2013
    Date of Patent: May 10, 2016
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Sameh W. Asaad, Parijat Dube, Hong Min, Donald W. Schmidt, Bharat Sukhwani, Mathew S. Thoennes
  • Patent number: 9336274
    Abstract: Embodiments include methods, systems and computer program products for offloading multiple processing operations to an accelerator. Aspects include receiving a database query from an application, performing an analysis on the query, and identifying a plurality of available accelerators. Aspects further include retrieving cost information for one or more templates available on each of the plurality of available accelerators, determining a query execution plan based on the cost information and the analysis on the query, and offloading one or more query operations to at least one of the plurality of accelerators based on the query execution plan.
    Type: Grant
    Filed: September 5, 2013
    Date of Patent: May 10, 2016
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Sameh W. Asaad, Parijat Dube, Hong Min, Bharat Sukhwani, Mathew S. Thoennes
  • Patent number: 9323677
    Abstract: A data processing system includes a plurality of virtual machines each having associated memory pages; a shared memory page cache that is accessible by each of the plurality of virtual machines; and a global hash map that is accessible by each of the plurality of virtual machines. The data processing system is configured such that, for a particular memory page stored in the shared memory page cache that is associated with two or more of the plurality of virtual machines, there is a single key stored in the global hash map that identifies at least a storage location in the shared memory page cache of the particular memory page. The system can be embodied at least partially in a cloud computing system.
    Type: Grant
    Filed: August 15, 2013
    Date of Patent: April 26, 2016
    Assignee: International Business Machines Corporation
    Inventors: Parijat Dube, Xavier R. Guerin, Seetharami R. Seelam
  • Patent number: 9317497
    Abstract: In an exemplary embodiment of this disclosure, a computer-implemented method includes determining that a database query warrants a first projection operation to project a plurality of input rows to a plurality of projected rows, where each of the plurality of input rows has one or more variable-length columns. A first projection control block is constructed, by a computer processor, to describe the first projection operation. The first projection operation is offloaded to a hardware accelerator. The first projection control block is provided to the hardware accelerator, and the first projection control block enables the hardware accelerator to perform the first projection operation at streaming rate.
    Type: Grant
    Filed: August 20, 2013
    Date of Patent: April 19, 2016
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Sameh W. Asaad, Parijat Dube, Hong Min, Bharat Sukhwani, Mathew S. Thoennes
  • Patent number: 9300552
    Abstract: A method for scaling a cloud infrastructure, comprises receiving at least one of resource-level metrics and application-level metrics, estimating parameters of at least one application based on the received metrics, automatically and dynamically determining directives for scaling application deployment based on the estimated parameters, and providing the directives to a cloud service provider to execute the scaling.
    Type: Grant
    Filed: November 26, 2014
    Date of Patent: March 29, 2016
    Assignee: International Business Machines Corporation
    Inventors: Parijat Dube, Anshul Gandhi, Alexei Karve, Andrzej Kochut, Li Zhang
  • Patent number: 9300553
    Abstract: A method for scaling a cloud infrastructure, comprises receiving at least one of resource-level metrics and application-level metrics, estimating parameters of at least one application based on the received metrics, automatically and dynamically determining directives for scaling application deployment based on the estimated parameters, and providing the directives to a cloud service provider to execute the scaling.
    Type: Grant
    Filed: July 7, 2015
    Date of Patent: March 29, 2016
    Assignee: International business Machines Corporation
    Inventors: Parijat Dube, Anshul Gandhi, Alexei Karve, Andrzej Kochut, Li Zhang
  • Publication number: 20160085587
    Abstract: In an approach for scheduling the execution of a workload in a computing environment, a computer receives a request for scheduling execution of a computing job, wherein the computing job includes a plurality of computing tasks to be executed in a sequence, and wherein at least one computing task requires access to a set of data. The computer identifies information related to the computing environment, wherein the information comprises at least processors available to execute each computing task of the plurality of computing tasks and storage device proximity to the processors. The computer determines an execution configuration for the computing job based, at least in part, on the received request, the information related to the computing environment, and current utilization of the processors' resources. The computer schedules execution of the execution configuration for the computing job.
    Type: Application
    Filed: September 18, 2014
    Publication date: March 24, 2016
    Inventors: Parijat Dube, Li Zhang
  • Patent number: 9274840
    Abstract: Methods and arrangements for dynamic memory management. Data are accepted for thread local storage, and memory usage is monitored in thread local storage. A memory block is allocated to thread local storage for storing accepted data, based on the monitored memory usage.
    Type: Grant
    Filed: March 15, 2013
    Date of Patent: March 1, 2016
    Assignee: International Business Machines Corporation
    Inventors: Michael H. Dawson, Parijat Dube, Liana L. Fo, Michel H. T. Hack, Graeme Johnson, Li Zhang
  • Publication number: 20160019262
    Abstract: In some embodiments, a query recipient is configured to determine that a database query warrants a first projection operation to project a plurality of input rows to a plurality of projected rows. Each of the input rows has one or more variable-length columns. A projection controller is configured to construct a first projection control block to describe the first projection operation. For this construction, the projection controller is configured to construct a plurality of projection control elements, each one corresponding to a corresponding column in the input rows, and a header to specify the order of the projection control elements. The projection controller is further configured to offload the first projection operation to a hardware accelerator, and to provide the first projection control block to the hardware accelerator, where the first projection control block enables the hardware accelerator to perform the first projection operation at streaming rate.
    Type: Application
    Filed: September 29, 2015
    Publication date: January 21, 2016
    Inventors: Sameh W. Asaad, Parijat Dube, Hong Min, Bharat Sukhwani, Matthew S. Thoennes
  • Publication number: 20160012107
    Abstract: Methods and arrangements for mapping a query operation to an accelerator are provided. The method includes receiving, by a processor, a query operation and determining the design logic of the query operation, receiving a configuration of one or more available accelerators and a design logic of each of the one or more available accelerators, and determining if the query operation can be offloaded to one or more of the one or more available accelerators. Based on a determination that the query operation can be offloaded to one or more of the one or more available accelerators, the method also includes creating software structures to interface with a selected accelerator and executing the query operation on the selected accelerator. Based on a determination that the query operation cannot be offloaded to one or more of the one or more available accelerators, the method further includes executing the query operation in software.
    Type: Application
    Filed: July 11, 2014
    Publication date: January 14, 2016
    Inventors: Sameh Asaad, Parijat Dube, Balakrishna R. Iyer, Hong Min, Bharat Sukhwani, Mathew S. Thoennes
  • Patent number: 9235564
    Abstract: In an exemplary embodiment of this disclosure, a computer-implemented method includes determining that a database query warrants a first projection operation to project a plurality of input rows to a plurality of projected rows, where each of the plurality of input rows has one or more variable-length columns. A first projection control block is constructed, by a computer processor, to describe the first projection operation. The first projection operation is offloaded to a hardware accelerator. The first projection control block is provided to the hardware accelerator, and the first projection control block enables the hardware accelerator to perform the first projection operation at streaming rate.
    Type: Grant
    Filed: July 19, 2013
    Date of Patent: January 12, 2016
    Assignee: International Business Machines Corporation
    Inventors: Sameh W. Asaad, Parijat Dube, Hong Min, Bharat Sukhwani, Mathew S. Thoennes
  • Patent number: 9229877
    Abstract: A method for configuring a large hybrid memory subsystem having a large cache size in a computing system where one or more performance metrics of the computing system are expressed as an explicit function of configuration parameters of the memory subsystem and workload parameters of the memory subsystem. The computing system hosts applications that utilize the memory subsystem, and the performance metrics cover the use of the memory subsystem by the applications. A performance goal containing values for the performance metric is identified for the computing system. These values for the performance metrics are used in the explicit function of performance metrics, configuration parameters and workload parameters to calculate values for the configuration parameters that achieve the identified performance goal. The calculated values of the configuration parameters are implemented in the memory subsystem.
    Type: Grant
    Filed: July 30, 2013
    Date of Patent: January 5, 2016
    Inventors: John Alan Bivens, Parijat Dube, Michael Mi Tsao, Li Zhang