Patents by Inventor Alekh Jindal

Alekh Jindal has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11934874
    Abstract: A serverless query processing system receives a query and determines whether the query is a recurring query or a non-recurring query. The system may predict, in response to determining that the query is the recurring query, a peak resource requirement during an execution of the query. The system may compute, in response to determining that the query is the non-recurring query, a tight resource requirement corresponding to an amount of resources that satisfy a performance requirement over the execution of the query, where the tight resource requirement is less than the peak resource requirement. The system allocates resources to the query based on an applicable one of the peak resource requirement or the tight resource requirement. The system then starts the execution of the query using the resources.
    Type: Grant
    Filed: August 24, 2022
    Date of Patent: March 19, 2024
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Hiren Shantilal Patel, Shi Qiao, Alekh Jindal, Malay Kumar Bag, Rathijit Sen, Carlo Aldo Curino
  • Publication number: 20240037097
    Abstract: Systems, methods, and devices are described for performing scalable data processing operations. A queue that includes a translatable portion comprising indications of data processing operations translatable to data queries and a non-translatable portion comprising indications of non-translatable data processing operations is maintained. A determination that a first data processing operation of a first code block statement is translatable to a database query is made. An indication of the first data processing operation is included in the translatable portion of the queue. Responsive to a determination that a second data processing operation of a second code block statement is undeferrable, the translatable portion of the queue is compiled into a database query. An execution of the database query to be executed by a database engine to generate a query result is caused. A result dataset corresponding to the query result is transmitted to an application configured to analyze the result dataset.
    Type: Application
    Filed: October 13, 2023
    Publication date: February 1, 2024
    Inventors: Kameswara Venkatesh EMANI, Avrilia FLORATOU, Carlo Aldo CURINO, Karthik Saligrama RAMACHANDRA, Alekh JINDAL
  • Publication number: 20230418819
    Abstract: In a set of data analytics queries, at least a one of the queries comprising more than one operator, and each query being at least one of i) a producer of data for an other query in the set, and ii) a consumer of data from an other query in the set. In such examples, one or more computing devices identify each producer/consumer relationship between the queries. The one or more computing devices identify one or more optimizations among the queries based on the identified relationships. The one or more computing devices then apply at least one identified optimization to at least one of the queries.
    Type: Application
    Filed: June 22, 2022
    Publication date: December 28, 2023
    Inventors: Jyoti LEEKA, Sunny GAKHAR, Hiren S. PATEL, Marc Todd FRIEDMAN, Brandon HAYNES, Shi QIAO, Alekh JINDAL
  • Patent number: 11847118
    Abstract: In a set of data analytics queries, at least a one of the queries comprising more than one operator, and each query being at least one of i) a producer of data for an other query in the set, and ii) a consumer of data from an other query in the set. In such examples, one or more computing devices identify each producer/consumer relationship between the queries. The one or more computing devices identify one or more optimizations among the queries based on the identified relationships. The one or more computing devices then apply at least one identified optimization to at least one of the queries.
    Type: Grant
    Filed: June 22, 2022
    Date of Patent: December 19, 2023
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Jyoti Leeka, Sunny Gakhar, Hiren S. Patel, Marc Todd Friedman, Brandon Haynes, Shi Qiao, Alekh Jindal
  • Patent number: 11829359
    Abstract: Systems, methods, and devices are described for performing scalable data processing operations. A queue that includes a translatable portion comprising indications of data processing operations translatable to data queries and a non-translatable portion comprising indications of non-translatable data processing operations is maintained. A determination that a first data processing operation of a first code block statement is translatable to a database query is made. An indication of the first data processing operation is included in the translatable portion of the queue. Responsive to a determination that a second data processing operation of a second code block statement is undeferrable, the translatable portion of the queue is compiled into a database query. An execution of the database query to be executed by a database engine to generate a query result is caused. A result dataset corresponding to the query result is transmitted to an application configured to analyze the result dataset.
    Type: Grant
    Filed: July 29, 2022
    Date of Patent: November 28, 2023
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Kameswara Venkatesh Emani, Avrilia Floratou, Carlo Aldo Curino, Karthik Saligrama Ramachandra, Alekh Jindal
  • Publication number: 20230350892
    Abstract: Embodiments described herein are directed to generating and returning materialized views for queries (or subexpressions thereof) having a particular relationship with each other. For instance, machine learning-based techniques may be utilized to identify query subexpressions that have at least one of a semantically equivalent relationship or a containment relationship with each other. Responsive to identifying such relationship(s), a materialized view may be generated for the identified subexpressions. When a query is subsequently received, machine learning-based techniques may be utilized to determine whether a subexpression of the query possesses at least one of a semantically equivalent relationship or a containment relationship with another subexpression for which a materialized view has been generated. Responsive to determining that such a subexpression of the query possesses one or more of such relationships, the materialized view generated for the other subexpression is returned.
    Type: Application
    Filed: April 30, 2022
    Publication date: November 2, 2023
    Inventors: Brandon Barry HAYNES, Jyoti LEEKA, Anna PAVLENKO, Rana ALOTAIBI, Alekh JINDAL
  • Publication number: 20230342359
    Abstract: Methods of machine learning for system deployments without performance regressions are performed by systems and devices. A performance safeguard system is used to design pre-production experiments for determining the production readiness of learned models based on a pre-production budget by leveraging big data processing infrastructure and deploying a large set of learned or optimized models for its query optimizer. A pipeline for learning and training differentiates the impact of query plans with and without the learned or optimized models, selects plan differences that are likely to lead to most dramatic performance difference, runs a constrained set of pre-production experiments to empirically observe the runtime performance, and finally picks the models that are expected to lead to consistently improved performance for deployment. The performance safeguard system enables safe deployment not just for learned or optimized models but also for additional of other ML-for-Systems features.
    Type: Application
    Filed: June 30, 2023
    Publication date: October 26, 2023
    Inventors: Irene Rogan SHAFFER, Remmelt Herbert Lieve AMMERLAAN, Gilbert ANTONIUS, Marc T. FRIEDMAN, Abhishek ROY, Lucas ROSENBLATT, Vijay Kumar RAMANI, Shi QIAO, Alekh JINDAL, Peter ORENBERG, H M Sajjad Hossain, Soundararajan Srinivasan, Hiren Shantilal PATEL, Markus WEIMER
  • Patent number: 11748350
    Abstract: Methods of machine learning for system deployments without performance regressions are performed by systems and devices. A performance safeguard system is used to design pre-production experiments for determining the production readiness of learned models based on a pre-production budget by leveraging big data processing infrastructure and deploying a large set of learned or optimized models for its query optimizer. A pipeline for learning and training differentiates the impact of query plans with and without the learned or optimized models, selects plan differences that are likely to lead to most dramatic performance difference, runs a constrained set of pre-production experiments to empirically observe the runtime performance, and finally picks the models that are expected to lead to consistently improved performance for deployment. The performance safeguard system enables safe deployment not just for learned or optimized models but also for additional of other ML-for-Systems features.
    Type: Grant
    Filed: April 3, 2020
    Date of Patent: September 5, 2023
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Irene Rogan Shaffer, Remmelt Herbert Lieve Ammerlaan, Gilbert Antonius, Marc T. Friedman, Abhishek Roy, Lucas Rosenblatt, Vijay Kumar Ramani, Shi Qiao, Alekh Jindal, Peter Orenberg, H M Sajjad Hossain, Soundararajan Srinivasan, Hiren Shantilal Patel, Markus Weimer
  • Patent number: 11693857
    Abstract: Implementations described herein relate to systems and methods to configure a data warehouse system.
    Type: Grant
    Filed: June 22, 2022
    Date of Patent: July 4, 2023
    Assignee: Keebo Inc.
    Inventors: Alekh Jindal, Barzan Mozafari, Yongjoo Park, David Wolfgang Grömling, Brian Westphal, Alan D. Cabrera
  • Publication number: 20230177053
    Abstract: Methods for optimization in query plans are performed by computing systems via a query optimizer advisor. A query optimizer advisor (QO-Advisor) is configured to steer a query plan optimizer towards more efficient plan choices by providing rule hints to improve navigation of the search space for each query in formulation of its query plan. The QO-Advisor receives historical information of a distributed data processing system as an input, and then generates a set of rule hint pairs based on the historical information. The QO-Advisor provides the set of rule hint pairs to a query plan optimizer, which then optimizes a query plan of an incoming query through application of a rule hint pair in the set. This application is based at least on a characteristic of the incoming query matching a portion of the rule hint pair.
    Type: Application
    Filed: March 28, 2022
    Publication date: June 8, 2023
    Inventors: Matteo INTERLANDI, Wangda ZHANG, Paul S. MINEIRO, Marc T. FRIEDMAN, Alekh JINDAL, Hiren S. PATEL, Rafah Aboul HOSN, Shi QIAO
  • Patent number: 11567936
    Abstract: Implementations described herein relate to systems and methods to provide platform agnostic query acceleration. In some implementations, a method includes receiving, at a processor associated with a query acceleration service, a request from an client/application, wherein the request conforms to a particular wire protocol of a plurality of supported wire protocols, and wherein the request includes header data and body content data, analyzing the request to identify at least one of a query and a command in the body content data, determining an optimal matched model of the one or more query acceleration models, rewriting the query based on the optimal matched model, transmitting the rewritten query to the query processing platform, receiving a response to the rewritten query or the query from the query processing platform, and transmitting the received response to the application, wherein the transmission is configured based on the particular wire protocol.
    Type: Grant
    Filed: September 19, 2022
    Date of Patent: January 31, 2023
    Assignee: Keebo Inc.
    Inventors: Alekh Jindal, Barzan Mozafari, Yongjoo Park, Brian Westphal, Shi Qiao, Matthew Larsen, Advait Abhay Dixit
  • Publication number: 20220413914
    Abstract: A serverless query processing system receives a query and determines whether the query is a recurring query or a non-recurring query. The system may predict, in response to determining that the query is the recurring query, a peak resource requirement during an execution of the query. The system may compute, in response to determining that the query is the non-recurring query, a tight resource requirement corresponding to an amount of resources that satisfy a performance requirement over the execution of the query, where the tight resource requirement is less than the peak resource requirement. The system allocates resources to the query based on an applicable one of the peak resource requirement or the tight resource requirement. The system then starts the execution of the query using the resources.
    Type: Application
    Filed: August 24, 2022
    Publication date: December 29, 2022
    Inventors: Hiren Shantilal PATEL, Shi QIAO, Alekh JINDAL, Malay Kumar BAG, Rathijit SEN, Carlo Aldo CURINO
  • Patent number: 11455192
    Abstract: A serverless query processing system receives a query and determines whether the query is a recurring query or a non-recurring query. The system may predict, in response to determining that the query is the recurring query, a peak resource requirement during an execution of the query. The system may compute, in response to determining that the query is the non-recurring query, a tight resource requirement corresponding to an amount of resources that satisfy a performance requirement over the execution of the query, where the tight resource requirement is less than the peak resource requirement. The system allocates resources to the query based on an applicable one of the peak resource requirement or the tight resource requirement. The system then starts the execution of the query using the resources.
    Type: Grant
    Filed: November 27, 2019
    Date of Patent: September 27, 2022
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Hiren Shantilal Patel, Shi Qiao, Alekh Jindal, Malay Kumar Bag, Rathijit Sen, Carlo Aldo Curino
  • Patent number: 11416487
    Abstract: Techniques are described herein that are capable of selecting checkpoints of a database job. For instance, at compile time, temporal indicators associated with the query plans of the database job are determined. Each temporal indicator indicates first and second subsets of stages of the respective query plan. Values of attributes of each stage in at least each first subset are predicted using a machine learning technique. At the compile time, candidate stage(s) for each query plan are identified based on the respective candidate stage being a child of stage(s) in the corresponding second subset or not being a child of another stage in the respective query plan. The candidate stage(s) for each query plan are selectively chosen as respective checkpoint(s) based on whether the values of the attributes of each stage in at least the first subset of the stages of the respective query plan satisfy one or more criteria.
    Type: Grant
    Filed: September 22, 2020
    Date of Patent: August 16, 2022
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Yiwen Zhu, Alekh Jindal, Malay Kumar Bag, Hiren Shantilal Patel
  • Publication number: 20220100763
    Abstract: Solutions for optimizing job runtimes via prediction-based token allocation includes receiving training data comprising historical run data, the historical run data comprising job characteristics, runtime results, and a token count for each of a plurality of prior jobs, and the job characteristics comprising an intermediate representation and job graph data; based at least on the training data, training a token estimator, the token estimator comprising a machine learning (ML) model; receiving job characteristics for a user-submitted job; based at least on the received job characteristics, generating, with the token estimator, token prediction data for the user-submitted job; selecting a token count for the user-submitted job, based at least on the token prediction data; identifying the selected token count to an execution environment; and executing, with the execution environment, the user-submitted job in accordance with the selected token count.
    Type: Application
    Filed: September 30, 2020
    Publication date: March 31, 2022
    Inventors: Rathijit SEN, Alekh JINDAL, Anish Yatin PIMPLEY, Shuo LI, Anubha SRIVASTAVA, Vishal Lalchand ROHRA, Yi ZHU, Hiren Shantilal PATEL, Shi QIAO, Marc Todd FRIEDMAN, Clemens Alden SZYPERSKI
  • Publication number: 20220092067
    Abstract: Techniques are described herein that are capable of selecting checkpoints of a database job. For instance, at compile time, temporal indicators associated with the query plans of the database job are determined. Each temporal indicator indicates first and second subsets of stages of the respective query plan. Values of attributes of each stage in at least each first subset are predicted using a machine learning technique. At the compile time, candidate stage(s) for each query plan are identified based on the respective candidate stage being a child of stage(s) in the corresponding second subset or not being a child of another stage in the respective query plan. The candidate stage(s) for each query plan are selectively chosen as respective checkpoint(s) based on whether the values of the attributes of each stage in at least the first subset of the stages of the respective query plan satisfy one or more criteria.
    Type: Application
    Filed: September 22, 2020
    Publication date: March 24, 2022
    Inventors: Yiwen Zhu, Alekh Jindal, Malay Kumar Bag, Hiren Shantilal Patel
  • Publication number: 20210263932
    Abstract: Methods of machine learning for system deployments without performance regressions are performed by systems and devices. A performance safeguard system is used to design pre-production experiments for determining the production readiness of learned models based on a pre-production budget by leveraging big data processing infrastructure and deploying a large set of learned or optimized models for its query optimizer. A pipeline for learning and training differentiates the impact of query plans with and without the learned or optimized models, selects plan differences that are likely to lead to most dramatic performance difference, runs a constrained set of pre-production experiments to empirically observe the runtime performance, and finally picks the models that are expected to lead to consistently improved performance for deployment. The performance safeguard system enables safe deployment not just for learned or optimized models but also for additional of other ML-for-Systems features.
    Type: Application
    Filed: April 3, 2020
    Publication date: August 26, 2021
    Inventors: Irene Rogan Shaffer, Remmelt Herbert Lieve Ammerlaan, Gilbert Antonius, Marc T. Friedman, Abhishek Roy, Lucas Rosenblatt, Vijay Kumar Ramani, Shi Qiao, Alekh Jindal, Peter Orenberg, H M Sajjad Hossain, Soundararajan Srinivasan, Hiren Shantilal Patel, Markus Weimer
  • Patent number: 11074256
    Abstract: Described herein is a system and method for training cardinality models in which workload data is analyzed to extract and compute features of subgraphs of queries. Using a machine learning algorithm, the cardinality models are trained based on the features and actual runtime statistics included in the workload data. The trained cardinality models are stored. Further described herein is a system and method of predicting cardinality of subgraphs of a query. Features for the subgraphs of the query are extracted and computed. Cardinality models are retrieved based on the features of the subgraphs of the query. Cardinalities of the subgraphs of the query are predicted using the retrieved cardinality models. One of the subgraphs of the query is selected to be utilized for execution of the query based on the predicted cardinalities.
    Type: Grant
    Filed: June 8, 2018
    Date of Patent: July 27, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Alekh Jindal, Hiren Patel, Saeed Amizadeh, Chenggang Wu
  • Patent number: 11068482
    Abstract: Described herein is a system and method for detecting and reusing overlapping computations. Overlapping subgraphs of the query are determined using a normalized signature for a particular subgraph that identifies a particular subgraph across recurring instances of data. A normalized signature for each overlapping subgraph for the determined overlapping subgraphs of the query is provided. For each overlapping subgraph determined to be materialized: whether or not the particular subgraph has been materialized is determined using a precise signature corresponding to a normalized signature of the particular overlapping subgraph. The precise signature identifies a particular subgraph corresponding to the normalized signature within a particular recurring instance of data. When the particular subgraph has not been materialized, the subgraph is materialized and used to respond to the query. When the particular subgraph has been materialized, the materialized subgraph is used to respond to the query.
    Type: Grant
    Filed: April 13, 2018
    Date of Patent: July 20, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Alekh Jindal, Hiren Patel, Shi Qiao, Jieming Di, Malay Kumar Bag, Zhicheng Yin
  • Publication number: 20210096915
    Abstract: A serverless query processing system receives a query and determines whether the query is a recurring query or a non-recurring query. The system may predict, in response to determining that the query is the recurring query, a peak resource requirement during an execution of the query. The system may compute, in response to determining that the query is the non-recurring query, a tight resource requirement corresponding to an amount of resources that satisfy a performance requirement over the execution of the query, where the tight resource requirement is less than the peak resource requirement. The system allocates resources to the query based on an applicable one of the peak resource requirement or the tight resource requirement. The system then starts the execution of the query using the resources.
    Type: Application
    Filed: November 27, 2019
    Publication date: April 1, 2021
    Inventors: Hiren Shantilal PATEL, Shi QIAO, Alekh JINDAL, Malay Kumar BAG, Rathijit SEN, Carlo Aldo CURINO