Patents by Inventor Alekh Jindal
Alekh Jindal has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11934874Abstract: A serverless query processing system receives a query and determines whether the query is a recurring query or a non-recurring query. The system may predict, in response to determining that the query is the recurring query, a peak resource requirement during an execution of the query. The system may compute, in response to determining that the query is the non-recurring query, a tight resource requirement corresponding to an amount of resources that satisfy a performance requirement over the execution of the query, where the tight resource requirement is less than the peak resource requirement. The system allocates resources to the query based on an applicable one of the peak resource requirement or the tight resource requirement. The system then starts the execution of the query using the resources.Type: GrantFiled: August 24, 2022Date of Patent: March 19, 2024Assignee: Microsoft Technology Licensing, LLCInventors: Hiren Shantilal Patel, Shi Qiao, Alekh Jindal, Malay Kumar Bag, Rathijit Sen, Carlo Aldo Curino
-
Publication number: 20240037097Abstract: Systems, methods, and devices are described for performing scalable data processing operations. A queue that includes a translatable portion comprising indications of data processing operations translatable to data queries and a non-translatable portion comprising indications of non-translatable data processing operations is maintained. A determination that a first data processing operation of a first code block statement is translatable to a database query is made. An indication of the first data processing operation is included in the translatable portion of the queue. Responsive to a determination that a second data processing operation of a second code block statement is undeferrable, the translatable portion of the queue is compiled into a database query. An execution of the database query to be executed by a database engine to generate a query result is caused. A result dataset corresponding to the query result is transmitted to an application configured to analyze the result dataset.Type: ApplicationFiled: October 13, 2023Publication date: February 1, 2024Inventors: Kameswara Venkatesh EMANI, Avrilia FLORATOU, Carlo Aldo CURINO, Karthik Saligrama RAMACHANDRA, Alekh JINDAL
-
Publication number: 20230418819Abstract: In a set of data analytics queries, at least a one of the queries comprising more than one operator, and each query being at least one of i) a producer of data for an other query in the set, and ii) a consumer of data from an other query in the set. In such examples, one or more computing devices identify each producer/consumer relationship between the queries. The one or more computing devices identify one or more optimizations among the queries based on the identified relationships. The one or more computing devices then apply at least one identified optimization to at least one of the queries.Type: ApplicationFiled: June 22, 2022Publication date: December 28, 2023Inventors: Jyoti LEEKA, Sunny GAKHAR, Hiren S. PATEL, Marc Todd FRIEDMAN, Brandon HAYNES, Shi QIAO, Alekh JINDAL
-
Patent number: 11847118Abstract: In a set of data analytics queries, at least a one of the queries comprising more than one operator, and each query being at least one of i) a producer of data for an other query in the set, and ii) a consumer of data from an other query in the set. In such examples, one or more computing devices identify each producer/consumer relationship between the queries. The one or more computing devices identify one or more optimizations among the queries based on the identified relationships. The one or more computing devices then apply at least one identified optimization to at least one of the queries.Type: GrantFiled: June 22, 2022Date of Patent: December 19, 2023Assignee: Microsoft Technology Licensing, LLCInventors: Jyoti Leeka, Sunny Gakhar, Hiren S. Patel, Marc Todd Friedman, Brandon Haynes, Shi Qiao, Alekh Jindal
-
Patent number: 11829359Abstract: Systems, methods, and devices are described for performing scalable data processing operations. A queue that includes a translatable portion comprising indications of data processing operations translatable to data queries and a non-translatable portion comprising indications of non-translatable data processing operations is maintained. A determination that a first data processing operation of a first code block statement is translatable to a database query is made. An indication of the first data processing operation is included in the translatable portion of the queue. Responsive to a determination that a second data processing operation of a second code block statement is undeferrable, the translatable portion of the queue is compiled into a database query. An execution of the database query to be executed by a database engine to generate a query result is caused. A result dataset corresponding to the query result is transmitted to an application configured to analyze the result dataset.Type: GrantFiled: July 29, 2022Date of Patent: November 28, 2023Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Kameswara Venkatesh Emani, Avrilia Floratou, Carlo Aldo Curino, Karthik Saligrama Ramachandra, Alekh Jindal
-
Publication number: 20230350892Abstract: Embodiments described herein are directed to generating and returning materialized views for queries (or subexpressions thereof) having a particular relationship with each other. For instance, machine learning-based techniques may be utilized to identify query subexpressions that have at least one of a semantically equivalent relationship or a containment relationship with each other. Responsive to identifying such relationship(s), a materialized view may be generated for the identified subexpressions. When a query is subsequently received, machine learning-based techniques may be utilized to determine whether a subexpression of the query possesses at least one of a semantically equivalent relationship or a containment relationship with another subexpression for which a materialized view has been generated. Responsive to determining that such a subexpression of the query possesses one or more of such relationships, the materialized view generated for the other subexpression is returned.Type: ApplicationFiled: April 30, 2022Publication date: November 2, 2023Inventors: Brandon Barry HAYNES, Jyoti LEEKA, Anna PAVLENKO, Rana ALOTAIBI, Alekh JINDAL
-
Publication number: 20230342359Abstract: Methods of machine learning for system deployments without performance regressions are performed by systems and devices. A performance safeguard system is used to design pre-production experiments for determining the production readiness of learned models based on a pre-production budget by leveraging big data processing infrastructure and deploying a large set of learned or optimized models for its query optimizer. A pipeline for learning and training differentiates the impact of query plans with and without the learned or optimized models, selects plan differences that are likely to lead to most dramatic performance difference, runs a constrained set of pre-production experiments to empirically observe the runtime performance, and finally picks the models that are expected to lead to consistently improved performance for deployment. The performance safeguard system enables safe deployment not just for learned or optimized models but also for additional of other ML-for-Systems features.Type: ApplicationFiled: June 30, 2023Publication date: October 26, 2023Inventors: Irene Rogan SHAFFER, Remmelt Herbert Lieve AMMERLAAN, Gilbert ANTONIUS, Marc T. FRIEDMAN, Abhishek ROY, Lucas ROSENBLATT, Vijay Kumar RAMANI, Shi QIAO, Alekh JINDAL, Peter ORENBERG, H M Sajjad Hossain, Soundararajan Srinivasan, Hiren Shantilal PATEL, Markus WEIMER
-
Patent number: 11748350Abstract: Methods of machine learning for system deployments without performance regressions are performed by systems and devices. A performance safeguard system is used to design pre-production experiments for determining the production readiness of learned models based on a pre-production budget by leveraging big data processing infrastructure and deploying a large set of learned or optimized models for its query optimizer. A pipeline for learning and training differentiates the impact of query plans with and without the learned or optimized models, selects plan differences that are likely to lead to most dramatic performance difference, runs a constrained set of pre-production experiments to empirically observe the runtime performance, and finally picks the models that are expected to lead to consistently improved performance for deployment. The performance safeguard system enables safe deployment not just for learned or optimized models but also for additional of other ML-for-Systems features.Type: GrantFiled: April 3, 2020Date of Patent: September 5, 2023Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Irene Rogan Shaffer, Remmelt Herbert Lieve Ammerlaan, Gilbert Antonius, Marc T. Friedman, Abhishek Roy, Lucas Rosenblatt, Vijay Kumar Ramani, Shi Qiao, Alekh Jindal, Peter Orenberg, H M Sajjad Hossain, Soundararajan Srinivasan, Hiren Shantilal Patel, Markus Weimer
-
Patent number: 11693857Abstract: Implementations described herein relate to systems and methods to configure a data warehouse system.Type: GrantFiled: June 22, 2022Date of Patent: July 4, 2023Assignee: Keebo Inc.Inventors: Alekh Jindal, Barzan Mozafari, Yongjoo Park, David Wolfgang Grömling, Brian Westphal, Alan D. Cabrera
-
Publication number: 20230177053Abstract: Methods for optimization in query plans are performed by computing systems via a query optimizer advisor. A query optimizer advisor (QO-Advisor) is configured to steer a query plan optimizer towards more efficient plan choices by providing rule hints to improve navigation of the search space for each query in formulation of its query plan. The QO-Advisor receives historical information of a distributed data processing system as an input, and then generates a set of rule hint pairs based on the historical information. The QO-Advisor provides the set of rule hint pairs to a query plan optimizer, which then optimizes a query plan of an incoming query through application of a rule hint pair in the set. This application is based at least on a characteristic of the incoming query matching a portion of the rule hint pair.Type: ApplicationFiled: March 28, 2022Publication date: June 8, 2023Inventors: Matteo INTERLANDI, Wangda ZHANG, Paul S. MINEIRO, Marc T. FRIEDMAN, Alekh JINDAL, Hiren S. PATEL, Rafah Aboul HOSN, Shi QIAO
-
Patent number: 11567936Abstract: Implementations described herein relate to systems and methods to provide platform agnostic query acceleration. In some implementations, a method includes receiving, at a processor associated with a query acceleration service, a request from an client/application, wherein the request conforms to a particular wire protocol of a plurality of supported wire protocols, and wherein the request includes header data and body content data, analyzing the request to identify at least one of a query and a command in the body content data, determining an optimal matched model of the one or more query acceleration models, rewriting the query based on the optimal matched model, transmitting the rewritten query to the query processing platform, receiving a response to the rewritten query or the query from the query processing platform, and transmitting the received response to the application, wherein the transmission is configured based on the particular wire protocol.Type: GrantFiled: September 19, 2022Date of Patent: January 31, 2023Assignee: Keebo Inc.Inventors: Alekh Jindal, Barzan Mozafari, Yongjoo Park, Brian Westphal, Shi Qiao, Matthew Larsen, Advait Abhay Dixit
-
Publication number: 20220413914Abstract: A serverless query processing system receives a query and determines whether the query is a recurring query or a non-recurring query. The system may predict, in response to determining that the query is the recurring query, a peak resource requirement during an execution of the query. The system may compute, in response to determining that the query is the non-recurring query, a tight resource requirement corresponding to an amount of resources that satisfy a performance requirement over the execution of the query, where the tight resource requirement is less than the peak resource requirement. The system allocates resources to the query based on an applicable one of the peak resource requirement or the tight resource requirement. The system then starts the execution of the query using the resources.Type: ApplicationFiled: August 24, 2022Publication date: December 29, 2022Inventors: Hiren Shantilal PATEL, Shi QIAO, Alekh JINDAL, Malay Kumar BAG, Rathijit SEN, Carlo Aldo CURINO
-
Patent number: 11455192Abstract: A serverless query processing system receives a query and determines whether the query is a recurring query or a non-recurring query. The system may predict, in response to determining that the query is the recurring query, a peak resource requirement during an execution of the query. The system may compute, in response to determining that the query is the non-recurring query, a tight resource requirement corresponding to an amount of resources that satisfy a performance requirement over the execution of the query, where the tight resource requirement is less than the peak resource requirement. The system allocates resources to the query based on an applicable one of the peak resource requirement or the tight resource requirement. The system then starts the execution of the query using the resources.Type: GrantFiled: November 27, 2019Date of Patent: September 27, 2022Assignee: Microsoft Technology Licensing, LLCInventors: Hiren Shantilal Patel, Shi Qiao, Alekh Jindal, Malay Kumar Bag, Rathijit Sen, Carlo Aldo Curino
-
Patent number: 11416487Abstract: Techniques are described herein that are capable of selecting checkpoints of a database job. For instance, at compile time, temporal indicators associated with the query plans of the database job are determined. Each temporal indicator indicates first and second subsets of stages of the respective query plan. Values of attributes of each stage in at least each first subset are predicted using a machine learning technique. At the compile time, candidate stage(s) for each query plan are identified based on the respective candidate stage being a child of stage(s) in the corresponding second subset or not being a child of another stage in the respective query plan. The candidate stage(s) for each query plan are selectively chosen as respective checkpoint(s) based on whether the values of the attributes of each stage in at least the first subset of the stages of the respective query plan satisfy one or more criteria.Type: GrantFiled: September 22, 2020Date of Patent: August 16, 2022Assignee: Microsoft Technology Licensing, LLCInventors: Yiwen Zhu, Alekh Jindal, Malay Kumar Bag, Hiren Shantilal Patel
-
Publication number: 20220100763Abstract: Solutions for optimizing job runtimes via prediction-based token allocation includes receiving training data comprising historical run data, the historical run data comprising job characteristics, runtime results, and a token count for each of a plurality of prior jobs, and the job characteristics comprising an intermediate representation and job graph data; based at least on the training data, training a token estimator, the token estimator comprising a machine learning (ML) model; receiving job characteristics for a user-submitted job; based at least on the received job characteristics, generating, with the token estimator, token prediction data for the user-submitted job; selecting a token count for the user-submitted job, based at least on the token prediction data; identifying the selected token count to an execution environment; and executing, with the execution environment, the user-submitted job in accordance with the selected token count.Type: ApplicationFiled: September 30, 2020Publication date: March 31, 2022Inventors: Rathijit SEN, Alekh JINDAL, Anish Yatin PIMPLEY, Shuo LI, Anubha SRIVASTAVA, Vishal Lalchand ROHRA, Yi ZHU, Hiren Shantilal PATEL, Shi QIAO, Marc Todd FRIEDMAN, Clemens Alden SZYPERSKI
-
Publication number: 20220092067Abstract: Techniques are described herein that are capable of selecting checkpoints of a database job. For instance, at compile time, temporal indicators associated with the query plans of the database job are determined. Each temporal indicator indicates first and second subsets of stages of the respective query plan. Values of attributes of each stage in at least each first subset are predicted using a machine learning technique. At the compile time, candidate stage(s) for each query plan are identified based on the respective candidate stage being a child of stage(s) in the corresponding second subset or not being a child of another stage in the respective query plan. The candidate stage(s) for each query plan are selectively chosen as respective checkpoint(s) based on whether the values of the attributes of each stage in at least the first subset of the stages of the respective query plan satisfy one or more criteria.Type: ApplicationFiled: September 22, 2020Publication date: March 24, 2022Inventors: Yiwen Zhu, Alekh Jindal, Malay Kumar Bag, Hiren Shantilal Patel
-
Publication number: 20210263932Abstract: Methods of machine learning for system deployments without performance regressions are performed by systems and devices. A performance safeguard system is used to design pre-production experiments for determining the production readiness of learned models based on a pre-production budget by leveraging big data processing infrastructure and deploying a large set of learned or optimized models for its query optimizer. A pipeline for learning and training differentiates the impact of query plans with and without the learned or optimized models, selects plan differences that are likely to lead to most dramatic performance difference, runs a constrained set of pre-production experiments to empirically observe the runtime performance, and finally picks the models that are expected to lead to consistently improved performance for deployment. The performance safeguard system enables safe deployment not just for learned or optimized models but also for additional of other ML-for-Systems features.Type: ApplicationFiled: April 3, 2020Publication date: August 26, 2021Inventors: Irene Rogan Shaffer, Remmelt Herbert Lieve Ammerlaan, Gilbert Antonius, Marc T. Friedman, Abhishek Roy, Lucas Rosenblatt, Vijay Kumar Ramani, Shi Qiao, Alekh Jindal, Peter Orenberg, H M Sajjad Hossain, Soundararajan Srinivasan, Hiren Shantilal Patel, Markus Weimer
-
Patent number: 11074256Abstract: Described herein is a system and method for training cardinality models in which workload data is analyzed to extract and compute features of subgraphs of queries. Using a machine learning algorithm, the cardinality models are trained based on the features and actual runtime statistics included in the workload data. The trained cardinality models are stored. Further described herein is a system and method of predicting cardinality of subgraphs of a query. Features for the subgraphs of the query are extracted and computed. Cardinality models are retrieved based on the features of the subgraphs of the query. Cardinalities of the subgraphs of the query are predicted using the retrieved cardinality models. One of the subgraphs of the query is selected to be utilized for execution of the query based on the predicted cardinalities.Type: GrantFiled: June 8, 2018Date of Patent: July 27, 2021Assignee: Microsoft Technology Licensing, LLCInventors: Alekh Jindal, Hiren Patel, Saeed Amizadeh, Chenggang Wu
-
Patent number: 11068482Abstract: Described herein is a system and method for detecting and reusing overlapping computations. Overlapping subgraphs of the query are determined using a normalized signature for a particular subgraph that identifies a particular subgraph across recurring instances of data. A normalized signature for each overlapping subgraph for the determined overlapping subgraphs of the query is provided. For each overlapping subgraph determined to be materialized: whether or not the particular subgraph has been materialized is determined using a precise signature corresponding to a normalized signature of the particular overlapping subgraph. The precise signature identifies a particular subgraph corresponding to the normalized signature within a particular recurring instance of data. When the particular subgraph has not been materialized, the subgraph is materialized and used to respond to the query. When the particular subgraph has been materialized, the materialized subgraph is used to respond to the query.Type: GrantFiled: April 13, 2018Date of Patent: July 20, 2021Assignee: Microsoft Technology Licensing, LLCInventors: Alekh Jindal, Hiren Patel, Shi Qiao, Jieming Di, Malay Kumar Bag, Zhicheng Yin
-
Publication number: 20210096915Abstract: A serverless query processing system receives a query and determines whether the query is a recurring query or a non-recurring query. The system may predict, in response to determining that the query is the recurring query, a peak resource requirement during an execution of the query. The system may compute, in response to determining that the query is the non-recurring query, a tight resource requirement corresponding to an amount of resources that satisfy a performance requirement over the execution of the query, where the tight resource requirement is less than the peak resource requirement. The system allocates resources to the query based on an applicable one of the peak resource requirement or the tight resource requirement. The system then starts the execution of the query using the resources.Type: ApplicationFiled: November 27, 2019Publication date: April 1, 2021Inventors: Hiren Shantilal PATEL, Shi QIAO, Alekh JINDAL, Malay Kumar BAG, Rathijit SEN, Carlo Aldo CURINO