Patents by Inventor Subramaniam Venkatraman Krishnan
Subramaniam Venkatraman Krishnan has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20250190280Abstract: Methods, systems, apparatuses, and computer-readable storage mediums described herein are directed to determining and recommending an optimal compute resource configuration for a cloud-based resource (e.g., a server, a virtual machine, etc.) for migrating a customer to the cloud. The embodiments described herein utilize a statistically robust approach that makes recommendations that are more flexible (elastic) and account for the full distribution of the amount of resource usage. Such an approach is utilized to develop a personalized rank of relevant recommendations to a customer. To determine which compute resource configuration to recommend to the customer, the customer's usage profile is matched to a set of customers that have already migrated to the cloud. The compute resource configuration that reaches the performance most similar to the performance of the configurations utilized by customers in the matched set is recommended to the user.Type: ApplicationFiled: February 18, 2025Publication date: June 12, 2025Inventors: Wenjing WANG, Joyce Yu CAHOON, Yiwen ZHU, Ya LIN, Subramaniam Venkatraman KRISHNAN, Neetu SINGH, Raymond TRUONG, XingYu LIU, Maria Alexandra CIORTEA, Sreraman NARASIMHAN, Pratyush RAWAT, Haitao SONG
-
Publication number: 20250156430Abstract: The description relates to executing an inference query relative to a database management system, such as a relational database management system. In one example a trained machine learning model can be stored within the database management system. An inference query can be received that applies the trained machine learning model on data local to the database management system. Analysis can be performed on the inference query and the trained machine learning model to generate a unified intermediate representation of the inference query and the trained model. Cross optimization can be performed on the unified intermediate representation. Based upon the cross-optimization, a first portion of the unified intermediate representation to be executed by a database engine of the database management system can be determined, and, a second portion of the unified intermediate representation to be executed by a machine learning runtime can be determined.Type: ApplicationFiled: January 15, 2025Publication date: May 15, 2025Applicant: Microsoft Technology Licensing, LLCInventors: Konstantinos KARANASOS, Matteo INTERLANDI, Fotios PSALLIDAS, Rathijit SEN, Kwanghyun PARK, Ivan POPIVANOV, Subramaniam VENKATRAMAN KRISHNAN, Markus WEIMER, Yuan YU, Raghunath RAMAKRISHNAN, Carlo Aldo CURINO, Doris Suiyi XIN, Karla Jean SAUR
-
Publication number: 20250123939Abstract: Example solutions provide an artificial intelligence (AI) agent for pre-build configuration of cloud services in order to enable the initial build of a computational resource (e.g., in a cloud service) to minimize the likelihood of excessive throttling or slack. Examples leverage prior-existing utilization data and project metadata to identify similar use cases. The utilization data includes capacity information and resource consumption information (e.g., throttling and slack) for prior-existing computational resources, and the project metadata includes information for hierarchically categorization, to identify similar resources. A pre-build configuration is generated for the customer's resource, which the customer may tune based upon the customer's preferences for a cost and performance balance point.Type: ApplicationFiled: October 13, 2023Publication date: April 17, 2025Inventors: Rajeev Sudhakar BHOPI, Yiwen ZHU, Helen Mary SERR, Jonah KARPMAN, Matthew Joseph GLEESON, Nicholas Kent GLAZE, Subramaniam Venkatraman KRISHNAN, Irwin Hollar MCNEELY, III
-
Patent number: 12260265Abstract: Methods, systems, apparatuses, and computer-readable storage mediums described herein are directed to determining and recommending an optimal compute resource configuration for a cloud-based resource (e.g., a server, a virtual machine, etc.) for migrating a customer to the cloud. The embodiments described herein utilize a statistically robust approach that makes recommendations that are more flexible (elastic) and account for the full distribution of the amount of resource usage. Such an approach is utilized to develop a personalized rank of relevant recommendations to a customer. To determine which compute resource configuration to recommend to the customer, the customer's usage profile is matched to a set of customers that have already migrated to the cloud. The compute resource configuration that reaches the performance most similar to the performance of the configurations utilized by customers in the matched set is recommended to the user.Type: GrantFiled: December 20, 2021Date of Patent: March 25, 2025Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Wenjing Wang, Joyce Yu Cahoon, Yiwen Zhu, Ya Lin, Subramaniam Venkatraman Krishnan, Neetu Singh, Raymond Truong, Xingyu Liu, Maria Alexandra Ciortea, Sreraman Narasimhan, Pratyush Rawat, Haitao Song
-
Publication number: 20250086202Abstract: Systems, methods and computer-readable memory devices are provided for greater efficiency in the configuration of a database cluster for performing a query workload. A database cluster configuration system is provided that includes a database cluster comprising one or more compute resources configured to perform database queries. A query workload comprising a plurality of queries is received. An initial workload-level configuration is applied. For each query of the query workload, a query-level configuration is generated using a query configuration model corresponding to each query in a contextual Bayesian optimization with centroid learning while also leveraging the query plan for each executing query for query characterization and including application of virtual operators. Query events are collected and used to update the corresponding query configuration model. The workload-level configuration is updated based on the query events and cached for use during a subsequent execution of the workload.Type: ApplicationFiled: September 13, 2023Publication date: March 13, 2025Inventors: Yiwen ZHU, Subramaniam Venkatraman KRISHNAN, Weihan TANG, Tengfei HUANG, Rui FANG, Rahul Kumar CHALLAPALLI, Mo LIU, Long TIAN, Karuna Sagar KRISHNA, Estera Zaneta KOT, Xin HE, Ashit R. GOSALIA, Dario Kikuchi BERNAL, Aditya LAKRA, Arshdeep SEKHON, Sule KAHRAMAN, Carlo Aldo CURINO, Brian Paul KROTH, Rathijit SEN, Andreas Christian MUELLER, Shaily Jignesh FOZDAR, Dhruv Harendra RELWANI, Xiang LI, Sergiy MATUSEVYCH
-
Patent number: 12242493Abstract: The description relates to executing an inference query relative to a database management system, such as a relational database management system. In one example a trained machine learning model can be stored within the database management system. An inference query can be received that applies the trained machine learning model on data local to the database management system. Analysis can be performed on the inference query and the trained machine learning model to generate a unified intermediate representation of the inference query and the trained model. Cross optimization can be performed on the unified intermediate representation. Based upon the cross-optimization, a first portion of the unified intermediate representation to be executed by a database engine of the database management system can be determined, and, a second portion of the unified intermediate representation to be executed by a machine learning runtime can be determined.Type: GrantFiled: August 11, 2020Date of Patent: March 4, 2025Assignee: Microsoft Technology Licensing, LLCInventors: Konstantinos Karanasos, Matteo Interlandi, Fotios Psallidas, Rathijit Sen, Kwanghyun Park, Ivan Popivanov, Subramaniam Venkatraman Krishnan, Markus Weimer, Yuan Yu, Raghunath Ramakrishnan, Carlo Aldo Curino, Doris Suiyi Xin, Karla Jean Saur
-
Publication number: 20240231927Abstract: The present application relates to a network, apparatus, and method for allocating clusters of computing nodes for programming jobs. A network includes a plurality of datacenters including computing resources configurable to instantiate nodes for executing programming jobs on a cluster. The computing resources at one of the datacenters are configured to: provision a live pool including a number of clusters, each cluster in the live pool including a plurality of nodes imaged with a configuration for executing the programming jobs in parallel on the cluster; receive a request from a user to execute a programming job; allocate a cluster from the live pool to the user for the programming job when the cluster is available; evict the cluster from the live pool; and provision a new cluster within the live pool to meet the number of clusters. The number of clusters may be optimized based on linear programming and machine-learning.Type: ApplicationFiled: January 10, 2023Publication date: July 11, 2024Inventors: Yiwen ZHU, Alex YEO, Harsha Nihanth NAGULAPALLI, Sumeet KHUSHALANI, Arijit TARAFDAR, Subramaniam VENKATRAMAN KRISHNAN, Deepak RAVIKUMAR, Andrew Francis FOGARTY, Steve D. SUH, Yoonjae PARK, Niharika DUTTA, Santhosh Kumar RAVINDRAN
-
Patent number: 12013853Abstract: The cloud-based query workload optimization system disclosed herein the cloud-based query workloads optimization system receives query logs from various query engines to a cloud data service, extracts various query entities from the query logs, parses query entities to generate a set of common workload features, generates intermediate representations of the query workloads, wherein the intermediate representations are agnostic to the language of the plurality of the queries, identifies a plurality of workload patterns based on the intermediate representations of the query workloads, categorizes the workloads in one or more workload type categories based on the workload patterns and the workload features, and selects an optimization scheme based on the category of workload pattern.Type: GrantFiled: September 25, 2019Date of Patent: June 18, 2024Assignee: Microsoft Technology Licensing, LLCInventors: Hiren S. Patel, Rathijit Sen, Zhicheng Yin, Shi Qiao, Abhishek Roy, Alekh Jindal, Subramaniam Venkatraman Krishnan, Carlo Aldo Curino
-
Publication number: 20240168948Abstract: Learned workload synthesis is disclosed. In an aspect of the present disclosure, a time series dataset corresponding to a target workload is received. A set of performance characteristics is determined from the time series dataset. A call is provided to a prediction model to determine a candidate query sequence based on the determined set of performance characteristics. A synthetic workload is generated based on the determined candidate query sequence. A synthetic workload is generated based on the determined candidate query sequence. A first similarity between a first performance profile of the synthetic workload and a second performance profile of the target workload meets a workload performance threshold condition. A performance insight is determined based on the synthetic workload. In a further aspect, the prediction model is trained to predict performance profiles based on workload profiles generated by executing benchmark queries using hardware and/or software configurations.Type: ApplicationFiled: November 23, 2022Publication date: May 23, 2024Inventors: Yiwen ZHU, Joyce CAHOON, Subramaniam Venkatraman KRISHNAN, Chengcheng WAN
-
Publication number: 20240111739Abstract: An automated tuning service is used to automatically tune, or modify, the operational parameters of a large-scale cloud infrastructure. The tuning service performs automated and fully data/model-driven configuration based from learning various real-time performance of the cloud infrastructure. Such performance is identified through monitoring various telemetric data of the cloud infrastructure. The tuning service leverages a mix of domain knowledge and principled data-science to capture the essence of our cluster dynamic behavior in a collection of descriptive machine learning (ML) models. The ML models power automated optimization procedures for parameter tuning, and inform administrators in most tactical and strategical engineering/capacity decisions (such as hardware and datacenter design, software investments, etc.).Type: ApplicationFiled: December 8, 2023Publication date: April 4, 2024Inventors: Yiwen ZHU, Subramaniam Venkatraman KRISHNAN, Konstantinos KARANASOS, Carlo CURINO, Isha TARTE, Sudhir DARBHA
-
Patent number: 11880347Abstract: An automated tuning service is used to automatically tune, or modify, the operational parameters of a large-scale cloud infrastructure. The tuning service performs automated and fully data/model-driven configuration based from learning various real-time performance of the cloud infrastructure. Such performance is identified through monitoring various telemetric data of the cloud infrastructure. The tuning service leverages a mix of domain knowledge and principled data-science to capture the essence of our cluster dynamic behavior in a collection of descriptive machine learning (ML) models. The ML models power automated optimization procedures for parameter tuning, and inform administrators in most tactical and strategical engineering/capacity decisions (such as hardware and datacenter design, software investments, etc.).Type: GrantFiled: April 2, 2021Date of Patent: January 23, 2024Assignee: Microsoft Technology Licensing, LLC.Inventors: Yiwen Zhu, Subramaniam Venkatraman Krishnan, Konstantinos Karanasos, Carlo Curino, Isha Tarte, Sudhir Darbha
-
Publication number: 20230394369Abstract: Embodiments described herein enable tracking machine learning (“ML”) model data provenance. In particular, a computing device is configured to accept ML model code that, when executed, instantiates and trains an ML model, to parse the ML model code into a workflow intermediate representation (WIR), to semantically annotate the WIR to provide an annotated WIR, and to identify, based on the annotated WIR and ML API corresponding to the ML model code, data from at least one data source that is relied upon by the ML model code when training the ML model. A WIR may be generated from an abstract syntax tree (AST) based on the ML model code, generating provenance relationships (PRs) based at least in part on relationships between nodes of the AST, wherein a PR comprises one or more input variables, an operation, a caller, and one or more output variables.Type: ApplicationFiled: August 21, 2023Publication date: December 7, 2023Inventors: Avrilia FLORATOU, Ashvin AGRAWAL, MohammadHossein NAMAKI, Subramaniam Venkatraman KRISHNAN, Fotios PSALLIDAS, Yinghui WU
-
Publication number: 20230385649Abstract: Linguistic schema mapping via semi-supervised learning is used to map a customer schema to a particular industry-specific schema (ISS). The customer schema is received and a corresponding ISS is identified. An attribute in the customer schema is selected for labeling. Candidate pairs are generated that include the first attribute and one or more second attributes which may describe the first attribute. A featurizer determines similarities between the first attribute and second attribute in each generated pair, one or more suggested labels are generated by a machine learning (ML) model, and one of the suggested labels is applied to the first attribute.Type: ApplicationFiled: May 28, 2022Publication date: November 30, 2023Inventors: Avrilia FLORATOU, Joyce Yu CAHOON, Subramaniam Venkatraman KRISHNAN, Andreas C. MUELLER, Dalitso Hansini BANDA, Fotis PSALLIDAS, Jignesh PATEL, Yunjia ZHANG
-
Patent number: 11775862Abstract: A system enables tracking machine learning (“ML”) model data provenance. In particular, a computing device is configured to accept ML model code that, when executed, instantiates and trains an ML model, to parse the ML model code into a workflow intermediate representation (WIR), to semantically annotate the WIR to provide an annotated WIR, and to identify, based on the annotated WIR and ML API corresponding to the ML model code, data from at least one data source that is relied upon by the ML model code when training the ML model. A WIR may be generated from an abstract syntax tree (AST) based on the ML model code, generating provenance relationships (PRs) based at least in part on relationships between nodes of the AST, wherein a PR comprises one or more input variables, an operation, a caller, and one or more output variables.Type: GrantFiled: January 14, 2020Date of Patent: October 3, 2023Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Avrilia Floratou, Ashvin Agrawal, MohammadHossein Namaki, Subramaniam Venkatraman Krishnan, Fotios Psallidas, Yinghui Wu
-
Publication number: 20230029888Abstract: Methods, systems, apparatuses, and computer-readable storage mediums described herein are directed to determining and recommending an optimal compute resource configuration for a cloud-based resource (e.g., a server, a virtual machine, etc.) for migrating a customer to the cloud. The embodiments described herein utilize a statistically robust approach that makes recommendations that are more flexible (elastic) and account for the full distribution of the amount of resource usage. Such an approach is utilized to develop a personalized rank of relevant recommendations to a customer. To determine which compute resource configuration to recommend to the customer, the customer’s usage profile is matched to a set of customers that have already migrated to the cloud. The compute resource configuration that reaches the performance most similar to the performance of the configurations utilized by customers in the matched set is recommended to the user.Type: ApplicationFiled: December 20, 2021Publication date: February 2, 2023Inventors: Wenjing WANG, Joyce Yu CAHOON, Yiwen ZHU, Ya LIN, Subramaniam Venkatraman KRISHNAN, Neetu SINGH, Raymond TRUONG, XingYu LIU, Maria Alexandra CIORTEA, Sreraman NARASIMHAN, Pratyush RAWAT, Haitao SONG
-
Publication number: 20220164327Abstract: An automated tuning service is used to automatically tune, or modify, the operational parameters of a large-scale cloud infrastructure. The tuning service performs automated and fully data/model-driven configuration based from learning various real-time performance of the cloud infrastructure. Such performance is identified through monitoring various telemetric data of the cloud infrastructure. The tuning service leverages a mix of domain knowledge and principled data-science to capture the essence of our cluster dynamic behavior in a collection of descriptive machine learning (ML) models. The ML models power automated optimization procedures for parameter tuning, and inform administrators in most tactical and strategical engineering/capacity decisions (such as hardware and datacenter design, software investments, etc.).Type: ApplicationFiled: April 2, 2021Publication date: May 26, 2022Inventors: Yiwen ZHU, Subramaniam Venkatraman KRISHNAN, Konstantinos KARANASOS, Carlo CURINO, Isha TARTE, Sudhir Darbha
-
Publication number: 20210216905Abstract: Embodiments described herein enable tracking machine learning (“ML”) model data provenance. In particular, a computing device is configured to accept ML model code that, when executed, instantiates and trains an ML model, to parse the ML model code into a workflow intermediate representation (WIR), to semantically annotate the WIR to provide an annotated WIR, and to identify, based on the annotated WIR and ML API corresponding to the ML model code, data from at least one data source that is relied upon by the ML model code when training the ML model. A WIR may be generated from an abstract syntax tree (AST) based on the ML model code, generating provenance relationships (PRs) based at least in part on relationships between nodes of the AST, wherein a PR comprises one or more input variables, an operation, a caller, and one or more output variables.Type: ApplicationFiled: January 14, 2020Publication date: July 15, 2021Inventors: Avrilia Floratou, Ashvin Agrawal, MohammadHossein Namaki, Subramaniam Venkatraman Krishnan, Fotios Psallidas, Yinghui Wu
-
Publication number: 20210124739Abstract: The description relates to executing an inference query relative to a database management system, such as a relational database management system. In one example a trained machine learning model can be stored within the database management system. An inference query can be received that applies the trained machine learning model on data local to the database management system. Analysis can be performed on the inference query and the trained machine learning model to generate a unified intermediate representation of the inference query and the trained model. Cross optimization can be performed on the unified intermediate representation. Based upon the cross-optimization, a first portion of the unified intermediate representation to be executed by a database engine of the database management system can be determined, and, a second portion of the unified intermediate representation to be executed by a machine learning runtime can be determined.Type: ApplicationFiled: August 11, 2020Publication date: April 29, 2021Applicant: Microsoft Technology Licensing, LLCInventors: Konstantinos KARANASOS, Matteo INTERLANDI, Fotios PSALLIDAS, Rathijit SEN, Kwanghyun PARK, Ivan POPIVANOV, Subramaniam VENKATRAMAN KRISHNAN, Markus WEIMER, Yuan YU, Raghunath RAMAKRISHNAN, Carlo Aldo CURINO, Doris Suiyi XIN, Karla Jean SAUR
-
Publication number: 20210089532Abstract: The cloud-based query workload optimization system disclosed herein the cloud-based query workloads optimization system receives query logs from various query engines to a cloud data service, extracts various query entities from the query logs, parses query entities to generate a set of common workload features, generates intermediate representations of the query workloads, wherein the intermediate representations are agnostic to the language of the plurality of the queries, identifies a plurality of workload patterns based on the intermediate representations of the query workloads, categorizes the workloads in one or more workload type categories based on the workload patterns and the workload features, and selects an optimization scheme based on the category of workload pattern.Type: ApplicationFiled: September 25, 2019Publication date: March 25, 2021Inventors: Hiren S. PATEL, Rathijit SEN, Zhicheng YIN, Shi QIAO, Abhishek ROY, Alekh JINDAL, Subramaniam Venkatraman KRISHNAN, Carlo Aldo CURINO
-
Patent number: 10936367Abstract: Described herein is a system and method for ranking and/or taking an action regarding execution of jobs of a shared computing cluster based upon predicted user impact. Information regarding previous executions of a plurality of jobs is obtained, for example, from job execution log(s). Data dependencies of the plurality of jobs are determined. Job impact of each of the plurality of jobs as a function of the determined data dependencies is calculated. User impact of each of the plurality of jobs as a function of the determined data dependencies, the calculated job impact, and time is calculated. The plurality of jobs are ranked in accordance with the calculated user impact. An action is taken in accordance with the ranking of the plurality of jobs. The action can include automatic scheduling of the jobs and/or providing information regarding the rankings to a user.Type: GrantFiled: January 22, 2019Date of Patent: March 2, 2021Assignee: Microsoft Technology Licensing, LLCInventors: Carlo Aldo Curino, Konstantinos Karanasos, Subramaniam Venkatraman Krishnan, Christopher William Douglas, Sriram S Rao, Andrew F Chung