Patents by Inventor Subramaniam Venkatraman Krishnan

Subramaniam Venkatraman Krishnan has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20250190280
    Abstract: Methods, systems, apparatuses, and computer-readable storage mediums described herein are directed to determining and recommending an optimal compute resource configuration for a cloud-based resource (e.g., a server, a virtual machine, etc.) for migrating a customer to the cloud. The embodiments described herein utilize a statistically robust approach that makes recommendations that are more flexible (elastic) and account for the full distribution of the amount of resource usage. Such an approach is utilized to develop a personalized rank of relevant recommendations to a customer. To determine which compute resource configuration to recommend to the customer, the customer's usage profile is matched to a set of customers that have already migrated to the cloud. The compute resource configuration that reaches the performance most similar to the performance of the configurations utilized by customers in the matched set is recommended to the user.
    Type: Application
    Filed: February 18, 2025
    Publication date: June 12, 2025
    Inventors: Wenjing WANG, Joyce Yu CAHOON, Yiwen ZHU, Ya LIN, Subramaniam Venkatraman KRISHNAN, Neetu SINGH, Raymond TRUONG, XingYu LIU, Maria Alexandra CIORTEA, Sreraman NARASIMHAN, Pratyush RAWAT, Haitao SONG
  • Publication number: 20250156430
    Abstract: The description relates to executing an inference query relative to a database management system, such as a relational database management system. In one example a trained machine learning model can be stored within the database management system. An inference query can be received that applies the trained machine learning model on data local to the database management system. Analysis can be performed on the inference query and the trained machine learning model to generate a unified intermediate representation of the inference query and the trained model. Cross optimization can be performed on the unified intermediate representation. Based upon the cross-optimization, a first portion of the unified intermediate representation to be executed by a database engine of the database management system can be determined, and, a second portion of the unified intermediate representation to be executed by a machine learning runtime can be determined.
    Type: Application
    Filed: January 15, 2025
    Publication date: May 15, 2025
    Applicant: Microsoft Technology Licensing, LLC
    Inventors: Konstantinos KARANASOS, Matteo INTERLANDI, Fotios PSALLIDAS, Rathijit SEN, Kwanghyun PARK, Ivan POPIVANOV, Subramaniam VENKATRAMAN KRISHNAN, Markus WEIMER, Yuan YU, Raghunath RAMAKRISHNAN, Carlo Aldo CURINO, Doris Suiyi XIN, Karla Jean SAUR
  • Publication number: 20250123939
    Abstract: Example solutions provide an artificial intelligence (AI) agent for pre-build configuration of cloud services in order to enable the initial build of a computational resource (e.g., in a cloud service) to minimize the likelihood of excessive throttling or slack. Examples leverage prior-existing utilization data and project metadata to identify similar use cases. The utilization data includes capacity information and resource consumption information (e.g., throttling and slack) for prior-existing computational resources, and the project metadata includes information for hierarchically categorization, to identify similar resources. A pre-build configuration is generated for the customer's resource, which the customer may tune based upon the customer's preferences for a cost and performance balance point.
    Type: Application
    Filed: October 13, 2023
    Publication date: April 17, 2025
    Inventors: Rajeev Sudhakar BHOPI, Yiwen ZHU, Helen Mary SERR, Jonah KARPMAN, Matthew Joseph GLEESON, Nicholas Kent GLAZE, Subramaniam Venkatraman KRISHNAN, Irwin Hollar MCNEELY, III
  • Patent number: 12260265
    Abstract: Methods, systems, apparatuses, and computer-readable storage mediums described herein are directed to determining and recommending an optimal compute resource configuration for a cloud-based resource (e.g., a server, a virtual machine, etc.) for migrating a customer to the cloud. The embodiments described herein utilize a statistically robust approach that makes recommendations that are more flexible (elastic) and account for the full distribution of the amount of resource usage. Such an approach is utilized to develop a personalized rank of relevant recommendations to a customer. To determine which compute resource configuration to recommend to the customer, the customer's usage profile is matched to a set of customers that have already migrated to the cloud. The compute resource configuration that reaches the performance most similar to the performance of the configurations utilized by customers in the matched set is recommended to the user.
    Type: Grant
    Filed: December 20, 2021
    Date of Patent: March 25, 2025
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Wenjing Wang, Joyce Yu Cahoon, Yiwen Zhu, Ya Lin, Subramaniam Venkatraman Krishnan, Neetu Singh, Raymond Truong, Xingyu Liu, Maria Alexandra Ciortea, Sreraman Narasimhan, Pratyush Rawat, Haitao Song
  • Publication number: 20250086202
    Abstract: Systems, methods and computer-readable memory devices are provided for greater efficiency in the configuration of a database cluster for performing a query workload. A database cluster configuration system is provided that includes a database cluster comprising one or more compute resources configured to perform database queries. A query workload comprising a plurality of queries is received. An initial workload-level configuration is applied. For each query of the query workload, a query-level configuration is generated using a query configuration model corresponding to each query in a contextual Bayesian optimization with centroid learning while also leveraging the query plan for each executing query for query characterization and including application of virtual operators. Query events are collected and used to update the corresponding query configuration model. The workload-level configuration is updated based on the query events and cached for use during a subsequent execution of the workload.
    Type: Application
    Filed: September 13, 2023
    Publication date: March 13, 2025
    Inventors: Yiwen ZHU, Subramaniam Venkatraman KRISHNAN, Weihan TANG, Tengfei HUANG, Rui FANG, Rahul Kumar CHALLAPALLI, Mo LIU, Long TIAN, Karuna Sagar KRISHNA, Estera Zaneta KOT, Xin HE, Ashit R. GOSALIA, Dario Kikuchi BERNAL, Aditya LAKRA, Arshdeep SEKHON, Sule KAHRAMAN, Carlo Aldo CURINO, Brian Paul KROTH, Rathijit SEN, Andreas Christian MUELLER, Shaily Jignesh FOZDAR, Dhruv Harendra RELWANI, Xiang LI, Sergiy MATUSEVYCH
  • Patent number: 12242493
    Abstract: The description relates to executing an inference query relative to a database management system, such as a relational database management system. In one example a trained machine learning model can be stored within the database management system. An inference query can be received that applies the trained machine learning model on data local to the database management system. Analysis can be performed on the inference query and the trained machine learning model to generate a unified intermediate representation of the inference query and the trained model. Cross optimization can be performed on the unified intermediate representation. Based upon the cross-optimization, a first portion of the unified intermediate representation to be executed by a database engine of the database management system can be determined, and, a second portion of the unified intermediate representation to be executed by a machine learning runtime can be determined.
    Type: Grant
    Filed: August 11, 2020
    Date of Patent: March 4, 2025
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Konstantinos Karanasos, Matteo Interlandi, Fotios Psallidas, Rathijit Sen, Kwanghyun Park, Ivan Popivanov, Subramaniam Venkatraman Krishnan, Markus Weimer, Yuan Yu, Raghunath Ramakrishnan, Carlo Aldo Curino, Doris Suiyi Xin, Karla Jean Saur
  • Publication number: 20240231927
    Abstract: The present application relates to a network, apparatus, and method for allocating clusters of computing nodes for programming jobs. A network includes a plurality of datacenters including computing resources configurable to instantiate nodes for executing programming jobs on a cluster. The computing resources at one of the datacenters are configured to: provision a live pool including a number of clusters, each cluster in the live pool including a plurality of nodes imaged with a configuration for executing the programming jobs in parallel on the cluster; receive a request from a user to execute a programming job; allocate a cluster from the live pool to the user for the programming job when the cluster is available; evict the cluster from the live pool; and provision a new cluster within the live pool to meet the number of clusters. The number of clusters may be optimized based on linear programming and machine-learning.
    Type: Application
    Filed: January 10, 2023
    Publication date: July 11, 2024
    Inventors: Yiwen ZHU, Alex YEO, Harsha Nihanth NAGULAPALLI, Sumeet KHUSHALANI, Arijit TARAFDAR, Subramaniam VENKATRAMAN KRISHNAN, Deepak RAVIKUMAR, Andrew Francis FOGARTY, Steve D. SUH, Yoonjae PARK, Niharika DUTTA, Santhosh Kumar RAVINDRAN
  • Patent number: 12013853
    Abstract: The cloud-based query workload optimization system disclosed herein the cloud-based query workloads optimization system receives query logs from various query engines to a cloud data service, extracts various query entities from the query logs, parses query entities to generate a set of common workload features, generates intermediate representations of the query workloads, wherein the intermediate representations are agnostic to the language of the plurality of the queries, identifies a plurality of workload patterns based on the intermediate representations of the query workloads, categorizes the workloads in one or more workload type categories based on the workload patterns and the workload features, and selects an optimization scheme based on the category of workload pattern.
    Type: Grant
    Filed: September 25, 2019
    Date of Patent: June 18, 2024
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Hiren S. Patel, Rathijit Sen, Zhicheng Yin, Shi Qiao, Abhishek Roy, Alekh Jindal, Subramaniam Venkatraman Krishnan, Carlo Aldo Curino
  • Publication number: 20240168948
    Abstract: Learned workload synthesis is disclosed. In an aspect of the present disclosure, a time series dataset corresponding to a target workload is received. A set of performance characteristics is determined from the time series dataset. A call is provided to a prediction model to determine a candidate query sequence based on the determined set of performance characteristics. A synthetic workload is generated based on the determined candidate query sequence. A synthetic workload is generated based on the determined candidate query sequence. A first similarity between a first performance profile of the synthetic workload and a second performance profile of the target workload meets a workload performance threshold condition. A performance insight is determined based on the synthetic workload. In a further aspect, the prediction model is trained to predict performance profiles based on workload profiles generated by executing benchmark queries using hardware and/or software configurations.
    Type: Application
    Filed: November 23, 2022
    Publication date: May 23, 2024
    Inventors: Yiwen ZHU, Joyce CAHOON, Subramaniam Venkatraman KRISHNAN, Chengcheng WAN
  • Publication number: 20240111739
    Abstract: An automated tuning service is used to automatically tune, or modify, the operational parameters of a large-scale cloud infrastructure. The tuning service performs automated and fully data/model-driven configuration based from learning various real-time performance of the cloud infrastructure. Such performance is identified through monitoring various telemetric data of the cloud infrastructure. The tuning service leverages a mix of domain knowledge and principled data-science to capture the essence of our cluster dynamic behavior in a collection of descriptive machine learning (ML) models. The ML models power automated optimization procedures for parameter tuning, and inform administrators in most tactical and strategical engineering/capacity decisions (such as hardware and datacenter design, software investments, etc.).
    Type: Application
    Filed: December 8, 2023
    Publication date: April 4, 2024
    Inventors: Yiwen ZHU, Subramaniam Venkatraman KRISHNAN, Konstantinos KARANASOS, Carlo CURINO, Isha TARTE, Sudhir DARBHA
  • Patent number: 11880347
    Abstract: An automated tuning service is used to automatically tune, or modify, the operational parameters of a large-scale cloud infrastructure. The tuning service performs automated and fully data/model-driven configuration based from learning various real-time performance of the cloud infrastructure. Such performance is identified through monitoring various telemetric data of the cloud infrastructure. The tuning service leverages a mix of domain knowledge and principled data-science to capture the essence of our cluster dynamic behavior in a collection of descriptive machine learning (ML) models. The ML models power automated optimization procedures for parameter tuning, and inform administrators in most tactical and strategical engineering/capacity decisions (such as hardware and datacenter design, software investments, etc.).
    Type: Grant
    Filed: April 2, 2021
    Date of Patent: January 23, 2024
    Assignee: Microsoft Technology Licensing, LLC.
    Inventors: Yiwen Zhu, Subramaniam Venkatraman Krishnan, Konstantinos Karanasos, Carlo Curino, Isha Tarte, Sudhir Darbha
  • Publication number: 20230394369
    Abstract: Embodiments described herein enable tracking machine learning (“ML”) model data provenance. In particular, a computing device is configured to accept ML model code that, when executed, instantiates and trains an ML model, to parse the ML model code into a workflow intermediate representation (WIR), to semantically annotate the WIR to provide an annotated WIR, and to identify, based on the annotated WIR and ML API corresponding to the ML model code, data from at least one data source that is relied upon by the ML model code when training the ML model. A WIR may be generated from an abstract syntax tree (AST) based on the ML model code, generating provenance relationships (PRs) based at least in part on relationships between nodes of the AST, wherein a PR comprises one or more input variables, an operation, a caller, and one or more output variables.
    Type: Application
    Filed: August 21, 2023
    Publication date: December 7, 2023
    Inventors: Avrilia FLORATOU, Ashvin AGRAWAL, MohammadHossein NAMAKI, Subramaniam Venkatraman KRISHNAN, Fotios PSALLIDAS, Yinghui WU
  • Publication number: 20230385649
    Abstract: Linguistic schema mapping via semi-supervised learning is used to map a customer schema to a particular industry-specific schema (ISS). The customer schema is received and a corresponding ISS is identified. An attribute in the customer schema is selected for labeling. Candidate pairs are generated that include the first attribute and one or more second attributes which may describe the first attribute. A featurizer determines similarities between the first attribute and second attribute in each generated pair, one or more suggested labels are generated by a machine learning (ML) model, and one of the suggested labels is applied to the first attribute.
    Type: Application
    Filed: May 28, 2022
    Publication date: November 30, 2023
    Inventors: Avrilia FLORATOU, Joyce Yu CAHOON, Subramaniam Venkatraman KRISHNAN, Andreas C. MUELLER, Dalitso Hansini BANDA, Fotis PSALLIDAS, Jignesh PATEL, Yunjia ZHANG
  • Patent number: 11775862
    Abstract: A system enables tracking machine learning (“ML”) model data provenance. In particular, a computing device is configured to accept ML model code that, when executed, instantiates and trains an ML model, to parse the ML model code into a workflow intermediate representation (WIR), to semantically annotate the WIR to provide an annotated WIR, and to identify, based on the annotated WIR and ML API corresponding to the ML model code, data from at least one data source that is relied upon by the ML model code when training the ML model. A WIR may be generated from an abstract syntax tree (AST) based on the ML model code, generating provenance relationships (PRs) based at least in part on relationships between nodes of the AST, wherein a PR comprises one or more input variables, an operation, a caller, and one or more output variables.
    Type: Grant
    Filed: January 14, 2020
    Date of Patent: October 3, 2023
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Avrilia Floratou, Ashvin Agrawal, MohammadHossein Namaki, Subramaniam Venkatraman Krishnan, Fotios Psallidas, Yinghui Wu
  • Publication number: 20230029888
    Abstract: Methods, systems, apparatuses, and computer-readable storage mediums described herein are directed to determining and recommending an optimal compute resource configuration for a cloud-based resource (e.g., a server, a virtual machine, etc.) for migrating a customer to the cloud. The embodiments described herein utilize a statistically robust approach that makes recommendations that are more flexible (elastic) and account for the full distribution of the amount of resource usage. Such an approach is utilized to develop a personalized rank of relevant recommendations to a customer. To determine which compute resource configuration to recommend to the customer, the customer’s usage profile is matched to a set of customers that have already migrated to the cloud. The compute resource configuration that reaches the performance most similar to the performance of the configurations utilized by customers in the matched set is recommended to the user.
    Type: Application
    Filed: December 20, 2021
    Publication date: February 2, 2023
    Inventors: Wenjing WANG, Joyce Yu CAHOON, Yiwen ZHU, Ya LIN, Subramaniam Venkatraman KRISHNAN, Neetu SINGH, Raymond TRUONG, XingYu LIU, Maria Alexandra CIORTEA, Sreraman NARASIMHAN, Pratyush RAWAT, Haitao SONG
  • Publication number: 20220164327
    Abstract: An automated tuning service is used to automatically tune, or modify, the operational parameters of a large-scale cloud infrastructure. The tuning service performs automated and fully data/model-driven configuration based from learning various real-time performance of the cloud infrastructure. Such performance is identified through monitoring various telemetric data of the cloud infrastructure. The tuning service leverages a mix of domain knowledge and principled data-science to capture the essence of our cluster dynamic behavior in a collection of descriptive machine learning (ML) models. The ML models power automated optimization procedures for parameter tuning, and inform administrators in most tactical and strategical engineering/capacity decisions (such as hardware and datacenter design, software investments, etc.).
    Type: Application
    Filed: April 2, 2021
    Publication date: May 26, 2022
    Inventors: Yiwen ZHU, Subramaniam Venkatraman KRISHNAN, Konstantinos KARANASOS, Carlo CURINO, Isha TARTE, Sudhir Darbha
  • Publication number: 20210216905
    Abstract: Embodiments described herein enable tracking machine learning (“ML”) model data provenance. In particular, a computing device is configured to accept ML model code that, when executed, instantiates and trains an ML model, to parse the ML model code into a workflow intermediate representation (WIR), to semantically annotate the WIR to provide an annotated WIR, and to identify, based on the annotated WIR and ML API corresponding to the ML model code, data from at least one data source that is relied upon by the ML model code when training the ML model. A WIR may be generated from an abstract syntax tree (AST) based on the ML model code, generating provenance relationships (PRs) based at least in part on relationships between nodes of the AST, wherein a PR comprises one or more input variables, an operation, a caller, and one or more output variables.
    Type: Application
    Filed: January 14, 2020
    Publication date: July 15, 2021
    Inventors: Avrilia Floratou, Ashvin Agrawal, MohammadHossein Namaki, Subramaniam Venkatraman Krishnan, Fotios Psallidas, Yinghui Wu
  • Publication number: 20210124739
    Abstract: The description relates to executing an inference query relative to a database management system, such as a relational database management system. In one example a trained machine learning model can be stored within the database management system. An inference query can be received that applies the trained machine learning model on data local to the database management system. Analysis can be performed on the inference query and the trained machine learning model to generate a unified intermediate representation of the inference query and the trained model. Cross optimization can be performed on the unified intermediate representation. Based upon the cross-optimization, a first portion of the unified intermediate representation to be executed by a database engine of the database management system can be determined, and, a second portion of the unified intermediate representation to be executed by a machine learning runtime can be determined.
    Type: Application
    Filed: August 11, 2020
    Publication date: April 29, 2021
    Applicant: Microsoft Technology Licensing, LLC
    Inventors: Konstantinos KARANASOS, Matteo INTERLANDI, Fotios PSALLIDAS, Rathijit SEN, Kwanghyun PARK, Ivan POPIVANOV, Subramaniam VENKATRAMAN KRISHNAN, Markus WEIMER, Yuan YU, Raghunath RAMAKRISHNAN, Carlo Aldo CURINO, Doris Suiyi XIN, Karla Jean SAUR
  • Publication number: 20210089532
    Abstract: The cloud-based query workload optimization system disclosed herein the cloud-based query workloads optimization system receives query logs from various query engines to a cloud data service, extracts various query entities from the query logs, parses query entities to generate a set of common workload features, generates intermediate representations of the query workloads, wherein the intermediate representations are agnostic to the language of the plurality of the queries, identifies a plurality of workload patterns based on the intermediate representations of the query workloads, categorizes the workloads in one or more workload type categories based on the workload patterns and the workload features, and selects an optimization scheme based on the category of workload pattern.
    Type: Application
    Filed: September 25, 2019
    Publication date: March 25, 2021
    Inventors: Hiren S. PATEL, Rathijit SEN, Zhicheng YIN, Shi QIAO, Abhishek ROY, Alekh JINDAL, Subramaniam Venkatraman KRISHNAN, Carlo Aldo CURINO
  • Patent number: 10936367
    Abstract: Described herein is a system and method for ranking and/or taking an action regarding execution of jobs of a shared computing cluster based upon predicted user impact. Information regarding previous executions of a plurality of jobs is obtained, for example, from job execution log(s). Data dependencies of the plurality of jobs are determined. Job impact of each of the plurality of jobs as a function of the determined data dependencies is calculated. User impact of each of the plurality of jobs as a function of the determined data dependencies, the calculated job impact, and time is calculated. The plurality of jobs are ranked in accordance with the calculated user impact. An action is taken in accordance with the ranking of the plurality of jobs. The action can include automatic scheduling of the jobs and/or providing information regarding the rankings to a user.
    Type: Grant
    Filed: January 22, 2019
    Date of Patent: March 2, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Carlo Aldo Curino, Konstantinos Karanasos, Subramaniam Venkatraman Krishnan, Christopher William Douglas, Sriram S Rao, Andrew F Chung