Patents by Inventor Karthikeyan Subramanian

Karthikeyan Subramanian has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11972301
    Abstract: The present disclosure relates to systems, methods, and computer readable media for predicting surplus capacity on a set of server nodes and determining a quantity of deferrable virtual machines (VMs) that may be scheduled over an upcoming period of time. This determination of VM quantity may be determined while minimizing risks associated with allocation failures on the set of server nodes. This disclosure described systems that facilitate features and functionality related to improving utilization of surplus resource capacity on a plurality of server nodes by implementing VMs having some flexibility in timing of deployment while also avoiding significant risk caused as a result of over-allocated storage and computing resources. In one or more embodiments, the quantity of deferrable VMs is determined and scheduled in accordance with rules of a scheduling policy.
    Type: Grant
    Filed: April 13, 2021
    Date of Patent: April 30, 2024
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Yuwen Yang, Gurpreet Virdi, Bo Qiao, Hang Dong, Karthikeyan Subramanian, Marko Lalic, Shandan Zhou, Si Qin, Thomas Moscibroda, Yunus Mohammed
  • Patent number: 11900171
    Abstract: A cloud computing capacity management system can include a fine-grained admission control layer, a policy engine, and an enforcement layer. The fine-grained admission control layer can be configured to ingest capacity signals and create a capacity mitigation policy, based at least in part on the capacity signals, to protect available capacity of a cloud computing system for prioritized users. The capacity mitigation policy can be directed to users of the cloud computing system. The policy engine can be configured to control how the capacity mitigation policy is applied to the cloud computing system. The enforcement layer can be configured to handle incoming resource requests and to enforce resource limits based on the capacity mitigation policy as applied by the policy engine.
    Type: Grant
    Filed: February 2, 2021
    Date of Patent: February 13, 2024
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Gurpreet Virdi, Fernando Gonzalez Todisco, Karthikeyan Subramanian, Sanjay Ramanujan, Sorin Iftimie, Xing wen Wang, Thomas Moscibroda, Yunus Mohammed, Vi Lam Nguyen, Rostislav Sudakov
  • Publication number: 20230396511
    Abstract: A computer implemented method includes receiving telemetry data corresponding to capacity health of nodes in a cloud based computing system. The received telemetry data is processed via a prediction engine to provide predictions of capacity health at multiple dimensions of the cloud based computing system. Node recoverability information is received and node recovery execution is initiated as a function of the representations of capacity health and node recoverability information.
    Type: Application
    Filed: June 6, 2022
    Publication date: December 7, 2023
    Inventors: Shandan ZHOU, Sam Prakash BHERI, Karthikeyan SUBRAMANIAN, Yancheng CHEN, Gaurav JAGTIANI, Abhay Sudhir KETKAR, Hemant MALIK, Thomas MOSCIBRODA, Shweta Balkrishna PATIL, Luke Rafael RODRIGUEZ, Dalianna Victoria VAYSMAN
  • Publication number: 20230359512
    Abstract: The present disclosure relates to systems, methods, and computer readable media for predicting expansion failures and implementing defragmentation instructions based on the predicted expansion failures and other signals. For example, systems disclosed herein may apply a failure prediction model to determine an expansion failure prediction associated with an estimated likelihood that deployment failures will occur on a node cluster. The systems disclosed herein may further generate defragmentation instructions indicating a severity level that a defragmentation engine may execute on a cluster level to prevent expansion failures while minimizing negative customer impacts. By uniquely generating defragmentation instructions for each node cluster, a cloud computing system can minimize expansion failures, increase resource capacity, reduce costs, and provide access to reliable services to customers.
    Type: Application
    Filed: July 19, 2023
    Publication date: November 9, 2023
    Inventors: Shandan ZHOU, Saurabh AGARWAL, Karthikeyan SUBRAMANIAN, Thomas MOSCIBRODA, Paul Naveen SELVARAJ, Sandeep RAMJI, Sorin IFTIMIE, Nisarg SHETH, Wanghai GU, Ajay MANI, Si QIN, Yong XU, Qingwei LIN
  • Patent number: 11726836
    Abstract: The present disclosure relates to systems, methods, and computer readable media for predicting expansion failures and implementing defragmentation instructions based on the predicted expansion failures and other signals. For example, systems disclosed herein may apply a failure prediction model to determine an expansion failure prediction associated with an estimated likelihood that deployment failures will occur on a node cluster. The systems disclosed herein may further generate defragmentation instructions indicating a severity level that a defragmentation engine may execute on a cluster level to prevent expansion failures while minimizing negative customer impacts. By uniquely generating defragmentation instructions for each node cluster, a cloud computing system can minimize expansion failures, increase resource capacity, reduce costs, and provide access to reliable services to customers.
    Type: Grant
    Filed: June 12, 2020
    Date of Patent: August 15, 2023
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Shandan Zhou, Saurabh Agarwal, Karthikeyan Subramanian, Thomas Moscibroda, Paul Naveen Selvaraj, Sandeep Ramji, Sorin Iftimie, Nisarg Sheth, Wanghai Gu, Ajay Mani, Si Qin, Yong Xu, Qingwei Lin
  • Patent number: 11652720
    Abstract: The present disclosure relates to systems, methods, and computer readable media for predicting deployment growth on one or more node clusters and selectively permitting deployment requests on a per cluster basis. For example, systems disclosed herein may apply tenant growth prediction system trained to output a deployment growth classification indicative of a predicted growth of deployments on a node cluster. The system disclosed herein may further utilize the deployment growth classification to determine whether a deployment request may be permitted while maintaining a sufficiently sized capacity buffer to avoid deployment failures for existing deployments previously implemented on the node cluster. By selectively permitting or denying deployments based on a variety of factors, the systems described herein can more efficiently utilize cluster resources on a per-cluster basis without causing a significant increase in deployment failures for existing customers.
    Type: Grant
    Filed: September 20, 2019
    Date of Patent: May 16, 2023
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Shandan Zhou, John Lawrence Miller, Christopher Cowdery, Thomas Moscibroda, Shanti Kemburu, Yong Xu, Si Qin, Qingwei Lin, Eli Cortez, Karthikeyan Subramanian
  • Patent number: 11550634
    Abstract: A method for minimizing allocation failures in a cloud computing system without overprovisioning may include determining a predicted supply for a virtual machine series in a system unit of the cloud computing system during an upcoming time period. The predicted supply may be based on a shared available current capacity and a shared available future added capacity for the virtual machine series in the system unit. The method may also include predicting an available capacity for the virtual machine series in the system unit during the upcoming time period. The predicted available capacity may be based at least in part on a predicted demand for the virtual machine series in the system unit during the upcoming time period and the predicted supply. The method may also include taking at least one mitigation action in response to determining that the predicted demand exceeds the predicted supply during the upcoming time period.
    Type: Grant
    Filed: March 8, 2019
    Date of Patent: January 10, 2023
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Saurabh Agarwal, Maitreyee Ramprasad Joshi, Vinayak Ramnath Karnataki, Neha Keshari, Gowtham Natarajan, Yash Purohit, Sanjay Ramanujan, Karthikeyan Subramanian, Ambrose Thomas Treacy, Shandan Zhou
  • Publication number: 20220327002
    Abstract: The present disclosure relates to systems, methods, and computer readable media for predicting surplus capacity on a set of server nodes and determining a quantity of deferrable virtual machines (VMs) that may be scheduled over an upcoming period of time. This determination of VM quantity may be determined while minimizing risks associated with allocation failures on the set of server nodes. This disclosure described systems that facilitate features and functionality related to improving utilization of surplus resource capacity on a plurality of server nodes by implementing VMs having some flexibility in timing of deployment while also avoiding significant risk caused as a result of over-allocated storage and computing resources. In one or more embodiments, the quantity of deferrable VMs is determined and scheduled in accordance with rules of a scheduling policy.
    Type: Application
    Filed: April 13, 2021
    Publication date: October 13, 2022
    Inventors: Yuwen YANG, Gurpreet VIRDI, Bo QIAO, Hang DONG, Karthikeyan SUBRAMANIAN, Marko LALIC, Shandan ZHOU, Si QIN, Thomas MOSCIBRODA, Yunus MOHAMMED
  • Publication number: 20220245001
    Abstract: A cloud computing capacity management system can include a fine-grained admission control layer, a policy engine, and an enforcement layer. The fine-grained admission control layer can be configured to ingest capacity signals and create a capacity mitigation policy, based at least in part on the capacity signals, to protect available capacity of a cloud computing system for prioritized users. The capacity mitigation policy can be directed to users of the cloud computing system. The policy engine can be configured to control how the capacity mitigation policy is applied to the cloud computing system. The enforcement layer can be configured to handle incoming resource requests and to enforce resource limits based on the capacity mitigation policy as applied by the policy engine.
    Type: Application
    Filed: February 2, 2021
    Publication date: August 4, 2022
    Inventors: Gurpreet VIRDI, Fernando GONZALEZ TODISCO, Karthikeyan SUBRAMANIAN, Sanjay RAMANUJAN, Sorin IFTIMIE, Xing wen WANG, Thomas MOSCIBRODA, Yunus MOHAMMED, Vi Lam NGUYEN, Rostislav SUDAKOV
  • Publication number: 20220206771
    Abstract: The present invention is directed towards a method 200 and system 100 to automate the entire lifecycle of Apigee API Management deployment via a templated approach, the method 200 comprising steps of receiving data related to application programming interfaces from the customer device 102a at the developer device 102b, validating the data for knowing the state of existing application programming interfaces (APIs), creating a template based on customer requirement, wherein the template is creating by generating an APIGEEĀ® application programming interface proxy 412 and customer integration or customer delivery (CI/CD) configuration 414 based on the data, storing the APIGEEĀ® application programming interface proxy and customer integration or customer delivery configuration at a data repository, generating a repeatable build job for deployment of the APIGEE R application programming interface proxy and customer integration or customer delivery configuration and deploying and executing the job on the on premises
    Type: Application
    Filed: December 24, 2020
    Publication date: June 30, 2022
    Inventor: Karthikeyan Subramanian
  • Publication number: 20210389894
    Abstract: The present disclosure relates to systems, methods, and computer readable media for predicting expansion failures and implementing defragmentation instructions based on the predicted expansion failures and other signals. For example, systems disclosed herein may apply a failure prediction model to determine an expansion failure prediction associated with an estimated likelihood that deployment failures will occur on a node cluster. The systems disclosed herein may further generate defragmentation instructions indicating a severity level that a defragmentation engine may execute on a cluster level to prevent expansion failures while minimizing negative customer impacts. By uniquely generating defragmentation instructions for each node cluster, a cloud computing system can minimize expansion failures, increase resource capacity, reduce costs, and provide access to reliable services to customers.
    Type: Application
    Filed: June 12, 2020
    Publication date: December 16, 2021
    Inventors: Shandan ZHOU, Saurabh AGARWAL, Karthikeyan SUBRAMANIAN, Thomas MOSCIBRODA, Paul Naveen SELVARAJ, Sandeep RAMJI, Sorin IFTIMIE, Nisarg SHETH, Wanghai GU, Ajay MANI, Si QIN, Yong XU, Qingwei LIN
  • Patent number: 11093266
    Abstract: A method for evaluating at least one potential policy for an IaaS system may include determining a predicted workload for the IaaS system based on at least one generative model corresponding to the IaaS system. The at least one potential policy for the IaaS system may be simulated based on the predicted workload, thereby producing one or more simulation metrics that indicate effects of the at least one potential policy. The performance of the IaaS system may be optimized based on the one or more simulation metrics.
    Type: Grant
    Filed: October 15, 2018
    Date of Patent: August 17, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Gowtham Natarajan, Karel Trueba Nobregas, Abhisek Pan, Karthikeyan Subramanian
  • Patent number: 10901824
    Abstract: Embodiments relate to determining whether to take a resource distribution unit (RDU) of a datacenter offline when the RDU becomes faulty. RDUs in a cloud or datacenter supply a resource such as power, network connectivity, and the like to respective sets of hosts that provide computing resources to tenant units such as virtual machines (VMs). When an RDU becomes faulty some of the hosts that it supplies may continue to function and others may become unavailable for various reasons. This can make a decision of whether to take the RDU offline for repair difficult, since in some situations countervailing requirements of the datacenter may be at odds. To decide whether to take an RDU offline, the potential impact on availability of tenant VMs, unused capacity of the datacenter, a number or ratio of unavailable hosts on the RDU, and other factors may be considered to make a balanced decision.
    Type: Grant
    Filed: July 20, 2018
    Date of Patent: January 26, 2021
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Saurabh Agarwal, Koon Hui Geoffrey Goh, Asad Yaqoob, Shandan Zhou, Karthikeyan Subramanian, Gowtham Natarajan, Vipin Kumar
  • Publication number: 20200387401
    Abstract: The present disclosure relates to systems, methods, and computer readable media for predicting deployment growth on one or more node clusters and selectively permitting deployment requests on a per cluster basis. For example, systems disclosed herein may apply tenant growth prediction system trained to output a deployment growth classification indicative of a predicted growth of deployments on a node cluster. The system disclosed herein may further utilize the deployment growth classification to determine whether a deployment request may be permitted while maintaining a sufficiently sized capacity buffer to avoid deployment failures for existing deployments previously implemented on the node cluster. By selectively permitting or denying deployments based on a variety of factors, the systems described herein can more efficiently utilize cluster resources on a per-cluster basis without causing a significant increase in deployment failures for existing customers.
    Type: Application
    Filed: September 20, 2019
    Publication date: December 10, 2020
    Inventors: Shandan ZHOU, John Lawrence MILLER, Christopher COWDERY, Thomas MOSCIBRODA, Shanti KEMBURU, Yong XU, Si QIN, Qingwei LIN, Eli CORTEZ, Karthikeyan SUBRAMANIAN
  • Publication number: 20200285525
    Abstract: A method for minimizing allocation failures in a cloud computing system without overprovisioning may include determining a predicted supply for a virtual machine series in a system unit of the cloud computing system during an upcoming time period. The predicted supply may be based on a shared available current capacity and a shared available future added capacity for the virtual machine series in the system unit. The method may also include predicting an available capacity for the virtual machine series in the system unit during the upcoming time period. The predicted available capacity may be based at least in part on a predicted demand for the virtual machine series in the system unit during the upcoming time period and the predicted supply. The method may also include taking at least one mitigation action in response to determining that the predicted demand exceeds the predicted supply during the upcoming time period.
    Type: Application
    Filed: March 8, 2019
    Publication date: September 10, 2020
    Inventors: Saurabh AGARWAL, Maitreyee Ramprasad JOSHI, Vinayak Ramnath KARNATAKI, Neha KESHARI, Gowtham NATARAJAN, Yash PUROHIT, Sanjay RAMANUJAN, Karthikeyan SUBRAMANIAN, Ambrose Thomas TREACY, Shandan ZHOU
  • Patent number: 10698926
    Abstract: Aspects extend to methods, systems, and computer program products for clustering streamed or batch data. Aspects of the invention include dynamic clustering and labeling of streamed data and/or batch data, including failures and error logs (user, platform, etc.), latency logs, warning logs, information logs, Virtual Machine (VM) creation data logs, template logs, etc., for use in analysis (e.g., error log analysis). A clustering system can learn from previously identified patterns and use that information to group newer information dynamically as it gets generated. The clustering system can leverage streamed data and/or batch data domain knowledge for preprocessing. In one aspect, a clustering system uses a similarity measure. Based on (e.g., users' configuration of) a similarity threshold, the cluster system (e.g., automatically) assigns/clusters streamed data and/or batch data into groups.
    Type: Grant
    Filed: April 20, 2017
    Date of Patent: June 30, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Shandan Zhou, Karthikeyan Subramanian, Murtaza Muidul Huda Chowdhury, Gowtham Natarajan
  • Publication number: 20200117492
    Abstract: A method for evaluating at least one potential policy for an IaaS system may include determining a predicted workload for the IaaS system based on at least one generative model corresponding to the IaaS system. The at least one potential policy for the IaaS system may be simulated based on the predicted workload, thereby producing one or more simulation metrics that indicate effects of the at least one potential policy. The performance of the IaaS system may be optimized based on the one or more simulation metrics.
    Type: Application
    Filed: October 15, 2018
    Publication date: April 16, 2020
    Inventors: Gowtham NATARAJAN, Karel TRUEBA NOBREGAS, Abhisek PAN, Karthikeyan SUBRAMANIAN
  • Patent number: 10565021
    Abstract: Techniques for automated capacity managed in distributed computing systems are disclosed herein. In one embodiment, a method includes receiving predicting one or more future usage levels of a computing resource in the distributed computing system based on received data representing historical usage levels of the computing resource and determining whether a currently available capacity of the computing resource in the distributed computing system is depleted beyond a threshold time period based on the one or more future usage levels. In response to determining that the currently available capacity of the computing resource in the distributed computing system is depleted before the threshold time period, the method includes immediately rebooting, reimaging, or performing other recovery actions on one or more out-for-repair hosts that provide the computing resource.
    Type: Grant
    Filed: November 30, 2017
    Date of Patent: February 18, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Shandan Zhou, Karthikeyan Subramanian, Zainab Hakim, Valentina Li, Michal Jama
  • Publication number: 20200026591
    Abstract: Embodiments relate to determining whether to take a resource distribution unit (RDU) of a datacenter offline when the RDU becomes faulty. RDUs in a cloud or datacenter supply a resource such as power, network connectivity, and the like to respective sets of hosts that provide computing resources to tenant units such as virtual machines (VMs). When an RDU becomes faulty some of the hosts that it supplies may continue to function and others may become unavailable for various reasons. This can make a decision of whether to take the RDU offline for repair difficult, since in some situations countervailing requirements of the datacenter may be at odds. To decide whether to take an RDU offline, the potential impact on availability of tenant VMs, unused capacity of the datacenter, a number or ratio of unavailable hosts on the RDU, and other factors may be considered to make a balanced decision.
    Type: Application
    Filed: July 20, 2018
    Publication date: January 23, 2020
    Inventors: Saurabh Agarwal, Koon Hui Geoffrey Goh, Asad Yaqoob, Shandan Zhou, Karthikeyan Subramanian, Gowtham Natarajan, Vipin Kumar
  • Patent number: 10510431
    Abstract: A method of detecting random telegraph noise defects in a memory includes initializing a first bit cell of the memory to a first value and reading the first value from the first bit cell. The method also includes writing a second value to the first bit cell and performing back to back read operations on a second bit cell adjacent to the first bit cell, after writing the second value. The method further includes attempting to read the second value from the first bit cell and determining whether the first bit cell is defective based on whether the second value was read from the first bit cell.
    Type: Grant
    Filed: September 22, 2017
    Date of Patent: December 17, 2019
    Assignee: QUALCOMM Incorporated
    Inventors: Sneha Revankar, Karthikeyan Subramanian