Patents by Inventor Yousef A. Khalidi

Yousef A. Khalidi has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20170111278
    Abstract: The ensuring of predictable and quantifiable networking performance includes adaptively throttling the rate of VM-to-VM traffic flow. A receiving hypervisor can detect congestion and communicate messages for throttling traffic flow to reduce congestion at the receiving hypervisor.
    Type: Application
    Filed: December 27, 2016
    Publication date: April 20, 2017
    Inventors: Changhoon Kim, Albert G. Greenberg, Alireza Dabagh, Yousef A. Khalidi, Deepak Bansal, Srikanth Kandula
  • Patent number: 9537773
    Abstract: The ensuring of predictable and quantifiable networking performance. Embodiments of the invention combine a congestion free network core with a hypervisor based (i.e., edge-based) throttling design to help insure quantitative and invariable subscription bandwidth rates. A lightweight shim layer in a hypervisor can adaptively throttle the rate of VM-to-VM traffic flow. A receiving hypervisor can detect congestion and communicate back to sending hypervisors that rates are to be regulated. In response, sending hypervisors can reduce transmission rate to mitigate congestion at the receiving hypervisor. In some embodiments, the principles are extended to any message processors communicating over a congestion free network.
    Type: Grant
    Filed: December 29, 2015
    Date of Patent: January 3, 2017
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Changhoon Kim, Albert G. Greenberg, Alireza Dabagh, Yousef A. Khalidi, Deepak Bansal, Srikanth Kandula
  • Patent number: 9507577
    Abstract: The provisioning of a host computing system by a controller located over a wide area network. The host computing system has power-on code that automatically executes upon powering up, and causes the host to notify the controller of the host address. In a first level of bootstrapping, the controller instructs the host to download a maintenance operating system. The host responds by downloading and installing a maintenance operating system, enabling further bootstrapping. The persistent memory may further have security data, such as a public key, that allows the host computing system to securely identify the source of the download instructions (and subsequent instructions) as originating from the controller. A second level of bootstrapping may accomplish the configuring of the host with a hypervisor and a host agent. A third level of bootstrapping may accomplish the provisioning of virtual machines on the host.
    Type: Grant
    Filed: June 21, 2012
    Date of Patent: November 29, 2016
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Yousef A. Khalidi, Deepak Bansal, Changhoon Kim, Srikanth Kandula, David A. Maltz
  • Publication number: 20160212011
    Abstract: Methods, systems, and computer-readable media for automating deployment of service applications by exposing environmental constraints in a service model are provided. In general, the methods are performed in the context of a general purpose platform configured as a server cloud to run various service applications distributed thereon. Accordingly, the general purpose platform may be flexibly configured to manage varying degrees of characteristics associated with each of the various service applications. Typically, these characteristics are provided in the service model that governs the environmental constraints under which each component program of the service application operates. As such, hosting environments are selected and adapted to satisfy the environmental constraints associated with each component program.
    Type: Application
    Filed: March 28, 2016
    Publication date: July 21, 2016
    Inventors: Jose Bernabeu-Auban, Yousef A. Khalidi
  • Publication number: 20160134538
    Abstract: The ensuring of predictable and quantifiable networking performance. Embodiments of the invention combine a congestion free network core with a hypervisor based (i.e., edge-based) throttling design to help insure quantitative and invariable subscription bandwidth rates. A lightweight shim layer in a hypervisor can adaptively throttle the rate of VM-to-VM traffic flow. A receiving hypervisor can detect congestion and communicate back to sending hypervisors that rates are to be regulated. In response, sending hypervisors can reduce transmission rate to mitigate congestion at the receiving hypervisor. In some embodiments, the principles are extended to any message processors communicating over a congestion free network.
    Type: Application
    Filed: December 29, 2015
    Publication date: May 12, 2016
    Inventors: Changhoon Kim, Albert G. Greenberg, Alireza Dabagh, Yousef A. Khalidi, Deepak Bansal, Srikanth Kandula
  • Patent number: 9300532
    Abstract: Methods, systems, and computer-readable media for automating deployment of service applications by exposing environmental constraints in a service model are provided. In general, the methods are performed in the context of a general purpose platform configured as a server cloud to run various service applications distributed thereon. Accordingly, the general purpose platform may be flexibly configured to manage varying degrees of characteristics associated with each of the various service applications. Typically, these characteristics are provided in the service model that governs the environmental constraints under which each component program of the service application operates. As such, hosting environments are selected and adapted to satisfy the environmental constraints associated with each component program.
    Type: Grant
    Filed: October 24, 2008
    Date of Patent: March 29, 2016
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Jose Bernabeu-Auban, Yousef A. Khalidi
  • Patent number: 9276860
    Abstract: A regional data center system corresponding to a particular region of the world. The data center system includes multiple data centers, each configured such that region-specific requests for services may be satisfied by any of the data centers. One or more of the data centers may also perform load balancing on received region-specific request. In order to perform load balancing, the receiving data center may itself process the requests, but in some cases may determine that the request is to be processed by another data center within the regional data center system.
    Type: Grant
    Filed: March 13, 2013
    Date of Patent: March 1, 2016
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Changhoon Kim, Yousef A. Khalidi, Guhan Suriyanarayanan
  • Patent number: 9262366
    Abstract: Software deployment to server nodes within large-scale networked systems is provided using image-based deployment. A mostly immutable image is provided at a central service and transferred from the central service to the server nodes. The image includes a base image, which is immutable, and a delta image, which is writable and provides role-specific and/or node-specific state to the base portion. When a server node receives an image from the central server, the server node stores the image intact without installation. The server node then runs software directly from the stored image.
    Type: Grant
    Filed: September 21, 2007
    Date of Patent: February 16, 2016
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Charles Thomas Lenzmeier, Yousef A. Khalidi, Anil Ingle, Saad Syed
  • Patent number: 9231869
    Abstract: The ensuring of predictable and quantifiable networking performance. Embodiments of the invention combine a congestion free network core with a hypervisor based (i.e., edge-based) throttling design to help insure quantitative and invariable subscription bandwidth rates. A lightweight shim layer in a hypervisor can adaptively throttle the rate of VM-to-VM traffic flow. A receiving hypervisor can detect congestion and communicate back to sending hypervisors that rates are to be regulated. In response, sending hypervisors can reduce transmission rate to mitigate congestion at the receiving hypervisor. In some embodiments, the principles are extended to any message processors communicating over a congestion free network.
    Type: Grant
    Filed: August 5, 2014
    Date of Patent: January 5, 2016
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Changhoon Kim, Albert G. Greenberg, Alireza Dabagh, Yousef A. Khalidi, Deepak Bansal, Srikanth Kandula
  • Patent number: 9038071
    Abstract: The resources needed by an application to execute are declared by the application. When the application is activated, only the declared resources are made available to the application because only the declared resources are connected to the execution environment. Accessibility to resources may be controlled by the operating system by making the resource visible or invisible to the executing software by mapping a local name used by the executing software to a global resource, possibly limiting the type of access allowed. Because the executing software relies on the mapping function performed by the operating system for access to resources, and the operating system only maps names declared by the software, the operating system can isolate the software, and prevent the application from accessing undeclared global resources.
    Type: Grant
    Filed: March 30, 2006
    Date of Patent: May 19, 2015
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Jose M. Bernabeu-Auban, Stephen E. Dossick, Frank V. Peschel-Gallee, Yousef A. Khalidi, Stephan J. Zachwieja
  • Publication number: 20140347998
    Abstract: The ensuring of predictable and quantifiable networking performance. Embodiments of the invention combine a congestion free network core with a hypervisor based (i.e., edge-based) throttling design to help insure quantitative and invariable subscription bandwidth rates. A lightweight shim layer in a hypervisor can adaptively throttle the rate of VM-to-VM traffic flow. A receiving hypervisor can detect congestion and communicate back to sending hypervisors that rates are to be regulated. In response, sending hypervisors can reduce transmission rate to mitigate congestion at the receiving hypervisor. In some embodiments, the principles are extended to any message processors communicating over a congestion free network.
    Type: Application
    Filed: August 5, 2014
    Publication date: November 27, 2014
    Inventors: Changhoon Kim, Albert G. Greenberg, Alireza Dabagh, Yousef A. Khalidi, Deepak Bansal, Srikanth Kandula
  • Patent number: 8893141
    Abstract: Systems, methods and computer storage media for operating a scalable computing platform are provided. A service description describing a requested service is received. Upon receiving the service description a determination of the required resources and the available resources is made. An instance description is produced. The resources required to sustain the deployment of the service are mapped to the available resources of the computing platform so the service may be deployed. The instance description is amended with each deployment of the service to allow for sustained deployment of the service.
    Type: Grant
    Filed: January 28, 2008
    Date of Patent: November 18, 2014
    Assignee: Microsoft Corporation
    Inventors: Jose Bernabeu-Auban, Yousef A. Khalidi
  • Publication number: 20140280818
    Abstract: A regional data center system corresponding to a particular region of the world. The data center system includes multiple data centers, each configured such that region-specific requests for services may be satisfied by any of the data centers. One or more of the data centers may also perform load balancing on received region-specific request. In order to perform load balancing, the receiving data center may itself process the requests, but in some cases may determine that the request is to be processed by another data center within the regional data center system.
    Type: Application
    Filed: March 13, 2013
    Publication date: September 18, 2014
    Applicant: MICROSOFT CORPORATION
    Inventors: Changhoon Kim, Yousef A. Khalidi, Guhan Suriyanarayanan
  • Patent number: 8804523
    Abstract: The ensuring of predictable and quantifiable networking performance. Embodiments of the invention combine a congestion free network core with a hypervisor based (i.e., edge-based) throttling design to help insure quantitative and invariable subscription bandwidth rates. A lightweight shim layer in a hypervisor can adaptively throttle the rate of VM-to-VM traffic flow. A receiving hypervisor can detect congestion and communicate back to sending hypervisors that rates are to be regulated. In response, sending hypervisors can reduce transmission rate to mitigate congestion at the receiving hypervisor. In some embodiments, the principles are extended to any message processors communicating over a congestion free network.
    Type: Grant
    Filed: June 21, 2012
    Date of Patent: August 12, 2014
    Assignee: Microsoft Corporation
    Inventors: Changhoon Kim, Albert G. Greenberg, Alireza Dabagh, Yousef A. Khalidi, Deepak Bansal, Srikanth Kandula
  • Publication number: 20140082048
    Abstract: A cloud computing environment providing a network service for a client computing entity. The network service is not an application level service, but rather a service that operates at or below the network layer in the protocol stack. For instance, the network service might be a network endpoint service such as a network address service (such as DNS) or a dynamic network service (such as DHCP), or a network traffic service such as a firewall service or a secure tunneling service (such as VPN). The service might also provide a pipeline of network services for network level traffic to and from the client computing entity. The cloud environment uses policy to determine which of a plurality of communication channels to use when exchanging cloud service data for the network service.
    Type: Application
    Filed: September 14, 2012
    Publication date: March 20, 2014
    Applicant: MICROSOFT CORPORATION
    Inventors: Yousef A. Khalidi, Deepak Bansal, Changhoon Kim, Srikanth Kandula, David A. Maltz, Parveen Kumar Patel, Albert Gordon Greenberg
  • Publication number: 20130343191
    Abstract: The ensuring of predictable and quantifiable networking performance. Embodiments of the invention combine a congestion free network core with a hypervisor based (i.e., edge-based) throttling design to help insure quantitative and invariable subscription bandwidth rates. A lightweight shim layer in a hypervisor can adaptively throttle the rate of VM-to-VM traffic flow. A receiving hypervisor can detect congestion and communicate back to sending hypervisors that rates are to be regulated. In response, sending hypervisors can reduce transmission rate to mitigate congestion at the receiving hypervisor. In some embodiments, the principles are extended to any message processors communicating over a congestion free network.
    Type: Application
    Filed: June 21, 2012
    Publication date: December 26, 2013
    Applicant: Microsoft Corporation
    Inventors: Changhoon Kim, Albert G. Greenberg, Alireza Dabagh, Yousef A. Khalidi, Deepak Bansal, Srikanth Kandula
  • Publication number: 20130346968
    Abstract: The provisioning of a host computing system by a controller located over a wide area network. The host computing system has power-on code that automatically executes upon powering up, and causes the host to notify the controller of the host address. In a first level of bootstrapping, the controller instructs the host to download a maintenance operating system. The host responds by downloading and installing a maintenance operating system, enabling further bootstrapping. The persistent memory may further have security data, such as a public key, that allows the host computing system to securely identify the source of the download instructions (and subsequent instructions) as originating from the controller. A second level of bootstrapping may accomplish the configuring of the host with a hypervisor and a host agent. A third level of bootstrapping may accomplish the provisioning of virtual machines on the host.
    Type: Application
    Filed: June 21, 2012
    Publication date: December 26, 2013
    Applicant: MICROSOFT CORPORATION
    Inventors: Yousef A. Khalidi, Deepak Bansal, Changhoon Kim, Srikanth Kandula, David A. Maltz
  • Publication number: 20130346558
    Abstract: A delivery controller for use in an enterprise environment that communicates with a cloud computing environment that is providing a service for the enterprise. As the cloud service processing progresses, some cloud service data is transferred from the cloud computing environment to the enterprise environment, and vice versa. The cloud service data may be exchanged over any one of a number of different types of communication channels. The delivery controller selects which communication channel to use to transfer specific data, depending on enterprise policy. Such policy might consider any business goals of the enterprise, and may be applied at the application level.
    Type: Application
    Filed: June 21, 2012
    Publication date: December 26, 2013
    Applicant: MICROSOFT CORPORATION
    Inventors: Yousef A. Khalidi, Deepak Bansal, Changhoon Kim, Srikanth Kandula, David A. Maltz, Parveen Patel
  • Patent number: 8539481
    Abstract: A containment mechanism provides for the grouping and isolation of multiple processes running on a single computer using a single instance of the operating system. A system is divided into one or more side-by-side and/or nested isolated environments enabling the partitioning and controlled sharing of resources by creating different views of hierarchical name spaces via virtual hierarchies.
    Type: Grant
    Filed: December 12, 2005
    Date of Patent: September 17, 2013
    Assignee: Microsoft Corporation
    Inventors: Frederick J. Smith, Jeff L. Havens, Madhusudhan Talluri, Yousef A. Khalidi
  • Patent number: 8533701
    Abstract: Systems, methods, and computer-readable media for updating a component utilized by an application within a distributed computing environment. An inventory of components relied on by applications within a distributed computing environment is created and maintained to facilitate identifying applications utilizing a particular component. A determination is made from the inventory of applications that utilize the particular component. An indication is received that an update is available for the particular component. An application image for an application utilizing the particular component is booted in an isolated computing environment to allow the component to be updated. A new image of the application is created to reflect the updated component. A user, such as a developer, of the application may be notified that the new image is available for future instantiations of the application.
    Type: Grant
    Filed: March 15, 2010
    Date of Patent: September 10, 2013
    Assignee: Microsoft Corporation
    Inventors: Yousef A. Khalidi, Jeffrey Lee Havens