Patents by Inventor Patrick G. Kutch
Patrick G. Kutch has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20250097306Abstract: An architecture to perform resource management among multiple network nodes and associated resources is disclosed. Example resource management techniques include those relating to: proactive reservation of edge computing resources; deadline-driven resource allocation; speculative edge QOS pre-allocation; and automatic QoS migration across edge computing nodes.Type: ApplicationFiled: September 24, 2024Publication date: March 20, 2025Inventors: Francesc Guim Bernat, Patrick Bohan, Kshitij Arun Doshi, Brinda Ganesh, Andrew J. Herdrich, Monica Kenguva, Karthik Kumar, Patrick G. Kutch, Felipe Pastor Beneyto, Rashmin Patel, Suraj Prabhakaran, Ned M. Smith, Petar Torre, Alexander Vul
-
Patent number: 12164977Abstract: An apparatus comprising a network interface controller comprising a queue for messages for a thread executing on a host computing system, wherein the queue is dedicated to the thread; and circuitry to send a notification to the host computing system to resume execution of the thread when a monitoring rule for the queue has been triggered.Type: GrantFiled: December 23, 2020Date of Patent: December 10, 2024Assignee: Intel CorporationInventors: Francesc Guim Bernat, Karthik Kumar, Patrick G. Kutch, Alexander Bachmutsky, Nicolae Octavian Popovici
-
Patent number: 12160368Abstract: Examples described herein relate to a device configured to allocate memory resources for packets received by the network interface based on received configuration settings. In some examples, the device is a network interface. Received configuration settings can include one or more of: latency, memory bandwidth, timing of when the content is expected to be accessed, or encryption parameters. In some examples, memory resources include one or more of: a cache, a volatile memory device, a storage device, or persistent memory. In some examples, based on a configuration settings not being available, the network interface is to perform one or more of: dropping a received packet, store the received packet in a buffer that does not meet the configuration settings, or indicate an error. In some examples, configuration settings are conditional where the settings are applied if one or more conditions is met.Type: GrantFiled: April 27, 2020Date of Patent: December 3, 2024Assignee: Intel CorporationInventors: Francesc Guim Bernat, Patrick Connor, Patrick G. Kutch, John J. Browne, Alexander Bachmutsky
-
Patent number: 12132790Abstract: An architecture to perform resource management among multiple network nodes and associated resources is disclosed. Example resource management techniques include those relating to: proactive reservation of edge computing resources; deadline-driven resource allocation; speculative edge QoS pre-allocation; and automatic QoS migration across edge computing nodes.Type: GrantFiled: July 28, 2022Date of Patent: October 29, 2024Assignee: Intel CorporationInventors: Francesc Guim Bernat, Patrick Bohan, Kshitij Arun Doshi, Brinda Ganesh, Andrew J. Herdrich, Monica Kenguva, Karthik Kumar, Patrick G Kutch, Felipe Pastor Beneyto, Rashmin Patel, Suraj Prabhakaran, Ned M. Smith, Petar Torre, Alexander Vul
-
Patent number: 12026116Abstract: Examples described herein include a system comprising: a processing unit package comprising: at least one core and at least one offload processing device communicatively coupled inline between the at least one core and a network interface controller, the at least one offload processing device configurable to perform packet processing. In some examples, the at least one offload processing device is to allow mapping of packet processing pipeline stages of networking applications among software running on the at least one core and the at least one offload processing device to permit flexible entry, exit, and re-entry points among the at least one core and the at least one offload processing device.Type: GrantFiled: December 26, 2020Date of Patent: July 2, 2024Assignee: Intel CorporationInventors: Patrick G. Kutch, Andrey Chilikin, Niall D. McDonnell, Brian A. Keating, Naveen Lakkakula, Ilango S. Ganga, Venkidesh Krishna Iyer, Patrick Fleming, Lokpraveen Mosur
-
Patent number: 11983131Abstract: Examples described herein include a system comprising: a processing unit package comprising: at least one core and at least one offload processing device communicatively coupled inline between the at least one core and a network interface controller, the at least one offload processing device configurable to perform packet processing. In some examples, the at least one offload processing device is to allow mapping of packet processing pipeline stages of networking applications among software running on the at least one core and the at least one offload processing device to permit flexible entry, exit, and re-entry points among the at least one core and the at least one offload processing device.Type: GrantFiled: December 26, 2020Date of Patent: May 14, 2024Assignee: Intel CorporationInventors: Patrick G. Kutch, Andrey Chilikin, Niall D. McDonnell, Brian A. Keating, Naveen Lakkakula, Ilango S. Ganga, Venkidesh Krishna Iyer, Patrick Fleming, Lokpraveen Mosur
-
Patent number: 11934330Abstract: Examples described herein relate to an offload processor to receive data for transmission using a network interface or received in a packet by a network interface. In some examples, the offload processor can include a packet storage controller to determine whether to store data in a buffer of the offload processing device or a system memory after processing by the offload processing device. In some examples, determine whether to store data in a buffer of the offload processor or a system memory is based on one or more of: available buffer space, latency limit associated with the data, priority associated with the data, or available bandwidth through an interface between the buffer and the system memory. In some examples, the offload processor is to receive a descriptor and specify a storage location of data in the descriptor, wherein the storage location is within the buffer or the system memory.Type: GrantFiled: May 13, 2020Date of Patent: March 19, 2024Assignee: Intel CorporationInventors: Patrick G. Kutch, Andrey Chilikin
-
Patent number: 11792132Abstract: Technologies for aligning network flows to processing resources include a computing device having multiple processing nodes, a network switch, and a network controller operating in a software-defined network. Each processing node of the computing device may include a processor, memory, and network adapter. The network switch may receive a network packet and request forwarding information from the network controller. The network controller may determine flow information corresponding to the network packet that indicates the application targeted by the network packet and the processing node executing the application. The flow information may be transmitted to the computing device, which may program a flow filter in the network adapter of the processing node executing the application.Type: GrantFiled: August 24, 2020Date of Patent: October 17, 2023Assignee: Intel CorporationInventors: Iosif Gasparakis, Brian P. Johnson, Patrick G. Kutch
-
Patent number: 11792271Abstract: Examples are disclosed for remote management of a computing device. In some examples, a secure communication link may be established between a network input/output device for a computing device and a remote management application. Commands may be received from the remote management application and management functions may be implemented at the network input/output device. Implementation of the management functions may enable the remote management application to manage or control at least some operating parameters of the computing device. Other examples are described and claimed.Type: GrantFiled: May 6, 2022Date of Patent: October 17, 2023Assignee: Intel CorporationInventor: Patrick G. Kutch
-
Publication number: 20230319133Abstract: Examples described herein relate to a network interface device that includes a network interface and circuitry. In some examples, the circuitry is to receive a request to perform a service and select a servicing node based on network latency and/or proximity of the requested service to the network interface device. In some examples, a proximity of the requested service includes execution in the network interface device.Type: ApplicationFiled: June 5, 2023Publication date: October 5, 2023Inventors: Patrick G. KUTCH, Carolyn WYBORNY, Ziye YANG
-
Publication number: 20230022620Abstract: An architecture to perform resource management among multiple network nodes and associated resources is disclosed. Example resource management techniques include those relating to: proactive reservation of edge computing resources; deadline-driven resource allocation; speculative edge QoS pre-allocation; and automatic QoS migration across edge computing nodes.Type: ApplicationFiled: July 28, 2022Publication date: January 26, 2023Inventors: Francesc Guim Bernat, Patrick Bohan, Kshitij Arun Doshi, Brinda Ganesh, Andrew J. Herdrich, Monica Kenguva, Karthik Kumar, Patrick G. Kutch, Felipe Pastor Beneyto, Rashmin Patel, Suraj Prabhakaran, Ned M. Smith, Petar Torre, Alexander Vul
-
Publication number: 20220272154Abstract: Examples are disclosed for remote management of a computing device. In some examples, a secure communication link may be established between a network input/output device for a computing device and a remote management application. Commands may be received from the remote management application and management functions may be implemented at the network input/output device. Implementation of the management functions may enable the remote management application to manage or control at least some operating parameters of the computing device. Other examples are described and claimed.Type: ApplicationFiled: May 6, 2022Publication date: August 25, 2022Inventor: Patrick G. Kutch
-
Patent number: 11412052Abstract: An architecture to perform resource management among multiple network nodes and associated resources is disclosed. Example resource management techniques include those relating to: proactive reservation of edge computing resources; deadline-driven resource allocation; speculative edge QoS pre-allocation; and automatic QoS migration across edge computing nodes.Type: GrantFiled: December 28, 2018Date of Patent: August 9, 2022Assignee: Intel CorporationInventors: Francesc Guim Bernat, Patrick Bohan, Kshitij Arun Doshi, Brinda Ganesh, Andrew J. Herdrich, Monica Kenguva, Karthik Kumar, Patrick G Kutch, Felipe Pastor Beneyto, Rashmin Patel, Suraj Prabhakaran, Ned M. Smith, Petar Torre, Alexander Vul
-
Publication number: 20220222117Abstract: Examples describe techniques to expose application telemetry in a virtualized execution environment. Examples include a plurality of application executing within the virtualized execution environment writing telemetry data to a memory associated with virtual devices of a hardware device. Examples also include an orchestrator to read the telemetry data from the memory and use the telemetry data to make resource allocation decisions.Type: ApplicationFiled: March 30, 2022Publication date: July 14, 2022Inventors: Patrick G. KUTCH, Francesc GUIM BERNAT, Andrey CHILIKIN, Brian P. JOHNSON
-
Patent number: 11381646Abstract: Examples are disclosed for remote management of a computing device. In some examples, a secure communication link may be established between a network input/output device for a computing device and a remote management application. Commands may be received from the remote management application and management functions may be implemented at the network input/output device. Implementation of the management functions may enable the remote management application to manage or control at least some operating parameters of the computing device. Other examples are described and claimed.Type: GrantFiled: December 7, 2020Date of Patent: July 5, 2022Assignee: Intel CorporationInventor: Patrick G. Kutch
-
Publication number: 20220197729Abstract: An apparatus comprising a network interface controller comprising a queue for messages for a thread executing on a host computing system, wherein the queue is dedicated to the thread; and circuitry to send a notification to the host computing system to resume execution of the thread when a monitoring rule for the queue has been triggered.Type: ApplicationFiled: December 23, 2020Publication date: June 23, 2022Applicant: Intel CorporationInventors: Francesc Guim Bernat, Karthik Kumar, Patrick G. Kutch, Alexander Bachmutsky, Nicolae Octavian Popovici
-
Publication number: 20220166846Abstract: Technologies for managing telemetry and sensor data on an edge networking platform are disclosed. According to one embodiment disclosed herein, a device monitors telemetry data associated with multiple services provided in the edge networking platform. The device identifies, for each of the services and as a function of the associated telemetry data, one or more service telemetry patterns. The device generates a profile including the identified service telemetry patterns.Type: ApplicationFiled: July 30, 2021Publication date: May 26, 2022Inventors: Ramanathan Sethuraman, Timothy Verrall, Ned M. Smith, Thomas Willhalm, Brinda Ganesh, Francesc Guim Bernat, Karthik Kumar, Evan Custodio, Suraj Prabhakaran, Ignacio Astilleros Diez, Nilesh K. Jain, Ravi Iyer, Andrew J. Herdrich, Alexander Vul, Patrick G. Kutch, Kevin Bohan, Trevor Cooper
-
Publication number: 20210349820Abstract: Examples described herein relate to an offload processor to receive data for transmission using a network interface or received in a packet by a network interface. In some examples, the offload processor can include a packet storage controller to determine whether to store data in a buffer of the offload processing device or a system memory after processing by the offload processing device. In some examples, determine whether to store data in a buffer of the offload processor or a system memory is based on one or more of: available buffer space, latency limit associated with the data, priority associated with the data, or available bandwidth through an interface between the buffer and the system memory. In some examples, the offload processor is to receive a descriptor and specify a storage location of data in the descriptor, wherein the storage location is within the buffer or the system memory.Type: ApplicationFiled: May 13, 2020Publication date: November 11, 2021Inventors: Patrick G. KUTCH, Andrey CHILIKIN
-
Publication number: 20210352148Abstract: Examples are disclosed for remote management of a computing device. In some examples, a secure communication link may be established between a network input/output device for a computing device and a remote management application. Commands may be received from the remote management application and management functions may be implemented at the network input/output device. Implementation of the management functions may enable the remote management application to manage or control at least some operating parameters of the computing device. Other examples are described and claimed.Type: ApplicationFiled: December 7, 2020Publication date: November 11, 2021Applicant: Intel CorporationInventor: Patrick G. Kutch
-
Publication number: 20210326221Abstract: Examples described herein relate to a network interface device that comprises circuitry, when operational, to select a platform to execute a function and based on load of the platform, selectively cause the function to execute on one or more other platforms to attempt to achieve or finish before the time-to-completion. In some examples, the circuitry is to detect progress of function execution to determine whether completion of execution of the function is predicted to not finish within the time-to-completion and cause the function to execute on one or more other platforms based on completion of execution of the function predicted to not finish within the time-to-completion. In some examples, the circuitry is to select the one or more other platforms to execute the function based on one or more of: processor computing utilization, available memory capacity, available cache capacity, network availability, or malfunction of a processor, memory, and/or cache.Type: ApplicationFiled: June 26, 2021Publication date: October 21, 2021Inventors: Francesc GUIM BERNAT, Karthik KUMAR, Alexander BACHMUTSKY, Patrick G. KUTCH, Marcos E. CARRANZA