Patents by Inventor Kannan Babu Ramia
Kannan Babu Ramia has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240106900Abstract: Methods, apparatus, systems, and articles of manufacture to add a non-native node to a cluster are disclosed. An example apparatus includes programmable circuitry to at least one of instantiate a first agent to interface with a management application to obtain a request from a node to join a cluster of nodes, the first agent to interface with the management application using a first protocol; and instantiate a second agent which employs a second protocol different than the first protocol: responsive to an authentication of an identity credential of the node, obtain a secret credential; and cause the first agent to pass the secret credential to the node via the management application to enable the node to join the cluster of nodes.Type: ApplicationFiled: December 7, 2023Publication date: March 28, 2024Inventors: Kannan Babu Ramia, Palaniappan Ramanathan, Deepak S, Bhavik Dhandhalya
-
Publication number: 20220224637Abstract: Methods, apparatus, systems, and articles of manufacture are disclosed for traffic control for application-independent service mesh. In one example, processor circuitry to perform operations to instantiate ingress traffic management circuitry to receive ingress traffic events, at least one of the ingress traffic events to request access to a target microservice running on the second endpoint. The processor circuitry further performs operations to instantiate virtual service authorization circuitry to perform a look up of an authorization policy to the target microservice in the microservice catalog. Finally, the processor circuitry performs operations to instantiate endpoint selection circuitry to select the second endpoint to service the ingress traffic event in response to the authorization policy allowing access to the target microservice.Type: ApplicationFiled: April 1, 2022Publication date: July 14, 2022Inventors: Deepak S, Kannan Babu Ramia, Palaniappan Ramanathan
-
Publication number: 20220116289Abstract: Various aspects of methods, systems, and use cases include using an edge device to control power states of nodes in an edge cluster. A method may include scheduling unscheduled application pods on a set of nodes of an edge cluster. The method may include determining a change to at least one power state of the set of power states based on the schedule. The method may include causing the edge cluster to operate according to the change.Type: ApplicationFiled: December 22, 2021Publication date: April 14, 2022Inventors: Palaniappan Ramanathan, Kannan Babu Ramia, Deepak S, Bhavik Dhandhalya, Neal Conrad Oliver, Francesc Guim Bernat, Thijs Metsch, Timothy Verrall
-
Publication number: 20220117036Abstract: Methods, apparatus, systems, and articles of manufacture are disclosed to improve mobile edge platform resiliency. An example apparatus includes memory, instructions, and processor circuitry to execute the instructions to initiate a search for one of an access point or a cluster head at a first time, in response to detection of the cluster head, initiate an intra-cluster mobile edge platform (MEP) mutual authentication associated with the cluster head, cause transmission of historical activity data to at least one MEP associated with the cluster head, and cause storage of an authentication metric in a secure storage, the authentication metric based on historical connectivity data.Type: ApplicationFiled: December 20, 2021Publication date: April 14, 2022Inventors: Hassnaa Moustafa, Francesc Guim Bernat, Kannan Babu Ramia, Christian Maciocco, Kshitij Arun Doshi, Nageen Himayat, Shilpa Talwar
-
Publication number: 20210119962Abstract: Systems and methods may be used to provide neutral host edge services in an edge network. An example method may include generating a virtual machine for a communication service provider at a compute device. The method may include receiving a user packet originated at a user device associated with the communication service provider and identifying dynamic route information related to the user packet using the virtual machine corresponding to the communication service provider. Data may be output corresponding to the user packet based on the dynamic route information.Type: ApplicationFiled: December 26, 2020Publication date: April 22, 2021Inventors: Kannan Babu Ramia, Deepak S, Palaniappan Ramanathan, Timothy Verrall, Francesc Guim Bernat
-
Patent number: 10158578Abstract: One embodiment provides a network device. The network device includes a a processor including at least one processor core; a network interface configured to transmit and receive packets at a line rate; a memory configured to store a scheduler hierarchical data structure; and a scheduler module. The scheduler module is configured to prefetch a next active pipe structure, the next active pipe structure included in the hierarchical data structure, update credits for a current pipe and an associated subport, identify a next active traffic class within the current pipe based, at least in part, on a current pipe data structure, select a next queue associated with the identified next active traffic class, and schedule a next packet from the selected next queue for transmission by the network interface if available traffic shaping token bucket credits and available traffic class credits are greater than or equal to a next packet credits.Type: GrantFiled: September 19, 2016Date of Patent: December 18, 2018Assignee: Intel CorporationInventors: Cristian Florin Dumitrescu, Andrey Chilikin, Pierre Laurent, Kannan Babu Ramia, Sravanthi Tangeda
-
Patent number: 10091122Abstract: One embodiment provides a network device. The network device includes a a processor including at least one processor core; a network interface configured to transmit and receive packets at a line rate; a memory configured to store a scheduler hierarchical data structure; and a scheduler module. The scheduler module is configured to prefetch a next active pipe structure, the next active pipe structure included in the hierarchical data structure, update credits for a current pipe and an associated subport, identify a next active traffic class within the current pipe based, at least in part, on a current pipe data structure, select a next queue associated with the identified next active traffic class, and schedule a next packet from the selected next queue for transmission by the network interface if available traffic shaping token bucket credits and available traffic class credits are greater than or equal to a next packet credits.Type: GrantFiled: December 31, 2016Date of Patent: October 2, 2018Assignee: Intel CorporationInventors: Cristian Florin Dumitrescu, Andrey Chilikin, Pierre Laurent, Kannan Babu Ramia, Sravanthi Tangeda
-
Patent number: 9882814Abstract: Technologies for bridging between coarse-grained and fine-grained load balancing include a computing node of a cluster computing device and a network controller. The computing node may add a flow entry to a local flow table based on flow information received from the network controller. The computing node may transmit a multicast network packet including the flow information and next hop information to other computing nodes of the cluster device. The computing node may also add a different flow entry to the local flow table and a next hop entry to a local next hop table based on a multicast network packet received from another computing node of the cluster device. The computing node may locally process a network packet received from a remote computing device or forward the received network packet to another computing node of the cluster device based on the flow entries added to the local flow table.Type: GrantFiled: September 25, 2014Date of Patent: January 30, 2018Assignee: Intel CorporationInventors: Kannan Babu Ramia, Christian Maciocco, Sameh Gobriel, Ashok Sunder Rajan
-
Publication number: 20170149678Abstract: One embodiment provides a network device. The network device includes a a processor including at least one processor core; a network interface configured to transmit and receive packets at a line rate; a memory configured to store a scheduler hierarchical data structure; and a scheduler module. The scheduler module is configured to prefetch a next active pipe structure, the next active pipe structure included in the hierarchical data structure, update credits for a current pipe and an associated subport, identify a next active traffic class within the current pipe based, at least in part, on a current pipe data structure, select a next queue associated with the identified next active traffic class, and schedule a next packet from the selected next queue for transmission by the network interface if available traffic shaping token bucket credits and available traffic class credits are greater than or equal to a next packet credits.Type: ApplicationFiled: December 31, 2016Publication date: May 25, 2017Applicant: INTEL CORPORATIONInventors: Cristian Florin Dumitrescu, Andrey Chilikin, Pierre Laurent, Kannan Babu Ramia, Sravanthi Tangeda
-
Publication number: 20170070356Abstract: One embodiment provides a network device. The network device includes a a processor including at least one processor core; a network interface configured to transmit and receive packets at a line rate; a memory configured to store a scheduler hierarchical data structure; and a scheduler module. The scheduler module is configured to prefetch a next active pipe structure, the next active pipe structure included in the hierarchical data structure, update credits for a current pipe and an associated subport, identify a next active traffic class within the current pipe based, at least in part, on a current pipe data structure, select a next queue associated with the identified next active traffic class, and schedule a next packet from the selected next queue for transmission by the network interface if available traffic shaping token bucket credits and available traffic class credits are greater than or equal to a next packet credits.Type: ApplicationFiled: September 19, 2016Publication date: March 9, 2017Applicant: INTEL CORPORATIONInventors: Cristian Florin Dumitrescu, Andrey Chilikin, Pierre Laurent, Kannan Babu Ramia, Sravanthi Tangeda
-
Patent number: 9450881Abstract: One embodiment provides a network device. The network device includes a a processor including at least one processor core; a network interface configured to transmit and receive packets at a line rate; a memory configured to store a scheduler hierarchical data structure; and a scheduler module. The scheduler module is configured to prefetch a next active pipe structure, the next active pipe structure included in the hierarchical data structure, update credits for a current pipe and an associated subport, identify a next active traffic class within the current pipe based, at least in part, on a current pipe data structure, select a next queue associated with the identified next active traffic class, and schedule a next packet from the selected next queue for transmission by the network interface if available traffic shaping token bucket credits and available traffic class credits are greater than or equal to a next packet credits.Type: GrantFiled: December 27, 2013Date of Patent: September 20, 2016Assignee: Intel CorporationInventors: Cristian Florin Dumitrescu, Andrey Chilikin, Pierre Laurent, Kannan Babu Ramia, Sravanthi Tangeda
-
Publication number: 20160094449Abstract: Technologies for bridging between coarse-grained and fine-grained load balancing include a computing node of a cluster computing device and a network controller. The computing node may add a flow entry to a local flow table based on flow information received from the network controller. The computing node may transmit a multicast network packet including the flow information and next hop information to other computing nodes of the cluster device. The computing node may also add a different flow entry to the local flow table and a next hop entry to a local next hop table based on a multicast network packet received from another computing node of the cluster device. The computing node may locally process a network packet received from a remote computing device or forward the received network packet to another computing node of the cluster device based on the flow entries added to the local flow table.Type: ApplicationFiled: September 25, 2014Publication date: March 31, 2016Inventors: Kannan Babu Ramia, Christian Maciocco, Sameh Gobriel, Ashok Sunder Rajan
-
Publication number: 20150016266Abstract: One embodiment provides a network device. The network device includes a a processor including at least one processor core; a network interface configured to transmit and receive packets at a line rate; a memory configured to store a scheduler hierarchical data structure; and a scheduler module. The scheduler module is configured to prefetch a next active pipe structure, the next active pipe structure included in the hierarchical data structure, update credits for a current pipe and an associated subport, identify a next active traffic class within the current pipe based, at least in part, on a current pipe data structure, select a next queue associated with the identified next active traffic class, and schedule a next packet from the selected next queue for transmission by the network interface if available traffic shaping token bucket credits and available traffic class credits are greater than or equal to a next packet credits.Type: ApplicationFiled: December 27, 2013Publication date: January 15, 2015Inventors: Cristian Florin Dumitrescu, Andrey Chilikin, Pierre Laurent, Kannan Babu Ramia, Sravanthi Tangeda
-
Patent number: 7174427Abstract: A device and method for handling Multiprotocol Label Switching (MPLS) label stacks. An incoming label mapping (ILM) table is stored in a first memory. A received packet's label stack is accessed, and an entry corresponding to a top label of the stack is read from the ILM table. A number of other entries are also read from the ILM table, and these other entries are cached in a second memory.Type: GrantFiled: December 5, 2003Date of Patent: February 6, 2007Assignee: Intel CorporationInventor: Kannan Babu Ramia