Patents by Inventor Mario Flajslik

Mario Flajslik has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20180006946
    Abstract: Technologies for adaptive routing based on network traffic pattern characterization include a network switch configured to receive a network packet via one of a plurality of input ports and identify a set of the plurality of output ports associated with a path usable to forward the received network packet to a destination computing device along. The network switch is further configured to adjust a total congestion value for each of the set of output ports based on a type of the path to which each of the set of output ports corresponds and a value of a minimal path counter to which each of the set of output ports corresponds and enqueue the received network packet into an output buffer queue of one of the set of output ports based on the total congestion value. Other embodiments are described herein.
    Type: Application
    Filed: July 1, 2016
    Publication date: January 4, 2018
    Inventors: Mario Flajslik, Eric R. Borch, Michael A. Parker
  • Publication number: 20180006945
    Abstract: Technologies for adaptive routing using throughput estimation that includes a network switch. The network switch is configured to determine an adjusted average saturation count for each output buffer queue as a function of a present value of a saturation counter of a corresponding output buffer queue and a weighted average saturation count and a running average saturation count for each of the plurality of output buffer queues as a function of the corresponding captured present value and the adjusted average saturation count. The network switch is further configured to determine a congestion rate value for each output buffer queue and a total congestion value as a function of the congestion rate values and a standard occupancy congestion corresponding to a respective one of the plurality of output buffer queues. Other embodiments are described herein.
    Type: Application
    Filed: July 1, 2016
    Publication date: January 4, 2018
    Inventors: Mario Flajslik, Eric R. Borch, Michael A. Parker, Scott S. Diesing
  • Publication number: 20180006950
    Abstract: Technologies for adaptive routing based on aggregated congestion information include a network switch that includes a plurality of output ports. The network switch is configured to determine a maximum local occupancy count for each output port based on a maximum local occupancy count of output buffer queues of each output port, a local congestion value based on the maximum local occupancy count, and a remote congestion value for a corresponding remote input buffer queue of a remote computing device communicatively coupled to a corresponding output port. The network switch is further configured to determine, for each output port, a total congestion value as a function of the local congestion value and the remote congestion value and enqueue the network packet into one of the output buffer queues of one of the output ports based on the total congestion values of the output ports. Other embodiments are described herein.
    Type: Application
    Filed: July 1, 2016
    Publication date: January 4, 2018
    Inventors: Mario Flajslik, Eric R. Borch, Michael A. Parker
  • Publication number: 20170289242
    Abstract: Technologies for dynamic work queue management include a producer computing device communicatively coupled to a consumer computing device. The consumer computing device is configured to transmit a pop request (e.g., a one-sided pull request) that includes consumption constraints indicating an amount of work (e.g., a range of acceptable fraction of work elements to return from a work queue of the producer computing device) to pull from the producer computing device. The producer computing device is configured to determine whether the pop request can be satisfied and generate a response that includes an indication of the result of the determination and one or more producer metrics usable by the consumer computing device to determine a subsequent action to be performed by the consumer computing device upon receipt of the response message. Other embodiments are described and claimed herein.
    Type: Application
    Filed: March 31, 2016
    Publication date: October 5, 2017
    Inventors: David Keppel, Ulf R. Hanebutte, Mario Flajslik, James Dinan
  • Publication number: 20170185561
    Abstract: In an example, there is disclosed a compute node, comprising: first one or more logic elements comprising a data producer engine to produce a datum; and a host fabric interface to communicatively couple the compute node to a fabric, the host fabric interface comprising second one or more logic elements comprising a data pulling engine, the data pulling engine to: publish the datum as available; receive a pull request for the datum, the pull request comprising a node identifier for a data consumer; and send the datum to the data consumer via the fabric. There is also disclosed a method of providing a data pulling engine.
    Type: Application
    Filed: December 24, 2015
    Publication date: June 29, 2017
    Applicant: Intel Corporation
    Inventors: James Dinan, Mario Flajslik, Keith Underwood, David Keppel, Ulf Rainer Hanebutte
  • Publication number: 20170090979
    Abstract: Technologies for integrated thread scheduling include a computing device having a network interface controller (NIC). The NIC is configured to detect and suspend a thread that is being blocked by one or more communication operations. A thread scheduling engine of the NIC is configured to move the suspended thread from a running queue of the system thread scheduler to a pending queue of the thread scheduling engine. The thread scheduling engine is further configured to move the suspended thread from the pending queue to a ready queue of the thread scheduling engine upon determining any dependencies and/or blocking communications operations have completed. Other embodiments are described and claimed.
    Type: Application
    Filed: September 25, 2015
    Publication date: March 30, 2017
    Inventors: James Dinan, Mario Flajslik, Tom St. John
  • Publication number: 20170093731
    Abstract: Technologies for estimating network round-trip times include a sender computing node in network communication with a set of neighboring computing nodes. The sender computing node is configured to determine the set of neighboring computing nodes, as well as a plurality of subsets of the set of neighboring computing nodes. Accordingly, the sender computing node generates a message queue for each of the plurality of subsets, each message queue including a probe message for each neighboring node in the subset to which the message queue corresponds. The sender computing node is further configured to determine a round-trip time for each message queue (i.e., subset of neighboring computing nodes) based on a duration of time between the first probe message of the message queue being transmitted and an acknowledgment being received in response to the last probe message of the message queue being transmitted.
    Type: Application
    Filed: September 25, 2015
    Publication date: March 30, 2017
    Inventors: Mario Flajslik, James Dinan
  • Publication number: 20170093770
    Abstract: Technologies for filtering a received message include a receiving computing device to receive messages and a sender computing device to send messages. The receiving computing device is configured to retrieve a descriptor from a received message and retrieve another descriptor from an inspection entry of a network port entry selected from a network port entry table by the receiving computing device based on the logical network port that received the message. The receiving computing device is further configured to compare the descriptors to determine whether the descriptors match. Upon finding a match, the receiving computing device is still further configured to perform an operation corresponding to the inspection entries whose descriptor matches the descriptor of the message. Other embodiments are described and claimed.
    Type: Application
    Filed: September 25, 2015
    Publication date: March 30, 2017
    Inventors: Mario Flajslik, James Dinan, Keith D. Underwood
  • Publication number: 20170085442
    Abstract: Technologies for aggregation-based message processing include multiple computing nodes in communication over a network. A computing node receives a message from a remote computing node, increments an event counter in response to receiving the message, determines whether an event trigger is satisfied in response to incrementing the counter, and writes a completion event to an event queue if the event trigger is satisfied. An application of the computing node monitors the event queue for the completion event. The application may be executed by a processor core of the computing node, and the other operations may be performed by a host fabric interface of the computing node. The computing node may be a target node and count one-sided messages received from an initiator node, or the computing node may be an initiator node and count acknowledgement messages received from a target node. Other embodiments are described and claimed.
    Type: Application
    Filed: September 23, 2015
    Publication date: March 23, 2017
    Inventors: James Dinan, Mario Flajslik, David Keppel, Ulf R. Hanebutte
  • Publication number: 20170085625
    Abstract: Technologies for handling message passing interface receive operations include a compute node to determine a plurality of parameters of a receive entry to be posted and determine whether the plurality of parameters includes a wildcard entry. The compute node generates a hash based on at least one parameter of the plurality of parameters in response to determining that the plurality of parameters does not include the wildcard entry and appends the receive entry to a list in a bin of a posted receive data structure, wherein the bin is determined based on the generated hash. The compute node further tracks the wildcard entry in the posted receive data structure in response to determining the plurality of parameters includes the wildcard entry and appends the receive entry to a wildcard list of the posted receive data structure in response to tracking the wildcard entry.
    Type: Application
    Filed: September 18, 2015
    Publication date: March 23, 2017
    Inventors: James Dinan, Mario Flajslik, Keith D. Underwood
  • Publication number: 20160381120
    Abstract: This disclosure is directed to a system for event dissemination. In general, a system may comprise a plurality of devices each including an event dissemination module (EDM) configured to disseminate events between the plurality of devices. New events may be generated during the normal course of operation in each of the plurality of devices. These events may be provided to at least one device designated as a network dispatch location. The network dispatch location may initiate the dissemination of the events. For example, each device may place received events into a local event queue within the device. The placement of an event into the local event queue may cause a counter in the EDM to increment. Incrementing the counter may, in turn, cause a trigger operation module in the EDM to perform at least one activity including, for example, forwarding the event to other devices within the plurality of devices.
    Type: Application
    Filed: June 24, 2015
    Publication date: December 29, 2016
    Applicant: INTEL CORPORATION
    Inventors: MARIO FLAJSLIK, JAMES DINAN, KEITH UNDERWOOD
  • Publication number: 20160314073
    Abstract: Technologies for one-side remote memory access communication include multiple computing nodes in communication over a network. A receiver computing node receives a message from a sender node and extracts a segment identifier from the message. The receiver computing node determines, based on the segment identifier, a segment start address associated with a partitioned global address space (PGAS) segment of its local memory. The receiver computing node may index a segment table stored in the local memory or in a host fabric interface. The receiver computing node determines a local destination address within the PGAS segment based on the segment start address and an offset included in the message. The receiver computing node performs a remote memory access operation at the local destination address. The receiver computing node may perform those operations in hardware by the host fabric interface of the receiver computing node. Other embodiments are described and claimed.
    Type: Application
    Filed: April 27, 2015
    Publication date: October 27, 2016
    Inventors: James Dinan, Mario Flajslik
  • Publication number: 20160283295
    Abstract: Systems, apparatuses and methods may provide for detecting an outbound communication and identifying a context of the outbound communication. Additionally, a completion status of the outbound communication may be tracked relative to the context. In one example, tracking the completion status includes incrementing a sent messages counter associated with the context in response to the outbound communication, detecting an acknowledgement of the outbound communication based on a network response to the outbound communication, incrementing a received acknowledgements counter associated with the context in response to the acknowledgement, comparing the sent messages counter to the received acknowledgements counter, and triggering a per-context memory ordering operation if the sent messages counter and the received acknowledgements counter have matching values.
    Type: Application
    Filed: March 27, 2015
    Publication date: September 29, 2016
    Applicant: INTEL CORPORATION
    Inventors: Mario Flajslik, James Dinan
  • Publication number: 20160100010
    Abstract: Generally, this disclosure provides systems, devices, methods and computer readable media for improved coordination between sender and receiver nodes in a one-sided memory access to a PGAS in a distributed computing environment. The system may include a transceiver module configured to receive a message over a network, the message comprising a data portion and a data size indicator and an offset handler module configured to calculate a destination address from a base address of a memory buffer and an offset counter. The transceiver module may further be configured to write the data portion to the memory buffer at the destination address; and the offset handler module may further be configured to update the offset counter based on the data size indicator.
    Type: Application
    Filed: December 24, 2014
    Publication date: April 7, 2016
    Applicant: Intel Corporation
    Inventors: Mario Flajslik, James Dinan