Patents by Inventor Norbert Egi
Norbert Egi has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 9459798Abstract: Embodiments of the present invention provide a data processing system and a data processing method. An MMIO address of a data request is obtained, where data stored in a PCIe storage device can be directly obtained from the MMIO address, and according to the MMIO address, a network adapter can directly read the data from the PCIe storage device of the data processing system, and transmit the data to a second data processing system, or directly write the data received from the second data processing system into the PCIe storage device. Therefore, the processing system can implement transmitting data directly from the PCIe storage device to the network adapter without using memory. During data transmission between the two data processing systems, a usage ratio of a resource, such as memory and a CPU, is reduced, and efficiency of data transmission is improved.Type: GrantFiled: December 11, 2014Date of Patent: October 4, 2016Assignee: Huawei Technologies Co., Ltd.Inventors: Jian He, Guangyu Shi, Xiaoke Ni, Norbert Egi, Xiancai Li, Yu Liu, Huawei Liu
-
Patent number: 9419918Abstract: The descriptions presented herein include explanation of high-dimensional PCI-Express (PCIe) network implementations. The new approaches can facilitate utilization of an efficient protocol (e.g., PCIe, etc.) while enabling implementation of various characteristics and features (e.g., characteristics and features similar to a fat-tree topology, CLOS topology, 2D and 3D topologies, etc.) that would otherwise not be compatible with the protocol. For example, implementation of alternative paths can be enabled and utilized while maintaining compliance with a protocol (e.g., PCIe, etc.) that would otherwise not be compatible with the use of alternative paths. The alternative paths can facilitate flexible topology implementation and network domain scaling while enabling improved communication latency.Type: GrantFiled: November 7, 2014Date of Patent: August 16, 2016Assignee: Futurewei Technologies, Inc.Inventors: Norbert Egi, Robert Lasater, Guangyu Shi
-
Publication number: 20160188643Abstract: Embodiments of the present invention pertain to a method and apparatus for a scalable sorting of a data set in a database on a computer system. A number of contiguous ranges spanning the data set are defined. Each individual data value of the data set is assigned to a range to which it falls into. The values in the ranges are then sorted. The sorting can be performed by different nodes in parallel. Once the sorting is completed, the results are stored in contiguous memory locations. This results the overall data set being sorted.Type: ApplicationFiled: December 31, 2014Publication date: June 30, 2016Inventors: Yan SUN, Norbert EGI, Edward CHING
-
Patent number: 9378167Abstract: A method implemented in a memory device, wherein the memory device comprises a first memory and a second memory, the method comprising receiving a direct memory access (DMA) write request from a first central processing unit (CPU) in a first computing system, wherein the DMA write request is for a plurality of bytes of data, in response to the DMA write request receiving the plurality of bytes of data from a memory in the first computing system without processing by the first CPU, and storing the plurality of bytes of data in the first memory, and upon completion of the storing, sending an interrupt message to a second CPU in a second computing system, wherein the interrupt message is configured to interrupt processing of the second CPU and initiate transfer of the plurality of bytes of data to a memory in the second computing system.Type: GrantFiled: August 19, 2013Date of Patent: June 28, 2016Assignee: Futurewei Technologies, Inc.Inventors: Norbert Egi, Guangyu Shi, Raju Joshi
-
Patent number: 9361238Abstract: Methods and apparatuses for insertion, searching, deletion, and load balancing using a hierarchical series of hash tables are described herein. The techniques disclosed provide nearly collision free or deterministic hash functions using a bitmap as a pre-filter. The hash functions have different priorities and one hashing result will be used to perform main memory access. For the hash functions, two hash bitmaps are used to store valid data and collision information. There is no collision allowed in the hash tables except for the hash table with the lowest priority. The hash tables and bitmaps may be stored in one or more caches in (e.g., a cache of a CPU, Block RAMs in FPGAs, etc.) which perform much faster than main memory.Type: GrantFiled: November 4, 2014Date of Patent: June 7, 2016Assignee: Futurewei Technologies, Inc.Inventors: Yan Sun, Norbert Egi
-
Publication number: 20160124864Abstract: Methods and apparatuses for insertion, searching, deletion, and load balancing using a hierarchical series of hash tables are described herein. The techniques disclosed provide nearly collision free or deterministic hash functions using a bitmap as a pre-filter. The hash functions have different priorities and one hashing result will be used to perform main memory access. For the hash functions, two hash bitmaps are used to store valid data and collision information. There is no collision allowed in the hash tables except for the hash table with the lowest priority. The hash tables and bitmaps may be stored in one or more caches in (e.g., a cache of a CPU, Block RAMs in FPGAs, etc.) which perform much faster than main memory.Type: ApplicationFiled: November 4, 2014Publication date: May 5, 2016Inventors: Yan SUN, Norbert EGI
-
Patent number: 9329783Abstract: Embodiments of the present invention provide a data processing system and a data processing method. An MMIO address of a data request is obtained, where data stored in a PCIe storage device can be directly obtained from the MMIO address, and according to the MMIO address, a network adapter can directly read the data from the PCIe storage device of the data processing system, and transmit the data to a second data processing system, or directly write the data received from the second data processing system into the PCIe storage device. Therefore, the processing system can implement transmitting data directly from the PCIe storage device to the network adapter without using memory. During data transmission between the two data processing systems, a usage ratio of a resource, such as memory and a CPU, is reduced, and efficiency of data transmission is improved.Type: GrantFiled: May 5, 2015Date of Patent: May 3, 2016Assignee: Huawei Technologies Co., Ltd.Inventors: Jian He, Guangyu Shi, Xiaoke Ni, Norbert Egi, Xiancai Li, Yu Liu, Huawei Liu
-
Publication number: 20160055119Abstract: A system comprising a unified interconnect network, a plurality of process memory modules, and a plurality of processor modules configured to share access to the memory modules via the unified interconnect network. Also disclosed is a method comprising communicating data between a plurality of processor modules and a plurality of shared resource pools via a unified interconnect network, wherein the communications comprise a protocol that is common to all resource pools, and wherein each resource pool comprises a plurality of resource modules each configured to perform a common function. Also disclosed is an apparatus comprising a network interface controller (NIC) module configured to receive data from a plurality of processor modules via a unified interconnect network, and provide core network connectivity to the processor modules.Type: ApplicationFiled: November 3, 2015Publication date: February 25, 2016Inventors: Norbert Egi, Guangyu Shi
-
Patent number: 9201837Abstract: A system comprising a unified interconnect network, a plurality of process memory modules, and a plurality of processor modules configured to share access to the memory modules via the unified interconnect network. Data may be communicated between a plurality of processor modules and a plurality of shared resource pools via a unified interconnect network, wherein the communications comprise a protocol that is common to all resource pools, and wherein each resource pool comprises a plurality of resource modules each configured to perform a common function. Further, a network interface controller (NIC) module may be configured to receive data from a plurality of processor modules via a unified interconnect network, and provide core network connectivity to the processor modules.Type: GrantFiled: March 13, 2013Date of Patent: December 1, 2015Assignee: Futurewei Technologies, Inc.Inventors: Norbert Egi, Guangyu Shi
-
Publication number: 20150234597Abstract: Embodiments of the present invention provide a data processing system and a data processing method. An MMIO address of a data request is obtained, where data stored in a PCIe storage device can be directly obtained from the MMIO address, and according to the MMIO address, a network adapter can directly read the data from the PCIe storage device of the data processing system, and transmit the data to a second data processing system, or directly write the data received from the second data processing system into the PCIe storage device. Therefore, the processing system can implement transmitting data directly from the PCIe storage device to the network adapter without using memory. During data transmission between the two data processing systems, a usage ratio of a resource, such as memory and a CPU, is reduced, and efficiency of data transmission is improved.Type: ApplicationFiled: May 5, 2015Publication date: August 20, 2015Inventors: Jian HE, Guangyu SHI, Xiaoke NI, Norbert EGI, Xiancai LI, Yu LIU, Huawei LIU
-
Publication number: 20150143016Abstract: An apparatus for initialization. The apparatus includes a management I/O device controller for managing initialization of a plurality of I/O devices coupled to a PCI-Express (PCIe) fabric. The management I/O device controller is configured for receiving a request to register a target interrupt register address of a first worker computing resource, wherein the target interrupt register address is associated with a first interrupt generated by a first I/O device coupled to the PCIe fabric. A mapping module of the management I/O device controller is configured for mapping the target interrupt register address to a mapped interrupt register address of a domain in which the first I/O device resides. A translating interrupt register table includes a plurality of mapped interrupt register addresses in the domain that is associated with a plurality of target interrupt register addresses of a plurality of worker computing resources.Type: ApplicationFiled: November 18, 2013Publication date: May 21, 2015Applicant: Futurewei Technologies, Inc.Inventors: Norbert EGI, Robert LASATER, Thomas BOYLE, John PETERS, Guangyu SHI
-
Publication number: 20150120969Abstract: Embodiments of the present invention provide a data processing system and a data processing method. An MMIO address of a data request is obtained, where data stored in a PCIe storage device can be directly obtained from the MMIO address, and according to the MMIO address, a network adapter can directly read the data from the PCIe storage device of the data processing system, and transmit the data to a second data processing system, or directly write the data received from the second data processing system into the PCIe storage device. Therefore, the processing system can implement transmitting data directly from the PCIe storage device to the network adapter without using memory.Type: ApplicationFiled: December 11, 2014Publication date: April 30, 2015Inventors: Jian HE, Guangyu SHI, Xiaoke NI, Norbert EGI, Xiancai LI, Yu LIU, Huawei LIU
-
Publication number: 20150052267Abstract: A method implemented in a memory device, wherein the memory device comprises a first memory and a second memory, the method comprising receiving a direct memory access (DMA) write request from a first central processing unit (CPU) in a first computing system, wherein the DMA write request is for a plurality of bytes of data, in response to the DMA write request receiving the plurality of bytes of data from a memory in the first computing system without processing by the first CPU, and storing the plurality of bytes of data in the first memory, and upon completion of the storing, sending an interrupt message to a second CPU in a second computing system, wherein the interrupt message is configured to interrupt processing of the second CPU and initiate transfer of the plurality of bytes of data to a memory in the second computing system.Type: ApplicationFiled: August 19, 2013Publication date: February 19, 2015Applicant: Futurewei Technologies, Inc.Inventors: Norbert Egi, Guangyu Shi, Raju Joshi
-
Publication number: 20150026380Abstract: A method of communicating data over a Peripheral Component Interconnect Express (PCIe) Non-Transparent Bridge (NTB) comprising transmitting a first posted write message to a remote processor via the NTB, wherein the first posted write message indicates an intent to transfer data to the remote processor, and receiving a second posted write message in response to the first posted write message, wherein the second posted write message indicates a destination address list for the data. Also disclosed is a method of communicating data over a PCIe NTB comprising transmitting a first posted write message to a remote processor via the NTB, wherein the first posted write message comprises a request to read data, and receiving a data transfer message comprising at least some of the data requested by the first posted write message.Type: ApplicationFiled: November 25, 2013Publication date: January 22, 2015Applicant: Futurewei Technologies, Inc.Inventors: Norbert Egi, Guangyu Shi
-
Publication number: 20150026385Abstract: Embodiments of the present invention provide a solution for managing inter-domain resource allocation in a Peripheral Component Interconnect-Express (PCIe) network. One processor among a plurality of link processors is elected as a management processor. The management processor obtains information about available resources of PCIe network. When a resource request from a request processor is received, the management processor allocates a resource of the available resources to the requesting processor. The management processor instructs one or more link processors to program one or more inter-domain NTBs through which the traffic between the allocated resource and the requesting processor is going to flow according to the memory address information of the allocated resource, to allow cross-domain resource access between the requesting processor and the allocated resource.Type: ApplicationFiled: June 25, 2014Publication date: January 22, 2015Applicant: FUTUREWEI TECHNOLOGIES, INC.Inventors: Norbert EGI, Robert LASATER, Thomas BOYLE, John PETERS, Guangyu SHI
-
Patent number: 8904028Abstract: Generally, this disclosure describes a scalable cluster router that includes a plurality of server-class computers interconnected together to form a router. Each server may be configured to independently schedule switching of packets to reduce the switch speed requirements on a per server basis. Each server may include a scheduler that independently load balances packet flows across servers of the cluster. Router capacity may be incrementally scaled by adding more servers, and router capacity may be increased by load balancing techniques within individual servers.Type: GrantFiled: July 19, 2010Date of Patent: December 2, 2014Assignee: Intel CorporationInventors: Gianluca Iannaccone, Sylvia Ratnasamy, Maziar Manesh, Katerina Argyraki, Byung-Gon Chun, Kevin Fall, Allan Knies, Norbert Egi, Mihai Dobrescu, Salman Baset
-
Publication number: 20140280687Abstract: A system comprising a unified interconnect network, a plurality of process memory modules, and a plurality of processor modules configured to share access to the memory modules via the unified interconnect network. Also disclosed is a method comprising communicating data between a plurality of processor modules and a plurality of shared resource pools via a unified interconnect network, wherein the communications comprise a protocol that is common to all resource pools, and wherein each resource pool comprises a plurality of resource modules each configured to perform a common function. Also disclosed is an apparatus comprising a network interface controller (NIC) module configured to receive data from a plurality of processor modules via a unified interconnect network, and provide core network connectivity to the processor modules.Type: ApplicationFiled: March 13, 2013Publication date: September 18, 2014Applicant: FUTUREWEI TECHNOLOGIES, INC.Inventors: Norbert Egi, Guangyu Shi
-
Publication number: 20140258577Abstract: A network element (NE) comprising a processor configured to receive a resource request via a Peripheral Component Interconnect (PCI) Express (PCI-e) network from a first device, wherein the first device is external to the NE, and query an access control list to determine whether the first device has permission to access a resource. The disclosure also includes an apparatus comprising a memory comprising instructions, and a processor configured to execute the instructions by allocating a resource of a shared device for use by an external device over a PCI-e network by updating a resource allocation table.Type: ApplicationFiled: March 11, 2013Publication date: September 11, 2014Applicant: FUTUREWEI TECHNOLOGIES, INC.Inventors: Norbert Egi, Raju Joshi, Guangyu Shi
-
Publication number: 20110016223Abstract: Generally, this disclosure describes a scalable cluster router that includes a plurality of server-class computers interconnected together to form a router. Each server may be configured to independently schedule switching of packets to reduce the switch speed requirements on a per server basis. Each server may include a scheduler that independently load balances packet flows across servers of the cluster. Router capacity may be incrementally scaled by adding more servers, and router capacity may be increased by load balancing techniques within individual servers.Type: ApplicationFiled: July 19, 2010Publication date: January 20, 2011Inventors: Gianluca Iannaccone, Sylvia Ratnasamy, Maziar Manesh, Katerina Argyraki, Byung-Gon Chun, Kevin Fall, Allan Knies, Norbert Egi, Mihai Dobrescu, Salman Baset