Patents by Inventor Naxin Zhang

Naxin Zhang has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 9225668
    Abstract: A method comprising advertising to a second node a total allocation of storage space of a buffer, wherein the total allocation is less than the capacity of the buffer, wherein the total allocation is partitioned into a plurality of allocations, wherein each of the plurality of allocations is advertised as being dedicated to a different packet type, and wherein a credit status for each packet type is used to manage the plurality of allocations, receiving a packet of a first packet type from the second node, and storing the packet to the buffer, wherein the space in the buffer occupied by the first packet type exceeds the advertised space for the first packet type due to the packet.
    Type: Grant
    Filed: June 17, 2013
    Date of Patent: December 29, 2015
    Assignee: Futurewei Technologies, Inc.
    Inventors: Iulin Lih, Chenghong He, Hongbo Shi, Naxin Zhang
  • Publication number: 20150234744
    Abstract: A reconfigurable cache architecture is provided. In processor design, as the density of on-chip components increases, a quantity and complexity of processing cores will increase as well. In order to take advantage of increased processing capabilities, many applications will take advantage of instruction level parallelism. The reconfigurable cache architecture provides a cache memory that in capable of being configured in a private mode and a fused mode for an associated multi-core processor. In the fused mode, individual cores of the multi-core processor can write and read data from certain cache banks of the cache memory with greater control over address routing. The cache architecture further includes control and configurability of the memory size and associativity of the cache memory itself.
    Type: Application
    Filed: February 18, 2014
    Publication date: August 20, 2015
    Applicants: National University of Singapore, Huawei Technologies Co., Ltd.
    Inventors: Mihai PRICOPI, Zhiguo GE, Yuan YAO, Tulika MITRA, Naxin ZHANG
  • Publication number: 20140052905
    Abstract: Disclosed herein is a processing network element (NE) comprising at least one receiver configured to receive a plurality of memory request messages from a plurality of memory nodes, wherein each memory request designates a source node, a destination node, and a memory location, and a plurality of response messages to the memory requests from the plurality of memory nodes, wherein each memory request designates a source node, a destination node, and a memory location, at least one transmitter configured to transmit the memory requests and memory responses to the plurality of memory nodes, and a controller coupled to the receiver and the transmitter and configured to enforce ordering such that memory requests and memory responses designating the same memory location and the same source node/destination node pair are transmitted by the transmitter in the same order received by the receiver.
    Type: Application
    Filed: August 2, 2013
    Publication date: February 20, 2014
    Applicant: Futurewei Technologies, Inc.
    Inventors: Iulin Lih, Chenghong He, Hongbo Shi, Naxin Zhang
  • Publication number: 20140052916
    Abstract: A processing network comprising a cache configured to store copies of memory data as a plurality of cache lines, a cache controller configured to receive data requests from a plurality of cache agents, and designate at least one of the cache agents as an owner of a first of the cache lines, and a directory configured to store cache ownership designations of the first cache line, and wherein the directory is encoded to support substantially simultaneous ownership of the first cache line by a plurality but less than all of the cache agents. Also disclosed is a method comprising receiving coherent transactions from a plurality of cache agents, and storing ownership designations of a plurality of cache lines by the cache agents in a directory, wherein the directory is configured to support storage of substantially simultaneous ownership designations for a plurality but less than all of the cache agents.
    Type: Application
    Filed: July 29, 2013
    Publication date: February 20, 2014
    Applicant: Futurewei Technologies, Inc.
    Inventors: Iulin Lih, Naxin Zhang, Chenghong He, Hongbo Shi
  • Publication number: 20140036919
    Abstract: A method comprising detecting at least one Quality of Service (QoS) requirement is met that indicates a very important packet (VIP) is outstanding from a source node in a multi-hop network comprising multiple nodes, sending an initiation message to an adjacent node in response to the detection that may activate a protocol in which a reserved channel is activated, and receiving the VIP via the reserved channel. Also, a method comprising receiving an initiation message from an adjacent node in a multi-hop network that comprises information identifying the VIP comprising a source node, a destination node, a packet type, wherein the initiation message activates a protocol in which a reserved channel is activated, searching for the VIP identified by the initiation message, and forwarding the VIP promptly if present via the reserved channel or forwarding an initiation message to adjacent nodes closer to the source node.
    Type: Application
    Filed: July 31, 2013
    Publication date: February 6, 2014
    Applicant: Futurewei Technologies, Inc.
    Inventors: Iulin Lih, Hongbo Shi, Chenghong He, Naxin Zhang
  • Publication number: 20140036930
    Abstract: A method comprising advertising to a second node a total allocation of storage space of a buffer, wherein the total allocation is less than the capacity of the buffer, wherein the total allocation is partitioned into a plurality of allocations, wherein each of the plurality of allocations is advertised as being dedicated to a different packet type, and wherein a credit status for each packet type is used to manage the plurality of allocations, receiving a packet of a first packet type from the second node, and storing the packet to the buffer, wherein the space in the buffer occupied by the first packet type exceeds the advertised space for the first packet type due to the packet.
    Type: Application
    Filed: June 17, 2013
    Publication date: February 6, 2014
    Inventors: Iulin Lih, Chenghong He, Hongbo Shi, Naxin Zhang
  • Publication number: 20140036680
    Abstract: A method comprising receiving a credit status from a second node comprising a plurality of credits used to manage the plurality of allocations of storage space in a buffer of the second node, wherein each of the plurality of allocations are dedicated to a different packet type, instructing the second node to use the credit dedicated to a second priority packet type for storing a first priority packet type, wherein the first priority is higher than the second priority, and wherein the credit status reflects the credits for the first priority packet type having reached a minimum value, and transmitting the first priority packet to the second node.
    Type: Application
    Filed: July 31, 2013
    Publication date: February 6, 2014
    Applicant: Futurewei Technologies, Inc.
    Inventors: Iulin Lih, Chenghong He, Hongbo Shi, Naxin Zhang
  • Publication number: 20140036929
    Abstract: A network node comprises a receiver configured to receive a first packet, a processor coupled to the receiver and configured to process the first packet, and prioritize the first packet according to a scheme, wherein the scheme assigns priority to packets based on phase, and a transmitter coupled to the processor and configured to transmit the first packet. An apparatus comprises a processor coupled to the memory and configured to generate instructions for a packet prioritization scheme, wherein the scheme assigns priority to packet transactions based on closeness to completion, and a memory coupled to the processor and configured to store the instructions. A method comprises receiving a first packet, processing the first packet, prioritizing the first packet according to a scheme, wherein the scheme assigns priority to packets based on phase, and transmitting the first packet.
    Type: Application
    Filed: July 30, 2013
    Publication date: February 6, 2014
    Applicant: Futurewei Technologies, Inc.
    Inventors: Iulin Lih, Chenghong He, Hongbo Shi, Naxin Zhang
  • Publication number: 20140040561
    Abstract: A method implemented by a computer system comprising a first memory agent and a second memory agent coupled to the first memory agent, wherein the second memory agent has access to a cache comprising a cache line, the method comprising changing a state of the cache line by the second memory agent, and sending a non-snoop message from the second memory agent to the first memory agent via a communication channel assigned to snoop responses, wherein the non-snoop message informs the first memory agent of the state change of the cache line.
    Type: Application
    Filed: May 22, 2013
    Publication date: February 6, 2014
    Inventors: Iulin Lih, Chenghong He, Hongbo Shi, Naxin Zhang
  • Publication number: 20140033209
    Abstract: A computing system for handling barrier commands includes a memory, an interface, and a processor. The memory is configured to store a pre-barrier spreading range that identifies a target computing system associated with a barrier command. The interface is coupled to the memory and is configured to send a pre-barrier computing probe to the target computing system identified in the pre-barrier spreading range and receive a barrier completion notification messages from the target computing system. The pre-barrier computing probe is configured to instruct the target computing system to monitor a status of a transaction that needs to be executed for the barrier command to be completed. The processor is coupled to the interface and is configured to determine a status of the barrier command based on the received barrier completion notification messages.
    Type: Application
    Filed: July 26, 2013
    Publication date: January 30, 2014
    Applicant: Futurewei Technologies, Inc.
    Inventors: Iulin Lih, Chenghong He, Hongbo Shi, Naxin Zhang
  • Publication number: 20140032731
    Abstract: An apparatus comprises an interconnection network comprising NK nodes, wherein N is an integer of two or greater and represents a degree of the network, wherein each node comprises N ports, wherein K is an integer of one or greater and represents a recursion level of the network, and wherein N ports are left available for recursion, and NK?1 clusters of nodes, wherein each cluster comprises N nodes, wherein each node within each cluster is directly connected to each remaining node in the cluster, and wherein each cluster is directly connected to at least one remaining cluster.
    Type: Application
    Filed: July 26, 2013
    Publication date: January 30, 2014
    Applicant: FUTUREWEI TECHNOLOGIES, INC.
    Inventors: Iulin Lih, Chenghong He, Hongbo Shi, Naxin Zhang
  • Publication number: 20140032854
    Abstract: A computer program product comprising computer executable instructions stored on a non-transitory medium that when executed by a processor cause the processor to perform the following: assign a first, second, third, and fourth coherence domain address to a cache data, wherein the first and second address provides the boundary for a first coherence domain, and wherein the third and fourth address provides the boundary for a second coherence domain, inform a first resource about the first coherence domain prior to the first resource executing a first task, and inform a second resource about the second coherence domain prior to the second resource executing a second task.
    Type: Application
    Filed: July 23, 2013
    Publication date: January 30, 2014
    Applicant: Futurewei Technologies, Inc.
    Inventors: Iulin Lih, Chenghong He, Hongbo Shi, Naxin Zhang
  • Publication number: 20140032853
    Abstract: A home node for selecting a source node using a cache coherency protocol, comprising a logic unit cluster coupled to a directory, wherein the logic unit cluster is configured to receive a request for data from a requesting cache node, determine a plurality of nodes that hold a copy of the requested data using the directory, select one of the nodes using one or more selection parameters as the source node, and transmit a message to the source node to determine whether the source node stores a copy of the requested data, wherein the source node forwards the requested data to the requesting cache node when the requested data is found within the source node, and wherein some of the nodes are marked as a Shared state corresponding to the cache coherency protocol.
    Type: Application
    Filed: June 17, 2013
    Publication date: January 30, 2014
    Inventors: Iulin Lih, Chenghong He, Hongbo Shi, Naxin Zhang