Patents by Inventor Tianchan GUAN

Tianchan GUAN has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240104360
    Abstract: Near memory processing systems for graph neural network processing can include a central core coupled to one or more memory units. The memory units can include one or more controllers and a plurality of memory devices. The system can be configured for offloading aggregation, concentrate and the like operations from the central core to the controllers of the one or more memory units. The central core can sample the graph neural network and schedule memory accesses for execution by the one or more memory units. The central core can also schedule aggregation, combination or the like operations associated with one or more memory accesses for execution by the controller. The controller can access data in accordance with the data access requests from the central core. One or more computation units of the controller can also execute the aggregation, combination or the like operations associated with one or more memory access.
    Type: Application
    Filed: December 2, 2020
    Publication date: March 28, 2024
    Inventors: Tianchan GUAN, Dimin NIU, Hongzhong ZHENG, Shuangchen LI
  • Publication number: 20240094922
    Abstract: A data processing system includes a first server and a second server. The first server includes a first processor group, a first memory space and a first interface circuit. The second server includes a second processor group, a second memory space and a second interface circuit. The first memory space and the second memory space are allocated to the first processor group. The first processor group is configured to perform memory error detection to generate an error log corresponding to a memory error. When the memory error occurs in the second memory space, the first interface circuit is configured to send the error log to the second interface circuit, and the second processor group is configured to log the memory error according to the error log received by the second interface circuit. The data processing system is capable of realizing memory reliability architecture supporting operations across different servers.
    Type: Application
    Filed: December 12, 2022
    Publication date: March 21, 2024
    Inventors: DIMIN NIU, TIANCHAN GUAN, YIJIN GUAN, SHUANGCHEN LI, HONGZHONG ZHENG
  • Publication number: 20240095179
    Abstract: A memory management method of a data processing system is provided. The memory management method includes: creating a first memory zone and a second memory zone related to a first node of a first server, wherein the first server is located in the data processing system, and the first node includes a processor and a first memory; mapping the first memory zone to the first memory; and mapping the second memory zone to a second memory of a second server, wherein the second server is located in the data processing system, and the processor is configured to access the second memory of the second server through an interface circuit of the first server and through an interface circuit of the second server.
    Type: Application
    Filed: December 13, 2022
    Publication date: March 21, 2024
    Inventors: DIMIN NIU, YIJIN GUAN, TIANCHAN GUAN, SHUANGCHEN LI, HONGZHONG ZHENG
  • Publication number: 20240078036
    Abstract: A memory module can include a hybrid media controller coupled to a volatile memory, a non-volatile memory, a non-volatile memory buffer and a set of memory mapped input/output (MMIO) register. The hybrid media controller can be configured for reading and writing data to a volatile memory of a memory mapped space of a memory module. The hybrid media controller can also be configured for reading and writing bulk data to a non-volatile memory of the memory mapped space. The hybrid media controller can also be configured for reading and writing data of a random-access granularity to the non-volatile memory of the memory mapped space. The hybrid media controller can also be configured for self-indexed moving data between the non-volatile memory and the volatile memory of the memory module.
    Type: Application
    Filed: December 24, 2020
    Publication date: March 7, 2024
    Inventors: Dimin NIU, Tianchan GUAN, Hongzhong ZHENG, Shuangchen LI
  • Publication number: 20240069755
    Abstract: The present application provides a computer system, a memory expansion device and a method for use in the computer system. The computer system includes multiple hosts and multiple memory expansion devices; the memory expansion devices correspond to the hosts in a one-to-one manner. Each host includes a CPU and a memory; each memory expansion device includes a first interface and multiple second interfaces. The first interface is configured to allow each memory expansion device to communicate with the corresponding CPU via a first coherence interconnection protocol, and the second interface is configured to allow each memory expansion device to communicate with a portion of memory expansion devices via a second coherence interconnection protocol. Any two memory expansion devices communicate with each other via at least two different paths, and the number of memory expansion devices that at least one of the two paths passes through is not more than one.
    Type: Application
    Filed: December 12, 2022
    Publication date: February 29, 2024
    Inventors: YIJIN GUAN, TIANCHAN GUAN, DIMIN NIU, HONGZHONG ZHENG
  • Publication number: 20240069954
    Abstract: The present application discloses a computing system and a memory sharing method for a computing system. The computing system includes a first host, a second host, a first memory extension device and a second memory extension device. The first memory extension device is coupled to the first host. The second memory extension device is coupled to the second host and the first memory extension device. The first host accesses a memory space of a memory of the second host through the first memory extension device and the second memory extension device according to a first physical address of the first host.
    Type: Application
    Filed: May 24, 2023
    Publication date: February 29, 2024
    Inventors: TIANCHAN GUAN, DIMIN NIU, YIJIN GUAN, HONGZHONG ZHENG
  • Publication number: 20240069754
    Abstract: The present application discloses a computing system and an associated method. The computing system includes a first host, a second host, a first memory extension device and a second memory extension device. The first host includes a first memory, and the second host includes a second memory. The first host has a plurality of first memory addresses corresponding to a plurality of memory spaces of the first memory, and a plurality of second memory addresses corresponding to a plurality of memory spaces of the second memory. The first memory extension device is coupled to the first host. The second memory extension device is coupled to the second host and the first memory extension device. The first host accesses the plurality of memory spaces of the second memory through the first memory extension device and the second memory extension device.
    Type: Application
    Filed: December 12, 2022
    Publication date: February 29, 2024
    Inventors: TIANCHAN GUAN, YIJIN GUAN, DIMIN NIU, HONGZHONG ZHENG
  • Publication number: 20240063200
    Abstract: The present disclosure relates to a hybrid bonding based integrated circuit (HBIC) device and its manufacturing method. In some embodiments, an exemplary HBIC device includes: a first die stack comprising one or more dies; and a second die stack integrated above the first die stack. The second die stack includes at least two memory dies communicatively connected to the first die stack by wire bonding.
    Type: Application
    Filed: February 6, 2020
    Publication date: February 22, 2024
    Inventors: Dimin NIU, Shuangchen LI, Tianchan GUAN, Hongzhong ZHENG
  • Patent number: 11886352
    Abstract: This specification describes methods and systems for accelerating attribute data access for graph neural network (GNN) processing. An example method includes: receiving a root node identifier corresponding to a node in a graph for GNN processing; determining one or more candidate node identifiers according to the root node identifier, wherein attribute data corresponding to the one or more candidate node identifiers are sequentially stored in a memory; and sampling one or more graph node identifiers at least from the one or more candidate node identifiers for the GNN processing.
    Type: Grant
    Filed: January 21, 2022
    Date of Patent: January 30, 2024
    Assignee: T-Head (Shanghai) Semiconductor Co., Ltd.
    Inventors: Heng Liu, Tianchan Guan, Shuangchen Li, Hongzhong Zheng
  • Publication number: 20240028554
    Abstract: A configurable processing unit including a core processing element and a plurality of assist processing elements can be coupled together by one or more networks. The core processing element can include a large processing logic, large non-volatile memory, input/output interfaces and multiple memory channels. The plurality of assist processing elements can each include smaller processing logic, smaller non-volatile memory and multiple memory channels. One or more bitstreams can be utilized to configure and reconfigure computation resources of the core processing element and memory management of the plurality of assist processing elements.
    Type: Application
    Filed: September 18, 2020
    Publication date: January 25, 2024
    Inventors: Dimin NIU, Tianchan GUAN, Hongzhong ZHENG, Shuangchen LI
  • Publication number: 20240020194
    Abstract: A system-in-package including a logic die and one or more memory dice can include a reliability availability serviceability (RAS) memory management unit (MMU) for memory error detection, memory error prediction and memory error handling. The RAS MMU can receive memory health information, on-die memory error information, system error information and read address information for the one or more memory dice. The RAS MMU can manage the memory blocks of the one or more memory dice based on the memory health information, on-die memory error type, system error type and read address. The RAS MMU can also further manage the memory blocks based on received on-die memory temperature information and or system temperature information.
    Type: Application
    Filed: November 4, 2020
    Publication date: January 18, 2024
    Inventors: Dimin NIU, Tianchan GUAN, Hongzhong ZHENG, Shuangchen LI
  • Patent number: 11841799
    Abstract: This application describes a hardware accelerator, a computer system, and a method for accelerating Graph Neural Network (GNN) node attribute fetching. The hardware accelerator comprises a GNN attribute processor; and a first memory, wherein the GNN attribute processor is configured to: receive a graph node identifier; determine a target memory address within the first memory based on the graph node identifier; determine, based on the received graph node identifier, whether attribute data corresponding to the received graph node identifier is cached in the first memory at the target memory address; and in response to determining that the attribute data is not cached in the first memory: fetch the attribute data from a second memory, and write the fetched attribute data into the first memory at the target memory address.
    Type: Grant
    Filed: January 21, 2022
    Date of Patent: December 12, 2023
    Assignee: T-Head (Shanghai) Semiconductor Co., Ltd.
    Inventors: Tianchan Guan, Heng Liu, Shuangchen Li, Hongzhong Zheng
  • Patent number: 11836188
    Abstract: A programmable device receives commands from a processor and, based on the commands: identifies a root node in a graph; identifies nodes in the graph that are neighbors of the root node; identifies nodes in the graph that are neighbors of the neighbors; retrieves data associated with the root node; retrieves data associated with at least a subset of the nodes that are neighbors of the root node and that are neighbors of the neighbor nodes; and writes the data that is retrieved into a memory.
    Type: Grant
    Filed: January 21, 2022
    Date of Patent: December 5, 2023
    Assignee: Alibaba Damo (Hangzhou) Technology Co., Ltd
    Inventors: Shuangchen Li, Tianchan Guan, Zhe Zhang, Heng Liu, Wei Han, Dimin Niu, Hongzhong Zheng
  • Publication number: 20230326905
    Abstract: Aspects of the present technology are directed toward three-dimensional (3D) stacked processing systems characterized by high memory capacity, high memory bandwidth, low power consumption and small form factor. The 3D stacked processing systems include a plurality of processor chiplets and input/output circuits directly coupled to each of the plurality of processor chiplets.
    Type: Application
    Filed: September 17, 2020
    Publication date: October 12, 2023
    Inventors: Dimin NIU, Wei HAN, Tianchan GUAN, Yuhao WANG, Shuangchen LI, Hongzhong ZHENG
  • Publication number: 20230281124
    Abstract: Apparatus, method, and system provided herein are directed to prioritizing cache line writing of compressed data. The memory controller comprises a cache line compression engine that receives raw data, compresses the raw data, determines a compression rate between the raw data and the compressed data, determines whether the compression rate is greater than a predetermined rate, and outputs the compressed data as data-to-be-written if the compression rate is greater than the predetermined rate. In response to determining that the compression rate is greater than the predetermined rate, the cache line compression engine generates a compression signal indicating the data-to-be-written is the compressed data and sends the compression signal to a scheduler of a command queue in the memory controller where writing of compressed data is prioritized.
    Type: Application
    Filed: August 6, 2020
    Publication date: September 7, 2023
    Inventors: Dimin Niu, Tianchan Guan, Lide Duan, Hongzhong Zheng
  • Publication number: 20230245711
    Abstract: The present invention provides systems and methods for efficiently and effectively priming and initializing a memory. In one embodiment, a memory controller includes a normal data path and a priming path. The normal data path directs storage operations during a normal memory read/write operation after power startup of a memory chip. The priming path includes a priming module, wherein the priming module directs memory priming operations during a power startup of the memory chip, including forwarding a priming pattern for storage in a write pattern mode register of a memory chip and selection of a memory address in the memory chip for initialization with the priming pattern. The priming pattern includes information corresponding to proper initial data values. The priming pattern can also include proper corresponding error correction code (ECC) values. The priming module can include a priming pattern register that stores the priming pattern.
    Type: Application
    Filed: January 19, 2021
    Publication date: August 3, 2023
    Inventors: Dimin NIU, Shuangchen LI, Tianchan GUAN, Hongzhong ZHENG
  • Publication number: 20230153568
    Abstract: This application describes an accelerator, a computer system, and a method for accelerating Graph Neural Network (GNN) node attribute fetching. The accelerator comprises a graph structure processor configured to obtain, according to a root node in a graph, a plurality of candidate nodes for GNN processing on the root node; a GNN sampler configured to generate sampled graph nodes based on the plurality of candidate nodes from the graph structure processor; and a GNN attribute processor configured to fetch attribute data of the sampled graph nodes received from the GNN sampler for the GNN processing on the root node.
    Type: Application
    Filed: January 12, 2022
    Publication date: May 18, 2023
    Inventors: Tianchan GUAN, Shuangchen LI, Heng LIU, Hongzhong ZHENG
  • Publication number: 20230153570
    Abstract: The present application discloses a computing system for implementing an artificial neural network model. The artificial neural network model has a structure of multiple layers. The computing system comprises a first processing unit, a second processing unit, and a third processing unit. The first processing unit performs computations of the first layer based on a first part of input data of the first layer to generate a first part of output data. The second processing unit performs computations of the first layer based on a second part of the input data of the first layer so as to generate a second part of the output data. The third processing unit performs computations of the second layer based on the first part and the second part of the output data. The first processing unit, the second processing unit, and the third processing unit have the same structure.
    Type: Application
    Filed: March 18, 2022
    Publication date: May 18, 2023
    Inventors: TIANCHAN GUAN, SHENGCHENG WANG, DIMIN NIU, HONGZHONG ZHENG
  • Publication number: 20230153250
    Abstract: This specification describes methods and systems for accelerating attribute data access for graph neural network (GNN) processing. An example method includes: receiving a root node identifier corresponding to a node in a graph for GNN processing; determining one or more candidate node identifiers according to the root node identifier, wherein attribute data corresponding to the one or more candidate node identifiers are sequentially stored in a memory; and sampling one or more graph node identifiers at least from the one or more candidate node identifiers for the GNN processing.
    Type: Application
    Filed: January 21, 2022
    Publication date: May 18, 2023
    Inventors: Heng LIU, Tianchan GUAN, Shuangchen LI, Hongzhong ZHENG
  • Publication number: 20230142254
    Abstract: The present application discloses a graph neural network processing method and associated machine and system. The graph neural network method is used for a master, wherein the master, a first worker and a second worker train the graph neural network in a distributed environment. The method includes: receiving a first request from the first worker and a second request from the second worker, wherein the first worker sends the first request to the master to obtain at least an attribute of a first requested node, and the second worker sends a second request to the master to obtain at least an attribute of a second requested node; determining whether the first requested node and the second requested node are the same nodes and generating a determination result accordingly; and selectively performing broadcast or unicast to the first worker and the second worker, at least based on the determination result.
    Type: Application
    Filed: January 25, 2022
    Publication date: May 11, 2023
    Inventors: YANHONG WANG, TIANCHAN GUAN, SHUANGCHEN LI, HONGZHONG ZHENG