Patents by Inventor Dimin Niu

Dimin Niu has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240004955
    Abstract: This application describes an accelerator, a computer system, and a method for memory optimization in sparse matrix-matrix multiplications (spGEMM). The memory optimization includes accurate memory pre-allocation for a to-be-generated output matrix of spGEMM between two sparse matrices. An exemplary method may include: sampling a plurality of first rows in the first sparse matrix; identifying, based on indices of non-zero data in the plurality of first rows, a plurality of second rows in a second sparse matrix; performing symbolic multiplication operations between the non-zero data in the plurality of first and second rows; determining an estimated compression ratio of the output matrix; determining an estimated mean row size for each row in the output matrix based on the estimated compression ratio; and allocating, according to the estimated mean row size and a total number of rows of the output matrix, a memory space in a hardware memory.
    Type: Application
    Filed: November 9, 2022
    Publication date: January 4, 2024
    Inventors: Zhaoyang DU, Yijin GUAN, Dimin NIU, Hongzhong ZHENG
  • Publication number: 20240005075
    Abstract: This application describes systems and methods for facilitating memory access for graph neural network (GNN) processing. An example method includes fetching, by an access engine circuitry implemented on a circuitry board, a portion of structure data of a graph from one or more of a plurality of flash memory drives implemented on the circuitry board; performing node sampling using the fetched portion of the structure data of the graph to select one or more sampled nodes; fetching a portion of attribute data of the graph from two or more of the plurality of memory drives in parallel according to the selected one or more sampled nodes; sending the fetched portion of the attribute data of the graph to a host outside of the circuitry board; and performing, by the host, GNN processing for the graph using the fetched portion of the attribute data of the graph.
    Type: Application
    Filed: November 30, 2022
    Publication date: January 4, 2024
    Inventors: Shuangchen LI, Dimin NIU, Hongzhong ZHENG
  • Patent number: 11847049
    Abstract: The total memory space that is logically available to a processor in a general-purpose graphics processing unit (GPGPU) module is increased to accommodate terabyte-sized amounts of data by utilizing the memory space in an external memory module, and by further utilizing a portion of the memory space in a number of other external memory modules.
    Type: Grant
    Filed: January 21, 2022
    Date of Patent: December 19, 2023
    Assignee: Alibaba Damo (Hangzhou) Technology Co., Ltd
    Inventors: Yuhao Wang, Dimin Niu, Yijin Guan, Shengcheng Wang, Shuangchen Li, Hongzhong Zheng
  • Patent number: 11836188
    Abstract: A programmable device receives commands from a processor and, based on the commands: identifies a root node in a graph; identifies nodes in the graph that are neighbors of the root node; identifies nodes in the graph that are neighbors of the neighbors; retrieves data associated with the root node; retrieves data associated with at least a subset of the nodes that are neighbors of the root node and that are neighbors of the neighbor nodes; and writes the data that is retrieved into a memory.
    Type: Grant
    Filed: January 21, 2022
    Date of Patent: December 5, 2023
    Assignee: Alibaba Damo (Hangzhou) Technology Co., Ltd
    Inventors: Shuangchen Li, Tianchan Guan, Zhe Zhang, Heng Liu, Wei Han, Dimin Niu, Hongzhong Zheng
  • Patent number: 11789610
    Abstract: A 3D-stacked memory device including: a base die including a plurality of switches to direct data flow and a plurality of arithmetic logic units (ALUs) to compute data; a plurality of memory dies stacked on the base die; and an interface to transfer signals to control the base die.
    Type: Grant
    Filed: June 21, 2021
    Date of Patent: October 17, 2023
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Mu-Tien Chang, Prasun Gera, Dimin Niu, Hongzhong Zheng
  • Publication number: 20230326905
    Abstract: Aspects of the present technology are directed toward three-dimensional (3D) stacked processing systems characterized by high memory capacity, high memory bandwidth, low power consumption and small form factor. The 3D stacked processing systems include a plurality of processor chiplets and input/output circuits directly coupled to each of the plurality of processor chiplets.
    Type: Application
    Filed: September 17, 2020
    Publication date: October 12, 2023
    Inventors: Dimin NIU, Wei HAN, Tianchan GUAN, Yuhao WANG, Shuangchen LI, Hongzhong ZHENG
  • Publication number: 20230281124
    Abstract: Apparatus, method, and system provided herein are directed to prioritizing cache line writing of compressed data. The memory controller comprises a cache line compression engine that receives raw data, compresses the raw data, determines a compression rate between the raw data and the compressed data, determines whether the compression rate is greater than a predetermined rate, and outputs the compressed data as data-to-be-written if the compression rate is greater than the predetermined rate. In response to determining that the compression rate is greater than the predetermined rate, the cache line compression engine generates a compression signal indicating the data-to-be-written is the compressed data and sends the compression signal to a scheduler of a command queue in the memory controller where writing of compressed data is prioritized.
    Type: Application
    Filed: August 6, 2020
    Publication date: September 7, 2023
    Inventors: Dimin Niu, Tianchan Guan, Lide Duan, Hongzhong Zheng
  • Patent number: 11729268
    Abstract: Various embodiments of the present disclosure relate to a computer-implemented method, a system, and a storage medium, where a graph stored in a computing system is logically divided into subgraphs, the subgraphs are stored on different interconnected (or coupled) devices in the computing system, and nodes of the subgraphs include hub nodes connected to adjacent subgraphs. Each device stores attributes and node structure information of the hub nodes of the subgraphs into other devices, and software or hardware prefetch engine on the device prefetches attributes and node structure information associated with a sampled node. A prefetcher on a device interfacing with the interconnected (or coupled) devices may further prefetch attributes and node structure information of nodes of the subgraphs on other devices. A traffic monitor is provided on an interface device to monitor traffic. When the traffic is small, the interface device prefetches node attributes and node structure information.
    Type: Grant
    Filed: June 8, 2022
    Date of Patent: August 15, 2023
    Assignee: Alibaba (China) Co., Ltd.
    Inventors: Wei Han, Shuangcheng Li, Hongzhong Zheng, Yawen Zhang, Heng Liu, Dimin Niu
  • Publication number: 20230245711
    Abstract: The present invention provides systems and methods for efficiently and effectively priming and initializing a memory. In one embodiment, a memory controller includes a normal data path and a priming path. The normal data path directs storage operations during a normal memory read/write operation after power startup of a memory chip. The priming path includes a priming module, wherein the priming module directs memory priming operations during a power startup of the memory chip, including forwarding a priming pattern for storage in a write pattern mode register of a memory chip and selection of a memory address in the memory chip for initialization with the priming pattern. The priming pattern includes information corresponding to proper initial data values. The priming pattern can also include proper corresponding error correction code (ECC) values. The priming module can include a priming pattern register that stores the priming pattern.
    Type: Application
    Filed: January 19, 2021
    Publication date: August 3, 2023
    Inventors: Dimin NIU, Shuangchen LI, Tianchan GUAN, Hongzhong ZHENG
  • Publication number: 20230229555
    Abstract: A method of correcting a memory error of a dynamic random-access memory module (DRAM) using a double data rate (DDR) interface, the method includes conducting a memory transaction including multiple bursts with a memory controller to send data from data chips of the DRAM to the memory controller, detecting one or more errors using an ECC chip of the DRAM, determining a number of the bursts having the errors using the ECC chip of the DRAM, determining whether the number of the bursts having the errors is greater than a threshold number, determining a type of the errors, and directing the memory controller based on the determined type of the errors, wherein the DRAM includes a single ECC chip per memory channel.
    Type: Application
    Filed: March 28, 2023
    Publication date: July 20, 2023
    Inventors: Dimin NIU, Mu-Tien CHANG, Hongzhong ZHENG, Hyun-Joong KIM, Won-hyung SONG, Jangseok CHOI
  • Patent number: 11658168
    Abstract: A flash memory device includes a plurality of flash memory cell arrays, wherein: a flash memory cell array in the plurality of flash memory cell arrays comprises a plurality of layers of flash memory cell planes; and a flash memory cell plane includes a plurality of flash memory cells. The flash memory device further includes a logic circuitry coupled to the plurality of flash memory cell arrays, configured to perform operations using the plurality of flash memory cell arrays; and a sensing circuitry configured to access a corresponding flash memory cell plane among the plurality of flash memory cell planes.
    Type: Grant
    Filed: August 5, 2020
    Date of Patent: May 23, 2023
    Inventors: Fei Xue, Shuangchen Li, Dimin Niu, Hongzhong Zheng
  • Publication number: 20230153570
    Abstract: The present application discloses a computing system for implementing an artificial neural network model. The artificial neural network model has a structure of multiple layers. The computing system comprises a first processing unit, a second processing unit, and a third processing unit. The first processing unit performs computations of the first layer based on a first part of input data of the first layer to generate a first part of output data. The second processing unit performs computations of the first layer based on a second part of the input data of the first layer so as to generate a second part of the output data. The third processing unit performs computations of the second layer based on the first part and the second part of the output data. The first processing unit, the second processing unit, and the third processing unit have the same structure.
    Type: Application
    Filed: March 18, 2022
    Publication date: May 18, 2023
    Inventors: TIANCHAN GUAN, SHENGCHENG WANG, DIMIN NIU, HONGZHONG ZHENG
  • Publication number: 20230144693
    Abstract: The total memory space that is logically available to a processor in a general-purpose graphics processing unit (GPGPU) module is increased to accommodate terabyte-sized amounts of data by utilizing the memory space in an external memory module, and by further utilizing a portion of the memory space in a number of other external memory modules.
    Type: Application
    Filed: January 21, 2022
    Publication date: May 11, 2023
    Inventors: Yuhao WANG, Dimin NIU, Yijin GUAN, Shengcheng WANG, Shuangchen LI, Hongzhong ZHENG
  • Publication number: 20230137162
    Abstract: A programmable device receives commands from a processor and, based on the commands: identifies a root node in a graph; identifies nodes in the graph that are neighbors of the root node; identifies nodes in the graph that are neighbors of the neighbors; retrieves data associated with the root node; retrieves data associated with at least a subset of the nodes that are neighbors of the root node and that are neighbors of the neighbor nodes; and writes the data that is retrieved into a memory.
    Type: Application
    Filed: January 21, 2022
    Publication date: May 4, 2023
    Inventors: Shuangchen LI, Tianchan GUAN, Zhe ZHANG, Heng LIU, Wei HAN, Dimin NIU, Hongzhong ZHENG
  • Publication number: 20230119291
    Abstract: A method of processing in-memory commands in a high-bandwidth memory (HBM) system includes sending a function-in-HBM instruction to the HBM by a HBM memory controller of a GPU. A logic component of the HBM receives the FIM instruction and coordinates the instructions execution using the controller, an ALU, and a SRAM located on the logic component.
    Type: Application
    Filed: December 14, 2022
    Publication date: April 20, 2023
    Inventors: Mu-Tien Chang, Krishna T. Malladi, Dimin Niu, Hongzhong Zheng
  • Patent number: 11625341
    Abstract: The systems and methods are configured to efficiently and effectively access memory. In one embodiment, a memory controller comprises a request queue, a buffer, a control component, and a data path system. The request queue receives memory access requests. The control component is configured to process information associated with access requests via a first narrow memory channel and a second narrow memory channel. The first narrow memory channel and the second narrow memory channel can have a portion of command/control communication lines and address communication lines that are included in and shared between the first narrow memory channel and the second narrow memory channel. The data path system can include a first data module and one set of unshared data lines associated with the first memory channel and a second data module and another set of unshared data lines associated with second memory channel.
    Type: Grant
    Filed: August 11, 2020
    Date of Patent: April 11, 2023
    Assignee: Alibaba Group Holding Limited
    Inventors: Jilan Lin, Dimin Niu, Shuangchen Li, Hongzhong Zheng, Yuan Xie
  • Patent number: 11625296
    Abstract: A method of correcting a memory error of a dynamic random-access memory module (DRAM) using a double data rate (DDR) interface, the method includes conducting a memory transaction including multiple bursts with a memory controller to send data from data chips of the DRAM to the memory controller, detecting one or more errors using an ECC chip of the DRAM, determining a number of the bursts having the errors using the ECC chip of the DRAM, determining whether the number of the bursts having the errors is greater than a threshold number, determining a type of the errors, and directing the memory controller based on the determined type of the errors, wherein the DRAM includes a single ECC chip per memory channel.
    Type: Grant
    Filed: May 13, 2021
    Date of Patent: April 11, 2023
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Dimin Niu, Mu-Tien Chang, Hongzhong Zheng, Hyun-Joong Kim, Won-hyung Song, Jangseok Choi
  • Publication number: 20230109617
    Abstract: A system for pruning weights during training of a neural network includes a configurable pruning hardware unit that is configured to: receive, from a neural network training engine, inputs including the weights, gradients associated with the weights, and a prune indicator per weight; select unpruned weights for pruning; prune the unpruned weights selected for pruning; update the prune indicator per weight for the weights that are selected and pruned; and provide the updated prune indicator to the training engine for the next iteration or epoch. The pruning hardware unit can be configured to perform incremental pruning or non-incremental pruning.
    Type: Application
    Filed: December 9, 2022
    Publication date: April 6, 2023
    Inventors: Tianchan GUAN, Yuan GAO, Hongzhong ZHENG, Minghai QIN, Chunsheng LIU, Dimin NIU
  • Publication number: 20230088939
    Abstract: The maximum capacity of a very fast memory in a system that requires very fast memory access times is increased by adding a memory with remote access times that are slower than required, and then moving infrequently accessed data from the memory with the very fast access times to the memory with the slow access times.
    Type: Application
    Filed: January 21, 2022
    Publication date: March 23, 2023
    Inventors: Yuhao WANG, Dimin NIU, Yijin GUAN, Shengcheng WANG, Shuangchen LI, Hongzhong ZHENG
  • Publication number: 20230087747
    Abstract: A high bandwidth memory system. In some embodiments, the system includes: a memory stack having a plurality of memory dies and eight 128-bit channels; and a logic die, the memory dies being stacked on, and connected to, the logic die; wherein the logic die may be configured to operate a first channel of the 128-bit channels in: a first mode, in which a first 64 bits operate in pseudo-channel mode, and a second 64 bits operate as two 32-bit fine-grain channels, or a second mode, in which the first 64 bits operate as two 32-bit fine-grain channels, and the second 64 bits operate as two 32-bit fine-grain channels.
    Type: Application
    Filed: November 28, 2022
    Publication date: March 23, 2023
    Inventors: Krishna T. MALLADI, Mu-Tien CHANG, Dimin NIU, Hongzhong ZHENG