Patents by Inventor Hongzhong Zheng

Hongzhong Zheng has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 12197354
    Abstract: A method of transferring data between a memory controller and at least one memory module via a primary data bus having a primary data bus width is disclosed. The method includes accessing a first one of a memory device group via a corresponding data bus path in response to a threaded memory request from the memory controller. The accessing results in data groups collectively forming a first data thread transferred across a corresponding secondary data bus path. Transfer of the first data thread across the primary data bus width is carried out over a first time interval, while using less than the primary data transfer continuous throughput during that first time interval. During the first time interval, at least one data group from a second data thread is transferred on the primary data bus.
    Type: Grant
    Filed: August 29, 2023
    Date of Patent: January 14, 2025
    Assignee: Rambus Inc.
    Inventors: Hongzhong Zheng, Frederick A Ware
  • Patent number: 12197726
    Abstract: A 3D-stacked memory device including: a base die including a plurality of switches to direct data flow and a plurality of arithmetic logic units (ALUs) to compute data; a plurality of memory dies stacked on the base die; and an interface to transfer signals to control the base die.
    Type: Grant
    Filed: September 15, 2023
    Date of Patent: January 14, 2025
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Mu-Tien Chang, Prasun Gera, Dimin Niu, Hongzhong Zheng
  • Patent number: 12189539
    Abstract: A memory management method of a data processing system is provided. The memory management method includes: creating a first memory zone and a second memory zone related to a first node of a first server, wherein the first server is located in the data processing system, and the first node includes a processor and a first memory; mapping the first memory zone to the first memory; and mapping the second memory zone to a second memory of a second server, wherein the second server is located in the data processing system, and the processor is configured to access the second memory of the second server through an interface circuit of the first server and through an interface circuit of the second server.
    Type: Grant
    Filed: December 13, 2022
    Date of Patent: January 7, 2025
    Assignee: ALIBABA (CHINA) CO., LTD.
    Inventors: Dimin Niu, Yijin Guan, Tianchan Guan, Shuangchen Li, Hongzhong Zheng
  • Patent number: 12189546
    Abstract: A memory module that includes a non-volatile memory and an asynchronous memory interface to interface with a memory controller is presented. The asynchronous memory interface may use repurposed pins of a double data rate (DDR) memory channel to send an asynchronous data to the memory controller. The asynchronous data may be device feedback indicating a status of the non-volatile memory.
    Type: Grant
    Filed: July 25, 2022
    Date of Patent: January 7, 2025
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Dimin Niu, Mu-Tien Chang, Hongzhong Zheng, Sun Young Lim, Indong Kim, Jangseok Choi, Craig Hanson
  • Publication number: 20250004658
    Abstract: A storage device and method of controlling a storage device are disclosed. The storage device includes a host, a logic die, and a high bandwidth memory stack including a memory die. A computation lookup table is stored on a memory array of the memory die. The host sends a command to perform an operation utilizing a kernel and a plurality of input feature maps, includes finding the product of a weight of the kernel and values of multiple input feature maps. The computation lookup table includes a row corresponding to a weight of the kernel, and a column corresponding to a value of the input feature maps. A result value stored at a position corresponding to a row and a column is the product of the weight corresponding to the row and the value corresponding to the column.
    Type: Application
    Filed: July 3, 2024
    Publication date: January 2, 2025
    Inventors: Peng Gu, Krishna T. Malladi, Hongzhong Zheng
  • Patent number: 12181987
    Abstract: According to one general aspect, an apparatus may include a plurality of stacked integrated circuit dies that include a memory cell die and a logic die. The memory cell die may be configured to store data at a memory address. The logic die may include an interface to the stacked integrated circuit dies and configured to communicate memory accesses between the memory cell die and at least one external device. The logic die may include a reliability circuit configured to ameliorate data errors within the memory cell die. The reliability circuit may include a spare memory configured to store data, and an address table configured to map a memory address associated with an error to the spare memory. The reliability circuit may be configured to determine if the memory access is associated with an error, and if so completing the memory access with the spare memory.
    Type: Grant
    Filed: October 12, 2021
    Date of Patent: December 31, 2024
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Dimin Niu, Krishna Malladi, Hongzhong Zheng
  • Patent number: 12164593
    Abstract: A general matrix-matrix multiplication (GEMM) dataflow accelerator circuit is disclosed that includes a smart 3D stacking DRAM architecture. The accelerator circuit includes a memory bank, a peripheral lookup table stored in the memory bank, and a first vector buffer to store a first vector that is used as a row address into the lookup table. The circuit includes a second vector buffer to store a second vector that is used as a column address into the lookup table, and lookup table buffers to receive and store lookup table entries from the lookup table. The circuit further includes adders to sum the first product and a second product, and an output buffer to store the sum. The lookup table buffers determine a product of the first vector and the second vector without performing a multiply operation. The embodiments include a hierarchical lookup architecture to reduce latency. Accumulation results are propagated in a systolic manner.
    Type: Grant
    Filed: July 13, 2021
    Date of Patent: December 10, 2024
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Peng Gu, Krishna Malladi, Hongzhong Zheng, Dimin Niu
  • Publication number: 20240394057
    Abstract: A reduced instruction set computer (RISC)-V vector extension (RVV) core communicated with one or more accelerators. The RVV core includes: a command queue configured to output commands; and an interface unit communicatively coupled to the command queue and having circuitry configured to generate an accelerator command to an accelerator of the one or more accelerators based on the output commands.
    Type: Application
    Filed: May 17, 2024
    Publication date: November 28, 2024
    Inventors: Qichen ZHANG, Zhe ZHANG, Linyong HUANG, Hongzhong ZHENG
  • Patent number: 12153646
    Abstract: An adaptive matrix multiplier. In some embodiments, the matrix multiplier includes a first multiplying unit a second multiplying unit, a memory load circuit, and an outer buffer circuit. The first multiplying unit includes a first inner buffer circuit and a second inner buffer circuit, and the second multiplying unit includes a first inner buffer circuit and a second inner buffer circuit. The memory load circuit is configured to load data from memory, in a single burst of a burst memory access mode, into the first inner buffer circuit of the first multiplying unit; and into the first inner buffer circuit of the second multiplying unit.
    Type: Grant
    Filed: October 17, 2022
    Date of Patent: November 26, 2024
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Dongyan Jiang, Dimin Niu, Hongzhong Zheng
  • Publication number: 20240385880
    Abstract: A task scheduling unit includes a progress detection subunit having circuitry configured to obtain first progress information of a first chip in which the task scheduling unit is located, the first progress information indicating a task execution progress of the first chip; a transmission subunit having circuitry configured to transmit the first progress information to a second chip, wherein the first chip and the second chip are located on a same wafer, and the task execution progress of the first chip is less than a task execution progress of the second chip; and a transfer subunit having circuitry configured to receive first request information transmitted by the second chip in response to the first progress information; and transfer at least some of tasks executed by the first chip to the second chip for execution based on the first request information.
    Type: Application
    Filed: May 17, 2024
    Publication date: November 21, 2024
    Inventors: Youwei ZHUO, Han XU, Zhe ZHANG, Shuangchen LI, Dimin NIU, Hongzhong ZHENG
  • Patent number: 12147360
    Abstract: A memory module that includes a non-volatile memory and an asynchronous memory interface to interface with a memory controller is presented. The asynchronous memory interface may use repurposed pins of a double data rate (DDR) memory channel to send an asynchronous data to the memory controller. The asynchronous data may be device feedback indicating a status of the non-volatile memory.
    Type: Grant
    Filed: July 25, 2022
    Date of Patent: November 19, 2024
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Dimin Niu, Mu-Tien Chang, Hongzhong Zheng, Sun Young Lim, Indong Kim, Jangseok Choi, Craig Hanson
  • Patent number: 12147341
    Abstract: Apparatus, method, and system provided herein are directed to prioritizing cache line writing of compressed data. The memory controller comprises a cache line compression engine that receives raw data, compresses the raw data, determines a compression rate between the raw data and the compressed data, determines whether the compression rate is greater than a predetermined rate, and outputs the compressed data as data-to-be-written if the compression rate is greater than the predetermined rate. In response to determining that the compression rate is greater than the predetermined rate, the cache line compression engine generates a compression signal indicating the data-to-be-written is the compressed data and sends the compression signal to a scheduler of a command queue in the memory controller where writing of compressed data is prioritized.
    Type: Grant
    Filed: August 6, 2020
    Date of Patent: November 19, 2024
    Assignee: Alibaba Group Holding Limited
    Inventors: Dimin Niu, Tianchan Guan, Lide Duan, Hongzhong Zheng
  • Patent number: 12141227
    Abstract: An adaptive matrix multiplier. In some embodiments, the matrix multiplier includes a first multiplying unit a second multiplying unit, a memory load circuit, and an outer buffer circuit. The first multiplying unit includes a first inner buffer circuit and a second inner buffer circuit, and the second multiplying unit includes a first inner buffer circuit and a second inner buffer circuit. The memory load circuit is configured to load data from memory, in a single burst of a burst memory access mode, into the first inner buffer circuit of the first multiplying unit; and into the first inner buffer circuit of the second multiplying unit.
    Type: Grant
    Filed: July 29, 2020
    Date of Patent: November 12, 2024
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Dongyan Jiang, Dimin Niu, Hongzhong Zheng
  • Patent number: 12141438
    Abstract: Zero skipping sparsity techniques for reduced data movement between memory and accelerators and reduced computational workload of accelerators. The techniques include detection of zero and near-zero values on the memory. The non-zero values are transferred to the accelerator for computation. The zero and near-zero values are written back within the memory as zero values.
    Type: Grant
    Filed: February 25, 2021
    Date of Patent: November 12, 2024
    Assignee: Alibaba Group Holding Limited
    Inventors: Fei Xue, Fei Sun, Yangjie Zhou, Lide Duan, Hongzhong Zheng
  • Patent number: 12142338
    Abstract: The present invention provides systems and methods for efficiently and effectively priming and initializing a memory. In one embodiment, a memory controller includes a normal data path and a priming path. The normal data path directs storage operations during a normal memory read/write operation after power startup of a memory chip. The priming path includes a priming module, wherein the priming module directs memory priming operations during a power startup of the memory chip, including forwarding a priming pattern for storage in a write pattern mode register of a memory chip and selection of a memory address in the memory chip for initialization with the priming pattern. The priming pattern includes information corresponding to proper initial data values. The priming pattern can also include proper corresponding error correction code (ECC) values. The priming module can include a priming pattern register that stores the priming pattern.
    Type: Grant
    Filed: January 19, 2021
    Date of Patent: November 12, 2024
    Assignee: Alibaba Group Holding Limited
    Inventors: Dimin Niu, Shuangchen Li, Tianchan Guan, Hongzhong Zheng
  • Publication number: 20240370168
    Abstract: The present disclosure provides a physical host including a memory, a first buffer, a second buffer, a third buffer and a processor. The first buffer stores a log regarding a plurality of dirty pages. The second buffer stores a dirty bitmap, where the dirty bitmap is written into the second buffer according to the log read from the first buffer. The third buffer stores the dirty bitmap. The processor obtains the current memory address to be migrated and a destination memory address, and marks a page table corresponding to the memory address to be migrated as a plurality of dirty pages and writes the log marked as the plurality of dirty pages into the first buffer when the memory address to be migrated is written. The processor includes a memory copy engine for reading the dirty bitmap from the third buffer, and copying the content corresponding to the plurality of dirty pages to the destination memory according to the dirty bitmap.
    Type: Application
    Filed: May 3, 2024
    Publication date: November 7, 2024
    Inventors: Jiacheng MA, Tianchan GUAN, Yijin GUAN, Dimin NIU, Hongzhong ZHENG
  • Publication number: 20240370374
    Abstract: The present disclosure relates to a computer system, a method for a computer system, and a computer-readable storage medium for executing the method for a computer system.
    Type: Application
    Filed: May 3, 2024
    Publication date: November 7, 2024
    Inventors: Jiacheng MA, Dimin NIU, Tianchan GUAN, Yijin GUAN, Hongzhong ZHENG
  • Publication number: 20240370384
    Abstract: This disclosure discloses a memory extension device, an operation method of the memory extension device, and a computer readable storage medium for executing the operation method. The method includes: converting local information received from a local host into local transaction layer information according to a first sub-protocol of a coherent interconnection protocol; converting the local transaction layer information into converted local transaction layer information according to a second sub-protocol of the coherent interconnection protocol, the converted local transaction layer information conforming to the second sub-protocol; packaging the converted local transaction layer information into a plurality of local data packets; and transmitting the plurality of local data packets to a remote memory extension device.
    Type: Application
    Filed: May 3, 2024
    Publication date: November 7, 2024
    Inventors: Tianchan GUAN, Yijin GUAN, Dimin NIU, Jiacheng MA, Zhaoyang DU, Hongzhong ZHENG
  • Patent number: 12130884
    Abstract: A general matrix-matrix multiplication (GEMM) dataflow accelerator circuit is disclosed that includes a smart 3D stacking DRAM architecture. The accelerator circuit includes a memory bank, a peripheral lookup table stored in the memory bank, and a first vector buffer to store a first vector that is used as a row address into the lookup table. The circuit includes a second vector buffer to store a second vector that is used as a column address into the lookup table, and lookup table buffers to receive and store lookup table entries from the lookup table. The circuit further includes adders to sum the first product and a second product, and an output buffer to store the sum. The lookup table buffers determine a product of the first vector and the second vector without performing a multiply operation. The embodiments include a hierarchical lookup architecture to reduce latency. Accumulation results are propagated in a systolic manner.
    Type: Grant
    Filed: July 13, 2021
    Date of Patent: October 29, 2024
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Peng Gu, Krishna Malladi, Hongzhong Zheng, Dimin Niu
  • Patent number: RE50205
    Abstract: A computing system includes: an adaptive back-up controller configured to calculate an adaptive back-up time based on a reserve power source for backing up a volatile memory to a nonvolatile memory; and a processor core, coupled to the adaptive back-up controller, configured to back up at least a portion of the volatile memory to the nonvolatile memory within the adaptive back-up time.
    Type: Grant
    Filed: September 5, 2020
    Date of Patent: November 12, 2024
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Hongzhong Zheng, Keith Chan, Wonseok Lee, Tackhwi Lee