Patents by Inventor Yingmin Li

Yingmin Li has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11263131
    Abstract: Embodiments of the disclosure provide systems and methods for allocating memory space in a memory device. The system can include: a memory device for providing the memory space; and a compiler component configured for: receiving a request for allocating a data array having a plurality of data elements in the memory device, wherein each of the plurality of data elements has a logical address; generating an instruction for allocating memory space for the data array in the memory device based on the request; generating device addresses for the plurality of data elements in the memory device based on logical addresses of the plurality of data elements; and allocating the memory space for the data array in the memory device based on the device addresses and the instruction.
    Type: Grant
    Filed: April 8, 2020
    Date of Patent: March 1, 2022
    Assignee: ALIBABA GROUP HOLDING LIMITED
    Inventors: Shuangchen Li, Dimin Niu, Fei Sun, Jingjun Chu, Hongzhong Zheng, Guoyang Chen, Yingmin Li, Weifeng Zhang, Xipeng Shen
  • Publication number: 20210318955
    Abstract: Embodiments of the disclosure provide systems and methods for allocating memory space in a memory device. The system can include: a memory device for providing the memory space; and a compiler component configured for: receiving a request for allocating a data array having a plurality of data elements in the memory device, wherein each of the plurality of data elements has a logical address; generating an instruction for allocating memory space for the data array in the memory device based on the request; generating device addresses for the plurality of data elements in the memory device based on logical addresses of the plurality of data elements; and allocating the memory space for the data array in the memory device based on the device addresses and the instruction.
    Type: Application
    Filed: April 8, 2020
    Publication date: October 14, 2021
    Inventors: Shuangchen LI, Dimin NIU, Fei SUN, Jingjun CHU, Hongzhong ZHENG, Guoyang CHEN, Yingmin LI, Weifeng ZHANG, Xipeng SHEN
  • Patent number: 10996976
    Abstract: The present disclosure relates to computer-implemented systems and methods for scheduling a neural network for execution. In one implementation, a system for scheduling a neural network for execution may include at least one memory storing instructions and at least one processor configured to execute the instructions to determine a profile for one or more applications co-scheduled with at least one neural network; determine a batch size for the at least one neural network based on the determined profile for the one or more applications; and scheduling the one or more applications and the at least one neural network based on the batch size.
    Type: Grant
    Filed: April 5, 2019
    Date of Patent: May 4, 2021
    Assignee: ALIBABA GROUP HOLDING LIMITED
    Inventors: Shuai Che, Guoyang Chen, Yingmin Li
  • Publication number: 20200319919
    Abstract: The present disclosure relates to computer-implemented systems and methods for scheduling a neural network for execution. In one implementation, a system for scheduling a neural network for execution may include at least one memory storing instructions and at least one processor configured to execute the instructions to determine a profile for one or more applications co-scheduled with at least one neural network; determine a batch size for the at least one neural network based on the determined profile for the one or more applications; and scheduling the one or more applications and the at least one neural network based on the batch size.
    Type: Application
    Filed: April 5, 2019
    Publication date: October 8, 2020
    Inventors: Shuai CHE, Guoyang CHEN, Yingmin LI
  • Publication number: 20200249998
    Abstract: The present disclosure relates to a method for scheduling a computation graph on a heterogeneous computing resource including one or more target devices for executing the computation graph. The computation graph includes a plurality of nodes and edges, each edge connecting two nodes among the plurality of nodes. The method comprises partitioning the computation graph into a plurality of subsets, each subset includes at least two nodes, and generating one or more task allocation models for each subset of the plurality of subsets. Wherein a task allocation model of the one or more task allocation models includes information of an execution order of operations represented by the at least two nodes of the corresponding subset and of a target device of the one or more target devices for executing each of the operations.
    Type: Application
    Filed: February 1, 2019
    Publication date: August 6, 2020
    Inventors: Shuai CHE, Yingmin LI, Ye YU
  • Publication number: 20200175361
    Abstract: Systems and methods are provided for improving the learning inference performance by partitioning the learning inference based on system fluctuations and available resources by parsing a trained neural network model of a neural network into a data flow graph with a plurality of nodes; generating a traversal order of the data flow graph; assigning a load level range to each edge device, an interconnect connecting the edge device and a cloud computing platform, and the cloud computing platform; profiling performance of each node over the load level range for the edge device and the cloud computing platform; and determining a partition point of the data flow graph based on the profiled performance of each node. By using a lookup table storing the profiled performance, the data flow diagram may be readily re-partitioned as needed for improving performance.
    Type: Application
    Filed: November 30, 2018
    Publication date: June 4, 2020
    Inventors: Shuai Che, Guoyang Chen, Yingmin Li