Patents by Inventor Hongsheng Wang

Hongsheng Wang has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240127027
    Abstract: Disclosed are an optimization method and apparatus for compiling computation graph. The optimization method includes the following steps: step S1: converting a computation graph into an intermediate representation; step S2: analyzing a dependency relationship; step S3: constructing a work stack; step S4: performing initialization to achieve a nonactivated state; step S5: popping out stack top node elements, and updating an input node set in a current round of iteration; step S6: adding the stack top node elements that depend on step S5 to a stack top position in sequence until the work stack is empty; step S7: implementing an intermediate representation in a fixed node state using a bit vector; and step S8: allocating registers for effective tensor variables contained in nodes of the intermediate representation in the fixed node state.
    Type: Application
    Filed: November 22, 2022
    Publication date: April 18, 2024
    Inventors: Hongsheng WANG, Shuibing HE, Guang CHEN
  • Publication number: 20240118897
    Abstract: Disclosed are an instruction execution method and apparatus for graph computation. The method includes the following steps: S1: sending operators of each node in a computational graph used for neural network computation to an operator interpreter; S2: building, by the operator interpreter, instructions in operation; S3: defining an instruction dependency relationship; S4: building an instruction dependency relationship graph; S5: building a topological order of parallel instructions; S6: scheduling the parallel instructions to hardware resources; S7: building shortest schedules for the parallel instructions: the shortest time required to execute the parallel instructions under the condition of limited hardware resources; and S8: releasing the completed instructions.
    Type: Application
    Filed: November 30, 2022
    Publication date: April 11, 2024
    Inventors: Hongsheng WANG, Guang CHEN, Lingfang ZENG, Aimin PAN
  • Publication number: 20240104341
    Abstract: A memory optimization method includes: compiling a neural network into a computational graph for neural network computation on a computer; transforming the computational graph into a topological graph; constructing a life cycle relationship graph of tensor variables in the computational graph; and analyzing a life cycle relationship among tensor variables in a node of the computational graph; iteratively merging those tensor variables connected by lines of the second type and caching into a memory any tensor variable that goes beyond a number of idle registers and is not allocated to a register, until all tensor variables that go beyond the number of the idle registers and are not allocated to registers are cached into the memory; caching any node of the life cycle relationship graph with a degree smaller than a number of registers into a stack.
    Type: Application
    Filed: November 22, 2022
    Publication date: March 28, 2024
    Inventors: Hongsheng WANG, Guang CHEN, Lingfang ZENG
  • Publication number: 20240104395
    Abstract: Disclosed are a memory optimization method and device oriented to neural network computing. The memory optimization method oriented to neural network computing includes the following steps: step S1: reconstructing a computation graph into a topological structure computation graph; step S2: constructing a life cycle interval about tensor variables; step S3: constructing a scanning line about the life cycle interval; step S4: allocating the tensor variables to idle registers; step S5: allocating to tensor variables exceeding the required number of registers; step S6: allocating registers allocated in the expired life cycle interval to tensor variables exceeding the required number of registers; and step S7: adding tensor variables transferred to a memory back to the life cycle interval in an activated state, and allocating idle registers for the tensor variables. According to the present disclosure, the memory of a data flow of a computation graph for neural network computing is optimized.
    Type: Application
    Filed: December 1, 2022
    Publication date: March 28, 2024
    Inventors: Hongsheng WANG, Guang CHEN
  • Publication number: 20240102793
    Abstract: An on-line measurement-error correction device and method for the inner profile of a special-shaped shell, including a fixing device. A vertical moving device is fixed at the top of the fixing device and connected with a horizontal moving device connected with a distance monitoring device, which is movable vertically and horizontally under the drive of the vertical and horizontal moving devices. The distance monitoring device includes a displacement monitoring element fixedly arranged on a fixing support hinged with an electric push rod configured to displace to drive the displacement monitoring element to deflect so as to change a monitoring direction. The displacement monitoring element is driven to deflect by the electric push rod of the distance monitoring device to change a monitoring direction, and a distance between each longitudinal section surface point on the inner surface of the special-shaped shell and the displacement monitoring element can be gradually measured.
    Type: Application
    Filed: March 15, 2023
    Publication date: March 28, 2024
    Applicants: SHANDONG UNIVERSITY, SHANDONG RESEARCH AND DESIGN INSTITUTE OF INDUSTRIAL CERAMICS CO., LTD.
    Inventors: Qinghua SONG, Xiaojuan WANG, Liping JIANG, Hongsheng WANG, Qiang LUAN, Zhanqiang LIU, Yicong DU
  • Publication number: 20240104016
    Abstract: The disclosure discloses an intermediate representation method for compiling computation graphs, including: step 1: compiling a neural network into a computation graph for neural network computation; step 2: constructing a node for each tensor variable in the computation graph; step 3: associating the node representing the tensor variable in the computation graph to a set of pointers to the tensor variable; step 4: analyzing constraint relationships between the tensor variables in the computation graph; step 5: iteratively constructing a topological graph of the intermediate representation based on the constraint relationships between the tensor variables in the computation graph; and step 6: analyzing the tensor variables with different aliases pointing to a same memory location based on the intermediate representation, and allocating a register for the tensor variables with different aliases.
    Type: Application
    Filed: November 30, 2022
    Publication date: March 28, 2024
    Inventors: Hongsheng WANG, Aimin PAN, Guang CHEN
  • Patent number: 11941532
    Abstract: Disclosed is a method for adapting a deep learning framework to a hardware device based on a unified backend engine, which comprises the following steps: S1, adding the unified backend engine to the deep learning framework; S2, adding the unified backend engine to the hardware device; S3, converting a computational graph, wherein the computational graph compiled and generated by the deep learning framework is converted into an intermediate representation of the unified backend engine; S4, compiling the intermediate representation, wherein the unified backend engine compiles the intermediate representation on the hardware device to generate an executable object; S5, running the executable object, wherein the deep learning framework runs the executable object on the hardware device; S6: managing memory of the unified backend engine.
    Type: Grant
    Filed: April 22, 2022
    Date of Patent: March 26, 2024
    Assignee: ZHEJIANG LAB
    Inventors: Hongsheng Wang, Wei Hua, Hujun Bao, Fei Yang
  • Patent number: 11941507
    Abstract: Disclosed are a data flow method and apparatus for neural network computation. The data flow method for neural network computation includes initializing the lifecycle of a variable in a computational graph; and defining a propagation rule for a variable in use to flow through a node. A definition of the variable is produced at a precursor node of the node, such that an input set of valid variables flowing through the node contains the variable. The method may be used on neural network computation in a deep learning training system.
    Type: Grant
    Filed: September 27, 2022
    Date of Patent: March 26, 2024
    Assignee: ZHEJIANG LAB
    Inventors: Hongsheng Wang, Guang Chen
  • Patent number: 11941514
    Abstract: The present disclosure discloses a method for execution of a computational graph in a neural network model and an apparatus thereof, including: creating task execution bodies on a native machine according to a physical computational graph compiled and generated by a deep learning framework, and designing a solution for allocating a plurality of idle memory blocks to each task execution body, so that the entire computational graph participates in deep learning training tasks of different batches of data in a pipelining and parallelizing manner.
    Type: Grant
    Filed: March 29, 2022
    Date of Patent: March 26, 2024
    Assignee: ZHEJIANG LAB
    Inventors: Hongsheng Wang, Hujun Bao, Guang Chen, Lingfang Zeng, Hongcai Cheng, Yong Li, Jian Zhu, Huanbo Zheng
  • Patent number: 11934887
    Abstract: The present disclosure discloses a distributed model compilation system. A master node of the system determines the logic calculation graph of the model based on model information, divides the logic calculation graph into multiple logic calculation sub-graphs, generates a distributing message for each logic calculation sub-graph, and then transmits the distributing message to a slave node. Each of the slave nodes allocates a local computing resource to compile the logic calculation sub-graph based on the received distributing message, and transmits compilation completion information to the master node. The master node determines the completion of model compilation based on the compilation completion information returned by each slave node, and executes the target work based on the compiled model.
    Type: Grant
    Filed: September 13, 2023
    Date of Patent: March 19, 2024
    Assignee: ZHEJIANG LAB
    Inventors: Hongsheng Wang, Fei Wu, Guang Chen, Feng Lin
  • Patent number: 11915135
    Abstract: The disclosure discloses a graph optimization method and apparatus for neural network computation. The graph optimization method includes the following steps: S1: converting a computation graph; S2: allocating a register; S3: defining a route selector for a redefined variable; S4: solving the route selector for the redefined variable; S5: defining a criterion of inserting the route selector for the redefined variable into a node; S6: analyzing a dominating edge set of the node for the redefined variable; S7: inserting the route selector for the redefined variable; and S8: renaming the redefined variable. The disclosure solves the problem of the corresponding route selection on a correct definition of the redefined variable when a node including the redefined variable in a computation graph in the compiling period flows through multiple paths of computation flow, reduces the memory cost and promotes the development of implementation application of a deep neural network model.
    Type: Grant
    Filed: September 21, 2022
    Date of Patent: February 27, 2024
    Assignee: ZHEJIANG LAB
    Inventors: Hongsheng Wang, Guang Chen
  • Publication number: 20240054319
    Abstract: Disclosed are a data flow method and apparatus for neural network computation. The method includes: step 1, initializing the lifecycle of a variable in a computational graph, i.e., initializing a time period from the start of a definition of the variable to the end of use as the lifecycle of the variable in the computational graph; and step 2, defining a propagation rule for a variable in use to flow through a node, i.e., defining that in the case that a variable at a certain node in the computational graph is used, a definition of the variable is produced at a precursor node of the node, such that an input set of valid variables flowing through the node contains the variable. The application discloses a data flow modeling method and apparatus for neural network computation in a deep learning training system.
    Type: Application
    Filed: September 27, 2022
    Publication date: February 15, 2024
    Inventors: Hongsheng WANG, Guang CHEN
  • Publication number: 20240028886
    Abstract: The disclosure discloses a graph optimization method and apparatus for neural network computation. The graph optimization method includes the following steps: S1: converting a computation graph; S2: allocating a register; S3: defining a route selector for a redefined variable; S4: solving the route selector for the redefined variable; S5: defining a criterion of inserting the route selector for the redefined variable into a node; S6: analyzing a dominating edge set of the node for the redefined variable; S7: inserting the route selector for the redefined variable; and S8: renaming the redefined variable. The disclosure solves the problem of the corresponding route selection on a correct definition of the redefined variable when a node including the redefined variable in a computation graph in the compiling period flows through multiple paths of computation flow, reduces the memory cost and promotes the development of implementation application of a deep neural network model.
    Type: Application
    Filed: September 21, 2022
    Publication date: January 25, 2024
    Inventors: Hongsheng WANG, Guang CHEN
  • Patent number: 11861505
    Abstract: The disclosure discloses a method of executing dynamic graph for neural network computation and the apparatus thereof. The method of executing dynamic graph includes the following steps: S1: constructing and distributing an operator and a tensor; S2: deducing an operator executing process by an operator interpreter; S3: constructing an instruction of a virtual machine at runtime by the operator interpreter; S4: sending the instruction to the virtual machine at runtime by the operator interpreter; S5: scheduling the instruction by the virtual machine; and S6: releasing an executed instruction by the virtual machine. According to the method of executing dynamic graph for neural network computation and the apparatus thereof provided by the disclosure, runtime is abstracted to be the virtual machine, and the virtual machine acquires a sub-graph of each step constructed by a user in real time through the interpreter and schedules, the virtual machines issues, and executes each sub-graph.
    Type: Grant
    Filed: June 6, 2022
    Date of Patent: January 2, 2024
    Assignee: ZHEJIANG LAB
    Inventors: Hongsheng Wang, Hujun Bao, Guang Chen
  • Publication number: 20230410560
    Abstract: Disclosed are a method and apparatus for constructing a three-dimensional data set of a pedestrian re-identification based on a neural radiation field. The method includes the following steps: S1: capturing images of pedestrians to be entered by a group of cameras at different viewing angles; S2: generating a three-dimensional spatial position point set by sampling through camera rays in the scenario, and converting observation directions of the cameras corresponding to the three-dimensional spatial position point set into three-dimensional Cartesian unit vectors; and S3: inputting, into a multi-layer sensor, the three-dimensional spatial position point set and the observation directions converted into the three-dimensional Cartesian unit vectors, to output corresponding densities and colors. The method and apparatus of the present disclosure gives a brand-new method for constructing a pedestrian re-identification data set, and provides a new idea of data set construction.
    Type: Application
    Filed: September 21, 2022
    Publication date: December 21, 2023
    Inventors: Hongsheng WANG, Guang CHEN, Hujun BAO
  • Patent number: 11823053
    Abstract: The disclosure discloses a method of neural network model computation-oriented intermediate representation and apparatus thereof. The method includes the following steps: S1, parsing an input model file so as to acquire topological structure information of a neural network; S2, constructing a logical computation graph; S21, inferring physical layout information of each operator in the logical computation graph; S22, inferring meta attributes of each operator in the logical computation graph; S23, inferring description information of input and output logical tensors of each operator in the logical computation graph; S3, constructing a physical computation graph; S31, generating a physical computation graph, etc.
    Type: Grant
    Filed: April 6, 2022
    Date of Patent: November 21, 2023
    Assignee: ZHEJIANG LAB
    Inventors: Hongsheng Wang, Wei Hua, Weiqiang Jia, Hujun Bao
  • Patent number: 11810366
    Abstract: Disclosed are a joint modeling method and apparatus for enhancing local features of pedestrians. The method includes the following steps: S1: acquiring an original surveillance video image data set, dividing the original surveillance video image data set into a training set and a test set in proportion; S2: cutting the surveillance video image training set to obtain image block vector sequences. In the present disclosure, local features of pedestrians in video images are extracted by a multi-head attention neural network, weight parameters of image channels are learned by channel convolution kernels, spatial features on the images are scanned through spatial convolution, local features of pedestrians are enhanced to improve the recognition rate of pedestrians, a feed-forward neural network and an activation function are adopted, so as to realize pedestrian re-recognition, thereby obtaining face images available.
    Type: Grant
    Filed: November 30, 2022
    Date of Patent: November 7, 2023
    Assignee: ZHEJIANG LAB
    Inventors: Hongsheng Wang, Guang Chen
  • Publication number: 20230353458
    Abstract: The present disclosure provides a neural network computing-oriented modeling method and apparatus for distributed data routing. The method includes the following steps: S1, designing the distributed attribute of a physical tensor: abstracting a mapping relationship between a logic tensor and the physical tensor into three distributed attributes including a broadcast attribute, a scatter attribute and a local reduction attribute; S2, deducing the distributed attribute of an output tensor: specifying the distributed attribute of an input tensor, and then deducing the legal distributed attribute of the output tensor according to the known distributed attribute of the input tensor; and S3, judging, according to the distributed attribute situation, whether an intermediate communication primitive needs to be inserted to obtain the distributed attribute of a local physical tensor.
    Type: Application
    Filed: June 23, 2022
    Publication date: November 2, 2023
    Inventors: Hongsheng WANG, Shuibing HE, Hujun BAO, Guang CHEN
  • Publication number: 20230351145
    Abstract: The present disclosure provides a pipelining and parallelizing graph execution method for neural network model computation and apparatus, and provides a pipelining and parallelizing graph execution method for neural network model computation and apparatus in a deep learning training system. The method includes the graph execution flow in a neural network model computation process and a process of cooperative work of all functional modules. The pipelining and parallelizing graph execution method for neural network model computation includes creating a graph executive on a native machine according to a physical computation graph compiled and generated by a deep learning framework.
    Type: Application
    Filed: June 13, 2022
    Publication date: November 2, 2023
    Inventors: Hongsheng WANG, Bowen TAN, Hujun BAO, Guang CHEN
  • Publication number: 20230351212
    Abstract: The disclosure provides a semi-supervised method and apparatus for public opinion text analysis. The semi-supervised method includes: first acquiring a public opinion data set, and preprocessing the data set; performing a data augmentation algorithm on preprocessed samples to generate data augmented samples; generating category labels for the unlabeled samples in the data set in an unsupervised extraction and clustering manner; calculating similarities of word vector latent semantic spaces and performing linear interpolation operation to generate, according to an operation result, similarity interpolation samples; constructing a final training sample set; adopting a semi-supervised method, inputting the final training sample set into a pre-trained language model to train the model to obtain a classification model; and predicting the test set by using the classification model to obtain a classification result.
    Type: Application
    Filed: June 10, 2022
    Publication date: November 2, 2023
    Inventors: Hongsheng WANG, Qing LIAO, Hujun BAO, Guang CHEN