Patents by Inventor Yuan Xie
Yuan Xie has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20250029269Abstract: A method of BIM reconstruction for a piping system is provided.Type: ApplicationFiled: November 21, 2022Publication date: January 23, 2025Applicants: NANYANG TECHNOLOGICAL UNIVERSITY, SURBANA JURONG PTE LTDInventors: Yuan XIE, Yiyu CAI
-
Patent number: 12190245Abstract: Methods and systems are provided for implementing training of learning models, including obtaining a pre-trained weight set for a learning model on a sample dataset and on a first loss function; selecting at least two tasks having heterogeneous features to be computed by a reference model; obtaining a reference dataset for the at least two tasks; designating a second loss function for feature embedding between the heterogeneous features of the at least two tasks; training the learning model on the first loss function and training the reference model on the second loss function, in turn; and updating the weight set based on a feature embedding learned by the learning model and a feature embedding learned by the reference model, in turn. Methods and systems of the present disclosure may alleviate computational overhead incurred by executing the learning model and loading different weight sets at a central network of the model.Type: GrantFiled: November 11, 2019Date of Patent: January 7, 2025Assignee: Alibaba Group Holding LimitedInventors: Chao Cheng, Xiaoxin Fan, Minghai Qin, Yuan Xie
-
Patent number: 12141699Abstract: The present disclosure relates to systems and methods for providing vector-wise sparsity in neural networks. In some embodiments, an exemplary method for providing vector-wise sparsity in a neural network, comprises: dividing a matrix associated with the neural network into a plurality of vectors; selecting a first subset of non-zero elements from the plurality of vectors to form a pruned matrix; and outputting the pruned matrix for executing the neural network using the pruned matrix.Type: GrantFiled: July 23, 2020Date of Patent: November 12, 2024Assignee: Alibaba Group Holding LimitedInventors: Maohua Zhu, Tao Zhang, Zhenyu Gu, Yuan Xie
-
Publication number: 20240348422Abstract: A privacy calculation unit includes a first calculation subunit, a storage subunit, and a communication subunit. The first calculation subunit includes circuitry to calculate first domain conversion ciphertexts sequentially. The storage subunit is configured to store the calculated first domain conversion ciphertexts received from the first calculation subunit. The first domain conversion ciphertext is an intermediate ciphertext when first to-be-converted data is converted from a first privacy-preserving computation domain to a second privacy-preserving computation domain.Type: ApplicationFiled: April 16, 2024Publication date: October 17, 2024Inventors: Zhaohui CHEN, Zhen GU, Yanheng LU, Dimin NIU, Ziyuan LIANG, Qi'ao JIN, Fan ZHANG, Yuan XIE
-
Patent number: 12099517Abstract: Systems and methods include causing presentation of a first cluster in association with an event of the first cluster, the first cluster from a first set of clusters of events. Each event includes a time stamp and event data. Based on the presentation of the first cluster, an extraction rule corresponding to the event of the first cluster is received from a user. Similarities in the event data between the events are determined based on the received extraction rule. The events are grouped into a second set of clusters based on the determined similarities. Presentation is caused of a second cluster in association with an event of the second cluster, where the second cluster is from the second set of clusters.Type: GrantFiled: April 14, 2023Date of Patent: September 24, 2024Assignee: Splunk Inc.Inventors: Jesse Brandau Miller, Katherine Kyle Feeney, Yuan Xie, Steve Zhang, Adam Jamison Oliner, Jindrich Dinga, Jacob Leverich
-
Patent number: 12073317Abstract: Embodiments of the disclosure provide methods and systems for processing a neural network associated with an input matrix having a first number of elements. The method can include: dividing the input matrix into a plurality of vectors, each vector having a second number of elements; grouping the plurality of vectors into a first group of vectors and a second group of vectors; and pruning the first group of vectors and the second group of vectors.Type: GrantFiled: January 7, 2020Date of Patent: August 27, 2024Assignee: Alibaba Group Holding LimitedInventors: Ao Ren, Tao Zhang, Yuhao Wang, Yuan Xie
-
Publication number: 20240176845Abstract: Methods and devices, the method including receiving a matrix of a neural network model; classifying at least a portion of the matrix as a first section based on a first distribution pattern of non-zero elements of the portion of the matrix; and identifying memory addresses of the non-zero elements in the first section of the matrix for loading, according to a first order determined based on the first distribution pattern, the non-zero elements in the first section into one or more vector registers.Type: ApplicationFiled: February 6, 2024Publication date: May 30, 2024Inventors: Guoyang CHEN, Yu PU, Yongzhi ZHANG, Weifeng ZHANG, Yuan XIE
-
Publication number: 20240160933Abstract: Methods and apparatus for reducing a size of a neural network model, the method including: compressing data of the neural network model; identifying structure information of a vector register, wherein the structure information includes a number of registers included in the vector register; comparing a number of elements in the compressed data with a first condition, wherein the first condition is determined based on the number of registers in the vector register; and in response to the number of elements satisfying the first condition, associating the compressed data with the vector register to enable loading the compressed data to the vector register.Type: ApplicationFiled: January 23, 2024Publication date: May 16, 2024Inventors: Weifeng ZHANG, Guoyang CHEN, Yu PU, Yongzhi ZHANG, Yuan XIE
-
Publication number: 20240161245Abstract: In an image optimization method, an image generation network, a to-be-optimized image, and a plurality of preset image features are obtained. A target feature is selected from the plurality of preset image features based on (i) the target feature and the to-be-optimized image and (ii) a preset similarity condition. The target feature and an initial offset parameter are input to the image generation network. The initial offset parameter is adjusted according to a difference between an output of the image generation network and the to-be-optimized image, to obtain a target offset parameter. The target feature and the target offset parameter are input to the image generation network, to generate an optimized image. Apparatus and non-transitory computer-readable storage medium counterpart embodiments are also contemplated.Type: ApplicationFiled: January 24, 2024Publication date: May 16, 2024Applicant: Tencent Technology (Shenzhen) Company LimitedInventors: Chuming LIN, Yanbo WANG, Donghao LUO, Ying TAI, Zhizhong ZHANG, Yuan XIE, Chengjie WANG
-
Patent number: 11947821Abstract: The present disclosure provides methods, systems, and non-transitory computer readable media for managing a primary storage unit of an accelerator. The methods include assessing activity of the accelerator; assigning, based on the assessed activity of the accelerator, a lease to a group of one or more pages of data on the primary storage unit, wherein the assigned lease indicates a lease duration; and marking, in response to the expiration of the lease duration indicated by the lease, the group of one or more pages of data as an eviction candidate.Type: GrantFiled: June 15, 2020Date of Patent: April 2, 2024Assignee: Alibaba Group Holding LimitedInventors: Yongbin Gu, Pengcheng Li, Tao Zhang, Yuan Xie
-
Patent number: 11921814Abstract: Methods and devices, the method including receiving a matrix of a neural network model; classifying at least a portion of the matrix as a first section based on a first distribution pattern of non-zero elements of the portion of the matrix; and identifying memory addresses of the non-zero elements in the first section of the matrix for loading, according to a first order determined based on the first distribution pattern, the non-zero elements in the first section into one or more vector registers.Type: GrantFiled: June 14, 2022Date of Patent: March 5, 2024Assignee: Alibaba Group Holding LimitedInventors: Guoyang Chen, Yu Pu, Yongzhi Zhang, Weifeng Zhang, Yuan Xie
-
Patent number: 11915138Abstract: Methods and apparatus for reducing a size of a neural network model, the method including: compressing data of the neural network model; identifying structure information of a vector register, wherein the structure information includes a number of registers included in the vector register; comparing a number of elements in the compressed data with a first condition, wherein the first condition is determined based on the number of registers in the vector register; and in response to the number of elements satisfying the first condition, associating the compressed data with the vector register to enable loading the compressed data to the vector register.Type: GrantFiled: February 18, 2020Date of Patent: February 27, 2024Assignee: Alibaba Group Holding LimitedInventors: Weifeng Zhang, Guoyang Chen, Yu Pu, Yongzhi Zhang, Yuan Xie
-
Patent number: 11900239Abstract: Systems and methods for dynamically executing sparse neural networks are provided. In one implementation, a system for providing dynamic sparsity in a neural network may include at least one memory storing instructions and at least one processor configured to execute the instructions to: reduce an input vector and a set of weights of the neural network, execute an input layer of the neural network using the reduced input vector and set of weights to generate a reduced output vector; expand the reduced output vector to a full output vector using first predictable output neurons (PONs); using a PON map, reduce a dimension of the full output vector; execute subsequent layers of the neural network using the reduced full output vector to produce a second reduced output vector; and expand the second reduced output vector to a second full output vector using second PONs.Type: GrantFiled: September 5, 2019Date of Patent: February 13, 2024Assignee: Alibaba Group Holding LimitedInventors: Zhenyu Gu, Liu Liu, Shuangchen Li, Yuan Xie
-
Publication number: 20230394300Abstract: This application describes methods, systems, and apparatus, for neural network-based program sampling (NPS). An example device may obtain an assembly code of a program and an execution trace of the program, and divide the assembly code into a plurality of execution intervals. The device may construct a plurality of code graphs respectively corresponding to the plurality of execution intervals, and for each of the plurality of code graphs: generate a plurality of graph snapshots based on the code graph and the execution trace of the program; embed, by using a Graph Neural Network, the plurality of graph snapshots into a plurality of vectors; and aggregate the plurality of vectors into an execution embedding. The device may cluster the plurality of execution embeddings into a plurality of clusters and select representative execution intervals of the program based on the plurality of clusters for execution.Type: ApplicationFiled: October 28, 2022Publication date: December 7, 2023Inventors: Yuanwei FANG, Jian CHEN, Yen-Kuang CHEN, Yuan XIE
-
Patent number: 11816574Abstract: An input weight pattern of a machine learning model may be received. The input weight pattern may be pruned to produce an output weight pattern based on a predetermined pruning algorithm. The pruning algorithm may include partitioning the input weight pattern into a plurality of sub-patterns, each row of the input weight pattern including sub-rows of a first number of sub-patterns, and each column of the input weight pattern including sub-columns of a second number of sub-patterns; and pruning sub-columns and sub-rows from the plurality of sub-patterns to achieve predetermined column and row sparsities respectively, with a constraint that at least one sub-row in each row of the input weight pattern is not pruned. The output weight pattern may further be compressed to produce a compact weight pattern. The compact weight pattern has lower memory and computational overheads as compared to the input weight pattern for the machine learning model.Type: GrantFiled: October 25, 2019Date of Patent: November 14, 2023Assignee: Alibaba Group Holding LimitedInventors: Ao Ren, Yuhao Wang, Tao Zhang, Yuan Xie
-
Patent number: 11755903Abstract: The present disclosure relates to systems and methods for providing block-wise sparsity in neural networks. In one implementation, a system for providing block-wise sparsity in a neural network may include at least one memory storing instructions and at least one processor configured to execute the instructions to: divide a matrix of weights associated with a neural network into a plurality of blocks; extract non-zero elements from one or more of the plurality of blocks; re-encode the extracted non-zero elements as vectors with associated coordinates of the extracted non-zero elements within the one or more blocks; enforce input sparsity in the neural network corresponding to the associated coordinates; and execute the neural network using the vectors and the enforced input sparsity.Type: GrantFiled: July 24, 2019Date of Patent: September 12, 2023Assignee: Alibaba Group Holding LimitedInventors: Maohua Zhu, Zhenyu Gu, Yuan Xie
-
Patent number: 11657065Abstract: Systems and methods include causing presentation of a first cluster in association with an event of the first cluster, the first cluster from a first set of clusters of events. Each event includes a time stamp and event data. Based on the presentation of the first cluster, an extraction rule corresponding to the event of the first cluster is received from a user. Similarities in the event data between the events are determined based on the received extraction rule. The events are grouped into a second set of clusters based on the determined similarities. Presentation is caused of a second cluster in association with an event of the second cluster, where the second cluster is from the second set of clusters.Type: GrantFiled: January 26, 2021Date of Patent: May 23, 2023Assignee: Splunk Inc.Inventors: Jesse Brandau Miller, Katherine Kyle Feeney, Yuan Xie, Steve Zhang, Adam Jamison Oliner, Jindrich Dinga, Jacob Leverich
-
Patent number: 11651012Abstract: A method includes in response to a user selection of a command of a coding language, causing display of a set of argument blocks in a text input region based on syntax of the command Each argument block allows the user to input a value of an argument of the command to the argument block. In response to a user selection to modify the set of argument blocks, an argument block is added to the set of argument blocks displayed in the text input region based on the syntax of the command. In response to receiving from the user the input of the value of the argument to the added argument block, the command is caused to be coded in the text input region with at least the argument having the value from the input to the added argument block.Type: GrantFiled: May 17, 2021Date of Patent: May 16, 2023Assignee: Splunk Inc.Inventors: Jindrich Dinga, Yuan Xie, Katherine Kyle Feeney, Jesse Miller
-
Publication number: 20230142598Abstract: An integrated circuit device includes a memory controller coupleable to a memory. The memory controller to schedule memory accesses to regions of the memory based on memory timing parameters specific to the regions. A method includes receiving a memory access request at a memory device. The method further includes accessing, from a timing data store of the memory device, data representing a memory timing parameter specific to a region of the memory cell circuitry targeted by the memory access request. The method also includes scheduling, at the memory controller, the memory access request based on the data.Type: ApplicationFiled: September 22, 2022Publication date: May 11, 2023Inventors: Yi XU, Nuwan S. JAYASENA, Yuan XIE
-
Patent number: 11625341Abstract: The systems and methods are configured to efficiently and effectively access memory. In one embodiment, a memory controller comprises a request queue, a buffer, a control component, and a data path system. The request queue receives memory access requests. The control component is configured to process information associated with access requests via a first narrow memory channel and a second narrow memory channel. The first narrow memory channel and the second narrow memory channel can have a portion of command/control communication lines and address communication lines that are included in and shared between the first narrow memory channel and the second narrow memory channel. The data path system can include a first data module and one set of unshared data lines associated with the first memory channel and a second data module and another set of unshared data lines associated with second memory channel.Type: GrantFiled: August 11, 2020Date of Patent: April 11, 2023Assignee: Alibaba Group Holding LimitedInventors: Jilan Lin, Dimin Niu, Shuangchen Li, Hongzhong Zheng, Yuan Xie