Patents by Inventor Lu Yuan
Lu Yuan has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20250148765Abstract: A method for annotating images to create a corpus for training a multi-task computer vision machine learning model is presented. The method comprises receiving, at one or more annotation specialist models, a plurality of images to be annotated. Via operation of the one or more annotation specialist models, pre-filtered annotations are generated for the plurality of images. Via operation of a data filtering and enhancement module, the pre-filtered annotations are filtered in accordance with predefined noise criteria so as to output candidate annotations for the plurality of images. The method further comprises, for each of one or more candidate annotations, selectively (1) storing the candidate annotation into the corpus as a final annotation for its associated image, or (2) adding the candidate annotation to its associated image using the one or more annotation specialist models and the data filtering and enhancement module for subsequent iterative annotation and filtering.Type: ApplicationFiled: January 30, 2024Publication date: May 8, 2025Applicant: Microsoft Technology Licensing, LLCInventors: Lu YUAN, Bin XIAO, Haiping WU, Weijian XU, Xiyang DAI, Houdong HU, Yumao LU, Nanshan ZENG, Ce Christopher LIU
-
Publication number: 20250123683Abstract: A method and an apparatus is for adjusting a display device and includes: obtaining a first facial image of a first user; determining first spatial coordinates of a facial feature point of the first user based on the first facial image; and adjusting, based on the first spatial coordinates, an orientation of a first side of an image displayed by the display device, where the first side of the image includes information to be communicated to the first user.Type: ApplicationFiled: November 8, 2024Publication date: April 17, 2025Inventors: Lu Yuan, Jingfang Zha
-
Patent number: 12223412Abstract: A computer device for automatic feature detection comprises a processor, a communication device, and a memory configured to hold instructions executable by the processor to instantiate a dynamic convolution neural network, receive input data via the communication network, and execute the dynamic convolution neural network to automatically detect features in the input data. The dynamic convolution neural network compresses the input data from an input space having a dimensionality equal to a predetermined number of channels into an intermediate space having a dimensionality less than the number of channels. The dynamic convolution neural network dynamically fuses the channels into an intermediate representation within the intermediate space and expands the intermediate representation from the intermediate space to an expanded representation in an output space having a higher dimensionality than the dimensionality of the intermediate space.Type: GrantFiled: December 16, 2020Date of Patent: February 11, 2025Assignee: Microsoft Technology Licensing, LLCInventors: Yinpeng Chen, Xiyang Dai, Mengchen Liu, Dongdong Chen, Lu Yuan, Zicheng Liu, Ye Yu, Mei Chen, Yunsheng Li
-
Publication number: 20250037252Abstract: The disclosure herein describes generating an inpainted image from a masked image using a patch-based encoder and an unquantized transformer. An image including a masked region and an unmasked region is received, and the received image is divided into a plurality of patches including masked patches. The plurality of patches is encoded into a plurality of feature vectors, wherein each patch is encoded to a feature vector. Using a transformer, a predicted token is generated for each masked patch using a feature vector encoded from the masked patch, and a quantized vector of the masked patch is determined using generated predicted token and a masked patch-specific codebook. The determined quantized vector of the masked patch is included into a set of quantized vectors associated with the plurality of patches, and an output image is generated from the set of quantized vectors using a decoder.Type: ApplicationFiled: October 11, 2024Publication date: January 30, 2025Inventors: Dongdong CHEN, Xiyang DAI, Yinpeng CHEN, Mengchen LIU, Lu YUAN
-
Patent number: 12190588Abstract: A system for tracking a target object across a plurality of image frames. The system comprises a logic machine and a storage machine. The storage machine holds instructions executable by the logic machine to calculate a trajectory for the target object over one or more previous frames occurring before a target frame. Responsive to assessing no detection of the target object in the target frame, the instructions are executable to predict an estimated region for the target object based on the trajectory, predict an occlusion center based on a set of candidate occluding locations for a set of other objects within a threshold distance of the estimated region, each location of the set of candidate occluding locations overlapping with the estimated region, and automatically estimate a bounding box for the target object in the target frame based on the occlusion center.Type: GrantFiled: June 4, 2021Date of Patent: January 7, 2025Assignee: Microsoft Technology Licensing, LLCInventors: Dongdong Chen, Qiankun Liu, Lu Yuan, Lei Zhang
-
Patent number: 12148131Abstract: The disclosure herein describes generating an inpainted image from a masked image using a patch-based encoder and an unquantized transformer. An image including a masked region and an unmasked region is received, and the received image is divided into a plurality of patches including masked patches. The plurality of patches is encoded into a plurality of feature vectors, wherein each patch is encoded to a feature vector. Using a transformer, a predicted token is generated for each masked patch using a feature vector encoded from the masked patch, and a quantized vector of the masked patch is determined using generated predicted token and a masked patch-specific codebook. The determined quantized vector of the masked patch is included into a set of quantized vectors associated with the plurality of patches, and an output image is generated from the set of quantized vectors using a decoder.Type: GrantFiled: April 29, 2022Date of Patent: November 19, 2024Assignee: Microsoft Technology Licensing, LLC.Inventors: Dongdong Chen, Xiyang Dai, Yinpeng Chen, Mengchen Liu, Lu Yuan
-
Publication number: 20240221128Abstract: The disclosure herein describes training an encoder network to inpaint images with masked portions. A primary encoding process is used to encode a visible portion of a masked input image into encoded token data. The encoded token data is then decoded into both pixel regression output and feature prediction output, wherein both outputs include inpainted image data associated with the masked portion of the masked input image. A pixel regression loss is determined using the pixel regression output and pixel data of an unmasked version of the masked input image. A feature prediction loss is determined using the feature prediction output and ground truth encoding output of the unmasked version of the masked input image. The primary encoding process is then trained using the pixel regression loss and the feature prediction loss, whereby the primary encoding process is trained to encode structural features of input images into encoded token data.Type: ApplicationFiled: May 19, 2022Publication date: July 4, 2024Inventors: Dongdong CHEN, Jianmin BAO, Ting ZHANG, Lu YUAN, Dong CHEN, Fang WEN, Xiaoyi DONG
-
Patent number: 12009894Abstract: Various aspects of the present disclosure generally relate to wireless communication. In some aspects, a network entity may receive a sounding reference signal (SRS) at a multi-panel system of the network entity, where the multi-panel system includes one or more sounded panels and one or more non-sounded panels. The network entity may estimate a channel to obtain channel state information (CSI) for the one or more sounded panels based at least in part on the SRS. The network entity may estimate CSI for the one or more non-sounded panels based at least in part on the CSI for the one or more sounded panels. The network entity may transmit or receive a communication based at least in part on the CSI for the one or more sounded panels and the CSI for the one or more non-sounded panels. Numerous other aspects are described.Type: GrantFiled: February 28, 2022Date of Patent: June 11, 2024Assignee: QUALCOMM IncorporatedInventors: Saeid Sahraei, Muhammad Sayed Khairy Abdelghaffar, Renqiu Wang, Lu Yuan, Joseph Patrick Burke, Tingfang Ji, Peter Gaal
-
Patent number: 11989956Abstract: Systems and methods for object detection generate a feature pyramid corresponding to image data, and rescaling the feature pyramid to a scale corresponding to a median level of the feature pyramid, wherein the rescaled feature pyramid is a four-dimensional (4D) tensor. The 4D tensor is reshaped into a three-dimensional (3D) tensor having individual perspectives including scale features, spatial features, and task features corresponding to different dimensions of the 3D tensor. The 3D tensor is used with a plurality of attention layers to update a plurality of feature maps associated with the image data. Object detection is performed on the image data using the updated plurality of feature maps.Type: GrantFiled: April 5, 2021Date of Patent: May 21, 2024Assignee: Microsoft Technology Licensing, LLCInventors: Xiyang Dai, Yinpeng Chen, Bin Xiao, Dongdong Chen, Mengchen Liu, Lu Yuan, Lei Zhang
-
Publication number: 20240078682Abstract: Training a multi-object tracking model includes: generating a plurality of training images based at least on scene generation information, each training image comprising a plurality of objects to be tracked; generating, for each training image, original simulated data based at least on the scene generation information, the original simulated data comprising tag data for a first object; locating, within the original simulated data, tag data for the first object, based on at least an anomaly alert (e.g., occlusion alert, proximity alert, motion alert) associated with the first object in the first training image; based at least on locating the tag data for the first object, modifying at least a portion of the tag data for the first object from the original simulated data, thereby generating preprocessed training data from the original simulated data; and training a multi-object tracking model with the preprocessed training data to produce a trained multi-object tracker.Type: ApplicationFiled: November 13, 2023Publication date: March 7, 2024Inventors: Ishani CHAKRABORTY, Jonathan C. HANZELKA, Lu YUAN, Pedro Urbina ESCOS, Thomas M. SOEMO
-
Patent number: 11880985Abstract: The disclosure herein enables tracking of multiple objects in a real-time video stream. For each individual frame received from the video stream, a frame type of the frame is determined. Based on the individual frame being an object detection frame type, a set of object proposals is detected in the individual frame, associations between the set of object proposals and a set of object tracks are assigned, and statuses of the set of object tracks are updated based on the assigned associations. Based on the individual frame being an object tracking frame type, single-object tracking is performed on the frame based on each object track of the set of object tracks and the set of object tracks is updated based on the performed single-object tracking. For each frame received, a real-time object location data stream is provided based on the set of object tracks.Type: GrantFiled: May 28, 2022Date of Patent: January 23, 2024Assignee: Microsoft Technology Licensing, LLC.Inventors: Ishani Chakraborty, Yi-Ling Chen, Lu Yuan
-
Patent number: 11854211Abstract: Training a multi-object tracking model includes: generating a plurality of training images based at least on scene generation information, each training image comprising a plurality of objects to be tracked; generating, for each training image, original simulated data based at least on the scene generation information, the original simulated data comprising tag data for a first object; locating, within the original simulated data, tag data for the first object, based on at least an anomaly alert (e.g., occlusion alert, proximity alert, motion alert) associated with the first object in the first training image; based at least on locating the tag data for the first object, modifying at least a portion of the tag data for the first object from the original simulated data, thereby generating preprocessed training data from the original simulated data; and training a multi-object tracking model with the preprocessed training data to produce a trained multi-object tracker.Type: GrantFiled: January 26, 2022Date of Patent: December 26, 2023Assignee: Microsoft Technology Licensing, LLC.Inventors: Ishani Chakraborty, Jonathan C. Hanzelka, Lu Yuan, Pedro Urbina Escos, Thomas M. Soemo
-
Publication number: 20230394855Abstract: Example solutions for image paragraph captioning use a first vision language model to generate visual information (comprising text) for an image. The visual information may include tags, an initial image caption, and information on objects within the image (e.g., further tags and captions, and object attributes and locations within the image). In some examples, the visual information further includes visual clues. A generative language model generates a plurality of image story caption candidates (e.g., descriptive paragraphs) from the visual information. A second vision language model evaluates the plurality of image story caption candidates and selects a caption as the final output caption.Type: ApplicationFiled: June 29, 2022Publication date: December 7, 2023Inventors: Yujia XIE, Lu Yuan, Nguyen BACH
-
Publication number: 20230351558Abstract: The disclosure herein describes generating an inpainted image from a masked image using a patch-based encoder and an unquantized transformer. An image including a masked region and an unmasked region is received, and the received image is divided into a plurality of patches including masked patches. The plurality of patches is encoded into a plurality of feature vectors, wherein each patch is encoded to a feature vector. Using a transformer, a predicted token is generated for each masked patch using a feature vector encoded from the masked patch, and a quantized vector of the masked patch is determined using generated predicted token and a masked patch-specific codebook. The determined quantized vector of the masked patch is included into a set of quantized vectors associated with the plurality of patches, and an output image is generated from the set of quantized vectors using a decoder.Type: ApplicationFiled: April 29, 2022Publication date: November 2, 2023Inventors: Dongdong CHEN, Xiyang DAI, Yinpeng CHEN, Mengchen LIU, Lu YUAN
-
Publication number: 20230285774Abstract: Systems and methods for generating a beam model for radiotherapy treatment planning are discussed. An exemplary system includes a memory to store a trained deep learning model, and a processor circuit to generate a beam model. The deep learning model can be trained to establish a relationship between machine scanning data and values of beam model parameters, and validated for accuracy. The processor circuit can receive machine scanning data indicative of a configuration or an operation status of the radiation therapy device, apply the machine scanning data to the trained deep learning model to determine values for the beam model parameters, and generate a beam model based on the determined values of the plurality of beam model parameters. The beam model may be provided to a user, or a treatment planning system.Type: ApplicationFiled: September 2, 2020Publication date: September 14, 2023Inventors: Shufei Chen, Lu Yuan
-
Patent number: 11751512Abstract: The present application discloses a woody rootstock for efficient grafting of solanaceous vegetables and an efficient grafting and seedling culture method thereof. According to the present application, a woody rootstock clone with high consistency is provided through tissue culture, efficient grafting is completed through sleeve grafting technology, and the grafting survival rate is improved by regulating the healing environment. A new idea for efficient industrial grafting of solanaceous vegetables is provided, scions are imparted with new features through distant grafting, and the problems of low grafting efficiency and low survival rate are solved. The method has the advantages of strong operability, simplicity, high efficiency and low cost, and provides a technical support for the industrial production of grafted seedlings of solanaceous vegetables.Type: GrantFiled: October 11, 2021Date of Patent: September 12, 2023Assignee: ZHEJIANG UNIVERSITYInventors: Liping Chen, Tingjin Wang, Lu Yuan, Ke Liu, Aijun Zhang, Yang Yang, Xuan Zhang, Yuzhuo Li, Zhenyu Qi
-
Publication number: 20230275635Abstract: Various aspects of the present disclosure generally relate to wireless communication. In some aspects, a network entity may receive a sounding reference signal (SRS) at a multi-panel system of the network entity, where the multi-panel system includes one or more sounded panels and one or more non-sounded panels. The network entity may estimate a channel to obtain channel state information (CSI) for the one or more sounded panels based at least in part on the SRS. The network entity may estimate CSI for the one or more non-sounded panels based at least in part on the CSI for the one or more sounded panels. The network entity may transmit or receive a communication based at least in part on the CSI for the one or more sounded panels and the CSI for the one or more non-sounded panels. Numerous other aspects are described.Type: ApplicationFiled: February 28, 2022Publication date: August 31, 2023Inventors: Saeid SAHRAEI, Muhammad Sayed Khairy ABDELGHAFFAR, Renqiu WANG, Lu YUAN, Joseph Patrick BURKE, Tingfang JI, Peter GAAL
-
Publication number: 20230229960Abstract: Some disclosed systems are configured to obtain a knowledge module configured to receive one or more knowledge inputs corresponding to one or more different modalities and generate a set of knowledge embeddings to be integrated with a set of multi-modal embeddings generated by a multi-modal main model. The systems receive a knowledge input at the knowledge module, identify a knowledge type associated with the knowledge input, and extract a knowledge unit from the knowledge input. The systems select a representation model that corresponds to the knowledge type and select a grounding type configured to ground the at least one knowledge unit into the representation model. The systems then ground the knowledge unit into the representation model according to the grounding type.Type: ApplicationFiled: January 19, 2022Publication date: July 20, 2023Inventors: Chenguang ZHU, Lu YUAN, Yao QIAN, Yu SHI, Nanshan ZENG, Xuedong David HUANG
-
Patent number: 11700156Abstract: An intelligent data and knowledge-driven method for modulation recognition includes the following steps: collecting spectrum data; constructing corresponding attribute vector labels for different modulation schemes; constructing and pre-training an attribute learning model based on the attribute vector labels for different modulation schemes; constructing and pre-training a visual model for modulation recognition; constructing a feature space transformation model, and constructing an intelligent data and knowledge-driven model for modulation recognition based on the attribute learning model and the visual model; transferring parameters of the pre-trained visual model and the pre-trained attribute learning model and retraining the transformation model; and determining whether training on a network is completed and outputting a classification result.Type: GrantFiled: September 2, 2022Date of Patent: July 11, 2023Assignee: Nanjing University of Aeronautics and AstronauticsInventors: Fuhui Zhou, Rui Ding, Ming Xu, Hao Zhang, Lu Yuan, Qihui Wu, Chao Dong
-
Patent number: 11593615Abstract: Image stylization is based on a learning network. A learning network is trained with a plurality of images and a reference image with a particular texture style. A plurality of different sub-networks of the learning network is trained, respectively. Specifically, one of the sub-networks is trained to extract one or more feature maps from the source image and transform the feature maps with the texture style applied thereon to a target image. Each of the feature maps indicates part of feature information of the source image. Another sub-network is trained to apply a specified texture style to the extracted feature maps, such that the target image generated based on the processed feature maps can embody the specified texture style.Type: GrantFiled: December 12, 2017Date of Patent: February 28, 2023Assignee: Microsoft Technology Licensing, LLCInventors: Gang Hua, Lu Yuan, Jing Liao, Dongdong Chen