Patents by Inventor Lu Yuan

Lu Yuan has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20250148765
    Abstract: A method for annotating images to create a corpus for training a multi-task computer vision machine learning model is presented. The method comprises receiving, at one or more annotation specialist models, a plurality of images to be annotated. Via operation of the one or more annotation specialist models, pre-filtered annotations are generated for the plurality of images. Via operation of a data filtering and enhancement module, the pre-filtered annotations are filtered in accordance with predefined noise criteria so as to output candidate annotations for the plurality of images. The method further comprises, for each of one or more candidate annotations, selectively (1) storing the candidate annotation into the corpus as a final annotation for its associated image, or (2) adding the candidate annotation to its associated image using the one or more annotation specialist models and the data filtering and enhancement module for subsequent iterative annotation and filtering.
    Type: Application
    Filed: January 30, 2024
    Publication date: May 8, 2025
    Applicant: Microsoft Technology Licensing, LLC
    Inventors: Lu YUAN, Bin XIAO, Haiping WU, Weijian XU, Xiyang DAI, Houdong HU, Yumao LU, Nanshan ZENG, Ce Christopher LIU
  • Publication number: 20250123683
    Abstract: A method and an apparatus is for adjusting a display device and includes: obtaining a first facial image of a first user; determining first spatial coordinates of a facial feature point of the first user based on the first facial image; and adjusting, based on the first spatial coordinates, an orientation of a first side of an image displayed by the display device, where the first side of the image includes information to be communicated to the first user.
    Type: Application
    Filed: November 8, 2024
    Publication date: April 17, 2025
    Inventors: Lu Yuan, Jingfang Zha
  • Patent number: 12223412
    Abstract: A computer device for automatic feature detection comprises a processor, a communication device, and a memory configured to hold instructions executable by the processor to instantiate a dynamic convolution neural network, receive input data via the communication network, and execute the dynamic convolution neural network to automatically detect features in the input data. The dynamic convolution neural network compresses the input data from an input space having a dimensionality equal to a predetermined number of channels into an intermediate space having a dimensionality less than the number of channels. The dynamic convolution neural network dynamically fuses the channels into an intermediate representation within the intermediate space and expands the intermediate representation from the intermediate space to an expanded representation in an output space having a higher dimensionality than the dimensionality of the intermediate space.
    Type: Grant
    Filed: December 16, 2020
    Date of Patent: February 11, 2025
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Yinpeng Chen, Xiyang Dai, Mengchen Liu, Dongdong Chen, Lu Yuan, Zicheng Liu, Ye Yu, Mei Chen, Yunsheng Li
  • Publication number: 20250037252
    Abstract: The disclosure herein describes generating an inpainted image from a masked image using a patch-based encoder and an unquantized transformer. An image including a masked region and an unmasked region is received, and the received image is divided into a plurality of patches including masked patches. The plurality of patches is encoded into a plurality of feature vectors, wherein each patch is encoded to a feature vector. Using a transformer, a predicted token is generated for each masked patch using a feature vector encoded from the masked patch, and a quantized vector of the masked patch is determined using generated predicted token and a masked patch-specific codebook. The determined quantized vector of the masked patch is included into a set of quantized vectors associated with the plurality of patches, and an output image is generated from the set of quantized vectors using a decoder.
    Type: Application
    Filed: October 11, 2024
    Publication date: January 30, 2025
    Inventors: Dongdong CHEN, Xiyang DAI, Yinpeng CHEN, Mengchen LIU, Lu YUAN
  • Patent number: 12190588
    Abstract: A system for tracking a target object across a plurality of image frames. The system comprises a logic machine and a storage machine. The storage machine holds instructions executable by the logic machine to calculate a trajectory for the target object over one or more previous frames occurring before a target frame. Responsive to assessing no detection of the target object in the target frame, the instructions are executable to predict an estimated region for the target object based on the trajectory, predict an occlusion center based on a set of candidate occluding locations for a set of other objects within a threshold distance of the estimated region, each location of the set of candidate occluding locations overlapping with the estimated region, and automatically estimate a bounding box for the target object in the target frame based on the occlusion center.
    Type: Grant
    Filed: June 4, 2021
    Date of Patent: January 7, 2025
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Dongdong Chen, Qiankun Liu, Lu Yuan, Lei Zhang
  • Patent number: 12148131
    Abstract: The disclosure herein describes generating an inpainted image from a masked image using a patch-based encoder and an unquantized transformer. An image including a masked region and an unmasked region is received, and the received image is divided into a plurality of patches including masked patches. The plurality of patches is encoded into a plurality of feature vectors, wherein each patch is encoded to a feature vector. Using a transformer, a predicted token is generated for each masked patch using a feature vector encoded from the masked patch, and a quantized vector of the masked patch is determined using generated predicted token and a masked patch-specific codebook. The determined quantized vector of the masked patch is included into a set of quantized vectors associated with the plurality of patches, and an output image is generated from the set of quantized vectors using a decoder.
    Type: Grant
    Filed: April 29, 2022
    Date of Patent: November 19, 2024
    Assignee: Microsoft Technology Licensing, LLC.
    Inventors: Dongdong Chen, Xiyang Dai, Yinpeng Chen, Mengchen Liu, Lu Yuan
  • Publication number: 20240221128
    Abstract: The disclosure herein describes training an encoder network to inpaint images with masked portions. A primary encoding process is used to encode a visible portion of a masked input image into encoded token data. The encoded token data is then decoded into both pixel regression output and feature prediction output, wherein both outputs include inpainted image data associated with the masked portion of the masked input image. A pixel regression loss is determined using the pixel regression output and pixel data of an unmasked version of the masked input image. A feature prediction loss is determined using the feature prediction output and ground truth encoding output of the unmasked version of the masked input image. The primary encoding process is then trained using the pixel regression loss and the feature prediction loss, whereby the primary encoding process is trained to encode structural features of input images into encoded token data.
    Type: Application
    Filed: May 19, 2022
    Publication date: July 4, 2024
    Inventors: Dongdong CHEN, Jianmin BAO, Ting ZHANG, Lu YUAN, Dong CHEN, Fang WEN, Xiaoyi DONG
  • Patent number: 12009894
    Abstract: Various aspects of the present disclosure generally relate to wireless communication. In some aspects, a network entity may receive a sounding reference signal (SRS) at a multi-panel system of the network entity, where the multi-panel system includes one or more sounded panels and one or more non-sounded panels. The network entity may estimate a channel to obtain channel state information (CSI) for the one or more sounded panels based at least in part on the SRS. The network entity may estimate CSI for the one or more non-sounded panels based at least in part on the CSI for the one or more sounded panels. The network entity may transmit or receive a communication based at least in part on the CSI for the one or more sounded panels and the CSI for the one or more non-sounded panels. Numerous other aspects are described.
    Type: Grant
    Filed: February 28, 2022
    Date of Patent: June 11, 2024
    Assignee: QUALCOMM Incorporated
    Inventors: Saeid Sahraei, Muhammad Sayed Khairy Abdelghaffar, Renqiu Wang, Lu Yuan, Joseph Patrick Burke, Tingfang Ji, Peter Gaal
  • Patent number: 11989956
    Abstract: Systems and methods for object detection generate a feature pyramid corresponding to image data, and rescaling the feature pyramid to a scale corresponding to a median level of the feature pyramid, wherein the rescaled feature pyramid is a four-dimensional (4D) tensor. The 4D tensor is reshaped into a three-dimensional (3D) tensor having individual perspectives including scale features, spatial features, and task features corresponding to different dimensions of the 3D tensor. The 3D tensor is used with a plurality of attention layers to update a plurality of feature maps associated with the image data. Object detection is performed on the image data using the updated plurality of feature maps.
    Type: Grant
    Filed: April 5, 2021
    Date of Patent: May 21, 2024
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Xiyang Dai, Yinpeng Chen, Bin Xiao, Dongdong Chen, Mengchen Liu, Lu Yuan, Lei Zhang
  • Publication number: 20240078682
    Abstract: Training a multi-object tracking model includes: generating a plurality of training images based at least on scene generation information, each training image comprising a plurality of objects to be tracked; generating, for each training image, original simulated data based at least on the scene generation information, the original simulated data comprising tag data for a first object; locating, within the original simulated data, tag data for the first object, based on at least an anomaly alert (e.g., occlusion alert, proximity alert, motion alert) associated with the first object in the first training image; based at least on locating the tag data for the first object, modifying at least a portion of the tag data for the first object from the original simulated data, thereby generating preprocessed training data from the original simulated data; and training a multi-object tracking model with the preprocessed training data to produce a trained multi-object tracker.
    Type: Application
    Filed: November 13, 2023
    Publication date: March 7, 2024
    Inventors: Ishani CHAKRABORTY, Jonathan C. HANZELKA, Lu YUAN, Pedro Urbina ESCOS, Thomas M. SOEMO
  • Patent number: 11880985
    Abstract: The disclosure herein enables tracking of multiple objects in a real-time video stream. For each individual frame received from the video stream, a frame type of the frame is determined. Based on the individual frame being an object detection frame type, a set of object proposals is detected in the individual frame, associations between the set of object proposals and a set of object tracks are assigned, and statuses of the set of object tracks are updated based on the assigned associations. Based on the individual frame being an object tracking frame type, single-object tracking is performed on the frame based on each object track of the set of object tracks and the set of object tracks is updated based on the performed single-object tracking. For each frame received, a real-time object location data stream is provided based on the set of object tracks.
    Type: Grant
    Filed: May 28, 2022
    Date of Patent: January 23, 2024
    Assignee: Microsoft Technology Licensing, LLC.
    Inventors: Ishani Chakraborty, Yi-Ling Chen, Lu Yuan
  • Patent number: 11854211
    Abstract: Training a multi-object tracking model includes: generating a plurality of training images based at least on scene generation information, each training image comprising a plurality of objects to be tracked; generating, for each training image, original simulated data based at least on the scene generation information, the original simulated data comprising tag data for a first object; locating, within the original simulated data, tag data for the first object, based on at least an anomaly alert (e.g., occlusion alert, proximity alert, motion alert) associated with the first object in the first training image; based at least on locating the tag data for the first object, modifying at least a portion of the tag data for the first object from the original simulated data, thereby generating preprocessed training data from the original simulated data; and training a multi-object tracking model with the preprocessed training data to produce a trained multi-object tracker.
    Type: Grant
    Filed: January 26, 2022
    Date of Patent: December 26, 2023
    Assignee: Microsoft Technology Licensing, LLC.
    Inventors: Ishani Chakraborty, Jonathan C. Hanzelka, Lu Yuan, Pedro Urbina Escos, Thomas M. Soemo
  • Publication number: 20230394855
    Abstract: Example solutions for image paragraph captioning use a first vision language model to generate visual information (comprising text) for an image. The visual information may include tags, an initial image caption, and information on objects within the image (e.g., further tags and captions, and object attributes and locations within the image). In some examples, the visual information further includes visual clues. A generative language model generates a plurality of image story caption candidates (e.g., descriptive paragraphs) from the visual information. A second vision language model evaluates the plurality of image story caption candidates and selects a caption as the final output caption.
    Type: Application
    Filed: June 29, 2022
    Publication date: December 7, 2023
    Inventors: Yujia XIE, Lu Yuan, Nguyen BACH
  • Publication number: 20230351558
    Abstract: The disclosure herein describes generating an inpainted image from a masked image using a patch-based encoder and an unquantized transformer. An image including a masked region and an unmasked region is received, and the received image is divided into a plurality of patches including masked patches. The plurality of patches is encoded into a plurality of feature vectors, wherein each patch is encoded to a feature vector. Using a transformer, a predicted token is generated for each masked patch using a feature vector encoded from the masked patch, and a quantized vector of the masked patch is determined using generated predicted token and a masked patch-specific codebook. The determined quantized vector of the masked patch is included into a set of quantized vectors associated with the plurality of patches, and an output image is generated from the set of quantized vectors using a decoder.
    Type: Application
    Filed: April 29, 2022
    Publication date: November 2, 2023
    Inventors: Dongdong CHEN, Xiyang DAI, Yinpeng CHEN, Mengchen LIU, Lu YUAN
  • Publication number: 20230285774
    Abstract: Systems and methods for generating a beam model for radiotherapy treatment planning are discussed. An exemplary system includes a memory to store a trained deep learning model, and a processor circuit to generate a beam model. The deep learning model can be trained to establish a relationship between machine scanning data and values of beam model parameters, and validated for accuracy. The processor circuit can receive machine scanning data indicative of a configuration or an operation status of the radiation therapy device, apply the machine scanning data to the trained deep learning model to determine values for the beam model parameters, and generate a beam model based on the determined values of the plurality of beam model parameters. The beam model may be provided to a user, or a treatment planning system.
    Type: Application
    Filed: September 2, 2020
    Publication date: September 14, 2023
    Inventors: Shufei Chen, Lu Yuan
  • Patent number: 11751512
    Abstract: The present application discloses a woody rootstock for efficient grafting of solanaceous vegetables and an efficient grafting and seedling culture method thereof. According to the present application, a woody rootstock clone with high consistency is provided through tissue culture, efficient grafting is completed through sleeve grafting technology, and the grafting survival rate is improved by regulating the healing environment. A new idea for efficient industrial grafting of solanaceous vegetables is provided, scions are imparted with new features through distant grafting, and the problems of low grafting efficiency and low survival rate are solved. The method has the advantages of strong operability, simplicity, high efficiency and low cost, and provides a technical support for the industrial production of grafted seedlings of solanaceous vegetables.
    Type: Grant
    Filed: October 11, 2021
    Date of Patent: September 12, 2023
    Assignee: ZHEJIANG UNIVERSITY
    Inventors: Liping Chen, Tingjin Wang, Lu Yuan, Ke Liu, Aijun Zhang, Yang Yang, Xuan Zhang, Yuzhuo Li, Zhenyu Qi
  • Publication number: 20230275635
    Abstract: Various aspects of the present disclosure generally relate to wireless communication. In some aspects, a network entity may receive a sounding reference signal (SRS) at a multi-panel system of the network entity, where the multi-panel system includes one or more sounded panels and one or more non-sounded panels. The network entity may estimate a channel to obtain channel state information (CSI) for the one or more sounded panels based at least in part on the SRS. The network entity may estimate CSI for the one or more non-sounded panels based at least in part on the CSI for the one or more sounded panels. The network entity may transmit or receive a communication based at least in part on the CSI for the one or more sounded panels and the CSI for the one or more non-sounded panels. Numerous other aspects are described.
    Type: Application
    Filed: February 28, 2022
    Publication date: August 31, 2023
    Inventors: Saeid SAHRAEI, Muhammad Sayed Khairy ABDELGHAFFAR, Renqiu WANG, Lu YUAN, Joseph Patrick BURKE, Tingfang JI, Peter GAAL
  • Publication number: 20230229960
    Abstract: Some disclosed systems are configured to obtain a knowledge module configured to receive one or more knowledge inputs corresponding to one or more different modalities and generate a set of knowledge embeddings to be integrated with a set of multi-modal embeddings generated by a multi-modal main model. The systems receive a knowledge input at the knowledge module, identify a knowledge type associated with the knowledge input, and extract a knowledge unit from the knowledge input. The systems select a representation model that corresponds to the knowledge type and select a grounding type configured to ground the at least one knowledge unit into the representation model. The systems then ground the knowledge unit into the representation model according to the grounding type.
    Type: Application
    Filed: January 19, 2022
    Publication date: July 20, 2023
    Inventors: Chenguang ZHU, Lu YUAN, Yao QIAN, Yu SHI, Nanshan ZENG, Xuedong David HUANG
  • Patent number: 11700156
    Abstract: An intelligent data and knowledge-driven method for modulation recognition includes the following steps: collecting spectrum data; constructing corresponding attribute vector labels for different modulation schemes; constructing and pre-training an attribute learning model based on the attribute vector labels for different modulation schemes; constructing and pre-training a visual model for modulation recognition; constructing a feature space transformation model, and constructing an intelligent data and knowledge-driven model for modulation recognition based on the attribute learning model and the visual model; transferring parameters of the pre-trained visual model and the pre-trained attribute learning model and retraining the transformation model; and determining whether training on a network is completed and outputting a classification result.
    Type: Grant
    Filed: September 2, 2022
    Date of Patent: July 11, 2023
    Assignee: Nanjing University of Aeronautics and Astronautics
    Inventors: Fuhui Zhou, Rui Ding, Ming Xu, Hao Zhang, Lu Yuan, Qihui Wu, Chao Dong
  • Patent number: 11593615
    Abstract: Image stylization is based on a learning network. A learning network is trained with a plurality of images and a reference image with a particular texture style. A plurality of different sub-networks of the learning network is trained, respectively. Specifically, one of the sub-networks is trained to extract one or more feature maps from the source image and transform the feature maps with the texture style applied thereon to a target image. Each of the feature maps indicates part of feature information of the source image. Another sub-network is trained to apply a specified texture style to the extracted feature maps, such that the target image generated based on the processed feature maps can embody the specified texture style.
    Type: Grant
    Filed: December 12, 2017
    Date of Patent: February 28, 2023
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Gang Hua, Lu Yuan, Jing Liao, Dongdong Chen