Patents by Inventor Jifeng DAI

Jifeng DAI has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20230316742
    Abstract: The embodiments of the present application discloses an image processing method, an image processing apparatus, an image processing device and a computer-readable storage medium.
    Type: Application
    Filed: February 2, 2023
    Publication date: October 5, 2023
    Applicant: Shanghai Artificial Intelligence Innovation Center
    Inventors: Hongyang LI, Zhiqi LI, Wenhai WANG, Chonghao SIMA, Li CHEN, Yang LI, Yu QIAO, Jifeng DAI
  • Patent number: 10152627
    Abstract: Various embodiments herein each include at least one of systems, methods, and software for feature flow for video recognition. Such embodiments generally include a fast and accurate framework for video recognition. One example method includes receiving a first frame captured by an imaging device and designating the first frame as a key frame. The method may then generate at least one feature map to identify features in the key frame and subsequently receive a second frame. The method also includes designating the second frame as a current frame and applying a flow estimation algorithm to the key frame and current frame to generate a flow field representing a flow from the key frame to the current frame. The method then propagates each of the at least one feature maps based on the flow field to approximate current locations of features identified within each of the at least one feature maps.
    Type: Grant
    Filed: March 20, 2017
    Date of Patent: December 11, 2018
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Yichen Wei, Lu Yuan, Jifeng Dai
  • Publication number: 20180268208
    Abstract: Various embodiments herein each include at least one of systems, methods, and software for feature flow for video recognition. Such embodiments generally include a fast and accurate framework for video recognition. One example method includes receiving a first frame captured by an imaging device and designating the first frame as a key frame. The method may then generate at least one feature map to identify features in the key frame and subsequently receive a second frame. The method also includes designating the second frame as a current frame and applying a flow estimation algorithm to the key frame and current frame to generate a flow field representing a flow from the key frame to the current frame. The method then propagates each of the at least one feature maps based on the flow field to approximate current locations of features identified within each of the at least one feature maps.
    Type: Application
    Filed: March 20, 2017
    Publication date: September 20, 2018
    Inventors: Yichen Wei, Lu Yuan, Jifeng Dai
  • Patent number: 9865042
    Abstract: In implementations of the subject matter described herein, the feature maps are obtained by convoluting an input image using a plurality of layers of convolution filters. The feature maps record semantic information for respective regions on the image and only need to be computed once. Segment features of the image are extracted from the convolutional feature maps. Particularly, the binary masks may be obtained from a set of candidate segments of the image. The binary masks are used to mask the feature maps instead of the raw image. The masked feature maps define the segment features. The semantic segmentation of the image is done by determining a semantic category for each pixel in the image at least in part based on the resulting segment features.
    Type: Grant
    Filed: July 17, 2015
    Date of Patent: January 9, 2018
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Jifeng Dai, Kaiming He, Jian Sun
  • Patent number: 9858525
    Abstract: Disclosed herein are technologies directed to training a neural network to perform semantic segmentation. A system receives a training image, and using the training image, candidate masks are generated. The candidate masks are ranked and a set of the ranked candidate masks are selected for further processing. One of the set of the ranked candidate masks is selected to train the neural network. The one of the set of the set of the ranked candidate masks is also used as an input to train the neural network in a further training evolution. In some examples, the one of the set of the ranked candidate masks is selected randomly to reduce the likelihood of ending up in poor local optima that result in poor training inputs.
    Type: Grant
    Filed: October 14, 2015
    Date of Patent: January 2, 2018
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Jifeng Dai, Kaiming He, Jian Sun
  • Publication number: 20170109625
    Abstract: Disclosed herein are technologies directed to training a neural network to perform semantic segmentation. A system receives a training image, and using the training image, candidate masks are generated. The candidate masks are ranked and a set of the ranked candidate masks are selected for further processing. One of the set of the ranked candidate masks is selected to train the neural network. The one of the set of the set of the ranked candidate masks is also used as an input to train the neural network in a further training evolution. In some examples, the one of the set of the ranked candidate masks is selected randomly to reduce the likelihood of ending up in poor local optima that result in poor training inputs.
    Type: Application
    Filed: October 14, 2015
    Publication date: April 20, 2017
    Inventors: Jifeng Dai, Kaiming He, Jian Sun
  • Publication number: 20160358337
    Abstract: In implementations of the subject matter described herein, the feature maps are obtained by convoluting an input image using a plurality of layers of convolution filters. The feature maps record semantic information for respective regions on the image and only need to be computed once. Segment features of the image are extracted from the convolutional feature maps. Particularly, the binary masks may be obtained from a set of candidate segments of the image. The binary masks are used to mask the feature maps instead of the raw image. The masked feature maps define the segment features. The semantic segmentation of the image is done by determining a semantic category for each pixel in the image at least in part based on the resulting segment features.
    Type: Application
    Filed: July 17, 2015
    Publication date: December 8, 2016
    Inventors: Jifeng DAI, Kaiming HE, Jian SUN