Patents by Inventor Xiaohui Shen

Xiaohui Shen has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20190333198
    Abstract: The present disclosure relates to training and utilizing an image exposure transformation network to generate a long-exposure image from a single short-exposure image (e.g., still image). In various embodiments, the image exposure transformation network is trained using adversarial learning, long-exposure ground truth images, and a multi-term loss function. In some embodiments, the image exposure transformation network includes an optical flow prediction network and/or an appearance guided attention network. Trained embodiments of the image exposure transformation network generate realistic long-exposure images from single short-exposure images without additional information.
    Type: Application
    Filed: April 25, 2018
    Publication date: October 31, 2019
    Inventors: Yilin Wang, Zhe Lin, Zhaowen Wang, Xin Lu, Xiaohui Shen, Chih-Yao Hsieh
  • Patent number: 10460154
    Abstract: Methods and systems for recognizing people in images with increased accuracy are disclosed. In particular, the methods and systems divide images into a plurality of clusters based on common characteristics of the images. The methods and systems also determine an image cluster to which an image with an unknown person instance most corresponds. One or more embodiments determine a probability that the unknown person instance is each known person instance in the image cluster using a trained cluster classifier of the image cluster. Optionally, the methods and systems determine context weights for each combination of an unknown person instance and each known person instance using a conditional random field algorithm based on a plurality of context cues associated with the unknown person instance and the known person instances. The methods and systems calculate a contextual probability based on the cluster-based probabilities and context weights to identify the unknown person instance.
    Type: Grant
    Filed: July 30, 2018
    Date of Patent: October 29, 2019
    Assignee: Adobe Inc.
    Inventors: Jonathan Brandt, Zhe Lin, Xiaohui Shen, Haoxiang Li
  • Patent number: 10460214
    Abstract: Systems, methods, and non-transitory computer-readable media are disclosed for segmenting objects in digital visual media utilizing one or more salient content neural networks. In particular, in one or more embodiments, the disclosed systems and methods train one or more salient content neural networks to efficiently identify foreground pixels in digital visual media. Moreover, in one or more embodiments, the disclosed systems and methods provide a trained salient content neural network to a mobile device, allowing the mobile device to directly select salient objects in digital visual media utilizing a trained neural network. Furthermore, in one or more embodiments, the disclosed systems and methods train and provide multiple salient content neural networks, such that mobile devices can identify objects in real-time digital visual media feeds (utilizing a first salient content neural network) and identify objects in static digital images (utilizing a second salient content neural network).
    Type: Grant
    Filed: October 31, 2017
    Date of Patent: October 29, 2019
    Assignee: Adobe Inc.
    Inventors: Xin Lu, Zhe Lin, Xiaohui Shen, Jimei Yang, Jianming Zhang, Jen-Chan Jeff Chien, Chenxi Liu
  • Publication number: 20190295223
    Abstract: Methods and systems are provided for generating enhanced image. A neural network system is trained where the training includes training a first neural network that generates enhanced images conditioned on content of an image undergoing enhancement and training a second neural network that designates realism of the enhanced images generated by the first neural network. The neural network system is trained by determine loss and accordingly adjusting the appropriate neural network(s). The trained neural network system is used to generate an enhanced aesthetic image from a selected image where the output enhanced aesthetic image has increased aesthetics when compared to the selected image.
    Type: Application
    Filed: March 22, 2018
    Publication date: September 26, 2019
    Inventors: Xiaohui Shen, Zhe Lin, Xin Lu, Sarah Aye Kong, I-Ming Pao, Yingcong Chen
  • Publication number: 20190287283
    Abstract: Certain embodiments involve using an image completion neural network to perform user-guided image completion. For example, an image editing application accesses an input image having a completion region to be replaced with new image content. The image editing application also receives a guidance input that is applied to a portion of a completion region. The image editing application provides the input image and the guidance input to an image completion neural network that is trained to perform image-completion operations using guidance input. The image editing application produces a modified image by replacing the completion region of the input image with the new image content generated with the image completion network. The image editing application outputs the modified image having the new image content.
    Type: Application
    Filed: March 15, 2018
    Publication date: September 19, 2019
    Inventors: Zhe Lin, Xin Lu, Xiaohui Shen, Jimei Yang, Jiahui Yu
  • Publication number: 20190279346
    Abstract: Certain embodiments involve blending images using neural networks to automatically generate alignment or photometric adjustments that control image blending operations. For instance, a foreground image and a background image data are provided to an adjustment-prediction network that has been trained, using a reward network, to compute alignment or photometric adjustments that optimize blending reward scores. An adjustment action (e.g., an alignment or photometric adjustment) is computed by applying the adjustment-prediction network to the foreground image and the background image data. A target background region is extracted from the background image data by applying the adjustment action to the background image data. The target background region is blended with the foreground image, and the resultant blended image is outputted.
    Type: Application
    Filed: March 7, 2018
    Publication date: September 12, 2019
    Inventors: Jianming Zhang, Zhe Lin, Xiaohui Shen, Wei-Chih Hung, Joon-Young Lee
  • Publication number: 20190279074
    Abstract: Semantic segmentation techniques and systems are described that overcome the challenges of limited availability of training data to describe the potentially millions of tags that may be used to describe semantic classes in digital images. In one example, the techniques are configured to train neural networks to leverage different types of training datasets using sequential neural networks and use of vector representations to represent the different semantic classes.
    Type: Application
    Filed: March 6, 2018
    Publication date: September 12, 2019
    Applicant: Adobe Inc.
    Inventors: Zhe Lin, Yufei Wang, Xiaohui Shen, Scott David Cohen, Jianming Zhang
  • Patent number: 10410351
    Abstract: The invention is directed towards segmenting images based on natural language phrases. An image and an n-gram, including a sequence of tokens, are received. An encoding of image features and a sequence of token vectors are generated. A fully convolutional neural network identifies and encodes the image features. A word embedding model generates the token vectors. A recurrent neural network (RNN) iteratively updates a segmentation map based on combinations of the image feature encoding and the token vectors. The segmentation map identifies which pixels are included in an image region referenced by the n-gram. A segmented image is generated based on the segmentation map. The RNN may be a convolutional multimodal RNN. A separate RNN, such as a long short-term memory network, may iteratively update an encoding of semantic features based on the order of tokens. The first RNN may update the segmentation map based on the semantic feature encoding.
    Type: Grant
    Filed: August 29, 2018
    Date of Patent: September 10, 2019
    Assignee: ADOBE INC.
    Inventors: Zhe Lin, Xin Lu, Xiaohui Shen, Jimei Yang, Chenxi Liu
  • Publication number: 20190258925
    Abstract: This disclosure covers methods, non-transitory computer readable media, and systems that learn attribute attention projections for attributes of digital images and parameters for an attention controlled neural network. By iteratively generating and comparing attribute-modulated-feature vectors from digital images, the methods, non-transitory computer readable media, and systems update attribute attention projections and parameters indicating either one (or both) of a correlation between some attributes of digital images and a discorrelation between other attributes of digital images. In certain embodiments, the methods, non-transitory computer readable media, and systems use the attribute attention projections in an attention controlled neural network as part of performing one or more tasks.
    Type: Application
    Filed: February 20, 2018
    Publication date: August 22, 2019
    Inventors: Haoxiang Li, Xiaohui Shen, Xiangyun Zhao
  • Patent number: 10387776
    Abstract: Provided are systems and techniques that provide an output phrase describing an image. An example method includes creating, with a convolutional neural network, feature maps describing image features in locations in the image. The method also includes providing a skeletal phrase for the image by processing the feature maps with a first long short-term memory (LSTM) neural network trained based on a first set of ground truth phrases which exclude attribute words. Then, attribute words are provided by processing the skeletal phrase and the feature maps with a second LSTM neural network trained based on a second set of ground truth phrases including words for attributes. Then, the method combines the skeletal phrase and the attribute words to form the output phrase.
    Type: Grant
    Filed: March 10, 2017
    Date of Patent: August 20, 2019
    Assignee: Adobe Inc.
    Inventors: Zhe Lin, Yufei Wang, Scott Cohen, Xiaohui Shen
  • Publication number: 20190252002
    Abstract: The present disclosure is directed toward systems and methods for tracking objects in videos. For example, one or more embodiments described herein utilize various tracking methods in combination with an image search index made up of still video frames indexed from a video. One or more embodiments described herein utilize a backward and forward tracking method that is anchored by one or more key frames in order to accurately track an object through the frames of a video, even when the video is long and may include challenging conditions.
    Type: Application
    Filed: April 25, 2019
    Publication date: August 15, 2019
    Inventors: Zhihong Ding, Zhe Lin, Xiaohui Shen, Michael Kaplan, Jonathan Brandt
  • Publication number: 20190244327
    Abstract: Image cropping suggestion using multiple saliency maps is described. In one or more implementations, component scores, indicative of visual characteristics established for visually-pleasing croppings, are computed for candidate image croppings using multiple different saliency maps. The visual characteristics on which a candidate image cropping is scored may be indicative of its composition quality, an extent to which it preserves content appearing in the scene, and a simplicity of its boundary. Based on the component scores, the croppings may be ranked with regard to each of the visual characteristics. The rankings may be used to cluster the candidate croppings into groups of similar croppings, such that croppings in a group are different by less than a threshold amount and croppings in different groups are different by at least the threshold amount. Based on the clustering, croppings may then be chosen, e.g., to present them to a user for selection.
    Type: Application
    Filed: April 15, 2019
    Publication date: August 8, 2019
    Applicant: Adobe Inc.
    Inventors: Zhe Lin, Radomir Mech, Xiaohui Shen, Brian L. Price, Jianming Zhang, Anant Gilra, Jen-Chan Jeff Chien
  • Patent number: 10361784
    Abstract: An example remote radio apparatus is provided, including a body, a mainboard, a mainboard heat sink, a maintenance cavity, an optical module, and an optical module heat sink. The maintenance cavity and the optical module heat sink are integrally connected, while the optical module is mounted on a bottom surface of the optical module heat sink. The maintenance cavity and the optical module heat sink are mounted on a side surface of the body, and the mainboard heat sink is mounted on and covers the mainboard. The mainboard heat sink and the mainboard are installed on a front surface of the body, and the mainboard heat sink and the optical module heat sink are spaced by a preset distance. The temperature of the optical module is controlled within a range required by a specification.
    Type: Grant
    Filed: June 7, 2018
    Date of Patent: July 23, 2019
    Assignee: Huawei Technologies Co., Ltd.
    Inventors: Xiaoming Shi, Xiaohui Shen, Dan Liang, Haigang Xiong, Haizheng Tang
  • Publication number: 20190213474
    Abstract: Various embodiments describe frame selection based on training and using a neural network. In an example, the neural network is a convolutional neural network trained with training pairs. Each training pair includes two training frames from a frame collection. The loss function relies on the estimated quality difference between the two training frames. Further, the definition of the loss function varies based on the actual quality difference between these two frames. In a further example, the neural network is trained by incorporating facial heatmaps generated from the training frames and facial quality scores of faces detected in the training frames. In addition, the training involves using a feature mean that represents an average of the features of the training frames belonging to the same frame collection. Once the neural network is trained, a frame collection is input thereto and a frame is selected based on generated quality scores.
    Type: Application
    Filed: January 9, 2018
    Publication date: July 11, 2019
    Inventors: Zhe Lin, Xiaohui Shen, Radomir Mech, Jian Ren
  • Patent number: 10346996
    Abstract: Image depth inference techniques and systems from semantic labels are described. In one or more implementations, a digital medium environment includes one or more computing devices to control a determination of depth within an image. Regions of the image are semantically labeled by the one or more computing devices. At least one of the semantically labeled regions is decomposed into a plurality of segments formed as planes generally perpendicular to a ground plane of the image. Depth of one or more of the plurality of segments is then inferred based on relationships of respective segments with respective locations of the ground plane of the image. A depth map is formed that describes depth for the at least one semantically labeled region based at least in part on the inferred depths for the one or more of the plurality of segments.
    Type: Grant
    Filed: August 21, 2015
    Date of Patent: July 9, 2019
    Assignee: Adobe Inc.
    Inventors: Xiaohui Shen, Zhe Lin, Scott D. Cohen, Brian L. Price
  • Patent number: 10346951
    Abstract: Image cropping suggestion using multiple saliency maps is described. In one or more implementations, component scores, indicative of visual characteristics established for visually-pleasing croppings, are computed for candidate image croppings using multiple different saliency maps. The visual characteristics on which a candidate image cropping is scored may be indicative of its composition quality, an extent to which it preserves content appearing in the scene, and a simplicity of its boundary. Based on the component scores, the croppings may be ranked with regard to each of the visual characteristics. The rankings may be used to cluster the candidate croppings into groups of similar croppings, such that croppings in a group are different by less than a threshold amount and croppings in different groups are different by at least the threshold amount. Based on the clustering, croppings may then be chosen, e.g., to present them to a user for selection.
    Type: Grant
    Filed: March 2, 2017
    Date of Patent: July 9, 2019
    Assignee: Adobe Inc.
    Inventors: Zhe Lin, Radomir Mech, Xiaohui Shen, Brian L. Price, Jianming Zhang, Anant Gilra, Jen-Chan Jeff Chien
  • Patent number: 10319412
    Abstract: The present disclosure is directed toward systems and methods for tracking objects in videos. For example, one or more embodiments described herein utilize various tracking methods in combination with an image search index made up of still video frames indexed from a video. One or more embodiments described herein utilize a backward and forward tracking method that is anchored by one or more key frames in order to accurately track an object through the frames of a video, even when the video is long and may include challenging conditions.
    Type: Grant
    Filed: November 16, 2016
    Date of Patent: June 11, 2019
    Assignee: ADOBE INC.
    Inventors: Zhihong Ding, Zhe Lin, Xiaohui Shen, Michael Kaplan, Jonathan Brandt
  • Patent number: 10311574
    Abstract: A digital medium environment includes an image processing application that performs object segmentation on an input image. An improved object segmentation method implemented by the image processing application comprises receiving an input image that includes an object region to be segmented by a segmentation process, processing the input image to provide a first segmentation that defines the object region, and processing the first segmentation to provide a second segmentation that provides pixel-wise label assignments for the object region. In some implementations, the image processing application performs improved sky segmentation on an input image containing a depiction of a sky.
    Type: Grant
    Filed: December 22, 2017
    Date of Patent: June 4, 2019
    Assignee: Adobe Inc.
    Inventors: Xiaohui Shen, Zhe Lin, Yi-Hsuan Tsai, Kalyan K. Sunkavalli
  • Publication number: 20190164261
    Abstract: Systems and techniques for estimating illumination from a single image are provided. An example system may include a neural network. The neural network may include an encoder that is configured to encode an input image into an intermediate representation. The neural network may also include an intensity decoder that is configured to decode the intermediate representation into an output light intensity map. An example intensity decoder is generated by a multi-phase training process that includes a first phase to train a light mask decoder using a set of low dynamic range images and a second phase to adjust parameters of the light mask decoder using a set of high dynamic range image to generate the intensity decoder.
    Type: Application
    Filed: November 28, 2017
    Publication date: May 30, 2019
    Inventors: Kalyan Sunkavalli, Mehmet Ersin Yumer, Marc-Andre Gardner, Xiaohui Shen, Jonathan Eisenmann, Emiliano Gambaretto
  • Publication number: 20190147224
    Abstract: Approaches are described for determining facial landmarks in images. An input image is provided to at least one trained neural network that determines a face region (e.g., bounding box of a face) of the input image and initial facial landmark locations corresponding to the face region. The initial facial landmark locations are provided to a 3D face mapper that maps the initial facial landmark locations to a 3D face model. A set of facial landmark locations are determined from the 3D face model. The set of facial landmark locations are provided to a landmark location adjuster that adjusts positions of the set of facial landmark locations based on the input image. The input image is presented on a user device using the adjusted set of facial landmark locations.
    Type: Application
    Filed: November 16, 2017
    Publication date: May 16, 2019
    Inventors: HAOXIANG LI, ZHE LIN, JONATHAN BRANDT, XIAOHUI SHEN