Patents by Inventor Hailin Jin

Hailin Jin has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11403739
    Abstract: Methods and apparatus for retargeting and prioritized interpolation of lens profiles. A lens profile file may include a set of lens sub-profiles. The camera body and/or settings described in the file may not exactly match that of camera body and/or settings used to capture a target image. A sub-profile processing module may perform a prioritized sub-profile sorting and interpolation method to generate an interpolated sub-profile that may be applied to the target image to correct aberrations including, but not limited to, geometric distortion, lateral chromatic aberration, and vignette. Thus, models generated for a reference camera at a variety of settings may be applied to a target image captured with the same type of lens but with a different camera and/or with different settings that are not exactly modeled in the lens profile file.
    Type: Grant
    Filed: April 12, 2010
    Date of Patent: August 2, 2022
    Assignee: Adobe Inc.
    Inventors: Simon Chen, Eric Chan, Hailin Jin, Jen-Chan Chien
  • Patent number: 11393158
    Abstract: Systems, methods, and non-transitory computer-readable media are disclosed for utilizing an encoder-decoder architecture to learn a volumetric 3D representation of an object using digital images of the object from multiple viewpoints to render novel views of the object. For instance, the disclosed systems can utilize patch-based image feature extraction to extract lifted feature representations from images corresponding to different viewpoints of an object. Furthermore, the disclosed systems can model view-dependent transformed feature representations using learned transformation kernels. In addition, the disclosed systems can recurrently and concurrently aggregate the transformed feature representations to generate a 3D voxel representation of the object. Furthermore, the disclosed systems can sample frustum features using the 3D voxel representation and transformation kernels.
    Type: Grant
    Filed: April 2, 2020
    Date of Patent: July 19, 2022
    Assignee: Adobe Inc.
    Inventors: Tong He, John Collomosse, Hailin Jin
  • Patent number: 11392806
    Abstract: Systems and methods provide for generating glyph initiations using a generative font system. A glyph variant may be generated based on an input vector glyph. A plurality of line segments may be approximated using a differentiable rasterizer with the plurality of line segments representing the contours of the glyph variant. A bitmap of the glyph variant may then be generated based on the line segments. The image loss between the bitmap and a rasterized representation of a vector glyph may be calculated and provided to the generative font system. Based on the image loss, a refined glyph variant may be provided to a user.
    Type: Grant
    Filed: February 12, 2020
    Date of Patent: July 19, 2022
    Assignee: Adobe Inc.
    Inventors: Zhaowen Wang, Zhifei Zhang, Xuan Li, Matthew Fisher, Hailin Jin
  • Publication number: 20220148325
    Abstract: The present disclosure relates to a tag-based font recognition system that utilizes a multi-learning framework to develop and improve tag-based font recognition using deep learning neural networks. In particular, the tag-based font recognition system jointly trains a font tag recognition neural network with an implicit font classification attention model to generate font tag probability vectors that are enhanced by implicit font classification information. Indeed, the font recognition system weights the hidden layers of the font tag recognition neural network with implicit font information to improve the accuracy and predictability of the font tag recognition neural network, which results in improved retrieval of fonts in response to a font tag query. Accordingly, using the enhanced tag probability vectors, the tag-based font recognition system can accurately identify and recommend one or more fonts in response to a font tag query.
    Type: Application
    Filed: January 26, 2022
    Publication date: May 12, 2022
    Inventors: Zhaowen Wang, Tianlang Chen, Ning Xu, Hailin Jin
  • Publication number: 20220122357
    Abstract: The present disclosure relates to systems, methods, and non-transitory computer-readable media for generating a response to a question received from a user during display or playback of a video segment by utilizing a query-response-neural network. The disclosed systems can extract a query vector from a question corresponding to the video segment using the query-response-neural network. The disclosed systems further generate context vectors representing both visual cues and transcript cues corresponding to the video segment using context encoders or other layers from the query-response-neural network. By utilizing additional layers from the query-response-neural network, the disclosed systems generate (i) a query-context vector based on the query vector and the context vectors, and (ii) candidate-response vectors representing candidate responses to the question from a domain-knowledge base or other source.
    Type: Application
    Filed: December 28, 2021
    Publication date: April 21, 2022
    Inventors: Wentian Zhao, Seokhwan Kim, Ning Xu, Hailin Jin
  • Publication number: 20220092108
    Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media for accurately and flexibly identifying digital images with similar style to a query digital image using fine-grain style determination via weakly supervised style extraction neural networks. For example, the disclosed systems can extract a style embedding from a query digital image using a style extraction neural network such as a novel two-branch autoencoder architecture or a weakly supervised discriminative neural network. The disclosed systems can generate a combined style embedding by combining complementary style embeddings from different style extraction neural networks. Moreover, the disclosed systems can search a repository of digital images to identify digital images with similar style to the query digital image.
    Type: Application
    Filed: September 18, 2020
    Publication date: March 24, 2022
    Inventors: John Collomosse, Zhe Lin, Saeid Motiian, Hailin Jin, Baldo Faieta, Alex Filipkowski
  • Patent number: 11244207
    Abstract: The present disclosure relates to a tag-based font recognition system that utilizes a multi-learning framework to develop and improve tag-based font recognition using deep learning neural networks. In particular, the tag-based font recognition system jointly trains a font tag recognition neural network with an implicit font classification attention model to generate font tag probability vectors that are enhanced by implicit font classification information. Indeed, the font recognition system weights the hidden layers of the font tag recognition neural network with implicit font information to improve the accuracy and predictability of the font tag recognition neural network, which results in improved retrieval of fonts in response to a font tag query. Accordingly, using the enhanced tag probability vectors, the tag-based font recognition system can accurately identify and recommend one or more fonts in response to a font tag query.
    Type: Grant
    Filed: November 23, 2020
    Date of Patent: February 8, 2022
    Assignee: Adobe Inc.
    Inventors: Zhaowen Wang, Tianlang Chen, Ning Xu, Hailin Jin
  • Patent number: 11244167
    Abstract: The present disclosure relates to systems, methods, and non-transitory computer-readable media for generating a response to a question received from a user during display or playback of a video segment by utilizing a query-response-neural network. The disclosed systems can extract a query vector from a question corresponding to the video segment using the query-response-neural network. The disclosed systems further generate context vectors representing both visual cues and transcript cues corresponding to the video segment using context encoders or other layers from the query-response-neural network. By utilizing additional layers from the query-response-neural network, the disclosed systems generate (i) a query-context vector based on the query vector and the context vectors, and (ii) candidate-response vectors representing candidate responses to the question from a domain-knowledge base or other source.
    Type: Grant
    Filed: February 6, 2020
    Date of Patent: February 8, 2022
    Assignee: Adobe Inc.
    Inventors: Wentian Zhao, Seokhwan Kim, Ning Xu, Hailin Jin
  • Patent number: 11238362
    Abstract: Modeling semantic concepts in an embedding space as distributions is described. In the embedding space, both images and text labels are represented. The text labels describe semantic concepts that are exhibited in image content. In the embedding space, the semantic concepts described by the text labels are modeled as distributions. By using distributions, each semantic concept is modeled as a continuous cluster which can overlap other clusters that model other semantic concepts. For example, a distribution for the semantic concept “apple” can overlap distributions for the semantic concepts “fruit” and “tree” since can refer to both a fruit and a tree. In contrast to using distributions, conventionally configured visual-semantic embedding spaces represent a semantic concept as a single point. Thus, unlike these conventionally configured embedding spaces, the embedding spaces described herein are generated to model semantic concepts as distributions, such as Gaussian distributions, Gaussian mixtures, and so on.
    Type: Grant
    Filed: January 15, 2016
    Date of Patent: February 1, 2022
    Assignee: Adobe Inc.
    Inventors: Hailin Jin, Zhou Ren, Zhe Lin, Chen Fang
  • Publication number: 20210409836
    Abstract: Systems, methods, and non-transitory computer-readable media are disclosed for automatic tagging of videos. In particular, in one or more embodiments, the disclosed systems generate a set of tagged feature vectors (e.g., tagged feature vectors based on action-rich digital videos) to utilize to generate tags for an input digital video. For instance, the disclosed systems can extract a set of frames for the input digital video and generate feature vectors from the set of frames. In some embodiments, the disclosed systems generate aggregated feature vectors from the feature vectors. Furthermore, the disclosed systems can utilize the feature vectors (or aggregated feature vectors) to identify similar tagged feature vectors from the set of tagged feature vectors. Additionally, the disclosed systems can generate a set of tags for the input digital videos by aggregating one or more tags corresponding to identified similar tagged feature vectors.
    Type: Application
    Filed: September 9, 2021
    Publication date: December 30, 2021
    Inventors: Bryan Russell, Ruppesh Nalwaya, Markus Woodson, Joon-Young Lee, Hailin Jin
  • Publication number: 20210319566
    Abstract: Technology is disclosed herein for learning motion in video. In an implementation, an artificial neural network extracts features from a video. A correspondence proposal (CP) module performs, for at least some of the features, a search for corresponding features in the video based on a semantic similarity of a given feature to others of the features. The CP module then generates a joint semantic vector for each of the features based at least on the semantic similarity of the given feature to one or more of the corresponding features and a spatiotemporal distance of the given feature to the one or more of the corresponding features. The artificial neural network is able to identify motion in the video using the joint semantic vectors generated for the features extracted from the video.
    Type: Application
    Filed: June 17, 2021
    Publication date: October 14, 2021
    Inventors: Xingyu Liu, Hailin Jin, Joonyoung Lee
  • Patent number: 11146862
    Abstract: Systems, methods, and non-transitory computer-readable media are disclosed for automatic tagging of videos. In particular, in one or more embodiments, the disclosed systems generate a set of tagged feature vectors (e.g., tagged feature vectors based on action-rich digital videos) to utilize to generate tags for an input digital video. For instance, the disclosed systems can extract a set of frames for the input digital video and generate feature vectors from the set of frames. In some embodiments, the disclosed systems generate aggregated feature vectors from the feature vectors. Furthermore, the disclosed systems can utilize the feature vectors (or aggregated feature vectors) to identify similar tagged feature vectors from the set of tagged feature vectors. Additionally, the disclosed systems can generate a set of tags for the input digital videos by aggregating one or more tags corresponding to identified similar tagged feature vectors.
    Type: Grant
    Filed: April 16, 2019
    Date of Patent: October 12, 2021
    Assignee: ADOBE INC.
    Inventors: Bryan Russell, Ruppesh Nalwaya, Markus Woodson, Joon-Young Lee, Hailin Jin
  • Publication number: 20210311936
    Abstract: Embodiments of the present invention provide systems, methods, and computer storage media for guided visual search. A visual search query can be represented as a sketch sequence that includes ordering information of the constituent strokes in the sketch. The visual search query can be encoded into a structural search encoding in a common search space by a structural neural network. Indexed visual search results can be identified in the common search space and clustered in an auxiliary semantic space. Sketch suggestions can be identified from a plurality of indexed sketches in the common search space. A sketch suggestion can be identified for each semantic cluster of visual search results and presented with the cluster to guide a user towards relevant content through an iterative search process. Selecting a sketch suggestion as a target sketch can automatically transform the visual search query to the target sketch via adversarial images.
    Type: Application
    Filed: June 17, 2021
    Publication date: October 7, 2021
    Inventors: Hailin Jin, John Collomosse
  • Publication number: 20210312698
    Abstract: Systems, methods, and non-transitory computer-readable media are disclosed for utilizing an encoder-decoder architecture to learn a volumetric 3D representation of an object using digital images of the object from multiple viewpoints to render novel views of the object. For instance, the disclosed systems can utilize patch-based image feature extraction to extract lifted feature representations from images corresponding to different viewpoints of an object. Furthermore, the disclosed systems can model view-dependent transformed feature representations using learned transformation kernels. In addition, the disclosed systems can recurrently and concurrently aggregate the transformed feature representations to generate a 3D voxel representation of the object. Furthermore, the disclosed systems can sample frustum features using the 3D voxel representation and transformation kernels.
    Type: Application
    Filed: April 2, 2020
    Publication date: October 7, 2021
    Inventors: Tong He, John Collomosse, Hailin Jin
  • Patent number: 11113599
    Abstract: The present disclosure includes methods and systems for generating captions for digital images. In particular, the disclosed systems and methods can train an image encoder neural network and a sentence decoder neural network to generate a caption from an input digital image. For instance, in one or more embodiments, the disclosed systems and methods train an image encoder neural network (e.g., a character-level convolutional neural network) utilizing a semantic similarity constraint, training images, and training captions. Moreover, the disclosed systems and methods can train a sentence decoder neural network (e.g., a character-level recurrent neural network) utilizing training sentences and an adversarial classifier.
    Type: Grant
    Filed: June 22, 2017
    Date of Patent: September 7, 2021
    Assignee: Adobe Inc.
    Inventors: Zhaowen Wang, Shuai Tang, Hailin Jin, Chen Fang
  • Publication number: 20210264236
    Abstract: Embodiments of the present disclosure are directed towards improved models trained using unsupervised domain adaptation. In particular, a style-content adaptation system provides improved translation during unsupervised domain adaptation by controlling the alignment of conditional distributions of a model during training such that content (e.g., a class) from a target domain is correctly mapped to content (e.g., the same class) in a source domain. The style-content adaptation system improves unsupervised domain adaptation using independent control over content (e.g., related to a class) as well as style (e.g., related to a domain) to control alignment when translating between the source and target domain. This independent control over content and style can also allow for images to be generated using the style-content adaptation system that contain desired content and/or style.
    Type: Application
    Filed: February 26, 2020
    Publication date: August 26, 2021
    Inventors: Ning XU, Bayram Safa CICEK, Hailin JIN, Zhaowen WANG
  • Publication number: 20210248376
    Abstract: The present disclosure relates to systems, methods, and non-transitory computer-readable media for generating a response to a question received from a user during display or playback of a video segment by utilizing a query-response-neural network. The disclosed systems can extract a query vector from a question corresponding to the video segment using the query-response-neural network. The disclosed systems further generate context vectors representing both visual cues and transcript cues corresponding to the video segment using context encoders or other layers from the query-response-neural network. By utilizing additional layers from the query-response-neural network, the disclosed systems generate (i) a query-context vector based on the query vector and the context vectors, and (ii) candidate-response vectors representing candidate responses to the question from a domain-knowledge base or other source.
    Type: Application
    Filed: February 6, 2020
    Publication date: August 12, 2021
    Inventors: Wentian Zhao, Seokhwan Kim, Ning Xu, Hailin Jin
  • Publication number: 20210248432
    Abstract: Systems and methods provide for generating glyph initiations using a generative font system. A glyph variant may be generated based on an input vector glyph. A plurality of line segments may be approximated using a differentiable rasterizer with the plurality of line segments representing the contours of the glyph variant. A bitmap of the glyph variant may then be generated based on the line segments. The image loss between the bitmap and a rasterized representation of a vector glyph may be calculated and provided to the generative font system. Based on the image loss, a refined glyph variant may be provided to a user.
    Type: Application
    Filed: February 12, 2020
    Publication date: August 12, 2021
    Inventors: Zhaowen Wang, Zhifei Zhang, Xuan Li, Matthew Fisher, Hailin Jin
  • Publication number: 20210241032
    Abstract: In implementations of recognizing text in images, text recognition systems are trained using noisy images that have nuisance factors applied, and corresponding clean images (e.g., without nuisance factors). Clean images serve as supervision at both feature and pixel levels, so that text recognition systems are trained to be feature invariant (e.g., by requiring features extracted from a noisy image to match features extracted from a clean image), and feature complete (e.g., by requiring that features extracted from a noisy image be sufficient to generate a clean image). Accordingly, text recognition systems generalize to text not included in training images, and are robust to nuisance factors. Furthermore, since clean images are provided as supervision at feature and pixel levels, training requires fewer training images than text recognition systems that are not trained with a supervisory clean image, thus saving time and resources.
    Type: Application
    Filed: April 26, 2021
    Publication date: August 5, 2021
    Applicant: Adobe Inc.
    Inventors: Zhaowen Wang, Hailin Jin, Yang Liu
  • Patent number: 11068493
    Abstract: Embodiments of the present invention provide systems, methods, and computer storage media for guided visual search. A visual search query can be represented as a sketch sequence that includes ordering information of the constituent strokes in the sketch. The visual search query can be encoded into a structural search encoding in a common search space by a structural neural network. Indexed visual search results can be identified in the common search space and clustered in an auxiliary semantic space. Sketch suggestions can be identified from a plurality of indexed sketches in the common search space. A sketch suggestion can be identified for each semantic cluster of visual search results and presented with the cluster to guide a user towards relevant content through an iterative search process. Selecting a sketch suggestion as a target sketch can automatically transform the visual search query to the target sketch via adversarial images.
    Type: Grant
    Filed: November 7, 2018
    Date of Patent: July 20, 2021
    Assignee: Adobe Inc.
    Inventors: Hailin Jin, John Collomosse