Patents by Inventor Joon-Young Lee

Joon-Young Lee has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11657230
    Abstract: A method, apparatus, and non-transitory computer readable medium for referring image segmentation are described. Embodiments of the method, apparatus, and non-transitory computer readable medium may extract an image feature vector from an input image, extract a plurality of language feature vectors for a referral expression, wherein each of the plurality of language feature vectors comprises a different number of dimensions, combine each of the language feature vectors with the image feature vector using a fusion module to produce a plurality of self-attention vectors, combine the plurality of self-attention vectors to produce a multi-modal feature vector, and decode the multi-modal feature vector to produce an image mask indicating a portion of the input image corresponding to the referral expression.
    Type: Grant
    Filed: June 12, 2020
    Date of Patent: May 23, 2023
    Assignee: ADOBE INC.
    Inventors: Joon-Young Lee, Seonguk Seo
  • Patent number: 11640714
    Abstract: Systems and methods for panoptic video segmentation are described. A method may include identifying a target frame and a reference frame from a video, generating target features for the target frame and reference features for the reference frame, combining the target features and the reference features to produce fused features for the target frame, generating a feature matrix comprising a correspondence between objects from the reference features and objects from the fused features; and generating panoptic segmentation information for the target frame based on the feature matrix.
    Type: Grant
    Filed: April 20, 2020
    Date of Patent: May 2, 2023
    Assignee: ADOBE INC.
    Inventors: Joon-Young Lee, Sanghyun Woo, Dahun Kim
  • Patent number: 11526698
    Abstract: Systems and methods for video object segmentation are described. Embodiments of systems and methods may receive a referral expression and a video comprising a plurality of image frames, generate a first image mask based on the referral expression and a first image frame of the plurality of image frames, generate a second image mask based on the referral expression, the first image frame, the first image mask, and a second image frame of the plurality of image frames, and generate annotation information for the video including the first image mask overlaid on the first image frame and the second image mask overlaid on the second image frame.
    Type: Grant
    Filed: June 5, 2020
    Date of Patent: December 13, 2022
    Assignee: ADOBE INC.
    Inventors: Joon-Young Lee, Seonguk Seo
  • Publication number: 20220375090
    Abstract: This disclosure describes one or more implementations of a panoptic segmentation system that generates panoptic segmented digital images that classify both known and unknown instances of digital images. For example, the panoptic segmentation system builds and utilizes a panoptic segmentation neural network to discover, cluster, and segment new unknown object subclasses for previously unknown object instances. In addition, the panoptic segmentation system can determine additional unknown object instances from additional digital images. Moreover, in some implementations, the panoptic segmentation system utilizes the newly generated unknown object subclasses to refine and tune the panoptic segmentation neural network to improve the detection of unknown object instances in input digital images.
    Type: Application
    Filed: May 13, 2021
    Publication date: November 24, 2022
    Inventors: Jaedong Hwang, Seoung Wug Oh, Joon-Young Lee
  • Patent number: 11449079
    Abstract: Systems and techniques are described that provide for generalizable approach policy learning and implementation for robotic object approaching. Described techniques provide fast and accurate approaching of a specified object, or type of object, in many different environments. The described techniques enable a robot to receive an identification of an object or type of object from a user, and then navigate to the desired object, without further control from the user. Moreover, the approach of the robot to the desired object is performed efficiently, e.g., with a minimum number of movements. Further, the approach techniques may be used even when the robot is placed in a new environment, such as when the same type of object must be approached in multiple settings.
    Type: Grant
    Filed: January 30, 2019
    Date of Patent: September 20, 2022
    Assignee: ADOBE INC.
    Inventors: Zhe Lin, Xin Ye, Joon-Young Lee, Jianming Zhang
  • Publication number: 20210409836
    Abstract: Systems, methods, and non-transitory computer-readable media are disclosed for automatic tagging of videos. In particular, in one or more embodiments, the disclosed systems generate a set of tagged feature vectors (e.g., tagged feature vectors based on action-rich digital videos) to utilize to generate tags for an input digital video. For instance, the disclosed systems can extract a set of frames for the input digital video and generate feature vectors from the set of frames. In some embodiments, the disclosed systems generate aggregated feature vectors from the feature vectors. Furthermore, the disclosed systems can utilize the feature vectors (or aggregated feature vectors) to identify similar tagged feature vectors from the set of tagged feature vectors. Additionally, the disclosed systems can generate a set of tags for the input digital videos by aggregating one or more tags corresponding to identified similar tagged feature vectors.
    Type: Application
    Filed: September 9, 2021
    Publication date: December 30, 2021
    Inventors: Bryan Russell, Ruppesh Nalwaya, Markus Woodson, Joon-Young Lee, Hailin Jin
  • Publication number: 20210390700
    Abstract: A method, apparatus, and non-transitory computer readable medium for referring image segmentation are described. Embodiments of the method, apparatus, and non-transitory computer readable medium may extract an image feature vector from an input image, extract a plurality of language feature vectors for a referral expression, wherein each of the plurality of language feature vectors comprises a different number of dimensions, combine each of the language feature vectors with the image feature vector using a fusion module to produce a plurality of self-attention vectors, combine the plurality of self-attention vectors to produce a multi-modal feature vector, and decode the multi-modal feature vector to produce an image mask indicating a portion of the input image corresponding to the referral expression.
    Type: Application
    Filed: June 12, 2020
    Publication date: December 16, 2021
    Inventors: JOON-YOUNG LEE, SEONGUK SEO
  • Patent number: 11200424
    Abstract: Certain aspects involve using a space-time memory network to locate one or more target objects in video content for segmentation or other object classification. In one example, a video editor generates a query key map and a query value map by applying a space-time memory network to features of a query frame from video content. The video editor retrieves a memory key map and a memory value map that are computed, with the space-time memory network, from a set of memory frames from the video content. The video editor computes memory weights by applying a similarity function to the memory key map and the query key map. The video editor classifies content in the query frame as depicting the target feature using a weighted summation that includes the memory weights applied to memory locations in the memory value map.
    Type: Grant
    Filed: March 5, 2019
    Date of Patent: December 14, 2021
    Assignee: Adobe Inc.
    Inventors: Joon-Young Lee, Ning Xu, Seoungwug Oh
  • Publication number: 20210383171
    Abstract: Systems and methods for video object segmentation are described. Embodiments of systems and methods may receive a referral expression and a video comprising a plurality of image frames, generate a first image mask based on the referral expression and a first image frame of the plurality of image frames, generate a second image mask based on the referral expression, the first image frame, the first image mask, and a second image frame of the plurality of image frames, and generate annotation information for the video including the first image mask overlaid on the first image frame and the second image mask overlaid on the second image frame.
    Type: Application
    Filed: June 5, 2020
    Publication date: December 9, 2021
    Inventors: JOON-YOUNG LEE, SEONGUK SEO
  • Patent number: 11176381
    Abstract: Various embodiments describe video object segmentation using a neural network and the training of the neural network. The neural network both detects a target object in the current frame based on a reference frame and a reference mask that define the target object and propagates the segmentation mask of the target object for a previous frame to the current frame to generate a segmentation mask for the current frame. In some embodiments, the neural network is pre-trained using synthetically generated static training images and is then fine-tuned using training videos.
    Type: Grant
    Filed: April 23, 2020
    Date of Patent: November 16, 2021
    Assignee: Adobe Inc.
    Inventors: Joon-Young Lee, Seoungwug Oh, Kalyan Krishna Sunkavalli
  • Patent number: 11157133
    Abstract: A portable terminal apparatus includes a bended display including a front display area and a side display area, and a controller that displays a UI element corresponding to an application on a side display area of the bended display if the cover is closed, and in response to a touch with respect to at least one UI element from among the UI elements and an opening of the cover being detected, executes an application corresponding to a UI element with the detected touch, and controls the bended display so that an execution screen of the application is displayed on the front display area.
    Type: Grant
    Filed: August 26, 2015
    Date of Patent: October 26, 2021
    Assignee: SAMSUNGN ELECTRONICS CO., LTD.
    Inventors: So-young Kim, Joon-young Lee, Moon-joo Lee, Shi-yun Cho
  • Publication number: 20210326638
    Abstract: Systems and methods for panoptic video segmentation are described. A method may include identifying a target frame and a reference frame from a video, generating target features for the target frame and reference features for the reference frame, combining the target features and the reference features to produce fused features for the target frame, generating a feature matrix comprising a correspondence between objects from the reference features and objects from the fused features; and generating panoptic segmentation information for the target frame based on the feature matrix.
    Type: Application
    Filed: April 20, 2020
    Publication date: October 21, 2021
    Inventors: Joon-Young Lee, Sanghyun Woo, Dahun Kim
  • Patent number: 11146862
    Abstract: Systems, methods, and non-transitory computer-readable media are disclosed for automatic tagging of videos. In particular, in one or more embodiments, the disclosed systems generate a set of tagged feature vectors (e.g., tagged feature vectors based on action-rich digital videos) to utilize to generate tags for an input digital video. For instance, the disclosed systems can extract a set of frames for the input digital video and generate feature vectors from the set of frames. In some embodiments, the disclosed systems generate aggregated feature vectors from the feature vectors. Furthermore, the disclosed systems can utilize the feature vectors (or aggregated feature vectors) to identify similar tagged feature vectors from the set of tagged feature vectors. Additionally, the disclosed systems can generate a set of tags for the input digital videos by aggregating one or more tags corresponding to identified similar tagged feature vectors.
    Type: Grant
    Filed: April 16, 2019
    Date of Patent: October 12, 2021
    Assignee: ADOBE INC.
    Inventors: Bryan Russell, Ruppesh Nalwaya, Markus Woodson, Joon-Young Lee, Hailin Jin
  • Patent number: 10991085
    Abstract: Embodiments herein describe a framework for classifying images. In some embodiments, it is determined whether an image includes synthetic image content. If it does, characteristics of the image are analyzed to determine if the image includes characteristics particular to panoramic images (e.g., possess a threshold equivalency of pixel values among the top and/or bottom boundaries of the image, or a difference between summed pixel values of the pixels comprising the right vertical boundary of the image and summed pixel values of the pixels comprising the left vertical boundary of the image being less than or equal to a threshold value). If the image includes characteristics particular to panoramic images, the image is classified as a synthetic panoramic image. If the image is determined to not include synthetic image content, a neural network is applied to the image and the image is classified as one of non-synthetic panoramic or non-synthetic non-panoramic.
    Type: Grant
    Filed: April 1, 2019
    Date of Patent: April 27, 2021
    Assignee: ADOBE INC.
    Inventors: Qi Sun, Li-Yi Wei, Joon-Young Lee, Jonathan Eisenmann, Jinwoong Jung, Byungmoon Kim
  • Publication number: 20210042944
    Abstract: Methods and systems are provided for performing material capture to determine properties of an imaged surface. A plurality of images can be received depicting a material surface. The plurality of images can be calibrated to align corresponding pixels of the images and determine reflectance information for at least a portion of the aligned pixels. After calibration, a set of reference materials from a material library can be selected using the calibrated images. The set of reference materials can be used to determine a material model that accurately represents properties of the material surface.
    Type: Application
    Filed: October 26, 2020
    Publication date: February 11, 2021
    Inventors: KALYAN KRISHNA SUNKAVALLI, SUNIL HADAP, JOON-YOUNG LEE, ZHUO HUI
  • Patent number: 10915798
    Abstract: Disclosed herein are embodiments of systems, methods, and products for a webly supervised training of a convolutional neural network (CNN) to predict emotion in images. A computer may query one or more image repositories using search keywords generated based on the tertiary emotion classes of Parrott's emotion wheel. The computer may filter images received in response to the query to generate a weakly labeled training dataset labels associated with the images that are noisy or wrong may be cleaned prior to training of the CNN. The computer may iteratively train the CNN leveraging the hierarchy of emotion classes by increasing the complexity of the labels (tags) for each iteration. Such curriculum guided training may generate a trained CNN that is more accurate than the conventionally trained neural networks.
    Type: Grant
    Filed: May 15, 2018
    Date of Patent: February 9, 2021
    Assignee: Adobe Inc.
    Inventors: Jianming Zhang, Rameswar Panda, Haoxiang Li, Joon-Young Lee, Xin Lu
  • Patent number: 10868635
    Abstract: Provided is a broadcast signal transmitting method including: generating a first bitstream including encoded data; generating a second bitstream by adding zero bits to the first bitstream, based on a difference between a modem data rate and a codec data rate; generating a third bitstream by encoding the second bitstream by using zero padding information about the zero bits; and transmitting a carrier signal generated by modulating the third bitstream.
    Type: Grant
    Filed: June 28, 2018
    Date of Patent: December 15, 2020
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Jung-hyun Park, Nam-hyun Kim, Min-ho Kim, Joon-young Lee, Jin-joo Chung, Doo-chan Hwang
  • Patent number: 10818022
    Abstract: Methods and systems are provided for performing material capture to determine properties of an imaged surface. A plurality of images can be received depicting a material surface. The plurality of images can be calibrated to align corresponding pixels of the images and determine reflectance information for at least a portion of the aligned pixels. After calibration, a set of reference materials from a material library can be selected using the calibrated images. The set of reference materials can be used to determine a material model that accurately represents properties of the material surface.
    Type: Grant
    Filed: December 21, 2018
    Date of Patent: October 27, 2020
    Assignee: ADOBE INC.
    Inventors: Kalyan Krishna Sunkavalli, Sunil Hadap, Joon-Young Lee, Zhuo Hui
  • Publication number: 20200336802
    Abstract: Systems, methods, and non-transitory computer-readable media are disclosed for automatic tagging of videos. In particular, in one or more embodiments, the disclosed systems generate a set of tagged feature vectors (e.g., tagged feature vectors based on action-rich digital videos) to utilize to generate tags for an input digital video. For instance, the disclosed systems can extract a set of frames for the input digital video and generate feature vectors from the set of frames. In some embodiments, the disclosed systems generate aggregated feature vectors from the feature vectors. Furthermore, the disclosed systems can utilize the feature vectors (or aggregated feature vectors) to identify similar tagged feature vectors from the set of tagged feature vectors. Additionally, the disclosed systems can generate a set of tags for the input digital videos by aggregating one or more tags corresponding to identified similar tagged feature vectors.
    Type: Application
    Filed: April 16, 2019
    Publication date: October 22, 2020
    Inventors: Bryan Russell, Ruppesh Nalwaya, Markus Woodson, Joon-Young Lee, Hailin Jin
  • Patent number: 10810435
    Abstract: In implementations of segmenting objects in video sequences, user annotations designate an object in any image frame of a video sequence, without requiring user annotations for all image frames. An interaction network generates a mask for an object in an image frame annotated by a user, and is coupled both internally and externally to a propagation network that propagates the mask to other image frames of the video sequence. Feature maps are aggregated for each round of user annotations and couple the interaction network and the propagation network internally. The interaction network and the propagation network are trained jointly using synthetic annotations in a multi-round training scenario, in which weights of the interaction network and the propagation network are adjusted after multiple synthetic annotations are processed, resulting in a trained object segmentation system that can reliably generate realistic object masks.
    Type: Grant
    Filed: November 7, 2018
    Date of Patent: October 20, 2020
    Assignee: Adobe Inc.
    Inventors: Joon-Young Lee, Seoungwug Oh, Ning Xu