Patents by Inventor Oliver Wang
Oliver Wang has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20250124212Abstract: In implementation of techniques for vector font generation based on cascaded diffusion, a computing device implements a glyph generation system to receive a sample glyph in a target font and a target glyph identifier. The glyph generation system generates a rasterized glyph in the target font using a raster diffusion model based on the sample glyph and the target glyph identifier, the rasterized glyph having a first level of resolution. The glyph generation system then generates a vector glyph using a vector diffusion model by vectorizing the rasterized glyph, the vector glyph having a second level of resolution different than the first level of resolution. The glyph generation system then displays the vector glyph in a user interface.Type: ApplicationFiled: November 13, 2023Publication date: April 17, 2025Applicant: Adobe Inc.Inventors: Difan Liu, Matthew David Fisher, Michaël Yanis Gharbi, Oliver Wang, Alec Stefan Jacobson, Vikas Thamizharasan, Evangelos Kalogerakis
-
Patent number: 12236640Abstract: Systems and methods for image dense field based view calibration are provided. In one embodiment, an input image is applied to a dense field machine learning model that generates a vertical vector dense field (VVF) and a latitude dense field (LDF) from the input image. The VVF comprises a vertical vector of a projected vanishing point direction for each of the pixels of the input image. The latitude dense field (LDF) comprises a projected latitude value for the pixels of the input image. A dense field map for the input image comprising the VVF and the LDF can be directly or indirectly used for a variety of image processing manipulations. The VVF and LDF can be optionally used to derive traditional camera calibration parameters from uncontrolled images that have undergone undocumented or unknown manipulations.Type: GrantFiled: March 28, 2022Date of Patent: February 25, 2025Assignee: Adobe Inc.Inventors: Jianming Zhang, Linyi Jin, Kevin Matzen, Oliver Wang, Yannick Hold-Geoffroy
-
Publication number: 20250061647Abstract: A scene modeling system accesses a set of input two-dimensional (2D) images of a three-dimensional (3D) environment, wherein the input 2D images captured from a plurality of camera orientations. The environment includes first content. The scene modeling system applies a scene generation model to the set of input 2D images to generate a 3D remix scene. Applying the scene generation model includes configuring the scene generation model using at least a 2D discriminator and a 3D discriminator. Applying the scene generation model includes transmitting, for display via a user interface, the 3D remix scene. The 3D remix scene includes second content that is different from the first content.Type: ApplicationFiled: August 14, 2023Publication date: February 20, 2025Inventors: Oliver Wang, Animesh Karnewar, Tobias Ritschel, Niloy Mitra
-
Publication number: 20240320872Abstract: A method, apparatus, non-transitory computer readable medium, and system for image generation include obtaining a text embedding of a text prompt and an image embedding of an image prompt. Some embodiments map the text embedding into a joint embedding space to obtain a joint text embedding and map the image embedding into the joint embedding space to obtain a joint image embedding. Some embodiments generate a synthetic image based on the joint text embedding and the joint image embedding.Type: ApplicationFiled: January 30, 2024Publication date: September 26, 2024Inventors: Tobias Hinz, Venkata Naveen Kumar Yadav Marri, Midhun Harikumar, Ajinkya Gorakhnath Kale, Zhe Lin, Oliver Wang, Jingwan Lu
-
Publication number: 20240320873Abstract: A method, apparatus, non-transitory computer readable medium, and system for image generation include obtaining a text prompt and encoding, using a text encoder jointly trained with an image generation model, the text prompt to obtain a text embedding. Some embodiments generate, using the image generation model, a synthetic image based on the text embedding.Type: ApplicationFiled: February 12, 2024Publication date: September 26, 2024Inventors: Tobias Hinz, Ali Aminian, Hao Tan, Kushal Kafle, Oliver Wang, Jingwan Lu
-
Publication number: 20240320789Abstract: A method, non-transitory computer readable medium, apparatus, and system for image generation include obtaining an input image having a first resolution, where the input image includes random noise, and generating a low-resolution image based on the input image, where the low-resolution image has the first resolution. The method, non-transitory computer readable medium, apparatus, and system further include generating a high-resolution image based on the low-resolution image, where the high-resolution image has a second resolution that is greater than the first resolution.Type: ApplicationFiled: February 23, 2024Publication date: September 26, 2024Inventors: Tobias Hinz, Taesung Park, Jingwan Lu, Elya Shechtman, Richard Zhang, Oliver Wang
-
Patent number: 12039657Abstract: Embodiments of the technology described herein, provide a view and time synthesis of dynamic scenes captured by a camera. The technology described herein represents a dynamic scene as a continuous function of both space and time. The technology may parameterize this function with a deep neural network (a multi-layer perceptron (MLP)), and perform rendering using volume tracing. At a very high level, a dynamic scene depicted in the video may be used to train the MLP. Once trained, the MLP is able to synthesize a view of the scene at a time and/or camera pose not found in the video through prediction. As used herein, a dynamic scene comprises one or more moving objects.Type: GrantFiled: March 17, 2021Date of Patent: July 16, 2024Assignee: Adobe Inc.Inventors: Oliver Wang, Simon Niklaus, Zhengqi Li
-
Publication number: 20240161327Abstract: Aspects of the methods, apparatus, non-transitory computer readable medium, and systems include obtaining a noise map and a global image code encoded from an original image and representing semantic content of the original image; generating a plurality of image patches based on the noise map and the global image code using a diffusion model; and combining the plurality of image patches to produce an output image including the semantic content.Type: ApplicationFiled: November 4, 2022Publication date: May 16, 2024Inventors: Yinbo Chen, Michaël Gharbi, Oliver Wang, Richard Zhang, Elya Shechtman
-
Patent number: 11936936Abstract: A method including receiving video of an event; generating an overlay for the video; generating an information message containing information enabling a receiver of the video and the overlay to selectively display or hide the overlay; and transmitting the video, the overlay, and the information message. The video is transmitted in a primary stream of a multi-stream transmission including a primary stream and one or more auxiliary streams. The overlay is transmitted in a first one of the auxiliary streams.Type: GrantFiled: October 9, 2014Date of Patent: March 19, 2024Assignee: DISNEY ENTERPRISES, INC.Inventors: Aljoscha Smolic, Nikolce Stefanoski, Oliver Wang
-
Patent number: 11930303Abstract: Systems and techniques for automatic digital parameter adjustment are described that leverage insights learned from an image set to automatically predict parameter values for an input item of digital visual content. To do so, the automatic digital parameter adjustment techniques described herein captures visual and contextual features of digital visual content to determine balanced visual output in a range of visual scenes and settings. The visual and contextual features of digital visual content are used to train a parameter adjustment model through machine learning techniques that captures feature patterns and interactions. The parameter adjustment model exploits these feature interactions to determine visually pleasing parameter values for an input item of digital visual content. The predicted parameter values are output, allowing further adjustment to the parameter values.Type: GrantFiled: November 15, 2021Date of Patent: March 12, 2024Assignee: Adobe Inc.Inventors: Pulkit Gera, Oliver Wang, Kalyan Krishna Sunkavalli, Elya Shechtman, Chetan Nanda
-
Patent number: 11908036Abstract: The technology described herein is directed to a cross-domain training framework that iteratively trains a domain adaptive refinement agent to refine low quality real-world image acquisition data, e.g., depth maps, when accompanied by corresponding conditional data from other modalities, such as the underlying images or video from which the image acquisition data is computed. The cross-domain training framework includes a shared cross-domain encoder and two conditional decoder branch networks, e.g., a synthetic conditional depth prediction branch network and a real conditional depth prediction branch network. The shared cross-domain encoder converts synthetic and real-world image acquisition data into synthetic and real compact feature representations, respectively.Type: GrantFiled: September 28, 2020Date of Patent: February 20, 2024Assignee: Adobe Inc.Inventors: Oliver Wang, Jianming Zhang, Dingzeyu Li, Zekun Hao
-
Patent number: 11900902Abstract: Embodiments are disclosed for determining an answer to a query associated with a graphical representation of data. In particular, in one or more embodiments, the disclosed systems and methods comprise receiving an input including an unprocessed audio sequence and a request to perform an audio signal processing effect on the unprocessed audio sequence. The one or more embodiments further include analyzing, by a deep encoder, the unprocessed audio sequence to determine parameters for processing the unprocessed audio sequence. The one or more embodiments further include sending the unprocessed audio sequence and the parameters to one or more audio signal processing effects plugins to perform the requested audio signal processing effect using the parameters and outputting a processed audio sequence after processing of the unprocessed audio sequence using the parameters of the one or more audio signal processing effects plugins.Type: GrantFiled: April 12, 2021Date of Patent: February 13, 2024Assignee: Adobe Inc.Inventors: Marco Antonio Martinez Ramirez, Nicholas J. Bryan, Oliver Wang, Paris Smaragdis
-
Publication number: 20240046430Abstract: One or more processing devices access a scene depicting a reference object that includes an annotation identifying a target region to be modified in one or more video frames. The one or more processing devices determine that a target pixel corresponds to a sub-region within the target region that includes hallucinated content. The one or more processing devices determine gradient constraints using gradient values of neighboring pixels in the hallucinated content, the neighboring pixels being adjacent to the target pixel and corresponding to four cardinal directions. The one or more processing devices update color data of the target pixel subject to the determined gradient constraints.Type: ApplicationFiled: September 29, 2023Publication date: February 8, 2024Inventors: Oliver Wang, John Nelson, Geoffrey Oxholm, Elya Shechtman
-
Patent number: 11893763Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media for generating a modified digital image from extracted spatial and global codes. For example, the disclosed systems can utilize a global and spatial autoencoder to extract spatial codes and global codes from digital images. The disclosed systems can further utilize the global and spatial autoencoder to generate a modified digital image by combining extracted spatial and global codes in various ways for various applications such as style swapping, style blending, and attribute editing.Type: GrantFiled: November 22, 2022Date of Patent: February 6, 2024Assignee: Adobe Inc.Inventors: Taesung Park, Richard Zhang, Oliver Wang, Junyan Zhu, Jingwan Lu, Elya Shechtman, Alexei A Efros
-
Patent number: 11880766Abstract: An improved system architecture uses a pipeline including a Generative Adversarial Network (GAN) including a generator neural network and a discriminator neural network to generate an image. An input image in a first domain and information about a target domain are obtained. The domains correspond to image styles. An initial latent space representation of the input image is produced by encoding the input image. An initial output image is generated by processing the initial latent space representation with the generator neural network. Using the discriminator neural network, a score is computed indicating whether the initial output image is in the target domain. A loss is computed based on the computed score. The loss is minimized to compute an updated latent space representation. The updated latent space representation is processed with the generator neural network to generate an output image in the target domain.Type: GrantFiled: July 23, 2021Date of Patent: January 23, 2024Assignee: Adobe Inc.Inventors: Cameron Smith, Ratheesh Kalarot, Wei-An Lin, Richard Zhang, Niloy Mitra, Elya Shechtman, Shabnam Ghadar, Zhixin Shu, Yannick Hold-Geoffrey, Nathan Carr, Jingwan Lu, Oliver Wang, Jun-Yan Zhu
-
Patent number: 11871145Abstract: Embodiments are disclosed for video image interpolation. In some embodiments, video image interpolation includes receiving a pair of input images from a digital video, determining, using a neural network, a plurality of spatially varying kernels each corresponding to a pixel of an output image, convolving a first set of spatially varying kernels with a first input image from the pair of input images and a second set of spatially varying kernels with a second input image from the pair of input images to generate filtered images, and generating the output image by performing kernel normalization on the filtered images.Type: GrantFiled: April 6, 2021Date of Patent: January 9, 2024Assignee: Adobe Inc.Inventors: Simon Niklaus, Oliver Wang, Long Mai
-
Patent number: 11854206Abstract: A Video Semantic Segmentation System (VSSS) is disclosed that performs accurate and fast semantic segmentation of videos using a set of temporally distributed neural networks. The VSSS receives as input a video signal comprising a contiguous sequence of temporally-related video frames. The VSSS extracts features from the video frames in the contiguous sequence and based upon the extracted features, selects, from a set of labels, a label to be associated with each pixel of each video frame in the video signal. In certain embodiments, a set of multiple neural networks are used to extract the features to be used for video segmentation and the extraction of features is distributed among the multiple neural networks in the set. A strong feature representation representing the entirety of the features is produced for each video frame in the sequence of video frames by aggregating the output features extracted by the multiple neural networks.Type: GrantFiled: May 3, 2022Date of Patent: December 26, 2023Assignee: Adobe Inc.Inventors: Federico Perazzi, Zhe Lin, Ping Hu, Oliver Wang, Fabian David Caba Heilbron
-
Patent number: 11823357Abstract: Certain aspects involve video inpainting in which content is propagated from a user-provided reference video frame to other video frames depicting a scene. One example method includes one or more processing devices that performs operations that include accessing a scene depicting a reference object that includes an annotation identifying a target region to be modified in one or more video frames. The operations also includes computing a target motion of a target pixel that is subject to a motion constraint. The motion constraint is based on a three-dimensional model of the reference object. Further, operations include determining color data of the target pixel to correspond to the target motion. The color data includes a color value and a gradient. Operations also include determining gradient constraints using gradient values of neighbor pixels. Additionally, the processing devices updates the color data of the target pixel subject to the gradient constraints.Type: GrantFiled: March 9, 2021Date of Patent: November 21, 2023Assignee: Adobe Inc.Inventors: Oliver Wang, John Nelson, Geoffrey Oxholm, Elya Shechtman
-
Patent number: 11798180Abstract: This disclosure describes one or more implementations of a depth prediction system that generates accurate depth images from single input digital images. In one or more implementations, the depth prediction system enforces different sets of loss functions across mix-data sources to generate a multi-branch architecture depth prediction model. For instance, in one or more implementations, the depth prediction model utilizes different data sources having different granularities of ground truth depth data to robustly train a depth prediction model. Further, given the different ground truth depth data granularities from the different data sources, the depth prediction model enforces different combinations of loss functions including an image-level normalized regression loss function and/or a pair-wise normal loss among other loss functions.Type: GrantFiled: February 26, 2021Date of Patent: October 24, 2023Assignee: Adobe Inc.Inventors: Wei Yin, Jianming Zhang, Oliver Wang, Simon Niklaus, Mai Long, Su Chen
-
Publication number: 20230306637Abstract: Systems and methods for image dense field based view calibration are provided. In one embodiment, an input image is applied to a dense field machine learning model that generates a vertical vector dense field (VVF) and a latitude dense field (LDF) from the input image. The VVF comprises a vertical vector of a projected vanishing point direction for each of the pixels of the input image. The latitude dense field (LDF) comprises a projected latitude value for the pixels of the input image. A dense field map for the input image comprising the VVF and the LDF can be directly or indirectly used for a variety of image processing manipulations. The VVF and LDF can be optionally used to derive traditional camera calibration parameters from uncontrolled images that have undergone undocumented or unknown manipulations.Type: ApplicationFiled: March 28, 2022Publication date: September 28, 2023Inventors: Jianming ZHANG, Linyi JIN, Kevin MATZEN, Oliver WANG, Yannick HOLD-GEOFFROY