Patents by Inventor Yijun Li
Yijun Li has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20250128760Abstract: Methods and apparatus to adjust a steering angle of a vehicle in a self-driving mode are described herein. An example vehicle disclosed herein includes a steering controller including instructions and programmable circuitry to execute the instructions to access a path follower (PF) angle request, generate a virtual boost curve (VBC) angle request based on a torque input to a steering wheel, determine an angle blending weight based on the torque input and a speed of the vehicle, determine a final angle request based on the PF angle request, the VBC angle request, and the angle blending weight, and convert the final angle request to a torque request to be used to adjust a steering angle of the vehicle via a motor.Type: ApplicationFiled: October 20, 2023Publication date: April 24, 2025Inventors: Ashrit Das, Yijun Li, Tomas Mawyin, Lodewijk Wijffels
-
Publication number: 20250117995Abstract: Methods, non-transitory computer readable media, apparatuses, and systems for image and depth map generation include receiving a prompt and encoding the prompt to obtain a guidance embedding. A machine learning model then generates an image and a depth map corresponding to the image based on the guidance embedding. The image and the depth map are each generated based on the guidance embedding.Type: ApplicationFiled: October 5, 2023Publication date: April 10, 2025Inventors: Yijun Li, Matheus Abrantes Gadelha, Krishna Kumar Singh, Soren Pirk
-
Patent number: 12272031Abstract: An image inpainting system is described that receives an input image that includes a masked region. From the input image, the image inpainting system generates a synthesized image that depicts an object in the masked region by selecting a first code that represents a known factor characterizing a visual appearance of the object and a second code that represents an unknown factor characterizing the visual appearance of the object apart from the known factor in latent space. The input image, the first code, and the second code are provided as input to a generative adversarial network that is trained to generate the synthesized image using contrastive losses. Different synthesized images are generated from the same input image using different combinations of first and second codes, and the synthesized images are output for display.Type: GrantFiled: April 21, 2022Date of Patent: April 8, 2025Assignee: Adobe Inc.Inventors: Krishna Kumar Singh, Yuheng Li, Yijun Li, Jingwan Lu, Elya Shechtman
-
Patent number: 12260530Abstract: The present disclosure relates to systems, methods, and non-transitory computer-readable media that modify digital images via scene-based editing using image understanding facilitated by artificial intelligence. For example, in one or more embodiments the disclosed systems utilize generative machine learning models to create modified digital images portraying human subjects. In particular, the disclosed systems generate modified digital images by performing infill modifications to complete a digital image or human inpainting for portions of a digital image that portrays a human. Moreover, in some embodiments, the disclosed systems perform reposing of subjects portrayed within a digital image to generate modified digital images. In addition, the disclosed systems in some embodiments perform facial expression transfer and facial expression animations to generate modified digital images or animations.Type: GrantFiled: March 27, 2023Date of Patent: March 25, 2025Assignee: Adobe Inc.Inventors: Krishna Kumar Singh, Yijun Li, Jingwan Lu, Duygu Ceylan Aksit, Yangtuanfeng Wang, Jimei Yang, Tobias Hinz, Qing Liu, Jianming Zhang, Zhe Lin
-
Patent number: 12249132Abstract: The present disclosure relates to systems, non-transitory computer-readable media, and methods for adapting generative neural networks to target domains utilizing an image translation neural network. In particular, in one or more embodiments, the disclosed systems utilize an image translation neural network to translate target results to a source domain for input in target neural network adaptation. For instance, in some embodiments, the disclosed systems compare a translated target result with a source result from a pretrained source generative neural network to adjust parameters of a target generative neural network to produce results corresponding in features to source results and corresponding in style to the target domain.Type: GrantFiled: July 27, 2022Date of Patent: March 11, 2025Assignee: Adobe Inc.Inventors: Yijun Li, Nicholas Kolkin, Jingwan Lu, Elya Shechtman
-
Publication number: 20250078327Abstract: The present disclosure relates to systems, methods, and non-transitory computer-readable media that utilize a text-image alignment loss to train a diffusion model to generate digital images from input text. In particular, in some embodiments, the disclosed systems generate a prompt noise representation form a text prompt with a first text concept and a second text concept using a denoising step of a diffusion neural network. Further, in some embodiments, the disclosed systems generate a first concept noise representation from the first text concept and a second concept noise representation from the second text concept. Moreover, the disclosed systems combine the first and second concept noise representation to generate a combined concept noise representation. Accordingly, in some embodiments, by comparing the combined concept noise representation and the prompt noise representation, the disclosed systems modify parameters of the diffusion neural network.Type: ApplicationFiled: August 29, 2023Publication date: March 6, 2025Inventors: Zhipeng Bao, Yijun Li, Krishna Kumar Singh
-
Publication number: 20250069204Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media for accurately, efficiently, and flexibly restoring degraded digital images utilizing a deep learning framework for repairing local defects, correcting global imperfections, and/or enhancing depicted faces. In particular, the disclosed systems can utilize a defect detection neural network to generate a segmentation map indicating locations of local defects within a digital image. In addition, the disclosed systems can utilize an inpainting algorithm to determine pixels for inpainting the local defects to reduce their appearance. In some embodiments, the disclosed systems utilize a global correction neural network to determine and repair global imperfections. Further, the disclosed systems can enhance one or more faces depicted within a digital image utilizing a face enhancement neural network as well.Type: ApplicationFiled: November 12, 2024Publication date: February 27, 2025Inventors: Ionut Mironica, Yijun Li
-
Patent number: 12230014Abstract: An image generation system enables user input during the process of training a generative model to influence the model's ability to generate new images with desired visual features. A source generative model for a source domain is fine-tuned using training images in a target domain to provide an adapted generative model for the target domain. Interpretable factors are determined for the source generative model and the adapted generative model. A user interface is provided that enables a user to select one or more interpretable factors. The user-selected interpretable factor(s) are used to generate a user-adapted generative model, for instance, by using a loss function based on the user-selected interpretable factor(s). The user-adapted generative model can be used to create new images in the target domain.Type: GrantFiled: February 25, 2022Date of Patent: February 18, 2025Assignee: ADOBE INC.Inventors: Yijun Li, Utkarsh Ojha, Richard Zhang, Jingwan Lu, Elya Shechtman, Alexei A. Efros
-
Publication number: 20250037431Abstract: Systems and methods for training a Generative Adversarial Network (GAN) using feature regularization are described herein. Embodiments are configured to generate a candidate image using a generator network of a GAN, classify the candidate image as real or generated using a discriminator network of the GAN, and train the GAN to generate realistic images based on the classifying of the candidate image. The training process includes regularizing a gradient with respect to features extracted using a discriminator network of the GAN.Type: ApplicationFiled: July 24, 2023Publication date: January 30, 2025Inventors: Min Jin Chong, Krishna Kumar Singh, Yijun Li, Jingwan Lu
-
Patent number: 12175641Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media for accurately, efficiently, and flexibly restoring degraded digital images utilizing a deep learning framework for repairing local defects, correcting global imperfections, and/or enhancing depicted faces. In particular, the disclosed systems can utilize a defect detection neural network to generate a segmentation map indicating locations of local defects within a digital image. In addition, the disclosed systems can utilize an inpainting algorithm to determine pixels for inpainting the local defects to reduce their appearance. In some embodiments, the disclosed systems utilize a global correction neural network to determine and repair global imperfections. Further, the disclosed systems can enhance one or more faces depicted within a digital image utilizing a face enhancement neural network as well.Type: GrantFiled: June 4, 2021Date of Patent: December 24, 2024Assignee: Adobe Inc.Inventors: Ionut Mironica, Yijun Li
-
Patent number: 12162466Abstract: Methods, systems, apparatus, and articles of manufacture to control a vehicle based on signal blending are disclosed. An example apparatus disclosed herein includes programmable circuitry at least determine a first yaw rate signal based on first signal data output by a yaw rate sensor of a vehicle, determine a second yaw rate signal based on second signal data output by a steering wheel angle sensor of the vehicle, determine a blended yaw rate signal based on the first yaw rate signal and the second yaw rate signal, and adjust a torque to be applied by a motor of the vehicle based on the blended yaw rate signal.Type: GrantFiled: November 30, 2023Date of Patent: December 10, 2024Assignee: Ford Global Technologies, LLCInventors: Ashrit Das, Joshua Guerra, Benjamin James Northrup, Lodewijk Maarten Erik Wijffels, Ziyu Ke, Ronald Loyd Chadwick, Yijun Li
-
Patent number: 12159413Abstract: In implementations of systems for image inversion using multiple latent spaces, a computing device implements an inversion system to generate a segment map that segments an input digital image into a first image region and a second image region and assigns the first image region to a first latent space and the second image region to a second latent space that corresponds to a layer of a convolutional neural network. An inverted latent representation of the input digital image is computed using a binary mask for the second image region. The inversion system modifies the inverted latent representation of the input digital image using an edit direction vector that corresponds to a visual feature. An output digital image is generated that depicts a reconstruction of the input digital image having the visual feature based on the modified inverted latent representation of the input digital image.Type: GrantFiled: March 14, 2022Date of Patent: December 3, 2024Assignee: Adobe Inc.Inventors: Gaurav Parmar, Krishna Kumar Singh, Yijun Li, Richard Zhang, Jingwan Lu
-
Publication number: 20240338799Abstract: The present disclosure relates to systems, non-transitory computer-readable media, and methods for utilizing machine learning models to generate modified digital images. In particular, in some embodiments, the disclosed systems generate image editing directions between textual identifiers of two visual features utilizing a language prediction machine learning model and a text encoder. In some embodiments, the disclosed systems generated an inversion of a digital image utilizing a regularized inversion model to guide forward diffusion of the digital image. In some embodiments, the disclosed systems utilize cross-attention guidance to preserve structural details of a source digital image when generating a modified digital image with a diffusion neural network.Type: ApplicationFiled: March 3, 2023Publication date: October 10, 2024Inventors: Yijun Li, Richard Zhang, Krishna Kumar Singh, Jingwan Lu, Gaurav Parmar, Jun-Yan Zhu
-
Publication number: 20240331236Abstract: The present disclosure relates to systems, non-transitory computer-readable media, and methods for utilizing machine learning models to generate modified digital images. In particular, in some embodiments, the disclosed systems generate image editing directions between textual identifiers of two visual features utilizing a language prediction machine learning model and a text encoder. In some embodiments, the disclosed systems generated an inversion of a digital image utilizing a regularized inversion model to guide forward diffusion of the digital image. In some embodiments, the disclosed systems utilize cross-attention guidance to preserve structural details of a source digital image when generating a modified digital image with a diffusion neural network.Type: ApplicationFiled: March 3, 2023Publication date: October 3, 2024Inventors: Yijun Li, Richard Zhang, Krishna Kumar Singh, Jingwan Lu, Gaurav Parmar, Jun-Yan Zhu
-
Patent number: 12083730Abstract: A rotating extrusion rheometer includes a control monitoring mechanism, a melt extrusion mechanism, a rotating extrusion rheology machine head, a sensor, a drive chain wheel, a coupler and an electric motor. The control monitoring mechanism, the melt extrusion mechanism, the rotating extrusion rheology machine head are sequentially connected. The rotating extrusion rheology machine head is formed by a connecting pipe (1), a flow dividing support (3), a lower machine neck (12), a machine head piece (15), an opening mold (17), an opening-mold driving chain wheel (20), a core bar (21) and a core-bar driving mechanism. The rheology measurement method comprises the steps where some parameter values of the rheometer are collected first, and then the rheological behaviors of the polymer melt in the rotating extrusion process are obtained by performing calculation by means of using the derived formula.Type: GrantFiled: November 22, 2018Date of Patent: September 10, 2024Assignee: SICHUAN UNIVERSITYInventors: Qi Wang, Min Nie, Lin Pi, Yijun Li, Shibing Bai
-
Publication number: 20240296607Abstract: The present disclosure relates to systems, non-transitory computer-readable media, and methods for utilizing machine learning models to generate modified digital images. In particular, in some embodiments, the disclosed systems generate image editing directions between textual identifiers of two visual features utilizing a language prediction machine learning model and a text encoder. In some embodiments, the disclosed systems generated an inversion of a digital image utilizing a regularized inversion model to guide forward diffusion of the digital image. In some embodiments, the disclosed systems utilize cross-attention guidance to preserve structural details of a source digital image when generating a modified digital image with a diffusion neural network.Type: ApplicationFiled: March 3, 2023Publication date: September 5, 2024Inventors: Yijun Li, Richard Zhang, Krishna Kumar Singh, Jingwan Lu, Gaurav Parmar, Jun-Yan Zhu
-
Publication number: 20240290022Abstract: Avatar generation from an image is performed using semi-supervised machine learning. An image space model undergoes unsupervised training from images to generate latent image vectors responsive to image inputs. An avatar parameter space model undergoes unsupervised training from avatar parameter values for avatar parameters to generate latent avatar parameter vectors responsive to avatar parameter value inputs. A cross-modal mapping model undergoes supervised training on image-avatar parameter pair inputs corresponding to the latent image vectors and the latent avatar parameter vectors. The trained image space model generates a latent image vector from an image input. The trained cross-modal mapping model translates the latent image vector to a latent avatar parameter vector. The trained avatar parameter space model generates avatar parameter values from the latent avatar parameter vector. The latent avatar parameter vector can be used to render an avatar having features corresponding to the input image.Type: ApplicationFiled: February 28, 2023Publication date: August 29, 2024Inventors: Yijun LI, Yannick HOLD-GEOFFROY, Manuel Rodriguez Ladron DE GUEVARA, Jose Ignacio Echevarria VALLESPI, Daichi ITO, Cameron Younger SMITH
-
Publication number: 20240233318Abstract: An image generation system implements a multi-branch GAN to generate images that each express visually similar content in a different modality. A generator portion of the multi-branch GAN includes multiple branches that are each tasked with generating one of the different modalities. A discriminator portion of the multi-branch GAN includes multiple fidelity discriminators, one for each of the generator branches, and a consistency discriminator, which constrains the outputs generated by the different generator branches to appear visually similar to one another. During training, outputs from each of the fidelity discriminators and the consistency discriminator are used to compute a non-saturating GAN loss. The non-saturating GAN loss is used to refine parameters of the multi-branch GAN during training until model convergence. The trained multi-branch GAN generates multiple images from a single input, where each of the multiple images depicts visually similar content expressed in a different modality.Type: ApplicationFiled: October 21, 2022Publication date: July 11, 2024Applicant: Adobe Inc.Inventors: Yijun Li, Zhixin Shu, Zhen Zhu, Krishna Kumar Singh
-
Publication number: 20240221252Abstract: Systems and methods for image processing are described. Embodiments of the present disclosure identify an original image depicting a face, identify a scribble image including a mask that indicates a portion of the original image for adding makeup to the face, and generate a target image depicting the face using a machine learning model based on the original image and the scribble image, where the target image includes the makeup in the portion indicated by the scribble image.Type: ApplicationFiled: January 4, 2023Publication date: July 4, 2024Inventors: Abhishek Lalwani, Xiaoyang Li, Yijun Li
-
Publication number: 20240169488Abstract: Systems and methods for synthesizing images with increased high-frequency detail are described. Embodiments are configured to identify an input image including a noise level and encode the input image to obtain image features. A diffusion model reduces a resolution of the image features at an intermediate stage of the model using a wavelet transform to obtain reduced image features at a reduced resolution, and generates an output image based on the reduced image features using the diffusion model. In some cases, the output image comprises a version of the input image that has a reduced noise level compared to the noise level of the input image.Type: ApplicationFiled: November 17, 2022Publication date: May 23, 2024Inventors: Nan Liu, Yijun Li, Michaƫl Yanis Gharbi, Jingwan Lu