Patents by Inventor Sai Bi

Sai Bi has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20250139883
    Abstract: Embodiments are configured to render 3D models using an importance sampling method. First, embodiments obtain a 3D model including a plurality of density values corresponding to a plurality of locations in a 3D space, respectively. Embodiments then sample the color information from within a random subset of the plurality of locations using a probability distribution based on the plurality of density values. Embodiments have a higher probability to sample each location within the random subset of locations if the location has a higher density probability. Embodiments then an image depicting a view of the 3D model based on the sampling within the random subset of the plurality of locations.
    Type: Application
    Filed: November 1, 2023
    Publication date: May 1, 2025
    Inventors: Milos Hasan, Iliyan Georgiev, Sai Bi, Julien Philip, Kalyan K. Sunkavalli, Xin Sun, Fujun Luan, Kevin James Blackburn-Matzen, Zexiang Xu, Kai Zhang
  • Publication number: 20250104349
    Abstract: A method, apparatus, non-transitory computer readable medium, and system for 3D model generation include obtaining a plurality of input images depicting an object and a set of 3D position embeddings, where each of the plurality of input images depicts the object from a different perspective, encoding the plurality of input images to obtain a plurality of 2D features corresponding to the plurality of input images, respectively, generating 3D features based on the plurality of 2D features and the set of 3D position embeddings, and generating a 3D model of the object based on the 3D features.
    Type: Application
    Filed: September 24, 2024
    Publication date: March 27, 2025
    Inventors: Sai Bi, Jiahao Li, Hao Tan, Kai Zhang, Zexiang Xu, Fujun Luan, Yinghao Xu, Yicong Hong, Kalyan K. Sunkavalli
  • Patent number: 12254570
    Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media that generate three-dimensional hybrid mesh-volumetric representations for digital objects. For instance, in one or more embodiments, the disclosed systems generate a mesh for a digital object from a plurality of digital images that portray the digital object using a multi-view stereo model. Additionally, the disclosed systems determine a set of sample points for a thin volume around the mesh. Using a neural network, the disclosed systems further generate a three-dimensional hybrid mesh-volumetric representation for the digital object utilizing the set of sample points for the thin volume and the mesh.
    Type: Grant
    Filed: May 3, 2022
    Date of Patent: March 18, 2025
    Assignee: Adobe Inc.
    Inventors: Sai Bi, Yang Liu, Zexiang Xu, Fujun Luan, Kalyan Sunkavalli
  • Publication number: 20250078393
    Abstract: Systems and methods for generating a 3D model from a single input image are described. Embodiments are configured to obtain an input image and camera view information corresponding to the input image; encode the input image to obtain 2D features comprising a plurality of 2D tokens corresponding to patches of the input image; decode the 2D features based on the camera view information to obtain 3D features comprising a plurality of 3D tokens corresponding to regions of a 3D representation; and generate a 3D model of the input image based on the 3D features.
    Type: Application
    Filed: September 5, 2023
    Publication date: March 6, 2025
    Inventors: HAO TAN, YICONG HONG, KAI ZHANG, JIUXIANG GU, SAI BI, YANG ZHOU, DIFAN LIU, FENG LIU, KALYAN K. SUNKAVALLI, TRUNG HUU BUI
  • Patent number: 12211225
    Abstract: A scene reconstruction system renders images of a scene with high-quality geometry and appearance and supports view synthesis, relighting, and scene editing. Given a set of input images of a scene, the scene reconstruction system trains a network to learn a volume representation of the scene that includes separate geometry and reflectance parameters. Using the volume representation, the scene reconstruction system can render images of the scene under arbitrary viewing (view synthesis) and lighting (relighting) locations. Additionally, the scene reconstruction system can render images that change the reflectance of objects in the scene (scene editing).
    Type: Grant
    Filed: April 15, 2021
    Date of Patent: January 28, 2025
    Assignee: ADOBE INC.
    Inventors: Sai Bi, Zexiang Xu, Kalyan Krishna Sunkavalli, Miloš Hašan, Yannick Hold-Geoffroy, David Jay Kriegman, Ravi Ramamoorthi
  • Patent number: 12211138
    Abstract: Embodiments of the present disclosure provide systems, methods, and computer storage media for generating editable synthesized views of scenes by inputting image rays into neural networks using neural basis decomposition. In embodiments, a set of input images of a scene depicting at least one object are collected and used to generate a plurality of rays of the scene. The rays each correspond to three-dimensional coordinates and viewing angles taken from the images. A volume density of the scene is determined by inputting the three-dimensional coordinates from the neural radiance fields into a first neural network to generate a 3D geometric representation of the object. An appearance decomposition is produced by inputting the three-dimensional coordinates and the viewing angles of the rays into a second neural network.
    Type: Grant
    Filed: December 13, 2022
    Date of Patent: January 28, 2025
    Assignee: Adobe Inc.
    Inventors: Zhengfei Kuang, Fujun Luan, Sai Bi, Zhixin Shu, Kalyan K. Sunkavalli
  • Patent number: 12190428
    Abstract: A method for providing a relightable avatar of a subject to a virtual reality application is provided. The method includes retrieving multiple images including multiple views of a subject and generating an expression-dependent texture map and a view-dependent texture map for the subject, based on the images. The method also includes generating, based on the expression-dependent texture map and the view-dependent texture map, a view of the subject illuminated by a light source selected from an environment in an immersive reality application, and providing the view of the subject to an immersive reality application running in a client device. A non-transitory, computer-readable medium storing instructions and a system that executes the instructions to perform the above method are also provided.
    Type: Grant
    Filed: June 13, 2023
    Date of Patent: January 7, 2025
    Assignee: Meta Platforms Technologies, LLC
    Inventors: Jason Saragih, Stephen Anthony Lombardi, Shunsuke Saito, Tomas Simon Kreuz, Shih-En Wei, Kevyn Alex Anthony McPhail, Yaser Sheikh, Sai Bi
  • Publication number: 20240404181
    Abstract: A scene modeling system receives a plurality of input two-dimensional (2D) images corresponding to a plurality of views of an object and a request to display a three-dimensional (3D) scene that includes the object. The scene modeling system generates an output 2D image for a view of the 3D scene by applying a scene representation model to the input 2D images. The scene representation model includes a point cloud generation model configured to generate, based on the input 2D images, a neural point cloud representing the 3D scene. The scene representation model includes a neural point volume rendering model configured to determine, for each pixel of the output image and using the neural point cloud and a volume rendering process, a color value. The scene modeling system transmits, responsive to the request, the output 2D image. Each pixel of the output image includes the respective determined color value.
    Type: Application
    Filed: August 9, 2024
    Publication date: December 5, 2024
    Inventors: Zexiang Xu, Zhixin Shu, Sai Bi, Qiangeng Xu, Kalyan Sunkavalli, Julien Philip
  • Publication number: 20240338915
    Abstract: Certain aspects and features of this disclosure relate to providing a controllable, dynamic appearance for neural 3D portraits. For example, a method involves projecting a color at points in a digital video portrait based on location, surface normal, and viewing direction for each respective point in a canonical space. The method also involves projecting, using the color, dynamic face normals for the points as changing according to an articulated head pose and facial expression in the digital video portrait. The method further involves disentangling, based on the dynamic face normals, a facial appearance in the digital video portrait into intrinsic components in the canonical space. The method additionally involves storing and/or rendering at least a portion of a head pose as a controllable, neural 3D portrait based on the digital video portrait using the intrinsic components.
    Type: Application
    Filed: April 7, 2023
    Publication date: October 10, 2024
    Inventors: Zhixin Shu, Zexiang Xu, Shahrukh Athar, Sai Bi, Kalyan Sunkavalli, Fujun Luan
  • Publication number: 20240312118
    Abstract: Embodiments are disclosed for fast large-scale radiance field reconstruction. A method of fast large-scale radiance field reconstruction may include receiving a sequence of input images that depict views of a scene and extracting, using an image encoder, image features from the sequence of input images. A first one or more machine learning models may generate a local volume based on the image features corresponding to one or more images from the sequence of input images. A second one or more machine learning models may generate a global volume based on the local volume. A novel view of the scene is synthesized based on the global volume.
    Type: Application
    Filed: March 16, 2023
    Publication date: September 19, 2024
    Inventors: Zexiang XU, Xiaoshuai ZHANG, Sai BI, Kalyan SUNKAVALLI, Hao SU
  • Patent number: 12073507
    Abstract: A scene modeling system receives a plurality of input two-dimensional (2D) images corresponding to a plurality of views of an object and a request to display a three-dimensional (3D) scene that includes the object. The scene modeling system generates an output 2D image for a view of the 3D scene by applying a scene representation model to the input 2D images. The scene representation model includes a point cloud generation model configured to generate, based on the input 2D images, a neural point cloud representing the 3D scene. The scene representation model includes a neural point volume rendering model configured to determine, for each pixel of the output image and using the neural point cloud and a volume rendering process, a color value. The scene modeling system transmits, responsive to the request, the output 2D image. Each pixel of the output image includes the respective determined color value.
    Type: Grant
    Filed: July 9, 2022
    Date of Patent: August 27, 2024
    Assignee: Adobe Inc.
    Inventors: Zexiang Xu, Zhixin Shu, Sai Bi, Qiangeng Xu, Kalyan Sunkavalli, Julien Philip
  • Publication number: 20240193850
    Abstract: Embodiments of the present disclosure provide systems, methods, and computer storage media for generating editable synthesized views of scenes by inputting image rays into neural networks using neural basis decomposition. In embodiments, a set of input images of a scene depicting at least one object are collected and used to generate a plurality of rays of the scene. The rays each correspond to three-dimensional coordinates and viewing angles taken from the images. A volume density of the scene is determined by inputting the three-dimensional coordinates from the neural radiance fields into a first neural network to generate a 3D geometric representation of the object. An appearance decomposition is produced by inputting the three-dimensional coordinates and the viewing angles of the rays into a second neural network.
    Type: Application
    Filed: December 13, 2022
    Publication date: June 13, 2024
    Inventors: Zhengfei Kuang, Fujun Luan, Sai Bi, Zhixin Shu, Kalyan K. Sunkavalli
  • Publication number: 20240127402
    Abstract: In some examples, a computing system accesses a field of view (FOV) image that has a field of view less than 360 degrees and has low dynamic range (LDR) values. The computing system estimates lighting parameters from a scene depicted in the FOV image and generates a lighting image based on the lighting parameters. The computing system further generates lighting features generated the lighting image and image features generated from the FOV image. These features are aggregated into aggregated features and a machine learning model is applied to the image features and the aggregated features to generate a panorama image having high dynamic range (HDR) values.
    Type: Application
    Filed: August 25, 2023
    Publication date: April 18, 2024
    Inventors: Mohammad Reza Karimi Dastjerdi, Yannick Hold-Geoffroy, Sai Bi, Jonathan Eisenmann, Jean-François Lalonde
  • Publication number: 20240013477
    Abstract: A scene modeling system receives a plurality of input two-dimensional (2D) images corresponding to a plurality of views of an object and a request to display a three-dimensional (3D) scene that includes the object. The scene modeling system generates an output 2D image for a view of the 3D scene by applying a scene representation model to the input 2D images. The scene representation model includes a point cloud generation model configured to generate, based on the input 2D images, a neural point cloud representing the 3D scene. The scene representation model includes a neural point volume rendering model configured to determine, for each pixel of the output image and using the neural point cloud and a volume rendering process, a color value. The scene modeling system transmits, responsive to the request, the output 2D image. Each pixel of the output image includes the respective determined color value.
    Type: Application
    Filed: July 9, 2022
    Publication date: January 11, 2024
    Inventors: Zexiang Xu, Zhixin Shu, Sai Bi, Qiangeng Xu, Kalyan Sunkavalli, Julien Philip
  • Publication number: 20230360327
    Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media that generate three-dimensional hybrid mesh-volumetric representations for digital objects. For instance, in one or more embodiments, the disclosed systems generate a mesh for a digital object from a plurality of digital images that portray the digital object using a multi-view stereo model. Additionally, the disclosed systems determine a set of sample points for a thin volume around the mesh. Using a neural network, the disclosed systems further generate a three-dimensional hybrid mesh-volumetric representation for the digital object utilizing the set of sample points for the thin volume and the mesh.
    Type: Application
    Filed: May 3, 2022
    Publication date: November 9, 2023
    Inventors: Sai Bi, Yang Liu, Zexiang Xu, Fujun Luan, Kalyan Sunkavalli
  • Publication number: 20230326112
    Abstract: A method for providing a relightable avatar of a subject to a virtual reality application is provided. The method includes retrieving multiple images including multiple views of a subject and generating an expression-dependent texture map and a view-dependent texture map for the subject, based on the images. The method also includes generating, based on the expression-dependent texture map and the view-dependent texture map, a view of the subject illuminated by a light source selected from an environment in an immersive reality application, and providing the view of the subject to an immersive reality application running in a client device. A non-transitory, computer-readable medium storing instructions and a system that executes the instructions to perform the above method are also provided.
    Type: Application
    Filed: June 13, 2023
    Publication date: October 12, 2023
    Inventors: Jason Saragih, Stephen Anthony Lombardi, Shunsuke Saito, Tomas Simon Kreuz, Shih-En Wei, Kevyn Alex Anthony McPhail, Yaser Sheikh, Sai Bi
  • Patent number: 11715248
    Abstract: A method for providing a relightable avatar of a subject to a virtual reality application is provided. The method includes retrieving multiple images including multiple views of a subject and generating an expression-dependent texture map and a view-dependent texture map for the subject, based on the images. The method also includes generating, based on the expression-dependent texture map and the view-dependent texture map, a view of the subject illuminated by a light source selected from an environment in an immersive reality application, and providing the view of the subject to an immersive reality application running in a client device. A non-transitory, computer-readable medium storing instructions and a system that executes the instructions to perform the above method are also provided.
    Type: Grant
    Filed: January 20, 2022
    Date of Patent: August 1, 2023
    Assignee: Meta Platforms Technologies, LLC
    Inventors: Jason Saragih, Stephen Anthony Lombardi, Shunsuke Saito, Tomas Simon Kreuz, Shih-En Wei, Kevyn Alex Anthony McPhail, Yaser Sheikh, Sai Bi
  • Patent number: 11669986
    Abstract: Enhanced methods and systems for generating both a geometry model and an optical-reflectance model (an object reconstruction model) for a physical object, based on a sparse set of images of the object under a sparse set of viewpoints. The geometry model is a mesh model that includes a set of vertices representing the object's surface. The reflectance model is SVBRDF that is parameterized via multiple channels (e.g., diffuse albedo, surface-roughness, specular albedo, and surface-normals). For each vertex of the geometry model, the reflectance model includes a value for each of the multiple channels. The object reconstruction model is employed to render graphical representations of a virtualized object (a VO based on the physical object) within a computation-based (e.g., a virtual or immersive) environment. Via the reconstruction model, the VO may be rendered from arbitrary viewpoints and under arbitrary lighting conditions.
    Type: Grant
    Filed: April 16, 2021
    Date of Patent: June 6, 2023
    Assignees: ADOBE INC., THE REGENTS OF THE UNIVERSITY OF CALIFORNIA
    Inventors: Sai Bi, Zexiang Xu, Kalyan Krishna Sunkavalli, David Jay Kriegman, Ravi Ramamoorthi
  • Publication number: 20220343522
    Abstract: Enhanced methods and systems for generating both a geometry model and an optical-reflectance model (an object reconstruction model) for a physical object, based on a sparse set of images of the object under a sparse set of viewpoints. The geometry model is a mesh model that includes a set of vertices representing the object's surface. The reflectance model is SVBRDF that is parameterized via multiple channels (e.g., diffuse albedo, surface-roughness, specular albedo, and surface-normals). For each vertex of the geometry model, the reflectance model includes a value for each of the multiple channels. The object reconstruction model is employed to render graphical representations of a virtualized object (a VO based on the physical object) within a computation-based (e.g., a virtual or immersive) environment. Via the reconstruction model, the VO may be rendered from arbitrary viewpoints and under arbitrary lighting conditions.
    Type: Application
    Filed: April 16, 2021
    Publication date: October 27, 2022
    Inventors: Sai Bi, Zexiang Xu, Kalyan Krishna Sunkavalli, David Jay Kriegman, Ravi Ramamoorthi
  • Publication number: 20220335636
    Abstract: A scene reconstruction system renders images of a scene with high-quality geometry and appearance and supports view synthesis, relighting, and scene editing. Given a set of input images of a scene, the scene reconstruction system trains a network to learn a volume representation of the scene that includes separate geometry and reflectance parameters. Using the volume representation, the scene reconstruction system can render images of the scene under arbitrary viewing (view synthesis) and lighting (relighting) locations. Additionally, the scene reconstruction system can render images that change the reflectance of objects in the scene (scene editing).
    Type: Application
    Filed: April 15, 2021
    Publication date: October 20, 2022
    Inventors: Sai Bi, Zexiang Xu, Kalyan Krishna Sunkavalli, Milos Hasan, Yannick Hold-Geoffroy, David Jay Kriegman, Ravi Ramamoorthi