Patents by Inventor Tomas SIMON
Tomas SIMON has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20220245910Abstract: A method for training a real-time, modeling for animating an avatar for a subject is provided. The method includes collecting multiple images of a subject. The method also includes selecting a plurality of vertex positions in a guide mesh, indicative of a volumetric primitive enveloping the subject, determining a geometric attribute for the volumetric primitive including a position, a rotation, and a scale factor of the volumetric primitive, determining a payload attribute for each of the volumetric primitive, the payload attribute including a color value and an opacity value for each voxel in a voxel grid defining the volumetric primitive, determining a loss factor for each point in the volumetric primitive based on the geometric attribute, the payload attribute and a ground truth value, and updating a three-dimensional model for the subject. A system and a non-transitory, computer-readable medium storing instructions to perform the above method are also provided.Type: ApplicationFiled: December 17, 2021Publication date: August 4, 2022Inventors: Stephen Anthony Lombardi, Tomas Simon Kreuz, Jason Saragih, Gabriel Bailowitz Schwartz, Michael Zollhoefer, Yaser Sheikh
-
Publication number: 20220237843Abstract: A method for providing a relightable avatar of a subject to a virtual reality application is provided. The method includes retrieving multiple images including multiple views of a subject and generating an expression-dependent texture map and a view-dependent texture map for the subject, based on the images. The method also includes generating, based on the expression-dependent texture map and the view-dependent texture map, a view of the subject illuminated by a light source selected from an environment in an immersive reality application, and providing the view of the subject to an immersive reality application running in a client device. A non-transitory, computer-readable medium storing instructions and a system that executes the instructions to perform the above method are also provided.Type: ApplicationFiled: January 20, 2022Publication date: July 28, 2022Inventors: Jason Saragih, Stephen Anthony Lombardi, Shunsuke Saito, Tomas Simon Kreuz, Shih-En Wei, Kevyn Alex Anthony McPhail, Yaser Sheikh, Sai Bi
-
Publication number: 20220207831Abstract: A method for simulating a solid body animation of a subject includes retrieving a first frame that includes a body image of a subject. The method also includes selecting, from the first frame, multiple key points within the body image of the subject that define a hull of a body part and multiple joint points that define a joint between two body parts, identifying a geometry, a speed, and a mass of the body part to include in a dynamic model of the subject, based on the key points and the joint points, determining, based on the dynamic model of the subject, a pose of the subject in a second frame after the first frame in a video stream, and providing the video stream to an immersive reality application running on a client device.Type: ApplicationFiled: December 20, 2021Publication date: June 30, 2022Inventors: Jason Saragih, Shih-En Wei, Tomas Simon Kreuz, Kris Makoto Kitani, Ye Yuan
-
Publication number: 20220198731Abstract: A method of forming a pixel-aligned volumetric avatar includes receiving multiple two-dimensional images having at least two or more fields of view of a subject. The method also includes extracting multiple image features from the two-dimensional images using a set of learnable weights, projecting the image features along a direction between a three-dimensional model of the subject and a selected observation point for a viewer, and providing, to the viewer, an image of the three-dimensional model of the subject. A system and a non-transitory, computer readable medium storing instructions to perform the above method, are also provided.Type: ApplicationFiled: December 20, 2021Publication date: June 23, 2022Inventors: Stephen Anthony Lombardi, Jason Saragih, Tomas Simon Kreuz, Shunsuke Saito, Michael Zollhoefer, Amit Raj, James Henry Hays
-
Publication number: 20220201273Abstract: A device for providing a reverse pass-through view of a user of a headset display to an onlooker includes an eyepiece comprising an optical surface configured to provide an image to a user on a first side of the optical surface. The device also includes a first camera configured to collect an image of a portion of a face of the user reflected from the optical surface in a first field of view, a display adjacent to the optical surface and configured to project forward an image of the face of the user, and a screen configured to receive light from the display and provide the image of the face of the user to an onlooker.Type: ApplicationFiled: December 17, 2021Publication date: June 23, 2022Inventors: Nathan Matsuda, Brian Wheelwright, Joel Hegland, Stephen Anthony Lombardi, Jason Saragih, Tomas Simon Kreuz, Shunsuke Saito, Michael Zollhoefer, Amit Raj, James Henry Hays
-
Patent number: 11182947Abstract: In one embodiment, a system may access a codec that encodes an appearance associated with a subject and comprise codec portions that respectively correspond to body parts of the subject. The system may generate a training codec that comprises a first subset of the codec portions (a first set of body parts) and a modified second subset of the codec portions (muted body parts). The system may decode the training codec using a machine-learning model to generate a mesh of the subject. The system may transform the mesh of the subject based on a predetermined pose. The system may update the machine-learning model based on a comparison between the transformed mesh and a target mesh of the subject having the predetermined pose. The system in the present application can train a machine-learning model to render an avatar with a pose using uncorrelated codec portions corresponding to different body parts.Type: GrantFiled: April 17, 2020Date of Patent: November 23, 2021Assignee: Facebook Technologies, LLC.Inventors: Chenglei Wu, Jason Saragih, Tomas Simon Kreuz, Takaaki Shiratori
-
Patent number: 11087521Abstract: The disclosed computer system may include an input module, an autoencoder, and a rendering module. The input module may receive geometry information and images of a subject. The geometry information may be indicative of variation in geometry of the subject over time. Each image may be associated with a respective viewpoint and may include a view-dependent texture map of the subject. The autoencoder may jointly encode texture information and the geometry information to provide a latent vector. The autoencoder may infer, using the latent vector, an inferred geometry and an inferred view-dependent texture of the subject for a predicted viewpoint. The rendering module may be configured to render a reconstructed image of the subject for the predicted viewpoint using the inferred geometry and the inferred view-dependent texture. Various other systems and methods are also disclosed.Type: GrantFiled: January 29, 2020Date of Patent: August 10, 2021Assignee: Facebook Technologies, LLCInventors: Stephen Anthony Lombardi, Jason Saragih, Yaser Sheikh, Takaaki Shiratori, Shoou-I Yu, Tomas Simon Kreuz, Chenglei Wu
-
Patent number: 11062502Abstract: In one embodiment, a method includes accessing a number of pictures of an object, constructing a modeling volume for three-dimensional modeling of the object by processing the number of pictures using a machine-learning framework, where the modeling volume is associated with a number of color and opacity information that are associated with a number of regions in the modeling volume, and rendering an image of the object from a view-point using the modeling volume, where each pixel of the image is rendered by projecting a virtual ray from the view-point and through the modeling volume, determining one or more of the number of regions in the modeling volume intersected by the virtual ray, and determining a color and an opacity of the pixel based on an accumulation of the color and opacity information associated with the one or more of the number of regions intersected by the virtual ray.Type: GrantFiled: April 9, 2019Date of Patent: July 13, 2021Assignee: Facebook Technologies, LLCInventors: Jason Saragih, Stephen Anthony Lombardi, Tomas Simon Kreuz, Gabriel Bailowitz Schwartz
-
Patent number: 11010951Abstract: In one embodiment, a system may capture one or more images of a user using one or more cameras, the one or more images depicting at least an eye and a face of the user. The system may determine a direction of a gaze of the user based on the eye depicted in the one or more images. The system may generate a facial mesh based on depth measurements of one or more features of the face depicted in the one or more images. The system may generate an eyeball texture for an eyeball mesh by processing the direction of the gaze and the facial mesh using a machine-learning model. The system may render an avatar of the user based on the eyeball mesh, the eyeball texture, the facial mesh, and a facial texture.Type: GrantFiled: January 9, 2020Date of Patent: May 18, 2021Assignee: Facebook Technologies, LLCInventors: Gabriel Bailowitz Schwartz, Jason Saragih, Tomas Simon Kreuz, Shih-En Wei, Stephen Anthony Lombardi
-
Publication number: 20200327718Abstract: In one embodiment, a method includes accessing a number of pictures of an object, constructing a modeling volume for three-dimensional modeling of the object by processing the number of pictures using a machine-learning framework, where the modeling volume is associated with a number of color and opacity information that are associated with a number of regions in the modeling volume, and rendering an image of the object from a view-point using the modeling volume, where each pixel of the image is rendered by projecting a virtual ray from the view-point and through the modeling volume, determining one or more of the number of regions in the modeling volume intersected by the virtual ray, and determining a color and an opacity of the pixel based on an accumulation of the color and opacity information associated with the one or more of the number of regions intersected by the virtual ray.Type: ApplicationFiled: April 9, 2019Publication date: October 15, 2020Inventors: Jason Saragih, Stephen Anthony Lombardi, Tomas Simon Kreuz, Gabriel Bailowitz Schwartz
-
Patent number: 10586370Abstract: The disclosed computer system may include an input module, an autoencoder, and a rendering module. The input module may receive geometry information and images of a subject. The geometry information may be indicative of variation in geometry of the subject over time. Each image may be associated with a respective viewpoint and may include a view-dependent texture map of the subject. The autoencoder may jointly encode texture information and the geometry information to provide a latent vector. The autoencoder may infer, using the latent vector, an inferred geometry and an inferred view-dependent texture of the subject for a predicted viewpoint. The rendering module may be configured to render a reconstructed image of the subject for the predicted viewpoint using the inferred geometry and the inferred view-dependent texture. Various other systems and methods are also disclosed.Type: GrantFiled: July 31, 2018Date of Patent: March 10, 2020Assignee: Facebook Technologies, LLCInventors: Stephen Anthony Lombardi, Jason Saragih, Yaser Sheikh, Takaaki Shiratori, Shoou-I Yu, Tomas Simon Kreuz, Chenglei Wu
-
Publication number: 20190213772Abstract: The disclosed computer system may include an input module, an autoencoder, and a rendering module. The input module may receive geometry information and images of a subject. The geometry information may be indicative of variation in geometry of the subject over time. Each image may be associated with a respective viewpoint and may include a view-dependent texture map of the subject. The autoencoder may jointly encode texture information and the geometry information to provide a latent vector. The autoencoder may infer, using the latent vector, an inferred geometry and an inferred view-dependent texture of the subject for a predicted viewpoint. The rendering module may be configured to render a reconstructed image of the subject for the predicted viewpoint using the inferred geometry and the inferred view-dependent texture. Various other systems and methods are also disclosed.Type: ApplicationFiled: July 31, 2018Publication date: July 11, 2019Inventors: Stephen Anthony Lombardi, Jason Saragih, Yaser Sheikh, Takaaki Shiratori, Shoou-I Yu, Tomas Simon Kreuz, Chenglei Wu
-
Patent number: 9883167Abstract: Embodiments disclosed herein relate to a method and apparatus for generating a three-dimensional surface. In one embodiment, there is a method for generating a three-dimensional surface. The method includes capturing a plurality of images of a target object with at least two cameras, the target object illuminated by at least two sets of red-green-blue (RGB) lights positioned in an array about the target object and generating a three-dimensional surface of the target object by iteratively reconstructing a surface estimate of the target object and aligning images of the target object using motion estimation until the images converge, wherein the images are processed in n-frame intervals.Type: GrantFiled: September 25, 2015Date of Patent: January 30, 2018Assignees: Disney Enterprises, Inc., Carnegie Mellon UniversityInventors: Iain Matthews, Paulo Urnau Gotardo, Tomas Simon, Yaser Sheikh
-
Publication number: 20170094253Abstract: Embodiments disclosed herein relate to a method and apparatus for generating a three-dimensional surface. In one embodiment, there is a method for generating a three-dimensional surface. The method includes capturing a plurality of images of a target object with at least two cameras, the target object illuminated by at least two sets of red-green-blue (RGB) lights positioned in an array about the target object and generating a three-dimensional surface of the target object by iteratively reconstructing a surface estimate of the target object and aligning images of the target object using motion estimation until the images converge, wherein the images are processed in n-frame intervals.Type: ApplicationFiled: September 25, 2015Publication date: March 30, 2017Inventors: Iain MATTHEWS, Paulo Urnau GOTARDO, Tomas SIMON, Yaser SHEIKH
-
Patent number: 9041718Abstract: Techniques are disclosed for generating a bilinear spatiotemporal basis model. A method includes the steps of predefining a trajectory basis for the bilinear spatiotemporal basis model, receiving three-dimensional spatiotemporal data for a training sequence, estimating a shape basis for the bilinear spatiotemporal basis model using the three-dimensional spatiotemporal data, and computing coefficients for the bilinear spatiotemporal basis model using the trajectory basis and the shape basis.Type: GrantFiled: March 20, 2012Date of Patent: May 26, 2015Assignee: Disney Enterprises, Inc.Inventors: Iain Matthews, Ijaz Akhter, Tomas Simon, Sohaib Khan, Yaser Sheikh
-
Publication number: 20130249905Abstract: Techniques are disclosed for generating a bilinear spatiotemporal basis model. A method includes the steps of predefining a trajectory basis for the bilinear spatiotemporal basis model, receiving three-dimensional spatiotemporal data for a training sequence, estimating a shape basis for the bilinear spatiotemporal basis model using the three-dimensional spatiotemporal data, and computing coefficients for the bilinear spatiotemporal basis model using the trajectory basis and the shape basis.Type: ApplicationFiled: March 20, 2012Publication date: September 26, 2013Applicant: DISNEY ENTERPRISES, INC.Inventors: Iain MATTHEWS, Ijaz AKHTER, Tomas SIMON, Sohaib KHAN, Yaser SHEIKH