Patents by Inventor Edilson De Aguiar
Edilson De Aguiar has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11164381Abstract: An improved system to provide a clothing model in augmented reality and/or virtual reality is disclosed. In one embodiment, the system provides a fully simulated garment, on the user's device, in an augmented reality or virtual reality display. In one embodiment, the process uses a parameterized smart garment model which enables real-time interaction with the garment on a device such as a mobile phone or tablet, with limited memory and processing power. In one embodiment, the system permits nearly instantaneous alteration of garment sizes. Because the garment is modeled to accurately reflect physical details, the lay of the garment changes as the size of the garment or body changes.Type: GrantFiled: November 2, 2018Date of Patent: November 2, 2021Assignee: GERBER TECHNOLOGY LLCInventors: James F O'Brien, David T Jackson, Carlo Camporesi, Daniel Ram, David Macy, Edilson de Aguiar, James L. Andrews, Justin Lee, Karen M Stritzinger, Kevin Armin Samii, Nathan Mitchell, Tobias Pfaff, Scott M Frankel
-
Publication number: 20210326955Abstract: The present invention relates to building a computer model of a garment based on a physical sample garment, and to the process of using the computer model of a garment to determine the garment's appearance on a human body.Type: ApplicationFiled: July 1, 2021Publication date: October 21, 2021Inventors: James F. O'Brien, David T. Jackson, Edilson de Aguiar, James L. Andrews, Carlo Camporesi, Scott M. Frankel, Zoran Kacic-Alesic, David Macy, Nathan Mitchell, Tobias Pfaff, Daniel Ram, Kevin Armin Samii
-
Publication number: 20200250892Abstract: The present invention relates to building a computer model of a garment based on a physical sample garment, and to the process of using the computer model of a garment to determine the garment's appearance on a human body.Type: ApplicationFiled: May 13, 2019Publication date: August 6, 2020Inventors: James F. O'Brien, David T. Jackson, Edilson de Aguiar, James L. Andrews, Carlo Camporesi, Scott M. Frankel, Zoran Kacic-Alesic, David Macy, Nathan Mitchell, Tobias Pfaff, Daniel Ram, Kevin Armin Samii
-
Publication number: 20190130649Abstract: An improved system to provide a clothing model in augmented reality and/or virtual reality is disclosed. In one embodiment, the system provides a fully simulated garment, on the user's device, in an augmented reality or virtual reality display. In one embodiment, the process uses a parameterized smart garment model which enables real-time interaction with the garment on a device such as a mobile phone or tablet, with limited memory and processing power. In one embodiment, the system permits nearly instantaneous alteration of garment sizes. Because the garment is modeled to accurately reflect physical details, the lay of the garment changes as the size of the garment or body changes.Type: ApplicationFiled: November 2, 2018Publication date: May 2, 2019Inventors: James F O'Brien, David T Jackson, Carlo Camporesi, Daniel Ram, David Macy, Edilson de Aguiar, James L Andrews, Justin Lee, Karen M Stritzinger, Kevin Armin Samii, Nathan Mitchell, Tobias Pfaff, Scott M Frankel
-
Patent number: 9704288Abstract: Techniques are disclosed for providing a learning-based clothing model that enables the simultaneous animation of multiple detailed garments in real-time. A simple conditional model learns and preserves key dynamic properties of cloth motions and folding details. Such a conditional model may be generated for each garment worn by a given character. Once generated, the conditional model may be used to determine complex body/cloth interactions in order to render the character and garment from frame-to-frame. The clothing model may be used for a variety of garments worn by male and female human characters (as well as non-human characters) while performing a varied set of motions typically used in video games (e.g., walking, running, jumping, turning, etc.).Type: GrantFiled: December 21, 2010Date of Patent: July 11, 2017Assignee: Disney Enterprises, Inc.Inventors: Edilson de Aguiar, Leonid Sigal, Adrien Treuille, Jessica K. Hodgins
-
Patent number: 9619914Abstract: Systems and methods are described for animating 3D characters using synthetic motion data generated by motion models in response to a high level description of a desired sequence of motion provided by an animator. In a number of embodiments, the synthetic motion data is streamed to a user device that includes a rendering engine and the user device renders an animation of a 3D character using the streamed synthetic motion data. In several embodiments, an animator can upload a custom model of a 3D character or a custom 3D character is generated by the server system in response to a high level description of a desired 3D character provided by the user and the synthetic motion data generated by the generative model is retargeted to animate the custom 3D character.Type: GrantFiled: December 2, 2013Date of Patent: April 11, 2017Assignee: FACEBOOK, INC.Inventors: Edilson de Aguiar, Emiliano Gambaretto, Stefano Corazza
-
Publication number: 20170046769Abstract: A method and apparatus to provide a depiction of a garment model of a body shape is described. The garment model takes into account the visual and mechanical characteristics of the fabric, as well as the drape of the fabric on a body shape.Type: ApplicationFiled: August 9, 2016Publication date: February 16, 2017Inventors: David T. Jackson, Zoran Kacic-Alesic, Scott M. Frankel, James F. O'Brien, Tobias Pfaff, James L. Andrews, Robert A. Backman, Edilson de Aguiar, Carlo Camporesi, Tyler N. Martin, Gregory R. Piesco-Putnam, Karen M. Stritzinger
-
Patent number: 9460539Abstract: Systems and methods are described for performing spatial and temporal compression of deformable mesh based representations of 3D character motion allowing the visualization of high-resolution 3D character animations in real time. In a number of embodiments, the deformable mesh based representation of the 3D character motion is used to automatically generate an interconnected graph based representation of the same 3D character motion. The interconnected graph based representation can include an interconnected graph that is used to drive mesh clusters during the rendering of a 3D character animation. The interconnected graph based representation provides spatial compression of the deformable mesh based representation, and further compression can be achieved by applying temporal compression processes to the time-varying behavior of the mesh clusters.Type: GrantFiled: June 6, 2014Date of Patent: October 4, 2016Assignee: ADOBE SYSTEMS INCORPORATEDInventors: Edilson de Aguiar, Stefano Corazza, Emiliano Gambaretto
-
Patent number: 9373185Abstract: Systems and methods are described for animating 3D characters using synthetic motion data generated by generative models in response to a high level description of a desired sequence of motion provided by an animator. In a number of embodiments, an animation system is accessible via a server system that utilizes the ability of generative models to generate synthetic motion data across a continuum to enable multiple animators to effectively reuse the same set of previously recorded motion capture data to produce a wide variety of desired animation sequences. One embodiment of the invention includes a server system configured to communicate with a database containing motion data including repeated sequences of motion, where the differences between the repeated sequences of motion are described using at least one high level characteristic.Type: GrantFiled: April 21, 2014Date of Patent: June 21, 2016Assignee: ADOBE SYSTEMS INCORPORATEDInventors: Graham Taylor, Stefano Corazza, Nazim Kareemi, Edilson de Aguiar
-
Publication number: 20140313207Abstract: Systems and methods are described for animating 3D characters using synthetic motion data generated by generative models in response to a high level description of a desired sequence of motion provided by an animator. In a number of embodiments, an animation system is accessible via a server system that utilizes the ability of generative models to generate synthetic motion data across a continuum to enable multiple animators to effectively reuse the same set of previously recorded motion capture data to produce a wide variety of desired animation sequences. One embodiment of the invention includes a server system configured to communicate with a database containing motion data including repeated sequences of motion, where the differences between the repeated sequences of motion are described using at least one high level characteristic.Type: ApplicationFiled: April 21, 2014Publication date: October 23, 2014Applicant: Mixamo, Inc.Inventors: Graham Taylor, Stefano Corazza, Nazim Kareemi, Edilson De Aguiar
-
Publication number: 20140285496Abstract: Systems and methods are described for performing spatial and temporal compression of deformable mesh based representations of 3D character motion allowing the visualization of high-resolution 3D character animations in real time. In a number of embodiments, the deformable mesh based representation of the 3D character motion is used to automatically generate an interconnected graph based representation of the same 3D character motion. The interconnected graph based representation can include an interconnected graph that is used to drive mesh clusters during the rendering of a 3D character animation. The interconnected graph based representation provides spatial compression of the deformable mesh based representation, and further compression can be achieved by applying temporal compression processes to the time-varying behavior of the mesh clusters.Type: ApplicationFiled: June 6, 2014Publication date: September 25, 2014Applicant: Mixamo, Inc.Inventors: Edilson de Aguiar, Stefano Corazza, Emiliano Gambaretto
-
Publication number: 20140160116Abstract: Systems and methods are described for animating 3D characters using synthetic motion data generated by motion models in response to a high level description of a desired sequence of motion provided by an animator. In a number of embodiments, the synthetic motion data is streamed to a user device that includes a rendering engine and the user device renders an animation of a 3D character using the streamed synthetic motion data. In several embodiments, an animator can upload a custom model of a 3D character or a custom 3D character is generated by the server system in response to a high level description of a desired 3D character provided by the user and the synthetic motion data generated by the generative model is retargeted to animate the custom 3D character.Type: ApplicationFiled: December 2, 2013Publication date: June 12, 2014Applicant: Mixamo, Inc.Inventors: Edilson de Aguiar, Emiliano Gambaretto, Stefano Corazza
-
Patent number: 8749556Abstract: Systems and methods are described for performing spatial and temporal compression of deformable mesh based representations of 3D character motion allowing the visualization of high-resolution 3D character animations in real time. In a number of embodiments, the deformable mesh based representation of the 3D character motion is used to automatically generate an interconnected graph based representation of the same 3D character motion. The interconnected graph based representation can include an interconnected graph that is used to drive mesh clusters during the rendering of a 3D character animation. The interconnected graph based representation provides spatial compression of the deformable mesh based representation, and further compression can be achieved by applying temporal compression processes to the time-varying behavior of the mesh clusters.Type: GrantFiled: October 14, 2009Date of Patent: June 10, 2014Assignee: Mixamo, Inc.Inventors: Edilson de Aguiar, Stefano Corazza, Emiliano Gambaretto
-
Patent number: 8737767Abstract: One embodiment of the present invention sets forth a technique for generating a three-dimensional (3D) model of an individual. The technique involves receiving pose data that includes image data associated with the individual being in a pose, selecting from a database a morphable 3D generic human model based on a gender of the individual, adjusting a pose of the morphable 3D generic model based on the image data, identifying, based on the image data, one or more 3D appearance parameter objects included in the database, and adjusting via one or more processors one or more properties of the morphable 3D generic model based on the image data as well as the one or more 3D appearance parameter objects to produce the 3D model of the individual.Type: GrantFiled: February 28, 2012Date of Patent: May 27, 2014Assignee: Disney Enterprises, Inc.Inventors: Jessica Kate Hodgins, Edilson De Aguiar, Lavanya Sharan, Moshe Benjamin Mahler, Ariel Shamir
-
Patent number: 8704832Abstract: Systems and methods are described for animating 3D characters using synthetic motion data generated by generative models in response to a high level description of a desired sequence of motion provided by an animator. An animation system is accessible via a server system that utilizes the ability of generative models to generate synthetic motion data across a continuum to enable multiple animators to effectively reuse the same set of previously recorded motion capture data to produce a wide variety of desired animation sequences. An animator can upload a custom model of a 3D character and the synthetic motion data generated by the generative model is retargeted to animate the custom 3D character.Type: GrantFiled: February 12, 2009Date of Patent: April 22, 2014Assignee: Mixamo, Inc.Inventors: Graham Taylor, Stefano Corazza, Nazim Kareemi, Edilson de Aguiar
-
Publication number: 20130226528Abstract: One embodiment of the present invention sets forth a technique for generating a three-dimensional (3D) model of an individual. The technique involves receiving pose data that includes image data associated with the individual being in a pose, selecting from a database a morphable 3D generic human model based on a gender of the individual, adjusting a pose of the morphable 3D generic model based on the image data, identifying, based on the image data, one or more 3D appearance parameter objects included in the database, and adjusting via one or more processors one or more properties of the morphable 3D generic model based on the image data as well as the one or more 3D appearance parameter objects to produce the 3D model of the individual.Type: ApplicationFiled: February 28, 2012Publication date: August 29, 2013Inventors: Jessica Kate HODGINS, Edilson DE AGUIAR, Lavanya SHARAN, Moshe Benjamin MAHLER, Ariel SHAMIR
-
Patent number: 8384714Abstract: A variety of methods, devices and storage mediums are implemented for creating digital representations of figures. According to one such computer implemented method, a volumetric representation of a figure is correlated with an image of the figure. Reference points are found that are common to each of two temporally distinct images of the figure, the reference points representing movement of the figure between the two images. A volumetric deformation is applied to the digital representation of the figure as a function of the reference points and the correlation of the volumetric representation of the figure. A fine deformation is applied as a function of the coarse/volumetric deformation. Responsive to the applied deformations, an updated digital representation of the figure is generated.Type: GrantFiled: December 24, 2008Date of Patent: February 26, 2013Assignees: The Board of Trustees of the Leland Stanford Junior University, Max-Planck-Gesellschaft zur Förderung der Wissenschaften e.V.Inventors: Edilson De Aguiar, Carsten Stoll, Christian Theobalt, Hans-Peter Seidel, Sebastian Thrun
-
Publication number: 20110273457Abstract: Techniques are disclosed for providing a learning-based clothing model that enables the simultaneous animation of multiple detailed garments in real-time. A simple conditional model learns and preserves key dynamic properties of cloth motions and folding details. Such a conditional model may be generated for each garment worn by a given character. Once generated, the conditional model may be used to determine complex body/cloth interactions in order to render the character and garment from frame-to-frame. The clothing model may be used for a variety of garments worn by male and female human characters (as well as non-human characters) while performing a varied set of motions typically used in video games (e.g., walking, running, jumping, turning, etc.).Type: ApplicationFiled: December 21, 2010Publication date: November 10, 2011Inventors: EDILSON DE AGUIAR, Leonid Sigal, Adrien Treuille, Jessica K. Hodgins
-
Publication number: 20100259547Abstract: Systems and methods are described for animating 3D characters using synthetic motion data generated by motion models in response to a high level description of a desired sequence of motion provided by an animator. In a number of embodiments, the synthetic motion data is streamed to a user device that includes a rendering engine and the user device renders an animation of a 3D character using the streamed synthetic motion data. In several embodiments, an animator can upload a custom model of a 3D character or a custom 3D character is generated by the server system in response to a high level description of a desired 3D character provided by the user and the synthetic motion data generated by the generative model is retargeted to animate the custom 3D character.Type: ApplicationFiled: April 1, 2010Publication date: October 14, 2010Applicant: Mixamo, Inc.Inventors: Edilson de Aguiar, Emiliano Gambaretto, Stefano Corazza
-
Publication number: 20100149179Abstract: Systems and methods are described for performing spatial and temporal compression of deformable mesh based representations of 3D character motion allowing the visualization of high-resolution 3D character animations in real time. In a number of embodiments, the deformable mesh based representation of the 3D character motion is used to automatically generate an interconnected graph based representation of the same 3D character motion. The interconnected graph based representation can include an interconnected graph that is used to drive mesh clusters during the rendering of a 3D character animation. The interconnected graph based representation provides spatial compression of the deformable mesh based representation, and further compression can be achieved by applying temporal compression processes to the time-varying behavior of the mesh clusters.Type: ApplicationFiled: October 14, 2009Publication date: June 17, 2010Inventors: Edilson de Aguiar, Stefano Corazza, Emiliano Gambaretto