Patents by Inventor Charles Tappan
Charles Tappan has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240144622Abstract: Techniques for aligning object representations for animation include analyzing a source object representation and a target object representation to identify a category of the source object and a category of the target object. Based on the category or categories of the objects, a feature extraction machine learning models is selected. The source object representation and the target object representation are provided as input to the selected feature extraction machine learning model to generate respective semantic descriptors and shape vectors for the source and target objects. Based on the semantic descriptors and the shape vectors for the source and target objects, an alignment machine learning model generates an aligned target object representation that is aligned with the source object representation and usable for animating the target object.Type: ApplicationFiled: October 31, 2022Publication date: May 2, 2024Applicants: Pixar, Disney Enterprises, Inc.Inventors: Hayko Riemenschneider, Clara Fernandez, Joan Massich, Christopher Richard Schroers, Daniel Teece, Rasmus Tamstorf, Charles Tappan, Mark Meyer, Antoine Milliez
-
Patent number: 10789757Abstract: There is provided a method for ray-mediated illumination control. The method includes identifying a first activation region corresponding to one of an origin and a destination of a ray, where the ray is described by a ray data associated with the ray. The method further includes identifying a second activation region corresponding to the other one of the origin and the destination of the ray, interpreting an illumination rule for the ray based on at least one of the first activation region and the second activation region, and modifying an illumination in one of the first activation region and the second activation region based on the illumination rule and the ray data.Type: GrantFiled: March 19, 2014Date of Patent: September 29, 2020Assignee: Disney Enterprises Inc.Inventors: Brent Burley, Christian Eisenacher, Sean Jenkins, Gregory Nichols, Andrew Selle, Charles Tappan, Henry A. Driskill, Brett W. Achorn
-
Publication number: 20140327675Abstract: There is provided a method for ray-mediated illumination control. The method includes identifying a first activation region corresponding to one of an origin and a destination of a ray, where the ray is described by a ray data associated with the ray. The method further includes identifying a second activation region corresponding to the other one of the origin and the destination of the ray, interpreting an illumination rule for the ray based on at least one of the first activation region and the second activation region, and modifying an illumination in one of the first activation region and the second activation region based on the illumination rule and the ray data.Type: ApplicationFiled: March 19, 2014Publication date: November 6, 2014Applicant: Disney Enterprises, Inc.Inventors: Brent Burley, Christian Eisenacher, Sean Jenkins, Gregory Nichols, Andrew Selle, Charles Tappan
-
Publication number: 20130055107Abstract: Methods and systems for dynamically and collaboratively generating software documentation are described. In particular, methods and systems provide for in-context authoring, identification, and presentation of software documentation based on one or more determined contexts. For example, methods and systems enable a computing device to identify one or more current contexts for the device, and to provide documentation content based on the identified context.Type: ApplicationFiled: August 26, 2011Publication date: February 28, 2013Applicant: Disney Enterprises, Inc.Inventors: Charles Tappan, Andrew Selle, Brent Burley, Dan Candela
-
Patent number: 8217940Abstract: A method for a computer system includes receiving a plurality of values for an illumination parameter associated with an illumination source for a scene, determining a plurality illumination contributions for a surface of an object within the scene, in response to the plurality of values, determining a plurality of texture maps associated with the surface of the object in response to the plurality of illumination contributions, wherein each of the plurality of texture maps is associated with each of the plurality of values for the illumination parameter, receiving a value for the illumination parameter at render time from a user, determining a texture map associated with the surface of the object, in response to the value for the illumination parameter and in response to the plurality of texture maps, rendering the surface of the object using the texture map to form an image, and displaying the image to the user.Type: GrantFiled: July 23, 2008Date of Patent: July 10, 2012Assignee: Disney Enterprises, Inc.Inventors: Brent Burley, Charles Tappan, Daniel Teece
-
Patent number: 8040355Abstract: Textures are transferred between different object models using a point cloud. In a first phase, a point cloud in 3-D space is created to represent a texture map as applied to a first, or “source,” object model. In a second phase, a value for a target texel of a texture map associated with a second, or “target,” object model, is determined by identifying the 3-D location on a surface defined by the target object model that maps to the location of the target texel and assigning a value based on the nearest point (or points) to that location in the 3-D point cloud. To the extent that differences between the source and target object models are minor, the texture transfer can be accomplished without loss of information or manual cleanup.Type: GrantFiled: July 22, 2008Date of Patent: October 18, 2011Assignee: Disney Enterprises, Inc.Inventors: Brent Burley, Charles Tappan, Daniel Teece
-
Publication number: 20090027412Abstract: Textures are transferred between different object models using a point cloud. In a first phase, a point cloud in 3-D space is created to represent a texture map as applied to a first, or “source,” object model. In a second phase, a value for a target texel of a texture map associated with a second, or “target,” object model, is determined by identifying the 3-D location on a surface defined by the target object model that maps to the location of the target texel and assigning a value based on the nearest point (or points) to that location in the 3-D point cloud. To the extent that differences between the source and target object models are minor, the texture transfer can be accomplished without loss of information or manual cleanup.Type: ApplicationFiled: July 22, 2008Publication date: January 29, 2009Applicant: Disney Enterprises, Inc.Inventors: Brent Burley, Charles Tappan, Daniel Teece
-
Publication number: 20090027391Abstract: A method for a computer system includes receiving a plurality of values for an illumination parameter associated with an illumination source for a scene, determining a plurality illumination contributions for a surface of an object within the scene, in response to the plurality of values, determining a plurality of texture maps associated with the surface of the object in response to the plurality of illumination contributions, wherein each of the plurality of texture maps is associated with each of the plurality of values for the illumination parameter, receiving a value for the illumination parameter at render time from a user, determining a texture map associated with the surface of the object, in response to the value for the illumination parameter and in response to the plurality of texture maps, rendering the surface of the object using the texture map to form an image, and displaying the image to the user.Type: ApplicationFiled: July 23, 2008Publication date: January 29, 2009Applicant: Disney Enterprises, Inc.Inventors: Brent Burley, Charles Tappan, Daniel Teece