Patents by Inventor Kaan YUCER
Kaan YUCER has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20230222792Abstract: Methods, systems, and apparatus for recognizing objects and providing content related to the recognized objects are described. In one aspect, a method includes detecting presence of one or more objects depicted in a viewfinder of a camera of the mobile device. In response to detecting the presence of the one or more objects, image data representing the one or more objects is sent to a content distribution system that selects content related to objects depicted in images. A location of each of the one or more objects in the viewfinder of the camera is tracked while waiting to receive content from the content distribution system. Content related to the one or more objects is received from the content distribution system. A current location of each object in the viewfinder is determined and the content related to the object is presented within the viewfinder at the current location of the object.Type: ApplicationFiled: March 22, 2023Publication date: July 13, 2023Inventors: Ibrahim Badr, Roland Peter Kehl, Kaan Yucer, Thomas Jonathan Stone, Gokhan H. Bakir, Adrian Zumbrunnen
-
Patent number: 11651589Abstract: Methods, systems, and apparatus for recognizing objects and providing content related to the recognized objects are described. In one aspect, a method includes detecting presence of one or more objects depicted in a viewfinder of a camera of the mobile device. In response to detecting the presence of the one or more objects, image data representing the one or more objects is sent to a content distribution system that selects content related to objects depicted in images. A location of each of the one or more objects in the viewfinder of the camera is tracked while waiting to receive content from the content distribution system. Content related to the one or more objects is received from the content distribution system. A current location of each object in the viewfinder is determined and the content related to the object is presented within the viewfinder at the current location of the object.Type: GrantFiled: May 7, 2018Date of Patent: May 16, 2023Assignee: GOOGLE LLCInventors: Ibrahim Badr, Roland Peter Kehl, Kaan Yucer, Thomas Jonathan Stone, Gokhan H. Bakir, Adrian Zumbrunnen
-
Patent number: 10887581Abstract: The present disclosure relates to techniques for reconstructing an object in three dimensions that is captured in a set of two-dimensional images. The object is reconstructed in three dimensions by computing depth values for edges of the object in the set of two-dimensional images. The set of two-dimensional images may be samples of a light field surrounding the object. The depth values may be computed by exploiting local gradient information in the set of two-dimensional images. After computing the depth values for the edges, depth values between the edges may be determined by identifying types of the edges (e.g., a texture edge, a silhouette edge, or other type of edge). Then, the depth values from the set of two-dimensional images may be aggregated in a three-dimensional space using a voting scheme, allowing the reconstruction of the object in three dimensions.Type: GrantFiled: October 31, 2017Date of Patent: January 5, 2021Assignees: DISNEY ENTERPRISES, INC., ETH ZÜRICH (EIDGENÖSSISCHE TECHNISCHE HOCHSCHULE ZÜRICH)Inventors: Kaan Yücer, Changil Kim, Alexander Sorkine-Hornung, Olga Sorkine-Hornung
-
Publication number: 20200265229Abstract: Methods, systems, and apparatus for recognizing objects and providing content related to the recognized objects are described. In one aspect, a method includes detecting presence of one or more objects depicted in a viewfinder of a camera of the mobile device. In response to detecting the presence of the one or more objects, image data representing the one or more objects is sent to a content distribution system that selects content related to objects depicted in images. A location of each of the one or more objects in the viewfinder of the camera is tracked while waiting to receive content from the content distribution system. Content related to the one or more objects is received from the content distribution system. A current location of each object in the viewfinder is determined and the content related to the object is presented within the viewfinder at the current location of the object.Type: ApplicationFiled: May 7, 2018Publication date: August 20, 2020Inventors: Ibrahim Badr, Roland Peter Kehl, Kaan Yucer, Thomas Jonathan Stone, Gokhan H. Bakir, Adrian Zumbrunnenn
-
Patent number: 10122994Abstract: The present disclosure relates to techniques for reconstructing an object in three dimensions that is captured in a set of two-dimensional images. The object is reconstructed in three dimensions by computing depth values for edges of the object in the set of two-dimensional images. The set of two-dimensional images may be samples of a light field surrounding the object. The depth values may be computed by exploiting local gradient information in the set of two-dimensional images. After computing the depth values for the edges, depth values between the edges may be determined by identifying types of the edges (e.g., a texture edge, a silhouette edge, or other type of edge). Then, the depth values from the set of two-dimensional images may be aggregated in a three-dimensional space using a voting scheme, allowing the reconstruction of the object in three dimensions.Type: GrantFiled: October 31, 2017Date of Patent: November 6, 2018Assignees: DISNEY ENTERPRISES, INC., ETH ZÜRICH (EIDGENÖSSISCHE TECHNISCHE HOCHSCHULE ZÜRICH)Inventors: Kaan Yücer, Changil Kim, Alexander Sorkine-Hornung, Olga Sorkine-Hornung
-
Publication number: 20180139436Abstract: The present disclosure relates to techniques for reconstructing an object in three dimensions that is captured in a set of two-dimensional images. The object is reconstructed in three dimensions by computing depth values for edges of the object in the set of two-dimensional images. The set of two-dimensional images may be samples of a light field surrounding the object. The depth values may be computed by exploiting local gradient information in the set of two-dimensional images. After computing the depth values for the edges, depth values between the edges may be determined by identifying types of the edges (e.g., a texture edge, a silhouette edge, or other type of edge). Then, the depth values from the set of two-dimensional images may be aggregated in a three-dimensional space using a voting scheme, allowing the reconstruction of the object in three dimensions.Type: ApplicationFiled: October 31, 2017Publication date: May 17, 2018Applicants: Disney Enterprises, Inc., ETH Zürich (Eidgenössische Technische Hochschule Zürich)Inventors: Kaan Yücer, Changil Kim, Alexander Sorkine-Hornung, Olga Sorkine-Hornung
-
Publication number: 20180137674Abstract: The present disclosure relates to techniques for reconstructing an object in three dimensions that is captured in a set of two-dimensional images. The object is reconstructed in three dimensions by computing depth values for edges of the object in the set of two-dimensional images. The set of two-dimensional images may be samples of a light field surrounding the object. The depth values may be computed by exploiting local gradient information in the set of two-dimensional images. After computing the depth values for the edges, depth values between the edges may be determined by identifying types of the edges (e.g., a texture edge, a silhouette edge, or other type of edge). Then, the depth values from the set of two-dimensional images may be aggregated in a three-dimensional space using a voting scheme, allowing the reconstruction of the object in three dimensions.Type: ApplicationFiled: October 31, 2017Publication date: May 17, 2018Applicants: Disney Enterprises, Inc., ETH Zürich (Eidgenössische Technische Hochschule Zürich)Inventors: Kaan Yücer, Changil Kim, Alexander Sorkine-Hornung, Olga Sorkine-Hornung
-
Patent number: 9747668Abstract: Systems and method for the reconstruction of an articulated object are disclosed herein, The articulated object can be reconstructed from image data collected by a moving camera over a period of time. A plurality of 2D feature points can be identified within the image data. These 2D feature points can be converted into three-dimensional space, which converted points are identified as 3D feature points. These 3D feature points can be used to identify one or several rigidity constrains and/or kinematic constraints. These rigidity and/or kinematic constraints can be applied to a model of the reconstructed articulated object.Type: GrantFiled: January 21, 2016Date of Patent: August 29, 2017Assignees: DISNEY ENTERPRISES, INC., ETH ZÜRICH (EIDGENÖESSISCHE TECHNISCHE HOCHSCHULE ZÜRICH)Inventors: Kaan Yücer, Alexander Sorkine Hornung, Oliver Wang, Olga Sorkine Hornung
-
Publication number: 20170213320Abstract: Systems and method for the reconstruction of an articulated object are disclosed herein, The articulated object can be reconstructed from image data collected by a moving camera over a period of time. A plurality of 2D feature points can be identified within the image data. These 2D feature points can be converted into three-dimensional space, which converted points are identified as 3D feature points. These 3D feature points can be used to identify one or several rigidity constrains and/or kinematic constraints. These rigidity and/or kinematic constraints can be applied to a model of the reconstructed articulated object.Type: ApplicationFiled: January 21, 2016Publication date: July 27, 2017Applicants: Disney Enterprises, Inc., ETH Zürich (Eidgenössische Technische Hochschule Zürich)Inventors: Kaan YÜCER, Alexander SORKINE HORNUNG, Oliver WANG, Olga SORKINE HORNUNG
-
Patent number: 9202431Abstract: The disclosure provides an approach for transferring image edits from a source image to target images. In one embodiment, a warp application receives a user-selected region of interest in a source image and determines for the region of interest content-aware bounded weight functions and seed locations for the same. For each of the target images, the warping application initializes a linear blend skinning subspace warp to a projection onto a feature space of a piecewise affine map from scale invariant feature transform features of the source image to the target image. After initializing the warps, the warping application iteratively optimizes the warps by applying the inverse compositional Lucas-Kanade procedure and using the content-aware weight functions in said procedure. Edits made to the source image may automatically be transferred to target images by warping those edits via the optimized warp function for the respective target images.Type: GrantFiled: October 17, 2012Date of Patent: December 1, 2015Assignee: DISNEY ENTERPRISES, INC.Inventors: Alexander Sorkine-Hornung, Kaan Yucer, Alec Stefan Jacobson, Olga Sorkine-Hornung
-
Publication number: 20140104295Abstract: The disclosure provides an approach for transferring image edits from a source image to target images. In one embodiment, a warp application receives a user-selected region of interest in a source image and determines for the region of interest content-aware bounded weight functions and seed locations for the same. For each of the target images, the warping application initializes a linear blend skinning subspace warp to a projection onto a feature space of a piecewise affine map from scale invariant feature transform features of the source image to the target image. After initializing the warps, the warping application iteratively optimizes the warps by applying the inverse compositional Lucas-Kanade procedure and using the content-aware weight functions in said procedure. Edits made to the source image may automatically be transferred to target images by warping those edits via the optimized warp function for the respective target images.Type: ApplicationFiled: October 17, 2012Publication date: April 17, 2014Applicant: Disney Enterprises, Inc.Inventors: Alexander SORKINE-HORNUNG, Kaan YUCER, Alec Stefan JACOBSON, Olga SORKINE-HORNUNG