Patents by Inventor Gabriel Takacs
Gabriel Takacs has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240087155Abstract: A scale and pose estimation method for a camera system is disclosed. Camera data for a scene acquired by the camera system is received. A scale prior parameter characterizing scale of the camera system is received. A cost of a cost function is calculated for a similarity transformation. The cost of the cost function is influenced at least by the scale prior parameter. Based at least on the cost function being less than a threshold cost, an estimated scale and pose of the camera system is output based on the similarity transformation.Type: ApplicationFiled: September 12, 2022Publication date: March 14, 2024Applicant: Microsoft Technology Licensing, LLCInventors: Victor M. FRAGOSO ROJAS, Mei CHEN, Gabriel TAKACS
-
Patent number: 11922560Abstract: One example provides a computing device configured to capture, via the camera, first image data imaging a first physical world location, create a first spatial representation of the first physical world location based on the first image data, receive a user input defining a pose of a first virtual spatial anchor point relative to a feature imaged in the first image data, track user movement to a second physical world location, capture second image data imaging the second physical world location, receive a user input defining a pose of a second virtual spatial anchor point relative to a feature imaged in the second image data, and send, to a remote computing device, data representing the first spatial representation, the pose of first spatial anchor point, the second spatial representation, the pose of second spatial anchor point, and a positional relationship between first and second spatial anchor points.Type: GrantFiled: April 26, 2021Date of Patent: March 5, 2024Assignee: Microsoft Technology Licensing, LLCInventors: Ali Reza Emami, Gabriel Takacs, Gavin Dean Lazarow, Skyler Mark Goodell
-
Publication number: 20230373847Abstract: In one example, a solver engine may execute a reverse calculation to determine a recipe ingredient set based on a goal descriptor describing a ceramic glaze. A descriptor interface of the solver engine may receive a goal descriptor describing a ceramic glaze. A model applicator of the solver engine may apply a glaze process model to the goal descriptor. The model applicator may automatically reverse calculate a glaze recipe describing a recipe ingredient set to produce the ceramic glaze described by the goal descriptor. A glaze mixing machine interface may direct a glaze mixing machine to mix the recipe ingredient set to produce the ceramic glaze.Type: ApplicationFiled: August 1, 2023Publication date: November 23, 2023Inventors: Andrea Wolf, Gabriel Takacs
-
Patent number: 11746040Abstract: In one example, a solver engine may execute a reverse calculation to determine a recipe ingredient set based on a goal descriptor describing a ceramic glaze. A descriptor interface of the solver engine may receive a goal descriptor describing a ceramic glaze. A model applicator of the solver engine may apply a glaze process model to the goal descriptor. The model applicator may automatically reverse calculate a glaze recipe describing a recipe ingredient set to produce the ceramic glaze described by the goal descriptor. A glaze mixing machine interface may direct a glaze mixing machine to mix the recipe ingredient set to produce the ceramic glaze.Type: GrantFiled: November 5, 2018Date of Patent: September 5, 2023Inventors: Andrea Wolf, Gabriel Takacs
-
Publication number: 20220351473Abstract: Systems, apparatuses and methods to provide image data, augmented with related data, to be displayed on a mobile computing device are disclosed. An example mobile device includes a camera to provide images of a scene from different angles to a server, at least one sensor to sense a position and an orientation of the camera, and a screen to present augmented reality data over the scene based on the position and the orientation of the camera and on a three-dimensional representation of the scene based on the images.Type: ApplicationFiled: July 15, 2022Publication date: November 3, 2022Inventors: Yi Wu, Gabriel Takacs, Maha El Choubassi, Igor V. Kozintsev
-
Patent number: 11443455Abstract: A scale and pose estimation method for a camera system is disclosed. Camera data for a scene acquired by the camera system is received. A rotation prior parameter characterizing a gravity direction is received. A scale prior parameter characterizing scale of the camera system is received. A cost of a cost function is calculated for a similarity transformation that is configured to encode a scale and pose of the camera system. The cost of the cost function is influenced by the rotation prior parameter and the scale prior parameter. A solved similarity transformation is determined upon calculating a cost for the cost function that is less than a threshold cost. An estimated scale and pose of the camera system is output based on the solved similarity transformation.Type: GrantFiled: January 15, 2020Date of Patent: September 13, 2022Assignee: Microsoft Technology Licensing, LLCInventors: Victor M. Fragoso Rojas, Mei Chen, Gabriel Takacs
-
Patent number: 11393173Abstract: Systems, apparatuses and methods to provide image data, augmented with related data, to be displayed on a mobile computing device are disclosed. An example mobile device includes a camera to provide images of a scene from different angles to a server, at least one sensor to sense a position and an orientation of the camera, and a screen to present augmented reality data over the scene based on the position and the orientation of the camera and on a three-dimensional representation of the scene based on the images.Type: GrantFiled: August 10, 2020Date of Patent: July 19, 2022Assignee: Intel CorporationInventors: Yi Wu, Gabriel Takacs, Maha El Choubassi, Igor V. Kozintsev
-
Publication number: 20210350612Abstract: One example provides a computing device configured to capture, via the camera, first image data imaging a first physical world location, create a first spatial representation of the first physical world location based on the first image data, receive a user input defining a pose of a first virtual spatial anchor point relative to a feature imaged in the first image data, track user movement to a second physical world location, capture second image data imaging the second physical world location, receive a user input defining a pose of a second virtual spatial anchor point relative to a feature imaged in the second image data, and send, to a remote computing device, data representing the first spatial representation, the pose of first spatial anchor point, the second spatial representation, the pose of second spatial anchor point, and a positional relationship between first and second spatial anchor points.Type: ApplicationFiled: April 26, 2021Publication date: November 11, 2021Applicant: Microsoft Technology Licensing, LLCInventors: Ali Reza EMAMI, Gabriel TAKACS, Gavin Dean LAZAROW, Skyler Mark GOODELL
-
Publication number: 20210192848Abstract: Systems, apparatuses and methods to provide image data, augmented with related data, to be displayed on a mobile computing device are disclosed. An example mobile device includes a camera to provide images of a scene from different angles to a server, at least one sensor to sense a position and an orientation of the camera, and a screen to present augmented reality data over the scene based on the position and the orientation of the camera and on a three-dimensional representation of the scene based on the images.Type: ApplicationFiled: August 10, 2020Publication date: June 24, 2021Inventors: Yi Wu, Gabriel Takacs, Maha El Choubassi, Igor V. Kozintsev
-
Patent number: 10997776Abstract: One example provides a computing device configured to capture, via the camera, first image data imaging a first physical world location, create a first spatial representation of the first physical world location based on the first image data, receive a user input defining a pose of a first virtual spatial anchor point relative to a feature imaged in the first image data, track user movement to a second physical world location, capture second image data imaging the second physical world location, receive a user input defining a pose of a second virtual spatial anchor point relative to a feature imaged in the second image data, and send, to a remote computing device, data representing the first spatial representation, the pose of first spatial anchor point, the second spatial representation, the pose of second spatial anchor point, and a positional relationship between first and second spatial anchor points.Type: GrantFiled: May 6, 2019Date of Patent: May 4, 2021Assignee: Microsoft Technology Licensing, LLCInventors: Ali Reza Emami, Gabriel Takacs, Gavin Dean Lazarow, Skyler Mark Goodell
-
Publication number: 20210125372Abstract: A scale and pose estimation method for a camera system is disclosed. Camera data for a scene acquired by the camera system is received. A rotation prior parameter characterizing a gravity direction is received. A scale prior parameter characterizing scale of the camera system is received. A cost of a cost function is calculated for a similarity transformation that is configured to encode a scale and pose of the camera system. The cost of the cost function is influenced by the rotation prior parameter and the scale prior parameter. A solved similarity transformation is determined upon calculating a cost for the cost function that is less than a threshold cost. An estimated scale and pose of the camera system is output based on the solved similarity transformation.Type: ApplicationFiled: January 15, 2020Publication date: April 29, 2021Applicant: Microsoft Technology Licensing, LLCInventors: Victor M. FRAGOSO ROJAS, Mei CHEN, Gabriel TAKACS
-
Patent number: 10848920Abstract: Described herein is a system and method of generating precise location information. Feature points of an image are received from a user device (e.g., mobile phone). Location information of the user device (e.g., GPS coordinates) and relative transform information from the user device are also received. An image anchor is identified based upon the received feature points, relative transform information, and the received location information. Precise location information of the identified image anchor is provided to the user device. The precise location information of the identified image anchor is based upon filtered feature points, location information and relative transform information received from a plurality of user devices over a period of time (e.g., hours, days, weeks, months). The precise location information of the identified image anchor is updated based upon the location information of the user device and the relative transform information received from the user device.Type: GrantFiled: September 17, 2019Date of Patent: November 24, 2020Assignee: Microsoft Technology Licensing, LLCInventors: Torfi Frans Olafsson, Jason Matthew Cahill, Jesse Dylan Merriam, David Richard Reed, Timothy Arthur Gill, Gabriel Takacs, Bogdan Cosmin Bucur, Andrew Nicholas Bond
-
Publication number: 20200273235Abstract: One example provides a computing device configured to capture, via the camera, first image data imaging a first physical world location, create a first spatial representation of the first physical world location based on the first image data, receive a user input defining a pose of a first virtual spatial anchor point relative to a feature imaged in the first image data, track user movement to a second physical world location, capture second image data imaging the second physical world location, receive a user input defining a pose of a second virtual spatial anchor point relative to a feature imaged in the second image data, and send, to a remote computing device, data representing the first spatial representation, the pose of first spatial anchor point, the second spatial representation, the pose of second spatial anchor point, and a positional relationship between first and second spatial anchor points.Type: ApplicationFiled: May 6, 2019Publication date: August 27, 2020Applicant: Microsoft Technology Licensing, LLCInventors: Ali Reza EMAMI, Gabriel TAKACS, Gavin Dean LAZAROW, Skyler Mark GOODELL
-
Patent number: 10740975Abstract: Systems, apparatuses and methods to provide image data, augmented with related data, to be displayed on a mobile computing device are disclosed. An example mobile device includes a camera to provide images of a scene from different angles to a server, at least one sensor to sense a position and an orientation of the camera, and a screen to present augmented reality data over the scene based on the position and the orientation of the camera and on a three-dimensional representation of the scene based on the images.Type: GrantFiled: March 30, 2018Date of Patent: August 11, 2020Assignee: Intel CorporationInventors: Yi Wu, Gabriel Takacs, Maha El Choubassi, Igor V. Kozintsev
-
Publication number: 20200140321Abstract: In one example, a solver engine may execute a reverse calculation to determine a recipe ingredient set based on a goal descriptor describing a ceramic glaze. A descriptor interface of the solver engine may receive a goal descriptor describing a ceramic glaze. A model applicator of the solver engine may apply a glaze process model to the goal descriptor. The model applicator may automatically reverse calculate a glaze recipe describing a recipe ingredient set to produce the ceramic glaze described by the goal descriptor. A glaze mixing machine interface may direct a glaze mixing machine to mix the recipe ingredient set to produce the ceramic glaze.Type: ApplicationFiled: November 5, 2018Publication date: May 7, 2020Inventors: Andrea Wolf, Gabriel Takacs
-
Patent number: 10531065Abstract: A computing device and method are provided for accelerating the coarse relocalization process of the computing device by generating a current fingerprint using signal data detected by sensors, statistically analyzing the current fingerprint for proximity to candidate fingerprints in fingerprint data to generate a ranked list, and subsequently retrieving map data corresponding to the closest matching fingerprints in the ranked list. The computing device may comprise a processor, a memory operatively coupled to the processor, and a fingerprint program stored in the memory and executed by the processor.Type: GrantFiled: June 15, 2017Date of Patent: January 7, 2020Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Gavin Dean Lazarow, Tushar Cyril Bhatnagar, David Maxwell Gedye, Gabriel Takacs, Jeroen Vanturennout, Erik Manfred Schulz, Jason Randall Wadsworth, Richard Renchi Ma, Gregg Wygonik, Jonathan David Hildebrandt, Evan Lincoln Tice
-
Patent number: 10504232Abstract: Mixed-reality systems are provided for using anchor data structures, such as anchor graphs, within a mixed-reality environment. These systems utilize anchor components, such as anchor vertexes, that comprise at least one first key frame, a first mixed-reality element, and at least one first transform connecting the at least one first key frame to the first mixed-reality element. Anchor connecting components, such as anchor edges, comprise transformations that connect the anchor components (e.g., anchor vertexes).Type: GrantFiled: February 19, 2019Date of Patent: December 10, 2019Assignee: Microsoft Technology Licensing, LLCInventors: Tracy K. Ard, Gabriel Takacs, David M. Gedye, James Fletcher, Matthew Sammis Ashman
-
Publication number: 20190180457Abstract: Mixed-reality systems are provided for using anchor data structures, such as anchor graphs, within a mixed-reality environment. These systems utilize anchor components, such as anchor vertexes, that comprise at least one first key frame, a first mixed-reality element, and at least one first transform connecting the at least one first key frame to the first mixed-reality element. Anchor connecting components, such as anchor edges, comprise transformations that connect the anchor components (e.g., anchor vertexes).Type: ApplicationFiled: February 19, 2019Publication date: June 13, 2019Inventors: Tracy K. Ard, Gabriel Takacs, David M. Gedye, James Fletcher, Matthew Sammis Ashman
-
Patent number: 10217231Abstract: Mixed-reality systems are provided for using anchor graphs within a mixed-reality environment. These systems utilize anchor vertexes that comprise at least one first key frame, a first mixed-reality element, and at least one first transform connecting the at least one first key frame to the first mixed-reality element. Anchor edges comprising transformations connect the anchor vertexes.Type: GrantFiled: May 31, 2016Date of Patent: February 26, 2019Assignee: Microsoft Technology Licensing, LLCInventors: Tracy K. Ard, Gabriel Takacs, David M. Gedye, James Fletcher, Matthew Sammis Ashman
-
Patent number: 10134196Abstract: Embodiments of the invention relate to systems, apparatuses and methods to provide image data, augmented with related data, to be displayed on a mobile computing device. Embodiments of the invention display a live view augmented with information identifying an object amongst other objects. Embodiments of the invention may utilize other related data, such as 3D point cloud data, image data and location data related to the object, to obtain a specific location of an object within the live view. Embodiments of the invention may further display a live view with augmented data three-dimensionally consistent with the position and orientation of the image sensor of the mobile computing device.Type: GrantFiled: March 21, 2017Date of Patent: November 20, 2018Assignee: INTEL CORPORATIONInventors: Yi Wu, Gabriel Takacs, Maha El Choubassi, Igor V. Kozintsev