Patents by Inventor Stephen Latta
Stephen Latta has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11127210Abstract: A system for automatically displaying virtual objects within a mixed reality environment is described. In some embodiments, a see-through head-mounted display device (HMD) identifies a real object (e.g., a person or book) within a field of view of the HMD, detects one or more interactions associated with real object, and automatically displays virtual objects associated with the real object if the one or more interactions involve touching or satisfy one or more social rules stored in a social rules database. The one or more social rules may be used to infer a particular social relationship by considering the distance to another person, the type of environment (e.g., at home or work), and particular physical interactions (e.g., handshakes or hugs). The virtual objects displayed on the HMD may depend on the particular social relationship inferred (e.g., a friend or acquaintance).Type: GrantFiled: December 22, 2016Date of Patent: September 21, 2021Assignee: Microsoft Technology Licensing, LLCInventors: Christopher Michael Novak, James Liu, Stephen Latta, Anton O. A. Andrews, Craig R. Maitlen, Sheridan Martin
-
Patent number: 10803670Abstract: Embodiments related to efficiently constructing an augmented reality environment with global illumination effects are disclosed. For example, one disclosed embodiment provides a method of displaying an augmented reality image via a display device. The method includes receiving image data, the image data capturing an image of a local environment of the display device, and identifying a physical feature of the local environment via the image data. The method further includes constructing an augmented reality image of a virtual structure for display over the physical feature in spatial registration with the physical feature from a viewpoint of a user, the augmented reality image comprising a plurality of modular virtual structure segments arranged in adjacent locations to form the virtual structure feature, each modular virtual structure segment comprising a pre-computed global illumination effect, and outputting the augmented reality image to the display device.Type: GrantFiled: January 18, 2019Date of Patent: October 13, 2020Assignee: Microsoft Technology Licensing, LLCInventors: Jonathan Steed, Aaron Krauss, Mike Scavezze, Wei Zhang, Arthur Tomlin, Tony Ambrus, Brian Mount, Stephen Latta, Ryan Hastings
-
Patent number: 10691216Abstract: Systems, methods and computer readable media are disclosed for gesture input beyond skeletal. A user's movement or body position is captured by a capture device of a system. Further, non-user-position data is received by the system, such as controller input by the user, an item that the user is wearing, a prop under the control of the user, or a second user's movement or body position. The system incorporates both the user-position data and the non-user-position data to determine one or more inputs the user made to the system.Type: GrantFiled: June 15, 2016Date of Patent: June 23, 2020Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Kevin Geisner, Stephen Latta, Relja Markovic, Gregory N. Snook
-
Patent number: 10510190Abstract: Embodiments that relate to interacting with a physical object in a mixed reality environment via a head-mounted display are disclosed. In one embodiment a mixed reality interaction program identifies an object based on an image from captured by the display. An interaction context for the object is determined based on an aspect of the mixed reality environment. A profile for the physical object is queried to determine interaction modes for the object. A selected interaction mode is programmatically selected based on the interaction context. A user input directed at the object is received via the display and interpreted to correspond to a virtual action based on the selected interaction mode. The virtual action is executed with respect to a virtual object associated with the physical object to modify an appearance of the virtual object. The modified virtual object is then displayed via the display.Type: GrantFiled: September 1, 2017Date of Patent: December 17, 2019Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Michael Scavezze, Jonathan Steed, Stephen Latta, Kevin Geisner, Daniel McCulloch, Brian Mount, Ryan Hastings, Phillip Charles Heckinger
-
Patent number: 10497175Abstract: A head-mounted display includes a see-through display and a virtual reality engine. The see-through display is configured to visually augment an appearance of a physical space to a user viewing the physical space through the see-through display. The virtual reality engine is configured to cause the see-through display to visually present a virtual monitor that appears to be integrated with the physical space to a user viewing the physical space through the see-through display.Type: GrantFiled: September 7, 2016Date of Patent: December 3, 2019Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Brian Mount, Stephen Latta, Adam Poulos, Daniel McCulloch, Darren Bennett, Ryan Hastings, Jason Scott
-
Publication number: 20190244430Abstract: Embodiments related to efficiently constructing an augmented reality environment with global illumination effects are disclosed. For example, one disclosed embodiment provides a method of displaying an augmented reality image via a display device. The method includes receiving image data, the image data capturing an image of a local environment of the display device, and identifying a physical feature of the local environment via the image data. The method further includes constructing an augmented reality image of a virtual structure for display over the physical feature in spatial registration with the physical feature from a viewpoint of a user, the augmented reality image comprising a plurality of modular virtual structure segments arranged in adjacent locations to form the virtual structure feature, each modular virtual structure segment comprising a pre-computed global illumination effect, and outputting the augmented reality image to the display device.Type: ApplicationFiled: January 18, 2019Publication date: August 8, 2019Applicant: Microsoft Technology Licensing, LLCInventors: Jonathan Steed, Aaron Krauss, Mike Scavezze, Wei Zhang, Arthur Tomlin, Tony Ambrus, Brian Mount, Stephen Latta, Ryan Hastings
-
Patent number: 10229544Abstract: Embodiments related to efficiently constructing an augmented reality environment with global illumination effects are disclosed. For example, one disclosed embodiment provides a method of displaying an augmented reality image via a display device. The method includes receiving image data, the image data capturing an image of a local environment of the display device, and identifying a physical feature of the local environment via the image data. The method further includes constructing an augmented reality image of a virtual structure for display over the physical feature in spatial registration with the physical feature from a viewpoint of a user, the augmented reality image comprising a plurality of modular virtual structure segments arranged in adjacent locations to form the virtual structure feature, each modular virtual structure segment comprising a pre-computed global illumination effect, and outputting the augmented reality image to the display device.Type: GrantFiled: November 16, 2017Date of Patent: March 12, 2019Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Jonathan Steed, Aaron Krauss, Mike Scavezze, Wei Zhang, Arthur Tomlin, Tony Ambrus, Brian Mount, Stephen Latta, Ryan Hastings
-
Patent number: 9977882Abstract: Embodiments are disclosed that relate to authenticating a user of a display device. For example, one disclosed embodiment includes displaying one or more virtual images on the display device, wherein the one or more virtual images include a set of augmented reality features. The method further includes identifying one or more movements of the user via data received from a sensor of the display device, and comparing the identified movements of the user to a predefined set of authentication information for the user that links user authentication to a predefined order of the augmented reality features. If the identified movements indicate that the user selected the augmented reality features in the predefined order, then the user is authenticated, and if the identified movements indicate that the user did not select the augmented reality features in the predefined order, then the user is not authenticated.Type: GrantFiled: July 22, 2015Date of Patent: May 22, 2018Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Mike Scavezze, Jason Scott, Jonathan Steed, Ian McIntyre, Aaron Krauss, Daniel McCulloch, Stephen Latta, Kevin Geisner, Brian Mount
-
Patent number: 9979809Abstract: Embodiments are disclosed herein that relate to the automatic tracking of objects. For example, one disclosed embodiment provides a method of operating a mobile computing device having an image sensor. The method includes acquiring image data, identifying an inanimate moveable object in the image data, determining whether the inanimate moveable object is a tracked object, if the inanimate moveable object is a tracked object, then storing information regarding a state of the inanimate moveable object, detecting a trigger to provide a notification of the state of the inanimate moveable object, and providing an output of the notification of the state of the inanimate moveable object.Type: GrantFiled: September 2, 2016Date of Patent: May 22, 2018Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Mike Scavezze, Jason Scott, Jonathan Steed, Ian McIntyre, Aaron Krauss, Daniel McCulloch, Stephen Latta
-
Patent number: 9934614Abstract: An example wearable display system includes a controller, a left display to display a left-eye augmented reality image with a left-eye display size at left-eye display coordinates, and a right display to display a right-eye augmented reality image with a right-eye display size at right-eye display coordinates, the left-eye and right-eye augmented reality images collectively forming an augmented reality object perceivable at an apparent real world depth by a wearer of the display system. The controller sets the left-eye display coordinates relative to the right-eye display coordinates as a function of the apparent real world depth of the augmented reality object. The function maintains an aspect of the left-eye and right-eye display sizes throughout a non-scaling range of apparent real world depths of the augmented reality object, and the function scales the left-eye and right-eye display sizes with changing apparent real world depth outside the non-scaling range.Type: GrantFiled: May 20, 2015Date of Patent: April 3, 2018Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Scott Ramsby, Dan Osborn, Shawn Wright, Anatolie Gavriliuc, Forest Woodcroft Gouin, Megan Saunders, Jesse Rapczak, Stephen Latta, Adam G. Poulos, Daniel McCulloch, Wei Zhang
-
Publication number: 20180075663Abstract: Embodiments related to efficiently constructing an augmented reality environment with global illumination effects are disclosed. For example, one disclosed embodiment provides a method of displaying an augmented reality image via a display device. The method includes receiving image data, the image data capturing an image of a local environment of the display device, and identifying a physical feature of the local environment via the image data. The method further includes constructing an augmented reality image of a virtual structure for display over the physical feature in spatial registration with the physical feature from a viewpoint of a user, the augmented reality image comprising a plurality of modular virtual structure segments arranged in adjacent locations to form the virtual structure feature, each modular virtual structure segment comprising a pre-computed global illumination effect, and outputting the augmented reality image to the display device.Type: ApplicationFiled: November 16, 2017Publication date: March 15, 2018Applicant: Microsoft Technology Licensing, LLCInventors: Jonathan Steed, Aaron Krauss, Mike Scavezze, Wei Zhang, Arthur Tomlin, Tony Ambrus, Brian Mount, Stephen Latta, Ryan Hastings
-
Patent number: 9910509Abstract: Systems, methods and computer readable media are disclosed for controlling perspective of a camera-controlled computer. A capture device captures user gestures and sends corresponding data to a recognizer engine. The recognizer engine analyzes the data with a plurality of filters, each filter corresponding to a gesture. Based on the output of those filters, a perspective control is determined, and a display device displays a new perspective corresponding to the perspective control.Type: GrantFiled: November 15, 2016Date of Patent: March 6, 2018Assignee: Microsoft Technology Licensing, LLCInventors: Relja Markovic, Gregory N. Snook, Stephen Latta, Kevin Geisner, Johnny Lee, Adam Jethro Langridge
-
Patent number: 9892562Abstract: Embodiments related to efficiently constructing an augmented reality environment with global illumination effects are disclosed. For example, one disclosed embodiment provides a method of displaying an augmented reality image via a display device. The method includes receiving image data, the image data capturing an image of a local environment of the display device, and identifying a physical feature of the local environment via the image data. The method further includes constructing an augmented reality image of a virtual structure for display over the physical feature in spatial registration with the physical feature from a viewpoint of a user, the augmented reality image comprising a plurality of modular virtual structure segments arranged in adjacent locations to form the virtual structure feature, each modular virtual structure segment comprising a pre-computed global illumination effect, and outputting the augmented reality image to the display device.Type: GrantFiled: October 20, 2016Date of Patent: February 13, 2018Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Jonathan Steed, Aaron Krauss, Mike Scavezze, Wei Zhang, Arthur Tomlin, Tony Ambrus, Brian Mount, Stephen Latta, Ryan Hastings
-
Patent number: 9874932Abstract: One embodiment provides a method to display video such as computer-rendered animation or other video. The method includes assembling a sequence of video frames featuring a moving object, each video frame including a plurality of subframes sequenced for display according to a schedule. The method also includes determining a vector-valued differential velocity of the moving object relative to a head of an observer of the video. At a time scheduled for display of a first subframe of a given frame, first-subframe image content transformed by a first transform is displayed. At a time scheduled for display of the second subframe of the given frame, second-subframe image content transformed by a second transform is displayed. The first and second transforms are computed based on the vector-valued differential velocity to mitigate artifacts.Type: GrantFiled: April 9, 2015Date of Patent: January 23, 2018Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Matthew Calbraith Crisler, Robert Thomas Held, Stephen Latta, Ashraf Ayman Michail, Martin Shetter, Arthur Tomlin
-
Publication number: 20180012412Abstract: Embodiments that relate to interacting with a physical object in a mixed reality environment via a head-mounted display are disclosed. In one embodiment a mixed reality interaction program identifies an object based on an image from captured by the display. An interaction context for the object is determined based on an aspect of the mixed reality environment. A profile for the physical object is queried to determine interaction modes for the object. A selected interaction mode is programmatically selected based on the interaction context. A user input directed at the object is received via the display and interpreted to correspond to a virtual action based on the selected interaction mode. The virtual action is executed with respect to a virtual object associated with the physical object to modify an appearance of the virtual object. The modified virtual object is then displayed via the display.Type: ApplicationFiled: September 1, 2017Publication date: January 11, 2018Applicant: Microsoft Technology Licensing, LLCInventors: Michael Scavezze, Jonathan Steed, Stephen Latta, Kevin Geisner, Daniel McCulloch, Brian Mount, Ryan Hastings, Phillip Charles Heckinger
-
Patent number: 9851787Abstract: A system and related methods for a resource management in a head-mounted display device are provided. In one example, the head-mounted display device includes a plurality of sensors and a display system for presenting holographic objects. A resource management program is configured to operate a selected sensor in a default power mode to achieve a selected fidelity. The program receives user-related information from one or more of the sensors, and determines whether target information is detected. Where target information is detected, the program adjusts the selected sensor to operate in a reduced power mode that uses less power than the default power mode.Type: GrantFiled: November 29, 2012Date of Patent: December 26, 2017Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Stephen Latta, Jedd Anthony Perry, Rod G. Fleck, Jack Clevenger, Frederik Schaffalitzky, Drew Steedly, Daniel McCulloch, Ian McIntyre, Alexandru Balan, Ben Sugden, Ryan Hastings, Brian Mount
-
Patent number: 9848106Abstract: Implementations for identifying, capturing, and presenting high-quality photo-representations of acts occurring during play of a game that employs motion tracking input technology are disclosed. As one example, a method is disclosed that includes capturing via an optical interface, a plurality of photographs of a player in a capture volume during play of the electronic game. The method further includes for each captured photograph of the plurality of captured photographs, comparing an event-based scoring parameter to an event depicted by or corresponding to the captured photograph. The method further includes assigning respective scores to the plurality of captured photographs based, at least in part, on the comparison to the even-based scoring parameter. The method further includes associating the captured photographs at an electronic storage media with the respective scores assigned to the captured photographs.Type: GrantFiled: December 21, 2010Date of Patent: December 19, 2017Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Mike Scavezze, Arthur Tomlin, Relja Markovic, Stephen Latta, Kevin Geisner
-
Patent number: 9836889Abstract: Embodiments for interacting with an executable virtual object associated with a real object are disclosed. In one example, a method for interacting with an executable virtual object associated with a real object includes receiving sensor input from one or more sensors attached to the portable see-through display device, and obtaining information regarding a location of the user based on the sensor input. The method also includes, if the location includes a real object comprising an associated executable virtual object, then determining an intent of the user to interact with the executable virtual object, and if the intent to interact is determined, then interacting with the executable object.Type: GrantFiled: March 1, 2017Date of Patent: December 5, 2017Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Ben Sugden, John Clavin, Ben Vaught, Stephen Latta, Kathryn Stone Perez, Daniel McCulloch, Jason Scott, Wei Zhang, Darren Bennett, Ryan Hastings, Arthur Tomlin, Kevin Geisner
-
Patent number: 9821224Abstract: Depth-image analysis is performed with a device that analyzes a human target within an observed scene by capturing depth-images that include depth information from the observed scene. The human target is modeled with a virtual skeleton including a plurality of joints. The virtual skeleton is used as an input for controlling a driving simulation.Type: GrantFiled: December 21, 2010Date of Patent: November 21, 2017Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Stephen Latta, Darren Bennett, Kevin Geisner, Relja Markovic, Kudo Tsunoda, Rhett Mathis, Matthew Monson, David Gierok, William Paul Giese, Darrin Brown, Cam McRae, David Seymour, William Axel Olsen, Matthew Searcy
-
Patent number: 9812046Abstract: A mixed reality accommodation system and related methods are provided. In one example, a head-mounted display device includes a plurality of sensors and a display system for presenting holographic objects. A mixed reality safety program is configured to receive a holographic object and associated content provider ID from a source. The program assigns a trust level to the object based on the content provider ID. If the trust level is less than a threshold, the object is displayed according to a first set of safety rules that provide a protective level of display restrictions. If the trust level is greater than or equal to the threshold, the object is displayed according to a second set of safety rules that provide a permissive level of display restrictions that are less than the protective level of display restrictions.Type: GrantFiled: January 10, 2013Date of Patent: November 7, 2017Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Stephen Latta, Peter Tobias Kinnebrew