Patents by Inventor Nathan Abercrombie
Nathan Abercrombie has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20250238112Abstract: A user, a manipulator such as a hand, and at least one entity such as a virtual or augmented reality object are in an interface such as a 3D environmental interface. The manipulation distance is the distance between a reference feature of the user and a manipulation feature of the manipulator. The entity distance is the distance between the reference feature and an entity feature of the entity. When the manipulation distance becomes greater than the entity distance, the entity is caused to fade, disappear, move out of the way, shrink, etc. so as to be less of an obstruction to the user's field of view, for example to avoid obstructing more distant entities. Other factors than the manipulation distance and entity distance may be considered in determining whether to reduce the obstructivity of the entity, and exceptions to the obstruction relation may be considered.Type: ApplicationFiled: September 4, 2024Publication date: July 24, 2025Inventors: Nathan Abercrombie, Theo Goguely
-
Publication number: 20250039350Abstract: A target is outputted to an ideal position in 3D space. A viewer indicates the apparent position of the target, and the indication is sensed. An offset between the ideal and apparent positions is determined, and an adjustment determined from the offset such that the apparent position of the ideal position with the adjustment matches the ideal position without the adjustment. The adjustment is made to the first entity and/or a second entity, such that the entities appear to the viewer in the ideal position. The indication may be monocular with a separate indication for each eye, or binocular with a single viewer indication for both eyes. The indication also may serve as communication, such as a PIN input, so that calibration is transparent to the viewer. The method may be continuous, intermittent, or otherwise ongoing over time.Type: ApplicationFiled: March 8, 2024Publication date: January 30, 2025Inventors: Sina Fateh, Nathan Abercrombie, Sleiman Itani
-
Publication number: 20240422304Abstract: To enhance a mono-output-only controller such as a mobile OS to support selective mono/stereo/mixed output, a stereo controller is instantiated in communication with the mono controller. The stereo controller coordinates stereo output, but calls and adapts functions already present in the mono controller for creating surface and image buffers, rendering, compositing, and/or merging. For content designated for 2D display, left and right surfaces are rendered from a mono perspective; for content designated for 3D display, left and right surfaces are rendered from left and right stereo perspectives, respectively. Some, all, or none of available content may be delivered to a stereo display in 3D, with a remainder delivered in 2D, and with comparable content still delivered in 2D to the mono display. The stereo controller is an add-on; the mono controller need not be replaced, removed, deactivated, or modified, facilitating transparency and backward compatibility.Type: ApplicationFiled: January 5, 2024Publication date: December 19, 2024Inventors: Mario Kosmiskas, Nathan Abercrombie, Sleiman Itani
-
Publication number: 20240414312Abstract: To enhance a mono-output-only controller such as a mobile OS to support selective mono/stereo/mixed output, a stereo controller is instantiated in communication with the mono controller. The stereo controller coordinates stereo output, but calls and adapts functions already present in the mono controller for creating surface and image buffers, rendering, compositing, and/or merging. For content designated for 2D display, left and right surfaces are rendered from a mono perspective; for content designated for 3D display, left and right surfaces are rendered from left and right stereo perspectives, respectively. Some, all, or none of available content may be delivered to a stereo display in 3D, with a remainder delivered in 2D, and with comparable content still delivered in 2D to the mono display. The stereo controller is an add-on; the mono controller need not be replaced, removed, deactivated, or modified, facilitating transparency and backward compatibility.Type: ApplicationFiled: January 12, 2024Publication date: December 12, 2024Inventors: Mario Kosmiskas, Nathan Abercrombie, Sleiman Itani
-
Publication number: 20240353931Abstract: A free space input standard is instantiated on a processor. Free space input is sensed and communicated to the processor. If the free space input satisfies the free space input standard, a touch screen input response is invoked in an operating system. The free space input may be sensed using continuous implicit, discrete implicit, active explicit, or passive explicit approaches. The touch screen input response may be invoked through communicating virtual touch screen input, a virtual input event, or a virtual command to or within the operating system. in this manner free space gestures may control existing touch screen interfaces and devices, without modifying those interfaces and devices directly to accept free space gestures.Type: ApplicationFiled: December 1, 2023Publication date: October 24, 2024Inventors: Shashwat Kandadai, Nathan Abercrombie, Yu-Hsiang Chen, Sleiman Itani
-
Publication number: 20240346776Abstract: A method, system, apparatus, and/or device for executing a translation instruction for a constructive movement. The method, system, apparatus, and/or device may include an input device configured to sense a first constructive movement input representative of a non-translational movement of a body of a user that does not move from a first point to a second point in the physical world environment. The method, system, apparatus, and/or device may include a processing device coupled to the input device, where the processing device is configured to execute a translational instruction associated with the first constructive movement input, execute a first resizing instruction to reduce a size of a portion of the physical world environment as displayed by a head-mounted display by an amount indicated by the first resizing stimulus, receive a second constructive movement input, and execute a second translational instruction associated with the second constructive translational movement input.Type: ApplicationFiled: December 1, 2023Publication date: October 17, 2024Inventors: Nathan Abercrombie, Iryna Issayeva, Greg James, Sleiman Itani
-
Patent number: 12086377Abstract: A user, a manipulator such as a hand, and at least one entity such as a virtual or augmented reality object are in an interface such as a 3D environmental interface. The manipulation distance is the distance between a reference feature of the user and a manipulation feature of the manipulator. The entity distance is the distance between the reference feature and an entity feature of the entity. When the manipulation distance becomes greater than the entity distance, the entity is caused to fade, disappear, move out of the way, shrink, etc. so as to be less of an obstruction to the user's field of view, for example to avoid obstructing more distant entities. Other factors than the manipulation distance and entity distance may be considered in determining whether to reduce the obstructivity of the entity, and exceptions to the obstruction relation may be considered.Type: GrantFiled: September 27, 2022Date of Patent: September 10, 2024Assignee: West Texas Technology Partners, LLCInventors: Nathan Abercrombie, Theo Goguely
-
Publication number: 20240241607Abstract: A method, system, apparatus, and/or device for sensing an input in an augmented reality construct. The method, system, apparatus, and/or device may include a mixed-reality device, comprises memory and one or more processors communicatively coupled to a sensor, a touch device, and a display device that is at least partially transparent and configured to display a virtual object. The store instructions executable by the one or more processors to generate virtual object data for displaying a virtual object by the display device; output the virtual object data to the display device; receive, from the sensor, interaction data corresponding to a free-space interaction by a hand of a user with the virtual object; generate touch input data based on the interaction data; and output the touch input data to the touch device.Type: ApplicationFiled: August 31, 2023Publication date: July 18, 2024Inventors: Shashwat Kandadai, Nathan Abercrombie, Yu-Hsiang Chen, Sleiman Itani
-
Patent number: 11876953Abstract: To enhance a mono-output-only controller such as a mobile OS to support selective mono/stereo/mixed output, a stereo controller is instantiated in communication with the mono controller. The stereo controller coordinates stereo output, but calls and adapts functions already present in the mono controller for creating surface and image buffers, rendering, compositing, and/or merging. For content designated for 2D display, left and right surfaces are rendered from a mono perspective; for content designated for 3D display, left and right surfaces are rendered from left and right stereo perspectives, respectively. Some, all, or none of available content may be delivered to a stereo display in 3D, with a remainder delivered in 2D, and with comparable content still delivered in 2D to the mono display. The stereo controller is an add-on; the mono controller need not be replaced, removed, deactivated, or modified, facilitating transparency and backward compatibility.Type: GrantFiled: February 21, 2022Date of Patent: January 16, 2024Assignee: West Texas Technology Partners LLCInventors: Mario Kosmiskas, Nathan Abercrombie, Sleiman Itani
-
Patent number: 11870971Abstract: To enhance a mono-output-only controller such as a mobile OS to support selective mono/stereo/mixed output, a stereo controller is instantiated in communication with the mono controller. The stereo controller coordinates stereo output, but calls and adapts functions already present in the mono controller for creating surface and image buffers, rendering, compositing, and/or merging. For content designated for 2D display, left and right surfaces are rendered from a mono perspective; for content designated for 3D display, left and right surfaces are rendered from left and right stereo perspectives, respectively. Some, all, or none of available content may be delivered to a stereo display in 3D, with a remainder delivered in 2D, and with comparable content still delivered in 2D to the mono display. The stereo controller is an add-on; the mono controller need not be replaced, removed, deactivated, or modified, facilitating transparency and backward compatibility.Type: GrantFiled: November 12, 2021Date of Patent: January 9, 2024Assignee: West Texas Technology Partners, LLCInventors: Mario Kosmiskas, Nathan Abercrombie, Sleiman Itani
-
Patent number: 11836295Abstract: A free space input standard is instantiated on a processor. Free space input is sensed and communicated to the processor. If the free space input satisfies the free space input standard, a touch screen input response is invoked in an operating system. The free space input may be sensed using continuous implicit, discrete implicit, active explicit, or passive explicit approaches. The touch screen input response may be invoked through communicating virtual touch screen input, a virtual input event, or a virtual command to or within the operating system. In this manner free space gestures may control existing touch screen interfaces and devices, without modifying those interfaces and devices directly to accept free space gestures.Type: GrantFiled: February 14, 2023Date of Patent: December 5, 2023Assignee: West Texas Technology Partners, LLCInventors: Shashwat Kandadai, Nathan Abercrombie, Yu-Hsiang Chen, Sleiman Itani
-
Publication number: 20230297173Abstract: A free space input standard is instantiated on a processor. Free space input is sensed and communicated to the processor. If the free space input satisfies the free space input standard, a touch screen input response is invoked in an operating system. The free space input may be sensed using continuous implicit, discrete implicit, active explicit, or passive explicit approaches. The touch screen input response may be invoked through communicating virtual touch screen input, a virtual input event, or a virtual command to or within the operating system. in this manner free space gestures may control existing touch screen interfaces and devices, without modifying those interfaces and devices directly to accept free space gestures.Type: ApplicationFiled: February 14, 2023Publication date: September 21, 2023Inventors: Shashwat Kandadai, Nathan Abercrombie, Yu-Hsiang Chen, Sleiman Itani
-
Publication number: 20230269358Abstract: A target is outputted to an ideal position in 3D space. A viewer indicates the apparent position of the target, and the indication is sensed. An offset between the ideal and apparent positions is determined, and an adjustment determined from the offset such that the apparent position of the ideal position with the adjustment matches the ideal position without the adjustment. The adjustment is made to the first entity and/or a second entity, such that the entities appear to the viewer in the ideal position. The indication may be monocular with a separate indication for each eye, or binocular with a single viewer indication for both eyes. The indication also may serve as communication, such as a PIN input, so that calibration is transparent to the viewer. The method may be continuous, intermittent, or otherwise ongoing over time.Type: ApplicationFiled: May 1, 2023Publication date: August 24, 2023Inventors: Sina Fateh, Nathan Abercrombie, Sleiman Itani
-
Patent number: 11579706Abstract: A free space input standard is instantiated on a processor. Free space input is sensed and communicated to the processor. If the free space input satisfies the free space input standard, a touch screen input response is invoked in an operating system. The free space input may be sensed using continuous implicit, discrete implicit, active explicit, or passive explicit approaches. The touch screen input response may be invoked through communicating virtual touch screen input, a virtual input event, or a virtual command to or within the operating system. In this manner free space gestures may control existing touch screen interfaces and devices, without modifying those interfaces and devices directly to accept free space gestures.Type: GrantFiled: March 4, 2022Date of Patent: February 14, 2023Assignee: WEST TEXAS TECHNOLOGY PARTNERS, LLCInventors: Shashwat Kandadai, Nathan Abercrombie, Yu-Hsiang Chen, Sleiman Itani
-
Publication number: 20230012770Abstract: A user, a manipulator such as a hand, and at least one entity such as a virtual or augmented reality object are in an interface such as a 3D environmental interface. The manipulation distance is the distance between a reference feature of the user and a manipulation feature of the manipulator. The entity distance is the distance between the reference feature and an entity feature of the entity. When the manipulation distance becomes greater than the entity distance, the entity is caused to fade, disappear, move out of the way, shrink, etc. so as to be less of an obstruction to the user's field of view, for example to avoid obstructing more distant entities. Other factors than the manipulation distance and entity distance may be considered in determining whether to reduce the obstructivity of the entity, and exceptions to the obstruction relation may be considered.Type: ApplicationFiled: September 27, 2022Publication date: January 19, 2023Inventors: Nathan Abercrombie, Theo Goguely
-
Patent number: 11455072Abstract: A user, a manipulator such as a hand, and at least one entity such as a virtual or augmented reality object are in an interface such as a 3D environmental interface. The manipulation distance is the distance between a reference feature of the user and a manipulation feature of the manipulator. The entity distance is the distance between the reference feature and an entity feature of the entity. When the manipulation distance becomes greater than the entity distance, the entity is caused to fade, disappear, move out of the way, shrink, etc. so as to be less of an obstruction to the user's field of view, for example to avoid obstructing more distant entities. Other factors than the manipulation distance and entity distance may be considered in determining whether to reduce the obstructivity of the entity, and exceptions to the obstruction relation may be considered.Type: GrantFiled: April 24, 2019Date of Patent: September 27, 2022Assignee: WEST TEXAS TECHNOLOGY PARTNERS, LLCInventors: Nathan Abercrombie, Theo Goguely
-
Publication number: 20220261086Abstract: A free space input standard is instantiated on a processor. Free space input is sensed and communicated to the processor. If the free space input satisfies the free space input standard, a touch screen input response is invoked in an operating system. The free space input may be sensed using continuous implicit, discrete implicit, active explicit, or passive explicit approaches. The touch screen input response may be invoked through communicating virtual touch screen input, a virtual input event, or a virtual command to or within the operating system. In this manner free space gestures may control existing touch screen interfaces and devices, without modifying those interfaces and devices directly to accept free space gestures.Type: ApplicationFiled: March 4, 2022Publication date: August 18, 2022Inventors: Shashwat Kandadai, Nathan Abercrombie, Yu-Hsiang Chen, Sleiman Itani
-
Publication number: 20220187945Abstract: A method, system, apparatus, and/or device for sensing an input in an augmented reality construct. The method, system, apparatus, and/or device may include a mixed-reality device, comprises memory and one or more processors communicatively coupled to a sensor, a touch device, and a display device that is at least partially transparent and configured to display a virtual object. The store instructions executable by the one or more processors to generate virtual object data for displaying a virtual object by the display device; output the virtual object data to the display device; receive, from the sensor, interaction data corresponding to a free-space interaction by a hand of a user with the virtual object; generate touch input data based on the interaction data; and output the touch input data to the touch device.Type: ApplicationFiled: March 7, 2022Publication date: June 16, 2022Inventors: Shashwat Kandadai, Nathan Abercrombie, Yu-Hsiang Chen, Sleiman Itani
-
Publication number: 20220182599Abstract: To enhance a mono-output-only controller such as a mobile OS to support selective mono/stereo/mixed output, a stereo controller is instantiated in communication with the mono controller. The stereo controller coordinates stereo output, but calls and adapts functions already present in the mono controller for creating surface and image buffers, rendering, compositing, and/or merging. For content designated for 2D display, left and right surfaces are rendered from a mono perspective; for content designated for 3D display, left and right surfaces are rendered from left and right stereo perspectives, respectively. Some, all, or none of available content may be delivered to a stereo display in 3D, with a remainder delivered in 2D, and with comparable content still delivered in 2D to the mono display. The stereo controller is an add-on; the mono controller need not be replaced, removed, deactivated, or modified, facilitating transparency and backward compatibility.Type: ApplicationFiled: February 21, 2022Publication date: June 9, 2022Inventors: Mario Kosmiskas, Nathan Abercrombie, Sleiman Itani
-
Publication number: 20220150465Abstract: To enhance a mono-output-only controller such as a mobile OS to support selective mono/stereo/mixed output, a stereo controller is instantiated in communication with the mono controller. The stereo controller coordinates stereo output, but calls and adapts functions already present in the mono controller for creating surface and image buffers, rendering, compositing, and/or merging. For content designated for 2D display, left and right surfaces are rendered from a mono perspective; for content designated for 3D display, left and right surfaces are rendered from left and right stereo perspectives, respectively. Some, all, or none of available content may be delivered to a stereo display in 3D, with a remainder delivered in 2D, and with comparable content still delivered in 2D to the mono display. The stereo controller is an add-on; the mono controller need not be replaced, removed, deactivated, or modified, facilitating transparency and backward compatibility.Type: ApplicationFiled: November 12, 2021Publication date: May 12, 2022Inventors: Mario Kosmiskas, Nathan Abercrombie, Sleiman Itani