Patents by Inventor Rajinder Sodhi
Rajinder Sodhi has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240028129Abstract: A artificial-reality system, including an optional head-wearable device to display an artificial-reality environment and a wrist-wearable device, identifies and detects shapes of surfaces, and, optionally, the distances of surfaces from the wrist-wearable device, and makes operations, including haptic events, available to users by detecting gestures performed by the user, causing respective operations to be performed based on the type of gestures, and optionally, the distance of the wrist-wearable device from surfaces during performance of the gestures. For example, the wrist-wearable device can be configured to cause performance of a first set of operations corresponding to in-air gestures, which are performed within an in-air threshold distance of a surface, a second set of operations corresponding to near-surface gestures performed within a surface threshold distance of a surface.Type: ApplicationFiled: May 1, 2023Publication date: January 25, 2024Inventors: Eric Whitmire, Rajinder Sodhi, Andre Levi, Austin Lee
-
Publication number: 20240019938Abstract: A method includes, while presenting an artificial-reality environment that includes a virtual object and a representation of a physical object, detecting, based on data from a first group of sensors of a wrist-wearable device, that the wrist-wearable device is within a first distance of an electronic device responsive to user gestures and within another threshold distance of the virtual object. The electronic device is represented within the artificial-reality environment. The method includes, while the wrist-wearable device is within the first activation-threshold distance of the electronic device and is also within a second activation-threshold distance of the virtual object, determining, based on data from a second set of sensors of the wrist-wearable device, that a gesture corresponds to an operation at the electronic device. And the method includes, based on determining that the user gesture corresponds to the operation, causing the physical electronic device to perform the operation.Type: ApplicationFiled: May 1, 2023Publication date: January 18, 2024Inventors: Eric Whitmire, Rajinder Sodhi, Andre Levi, Austin Lee, Szeyin Lee
-
Patent number: 11455739Abstract: One variation of method includes: serving setup frames to a projector facing a scene; at a peripheral control module comprising a camera facing the scene, recording a set of images during projection of corresponding setup frames onto the scene by the projector and a baseline image depicting the scene in the field of view of the camera; calculating a pixel correspondence map based on the set of images and the setup frames; transforming the baseline image into a corrected color image—depicting the scene in the field of view of the camera—based on the pixel correspondence map; linking visual assets to discrete regions in the corrected color image; generating augmented reality frames depicting the visual assets aligned with these discrete regions; and serving the augmented reality frames to the projector to cast depictions of the visual assets onto surfaces, in the scene, corresponding to these discrete regions.Type: GrantFiled: May 11, 2020Date of Patent: September 27, 2022Assignee: Lightform, Inc.Inventors: Kevin Karsch, Rajinder Sodhi, Brett Jones, Pulkit Budhiraja, Phil Reyneri, Douglas Rieck, Andrew Kilkenny, Ehsan Noursalehi, Derek Nedelman, Laura LaPerche, Brittany Factura
-
Patent number: 11245883Abstract: One variation of a method for augmenting surfaces in a space includes: at a projection system, recording a sequence of scans of the space; aggregating the sequence of scans into a projector-domain image of the space; detecting a set of objects in the projector-domain image; identifying an object, in the set of objects, as of a first type; detecting a surface proximal the object in the projector-domain image; defining an association between the surface and a content source based on the first type of the object; based on the association, warping visual content output by the content source according to a profile of the surface extracted from the projector-domain image; actuating a projector assembly in the projection system to locate the surface in the field of view of a light projector in the projection system; and projecting the visual content, via the light projector, toward the surface.Type: GrantFiled: January 29, 2020Date of Patent: February 8, 2022Assignee: Lightform, Inc.Inventors: Rajinder Sodhi, Brett Jones, Kevin Karsch, Pulkit Budhiraja, Phil Reyneri, Douglas Rieck
-
Patent number: 11223807Abstract: One variation of a method for augmenting surfaces within spaces with projected light includes: at a projector system during a first time period, projecting visual content onto nearby surfaces via a light projector integrated into the projector system and capturing a first scan of nearby surfaces, illuminated by the light projector, via an optical sensor integrated into the projector system; identifying a first space occupied by the projector system during the first time period based on features detected in the first scan; selecting a first augmented content source, from a first set of augmented content sources affiliated with the first space, associated with a first surface in the first space; articulating the light projector to locate the first surface in a field of view of the light projector; accessing a frame from the first augmented content source; and projecting the frame onto the first surface via the light projector.Type: GrantFiled: September 10, 2020Date of Patent: January 11, 2022Assignee: Lightform, Inc.Inventors: Rajinder Sodhi, Brett Jones, Kevin Karsch, Pulkit Budhiraja, Phil Reyneri, Douglas Rieck, Andrew Kilkenny
-
Publication number: 20200413016Abstract: One variation of a method for augmenting surfaces within spaces with projected light includes: at a projector system during a first time period, projecting visual content onto nearby surfaces via a light projector integrated into the projector system and capturing a first scan of nearby surfaces, illuminated by the light projector, via an optical sensor integrated into the projector system; identifying a first space occupied by the projector system during the first time period based on features detected in the first scan; selecting a first augmented content source, from a first set of augmented content sources affiliated with the first space, associated with a first surface in the first space; articulating the light projector to locate the first surface in a field of view of the light projector; accessing a frame from the first augmented content source; and projecting the frame onto the first surface via the light projector.Type: ApplicationFiled: September 10, 2020Publication date: December 31, 2020Inventors: Rajinder Sodhi, Brett Jones, Kevin Karsch, Pulkit Budhiraja, Phil Reyneri, Douglas Rieck, Andrew Kilkenny
-
Publication number: 20200402247Abstract: One variation of method includes: serving setup frames to a projector facing a scene; at a peripheral control module comprising a camera facing the scene, recording a set of images during projection of corresponding setup frames onto the scene by the projector and a baseline image depicting the scene in the field of view of the camera; calculating a pixel correspondence map based on the set of images and the setup frames; transforming the baseline image into a corrected color image—depicting the scene in the field of view of the camera—based on the pixel correspondence map; linking visual assets to discrete regions in the corrected color image; generating augmented reality frames depicting the visual assets aligned with these discrete regions; and serving the augmented reality frames to the projector to cast depictions of the visual assets onto surfaces, in the scene, corresponding to these discrete regions.Type: ApplicationFiled: May 11, 2020Publication date: December 24, 2020Inventors: Kevin Karsch, Rajinder Sodhi, Brett Jones, Pulkit Budhiraja, Phil Reyneri, Douglas Rieck, Andrew Kilkenny, Ehsan Noursalehi, Derek Nedelman, Laura LaPerche, Brittany Factura
-
Publication number: 20200374498Abstract: One variation of a method for augmenting surfaces in a space includes: at a projection system, recording a sequence of scans of the space; aggregating the sequence of scans into a projector-domain image of the space; detecting a set of objects in the projector-domain image; identifying an object, in the set of objects, as of a first type; detecting a surface proximal the object in the projector-domain image; defining an association between the surface and a content source based on the first type of the object; based on the association, warping visual content output by the content source according to a profile of the surface extracted from the projector-domain image; actuating a projector assembly in the projection system to locate the surface in the field of view of a light projector in the projection system; and projecting the visual content, via the light projector, toward the surface.Type: ApplicationFiled: January 29, 2020Publication date: November 26, 2020Inventors: Rajinder Sodhi, Brett Jones, Kevin Karsch, Pulkit Budhiraja, Phil Reyneri, Douglas Rieck
-
Patent number: 10805585Abstract: One variation of a method for augmenting surfaces within spaces with projected light includes: at a projector system during a first time period, projecting visual content onto nearby surfaces via a light projector integrated into the projector system and capturing a first scan of nearby surfaces, illuminated by the light projector, via an optical sensor integrated into the projector system; identifying a first space occupied by the projector system during the first time period based on features detected in the first scan; selecting a first augmented content source, from a first set of augmented content sources affiliated with the first space, associated with a first surface in the first space; articulating the light projector to locate the first surface in a field of view of the light projector; accessing a frame from the first augmented content source; and projecting the frame onto the first surface via the light projector.Type: GrantFiled: December 2, 2019Date of Patent: October 13, 2020Assignee: Lightform, Inc.Inventors: Rajinder Sodhi, Brett Jones, Kevin Karsch, Pulkit Budhiraja, Phil Reyneri, Douglas Rieck, Andrew Kilkenny
-
Patent number: 10692233Abstract: One variation of method includes: serving setup frames to a projector facing a scene; at a peripheral control module comprising a camera facing the scene, recording a set of images during projection of corresponding setup frames onto the scene by the projector and a baseline image depicting the scene in the field of view of the camera; calculating a pixel correspondence map based on the set of images and the setup frames; transforming the baseline image into a corrected color image—depicting the scene in the field of view of the camera—based on the pixel correspondence map; linking visual assets to discrete regions in the corrected color image; generating augmented reality frames depicting the visual assets aligned with these discrete regions; and serving the augmented reality frames to the projector to cast depictions of the visual assets onto surfaces, in the scene, corresponding to these discrete regions.Type: GrantFiled: June 20, 2019Date of Patent: June 23, 2020Assignee: Lightform, Inc.Inventors: Kevin Karsch, Rajinder Sodhi, Brett Jones, Pulkit Budhiraja, Phil Reyneri, Douglas Rieck, Andrew Kilkenny, Ehsan Noursalehi, Derek Nedelman, Laura LaPerche, Brittany Factura
-
Publication number: 20200195900Abstract: One variation of a method for augmenting surfaces within spaces with projected light includes: at a projector system during a first time period, projecting visual content onto nearby surfaces via a light projector integrated into the projector system and capturing a first scan of nearby surfaces, illuminated by the light projector, via an optical sensor integrated into the projector system; identifying a first space occupied by the projector system during the first time period based on features detected in the first scan; selecting a first augmented content source, from a first set of augmented content sources affiliated with the first space, associated with a first surface in the first space; articulating the light projector to locate the first surface in a field of view of the light projector; accessing a frame from the first augmented content source; and projecting the frame onto the first surface via the light projector.Type: ApplicationFiled: December 2, 2019Publication date: June 18, 2020Inventors: Rajinder Sodhi, Brett Jones, Kevin Karsch, Pulkit Budhiraja, Phil Reyneri, Douglas Rieck, Andrew Kilkenny
-
Publication number: 20200105006Abstract: One variation of method includes: serving setup frames to a projector facing a scene; at a peripheral control module comprising a camera facing the scene, recording a set of images during projection of corresponding setup frames onto the scene by the projector and a baseline image depicting the scene in the field of view of the camera; calculating a pixel correspondence map based on the set of images and the setup frames; transforming the baseline image into a corrected color image—depicting the scene in the field of view of the camera—based on the pixel correspondence map; linking visual assets to discrete regions in the corrected color image; generating augmented reality frames depicting the visual assets aligned with these discrete regions; and serving the augmented reality frames to the projector to cast depictions of the visual assets onto surfaces, in the scene, corresponding to these discrete regions.Type: ApplicationFiled: June 20, 2019Publication date: April 2, 2020Inventors: Kevin Karsch, Rajinder Sodhi, Brett Jones, Pulkit Budhiraja, Phil Reyneri, Douglas Rieck, Andrew Kilkenny, Ehsan Noursalehi, Derek Nedelman, Laura LaPerche, Brittany Factura
-
Patent number: 10373325Abstract: One variation of method includes: serving setup frames to a projector facing a scene; at a peripheral control module comprising a camera facing the scene, recording a set of images during projection of corresponding setup frames onto the scene by the projector and a baseline image depicting the scene in the field of view of the camera; calculating a pixel correspondence map based on the set of images and the setup frames; transforming the baseline image into a corrected color image—depicting the scene in the field of view of the camera—based on the pixel correspondence map; linking visual assets to discrete regions in the corrected color image; generating augmented reality frames depicting the visual assets aligned with these discrete regions; and serving the augmented reality frames to the projector to cast depictions of the visual assets onto surfaces, in the scene, corresponding to these discrete regions.Type: GrantFiled: September 28, 2018Date of Patent: August 6, 2019Assignee: Lightform, Inc.Inventors: Kevin Karsch, Rajinder Sodhi, Brett Jones, Pulkit Budhiraja, Phil Reyneri, Douglas Rieck, Andrew Kilkenny, Ehsan Noursalehi, Derek Nedelman, Laura LaPerche, Brittany Factura
-
Patent number: 9619027Abstract: To convey tactile sensations over an open space, a system may use a vortex generator to direct one or more vortices at an object in 3-D space. Once a vortex strikes an object—e.g., a user's hand—it applies a force. The vortex generator can control the frequency and intensity of the vortices in order to provide different tactile sensations that correspond to virtual objects or events in a visual presentation. The system may identify and track objects in the real-world environment, and based on information provided by a device displaying the visual presentation, transmit instructions to the vortex generator to discharge vortices that convey a tactile sensation corresponding to the virtual object or event in the visual presentation. By doing so, the vortices augment the real-world environment to immerse the user in the visual presentation.Type: GrantFiled: July 12, 2013Date of Patent: April 11, 2017Assignee: Disney Enterprises, Inc.Inventors: Rajinder Sodhi, Ivan Poupyrev, Matthew Feld Glisson
-
Patent number: 9210401Abstract: Physical movement of a human subject may be guided by a visual cue. A physical environment may be observed to identify a current position of a body portion of the human subject. A model path of travel may be obtained for the body portion of the human subject. The visual cue may be projected onto the human subject and/or into a field of view of the human subject. The visual cue may indicate the model path of travel for the body portion of the human subject.Type: GrantFiled: May 3, 2012Date of Patent: December 8, 2015Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Andrew Wilson, Hrvoje Benko, Rajinder Sodhi
-
Publication number: 20150015607Abstract: To convey tactile sensations over an open space, a system may use a vortex generator to direct one or more vortices at an object in 3-D space. Once a vortex strikes an object—e.g., a user's hand—it applies a force. The vortex generator can control the frequency and intensity of the vortices in order to provide different tactile sensations that correspond to virtual objects or events in a visual presentation. The system may identify and track objects in the real-world environment, and based on information provided by a device displaying the visual presentation, transmit instructions to the vortex generator to discharge vortices that convey a tactile sensation corresponding to the virtual object or event in the visual presentation. By doing so, the vortices augment the real-world environment to immerse the user in the visual presentation.Type: ApplicationFiled: July 12, 2013Publication date: January 15, 2015Inventors: Rajinder Sodhi, Ivan Poupyrev, Matthew Feld Glisson
-
Publication number: 20130295539Abstract: Physical movement of a human subject may be guided by a visual cue. A physical environment may be observed to identify a current position of a body portion of the human subject. A model path of travel may be obtained for the body portion of the human subject. The visual cue may be projected onto the human subject and/or into a field of view of the human subject. The visual cue may indicate the model path of travel for the body portion of the human subject.Type: ApplicationFiled: May 3, 2012Publication date: November 7, 2013Applicant: MICROSOFT CORPORATIONInventors: Andrew Wilson, Hrvoje Benko, Rajinder Sodhi