Patents by Inventor Robert Konrad

Robert Konrad has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 12367234
    Abstract: A system enables a user to query based on a user's gaze by receiving a query from the user and capturing, via an eye tracking system on a headset, the user's gaze location near an object in a local area. The system captures one or more images of the local area with the object and formats the images based in part on a region of interest in the one or more images that includes the object. The system generates a formatted query based in part on the query. The formatted query is provided to a search engine. Information describing the object determined from the one or more formatted images and information describing the query are used by the search engine to determine an answer to the query about the object. The system presents the answer to the query about the object.
    Type: Grant
    Filed: January 29, 2024
    Date of Patent: July 22, 2025
    Assignee: Sesame AI, Inc.
    Inventors: Robert Konrad Konrad, Gordon Wetzstein, Kevin Conlon Boyle, John Gabriel Buckmaster, Nitish Padmanaban
  • Patent number: 12346495
    Abstract: A differential camera system for object tracking. The system includes a co-aligned light source camera assembly (LSCA) and a controller. The co-aligned LSCA includes a light source and a differential camera sensor. The light source is configured to emit light along an optical path that is directed towards an eye box including an eye of a user. The differential camera sensor is configured to detect a change in brightness of the eye caused in part by the emitted light, asynchronously output data samples corresponding to the detected change in brightness, wherein the optical path is substantially co-aligned with an optical path of the differential camera sensor. The controller is configured to identify a pupil of the eye based on data samples output from the differential camera sensor resulting from the emitted light, and determine a gaze location of the user based in part on the identified pupil.
    Type: Grant
    Filed: September 15, 2023
    Date of Patent: July 1, 2025
    Assignee: Sesame AI, Inc.
    Inventors: Robert Konrad Konrad, Kevin Conlon Boyle, Gordon Wetzstein, Nitish Padmanaban, John Gabriel Buckmaster
  • Patent number: 12287915
    Abstract: An event camera system for pupil detection may include a camera assembly and a controller, and may also include one or more off-axis light sources. The camera assembly may include one or more infrared (IR) light sources and an event camera. The one or more IR light sources are configured to emit pulses of IR light along an optical path toward an eyebox. The IR light is reflected from an eye in the eyebox, and the reflected light propagates back along the optical path toward the event camera for detection. The controller is configured to determine an orientation of the eye using data output from the event camera.
    Type: Grant
    Filed: November 15, 2022
    Date of Patent: April 29, 2025
    Assignee: Sesame AI, Inc.
    Inventors: Kevin Boyle, Robert Konrad, Nitish Padmanaban
  • Patent number: 12204532
    Abstract: A device configured to perform data analytics comprising a memory and a processor may be configured to perform the techniques described in this disclosure. The memory may store multi-dimensional data. The processor may receive a sequence of inputs defining a recipe for analyzing the multi-dimensional data according to a language sub-surface specifying a natural language containment hierarchy defining a grammar for a natural language as a hierarchical arrangement of a plurality of language sub-surfaces. The processor may also receive data indicative of a summarized narration of the recipe and parameterize a field of the summarized narration to insert a user adjustable parameter that enables manipulation of the underlying recipe and obtain a parameterized summary that includes the user adjustable parameter. The processor may next present, via a first user interface, the parameterized summary.
    Type: Grant
    Filed: December 9, 2021
    Date of Patent: January 21, 2025
    Assignee: DataChat.ai
    Inventors: Jignesh Patel, Robert Konrad Claus, Amos Kendall, Rogers Jeffrey Leo John, Ushmal Ramesh, Jiatong Li
  • Publication number: 20240419721
    Abstract: A system enables a user to query based on a user's gaze by receiving a query from the user and capturing, via an eye tracking system on a headset, the user's gaze location near an object in a local area. The system captures one or more images of the local area with the object and formats the images based in part on a region of interest in the one or more images that includes the object. The system generates a formatted query based in part on the query. The formatted query is provided to a search engine. Information describing the object determined from the one or more formatted images and information describing the query are used by the search engine to determine an answer to the query about the object. The system presents the answer to the query about the object.
    Type: Application
    Filed: January 29, 2024
    Publication date: December 19, 2024
    Inventors: Robert Konrad Konrad, Gordon Wetzstein, Kevin Conlon Boyle, John Gabriel Buckmaster, Nitish Padmanaban
  • Patent number: 12124624
    Abstract: An event camera system for pupil detection may include a camera assembly and a controller, and may also include one or more off-axis light sources. The camera assembly may include one or more infrared (IR) light sources and an event camera. The one or more IR light sources are configured to emit pulses of IR light along an optical path toward an eyebox. The IR light is reflected from an eye in the eyebox, and the reflected light propagates back along the optical path toward the event camera for detection. The controller is configured to determine an orientation of the eye using data output from the event camera.
    Type: Grant
    Filed: November 15, 2022
    Date of Patent: October 22, 2024
    Assignee: Sesame AI, Inc.
    Inventors: Kevin Boyle, Robert Konrad, Nitish Padmanaban
  • Publication number: 20240256529
    Abstract: A device comprising a processor and a memory may be configured to perform the techniques described in this disclosure. The processor may present, via one or more portions of a first, second, or third user interface, one or more of either an interactive text box or interactive search bar in which a user may enter data indicative of a current input, an interactive log of previous inputs, a graphical representation of result data obtained responsive to the data indicative of the current input, one or more datasets, and at least a portion of the multi-dimensional data included in the one or more datasets. Various portions of the various user interfaces are separately scrollable but coupled such that interactions in the various portions of the various user interfaces synchronize the various portions of the various user interfaces. The memory is configured to store the data indicative of the current input.
    Type: Application
    Filed: January 26, 2023
    Publication date: August 1, 2024
    Inventors: Jignesh Patel, Rogers Jeffrey Leo John, Robert Konrad Claus, Jiatong Li, Sulong Zhou, Yukiko Suzuki
  • Publication number: 20240160012
    Abstract: An event camera system for pupil detection may include a camera assembly and a controller, and may also include one or more off-axis light sources. The camera assembly may include one or more infrared (IR) light sources and an event camera. The one or more IR light sources are configured to emit pulses of IR light along an optical path toward an eyebox. The IR light is reflected from an eye in the eyebox, and the reflected light propagates back along the optical path toward the event camera for detection. The controller is configured to determine an orientation of the eye using data output from the event camera.
    Type: Application
    Filed: November 15, 2022
    Publication date: May 16, 2024
    Inventors: Kevin Boyle, Robert Konrad, Nitish Padmanaban
  • Publication number: 20240098360
    Abstract: A tracking system for object detection and tracking. The system may include a plurality of light sources, a differential camera, and a controller. The plurality of light sources is positioned at different locations on a device and are configured to emit pulses of light that illuminate an object. The differential camera has an optical axis, and at least some of the plurality of light sources are off-axis relative to the optical axis. The differential camera is configured to detect a change in brightness of the object caused in part by one or more of the pulses of light, and asynchronously output data samples corresponding to the detected change in brightness. The controller is configured to track the object based in part on the data samples output by the differential camera.
    Type: Application
    Filed: September 15, 2023
    Publication date: March 21, 2024
    Inventors: Robert Konrad Konrad, Kevin Conlon Boyle, Gordon Wetzstein
  • Publication number: 20240094811
    Abstract: A differential camera system for object tracking. The system includes a co-aligned light source camera assembly (LSCA) and a controller. The co-aligned LSCA includes a light source and a differential camera sensor. The light source is configured to emit light along an optical path that is directed towards an eye box including an eye of a user. The differential camera sensor is configured to detect a change in brightness of the eye caused in part by the emitted light, asynchronously output data samples corresponding to the detected change in brightness, wherein the optical path is substantially co-aligned with an optical path of the differential camera sensor. The controller is configured to identify a pupil of the eye based on data samples output from the differential camera sensor resulting from the emitted light, and determine a gaze location of the user based in part on the identified pupil.
    Type: Application
    Filed: September 15, 2023
    Publication date: March 21, 2024
    Inventors: Robert Konrad Konrad, Kevin Conlon Boyle, Gordon Wetzstein, Nitish Padmanaban, John Gabriel Buckmaster
  • Publication number: 20240045215
    Abstract: An augmented reality display system is configured to direct a plurality of parallactically-disparate intra-pupil images into a viewer's eye. The parallactically-disparate intra-pupil images provide different parallax views of a virtual object, and impinge on the pupil from different angles. In the aggregate, the wavefronts of light forming the images approximate a continuous divergent wavefront and provide selectable accommodation cues for the user, depending on the amount of parallax disparity between the intra-pupil images. The amount of parallax disparity is selected using a light source that outputs light for different images from different locations, with spatial differences in the locations of the light output providing differences in the paths that the light takes to the eye, which in turn provide different amounts of parallax disparity.
    Type: Application
    Filed: October 19, 2023
    Publication date: February 8, 2024
    Inventors: Michael Anthony KLUG, Robert KONRAD, Gordon WETZSTEIN, Brian T. SCHOWENGERDT, Michal Beau Dennison VAUGHN
  • Patent number: 11835724
    Abstract: An augmented reality display system is configured to direct a plurality of parallactically-disparate intra-pupil images into a viewer's eye. The parallactically-disparate intra-pupil images provide different parallax views of a virtual object, and impinge on the pupil from different angles. In the aggregate, the wavefronts of light forming the images approximate a continuous divergent wavefront and provide selectable accommodation cues for the user, depending on the amount of parallax disparity between the intra-pupil images. The amount of parallax disparity is selected using a light source that outputs light for different images from different locations, with spatial differences in the locations of the light output providing differences in the paths that the light takes to the eye, which in turn provide different amounts of parallax disparity.
    Type: Grant
    Filed: February 13, 2023
    Date of Patent: December 5, 2023
    Assignee: Magic Leap, Inc.
    Inventors: Michael Anthony Klug, Robert Konrad, Gordon Wetzstein, Brian T. Schowengerdt, Michal Beau Dennison Vaughn
  • Publication number: 20230240606
    Abstract: Embodiments are related to a system with a headset capable of monitoring psychomotor performance of a user of the headset based on eyelid tracking information. The headset includes a sensor assembly coupled to a frame of the headset, and a transceiver coupled to the sensor assembly. The sensor assembly is configured to track an eyelid of an eye of the user and capture eyelid tracking information. The transceiver is configured to obtain the eyelid tracking information from the sensor assembly and communicate the eyelid tracking information to a secondary device coupled to the headset for processing the eyelid tracking information and determination of sleep information for the user based in part on the processed eyelid tracking information.
    Type: Application
    Filed: January 19, 2023
    Publication date: August 3, 2023
    Inventors: Kevin Boyle, Robert Konrad, Nitish Padmanaban
  • Publication number: 20230194879
    Abstract: An augmented reality display system is configured to direct a plurality of parallactically-disparate intra-pupil images into a viewer's eye. The parallactically-disparate intra-pupil images provide different parallax views of a virtual object, and impinge on the pupil from different angles. In the aggregate, the wavefronts of light forming the images approximate a continuous divergent wavefront and provide selectable accommodation cues for the user, depending on the amount of parallax disparity between the intra-pupil images. The amount of parallax disparity is selected using a light source that outputs light for different images from different locations, with spatial differences in the locations of the light output providing differences in the paths that the light takes to the eye, which in turn provide different amounts of parallax disparity.
    Type: Application
    Filed: February 13, 2023
    Publication date: June 22, 2023
    Inventors: Michael Anthony Klug, Robert Konrad, Gordon Wetzstein, Brian T. Schowengerdt, Michal Beau Dennison Vaughn
  • Publication number: 20230195220
    Abstract: An event camera system for pupil detection may include a camera assembly and a controller, and may also include one or more off-axis light sources. The camera assembly may include one or more infrared (IR) light sources and an event camera. The one or more IR light sources are configured to emit pulses of IR light along an optical path toward an eyebox. The IR light is reflected from an eye in the eyebox, and the reflected light propagates back along the optical path toward the event camera for detection. The controller is configured to determine an orientation of the eye using data output from the event camera.
    Type: Application
    Filed: November 15, 2022
    Publication date: June 22, 2023
    Inventors: Kevin Boyle, Robert Konrad, Nitish Padmanaban
  • Publication number: 20230185798
    Abstract: A device configured to perform data analytics comprising a memory and a processor may be configured to perform the techniques described in this disclosure. The memory may store multi-dimensional data. The processor may receive a sequence of inputs defining a recipe for analyzing the multi-dimensional data according to a language sub-surface specifying a natural language containment hierarchy defining a grammar for a natural language as a hierarchical arrangement of a plurality of language sub-surfaces. The processor may also receive data indicative of a summarized narration of the recipe and parameterize a field of the summarized narration to insert a user adjustable parameter that enables manipulation of the underlying recipe and obtain a parameterized summary that includes the user adjustable parameter. The processor may next present, via a first user interface, the parameterized summary.
    Type: Application
    Filed: December 9, 2021
    Publication date: June 15, 2023
    Inventors: Jignesh Patel, Robert Konrad Claus, Amos Kendall, Rogers Jeffrey Leo John, Ushmal Ramesh, Jiatong Li
  • Patent number: 11662574
    Abstract: A device includes a camera assembly and a controller. The camera assembly is configured to capture images of both eyes of a user. Using the captured images, the controller determines a location for each pupil of each eye of the user. The determined pupil locations and captured images are used to determine eye tracking parameters which are used to compute values of eye tracking functions. With the computed values and a model that maps the eye tracking functions to gaze depths, a gaze depth of the user is determined. An action is performed based on the determined gaze depth.
    Type: Grant
    Filed: November 9, 2021
    Date of Patent: May 30, 2023
    Assignee: Zinn Labs, Inc.
    Inventors: Kevin Boyle, Robert Konrad, Nitish Padmanaban, Gordon Wetzstein
  • Publication number: 20230142618
    Abstract: Embodiments relate to an eye tracking system. A headset of the system includes an eye tracking sensor that captures eye tracking data indicating positions and movements of a user's eye. A controller (e.g., in the headset) of the tracking system analyzes eye tracking data from the sensors to determine eye tracking feature values of the eye during a time period. The controller determines an activity of the user during the time period based on the eye tracking feature values. The controller updates an activity history of the user with the determined activity.
    Type: Application
    Filed: November 3, 2022
    Publication date: May 11, 2023
    Inventors: Kevin Conlon Boyle, Robert Konrad Konrad, Nitish Padmanaban
  • Patent number: 11625095
    Abstract: Embodiments are related to a plurality of gaze sensors embedded into a frame of a headset for detection of a gaze vector of a user wearing the headset and user's control at the headset. The gaze vector for an eye of the user can be within a threshold distance from one of the gaze sensors. By monitoring signals detected by the gaze sensors, it can be determined that the gaze vector is within the threshold distance from the gaze sensor. Based on this determination, at least one action associated with the headset is initiated.
    Type: Grant
    Filed: January 20, 2022
    Date of Patent: April 11, 2023
    Assignee: Zinn Labs, Inc.
    Inventors: Robert Konrad, Kevin Boyle, Nitish Padmanaban, Gordon Wetzstein
  • Patent number: 11614628
    Abstract: An augmented reality display system is configured to direct a plurality of parallactically-disparate intra-pupil images into a viewer's eye. The parallactically-disparate intra-pupil images provide different parallax views of a virtual object, and impinge on the pupil from different angles. In the aggregate, the wavefronts of light forming the images approximate a continuous divergent wavefront and provide selectable accommodation cues for the user, depending on the amount of parallax disparity between the intra-pupil images. The amount of parallax disparity is selected using a light source that outputs light for different images from different locations, with spatial differences in the locations of the light output providing differences in the paths that the light takes to the eye, which in turn provide different amounts of parallax disparity.
    Type: Grant
    Filed: January 21, 2022
    Date of Patent: March 28, 2023
    Assignee: Magic Leap, Inc.
    Inventors: Michael Anthony Klug, Robert Konrad, Gordon Wetzstein, Brian T. Schowengerdt, Michal Beau Dennison Vaughn