Patents by Inventor Leonard Zerman
Leonard Zerman has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240122449Abstract: A system is disclosed that can have a first device and a second device. The first device can have a tube, a deflectable section, and a camera. The second device can have a first tube. A stabilizer can be advanceable from the first tube. The second device can have a second tube. A third tube can be advanceable from the second tube. The first device can be an endoscope. The first device can be attachable to the second device via a connector.Type: ApplicationFiled: December 21, 2023Publication date: April 18, 2024Applicant: Grumpy Innovation, Inc.Inventors: Sameer SHARMA, Luke Emerson HESS, Emily Marie ROBIDEAU, Leonard ZERMAN, Anthony ASAAD
-
Publication number: 20230274516Abstract: A method and system of a user interface device with dual-sided display which may include a system using a brain computer interface with an Augmented Reality (AR) headset. The user's intent is sent to the system, which processes, analyzes and maps the user's intent. An output corresponding to the user's intent is projected using the user interface device. This output is displayed on the user's side of the display. An image corresponding to the output is displayed on the observer's side of the display.Type: ApplicationFiled: May 9, 2023Publication date: August 31, 2023Applicant: Cognixion CorporationInventors: Joseph Andreas Forsland, Leonard Zerman
-
Publication number: 20220326772Abstract: An apparatus, system, and method of a brain computer interface in a headset including an augmented reality display, one or more sensors, a processing module, at least one biofeedback device, and a battery. The interface may include a printed circuit board that has the sensors to read bio-signals, provides biofeedback, and performs the processing, analyzing, and mapping of bio-signals into output. The output provides feedback via stimulation of multiple sensory brain systems of a user, including audio and visual on the augmented reality display, or audio and haptic in terms of vibration patterns that a human user may feel. All together this forms a closed-loop system, by detecting the bio-signal, then providing sensory-feedback, which in turn enhances the bio-signal.Type: ApplicationFiled: June 23, 2022Publication date: October 13, 2022Applicant: Cognixion CorporationInventors: Andreas Forsland, Leonard Zerman
-
Patent number: 11402909Abstract: An apparatus, system, and method of a brain computer interface in a headset including an augmented reality display, one or more sensors, a processing module, at least one biofeedback device, and a battery. The interface may include a printed circuit board that has the sensors to read bio-signals, provides biofeedback, and performs the processing, analyzing, and mapping of bio-signals into output. The output provides feedback via stimulation of multiple sensory brain systems of a user, including audio and visual on the augmented reality display, or audio and haptic in terms of vibration patterns that a human user may feel. All together this forms a closed-loop system, by detecting the bio-signal, then providing sensory-feedback, which in turn enhances the bio-signal.Type: GrantFiled: April 5, 2021Date of Patent: August 2, 2022Assignee: CognixionInventors: Andreas Forsland, Leonard Zerman
-
Publication number: 20210223864Abstract: An apparatus, system, and method of a brain computer interface in a headset including an augmented reality display, one or more sensors, a processing module, at least one biofeedback device, and a battery. The interface may include a printed circuit board that has the sensors to read bio-signals, provides biofeedback, and performs the processing, analyzing, and mapping of bio-signals into output. The output provides feedback via stimulation of multiple sensory brain systems of a user, including audio and visual on the augmented reality display, or audio and haptic in terms of vibration patterns that a human user may feel. All together this forms a closed-loop system, by detecting the bio-signal, then providing sensory-feedback, which in turn enhances the bio-signal.Type: ApplicationFiled: April 5, 2021Publication date: July 22, 2021Applicant: CognixionInventors: Andreas Forsland, Leonard Zerman
-
Patent number: 10990175Abstract: A method and system of a brain computer interface in a headset including an augmented reality display, one or more sensors, a processing module, at least one biofeedback device, and a battery. The interface may include a printed circuit board that has the sensors to read bio-signals, provides biofeedback, and performs the processing, analyzing, and mapping of bio-signals into output. The output provides feedback via stimulation of multiple sensory brain systems of a user, including audio and visual on the augmented reality display, or audio and haptic in terms of vibration patterns that a human user may feel. All together this forms a closed-loop system, by detecting the bio-signal, then providing sensory-feedback, which in turn enhances the bio-signal.Type: GrantFiled: January 9, 2019Date of Patent: April 27, 2021Assignee: CognixionInventors: Andreas Forsland, Leonard Zerman
-
Patent number: 10860095Abstract: A method of real-time eye tracking feedback with an eye-movement tracking camera includes receiving a left eye movement transform, a right eye movement transform, and gaze direction information from a user's face and user's eyes. An eye tracking map is constructed including the left eye movement transform and the right eye movement transform. The eye tracking map is displayed with the left eye movement information, the right eye movement information, and the gaze direction information on a device screen. Feedback is provided to the user related the left eye movement transform, the right eye movement transform, and the gaze direction information.Type: GrantFiled: May 2, 2019Date of Patent: December 8, 2020Assignee: CognixionInventors: Leonard Zerman, Andreas Forsland
-
Publication number: 20200348752Abstract: A method of real-time eye tracking feedback with an eye-movement tracking camera includes receiving a left eye movement transform, a right eye movement transform, and gaze direction information from a user's face and user's eyes. An eye tracking map is constructed including the left eye movement transform and the right eye movement transform. The eye tracking map is displayed with the left eye movement information, the right eye movement information, and the gaze direction information on a device screen. Feedback is provided to the user related the left eye movement transform, the right eye movement transform, and the gaze direction information.Type: ApplicationFiled: May 2, 2019Publication date: November 5, 2020Inventors: Leonard Zerman, Andreas Forsland
-
Publication number: 20200234503Abstract: A method and system of a user interface device with dual-sided display which may include a system using a brain computer interface with an Augmented Reality (AR) headset. The user's intent is sent to the system, which processes, analyzes and maps the user's intent. An output corresponding to the user's intent is projected using the user interface device. This output is displayed on the user's side of the display. An image corresponding to the output is displayed on the observer's side of the display.Type: ApplicationFiled: January 22, 2020Publication date: July 23, 2020Applicant: CognixionInventors: Andreas Forsland, Leonard Zerman
-
Publication number: 20200133393Abstract: A method and system of a brain computer interface in a headset including an augmented reality display, one or more sensors, a processing module, at least one biofeedback device, and a battery. The interface may include a printed circuit board that has the sensors to read bio-signals, provides biofeedback, and performs the processing, analyzing, and mapping of bio-signals into output. The output provides feedback via stimulation of multiple sensory brain systems of a user, including audio and visual on the augmented reality display, or audio and haptic in terms of vibration patterns that a human user may feel. All together this forms a closed-loop system, by detecting the bio-signal, then providing sensory-feedback, which in turn enhances the bio-signal.Type: ApplicationFiled: January 9, 2019Publication date: April 30, 2020Inventors: Andreas Forsland, Leonard Zerman