Patents by Inventor Helmut Grabner

Helmut Grabner has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11568549
    Abstract: Methods and systems for determining a displacement of a peripheral device are provided. In one example, a peripheral device comprises: an image sensor, and a hardware processor configured to: control the image sensor to capture a first image of a surface when the peripheral device is at a first location on the surface, the first image comprising a feature of the first location of the surface; execute a trained machine learning model using data derived from the first image to estimate a displacement of the feature between the first image and a reference image captured at a second location of the surface; and determine a displacement of the peripheral device based on the estimated displacement of the feature.
    Type: Grant
    Filed: May 27, 2021
    Date of Patent: January 31, 2023
    Assignee: Logitech Europe S.A.
    Inventors: Nicolas Chauvin, Francois Morier, Helmut Grabner
  • Patent number: 11182962
    Abstract: Methods and systems for providing a mixed reality (MR) interaction are provided. In one example, the method comprises: capturing a first image of a user's appendage and a physical peripheral input device, wherein the user's appendage covers at least a part of the physical peripheral input device; performing a segmentation of the first image into a first region of pixels corresponding to the user's appendage, wherein a second region of pixels corresponds to the physical peripheral input device; generating a composite image from the first image based on rendering, in place of the second region of pixels, pixels representing a virtual peripheral input device, such that the virtual peripheral input device appears in the composite image as at a same spatial location as the physical peripheral input device and covered at least partially by the user's appendage; and displaying the composite image to the user.
    Type: Grant
    Filed: February 4, 2019
    Date of Patent: November 23, 2021
    Assignee: Logitech Europe S.A.
    Inventors: Mario Gutierrez, Thomas Rouvinez, Sidney Bovet, Helmut Grabner, Mathieu Meisser
  • Publication number: 20210287380
    Abstract: Methods and systems for determining a displacement of a peripheral device are provided. In one example, a peripheral device comprises: an image sensor, and a hardware processor configured to: control the image sensor to capture a first image of a surface when the peripheral device is at a first location on the surface, the first image comprising a feature of the first location of the surface; execute a trained machine learning model using data derived from the first image to estimate a displacement of the feature between the first image and a reference image captured at a second location of the surface; and determine a displacement of the peripheral device based on the estimated displacement of the feature.
    Type: Application
    Filed: May 27, 2021
    Publication date: September 16, 2021
    Inventors: Nicolas Chauvin, Francois Morier, Helmut Grabner
  • Patent number: 11042992
    Abstract: Methods and systems for determining a displacement of a peripheral device are provided. In one example, a peripheral device comprises: an image sensor, and a hardware processor configured to: control the image sensor to capture a first image of a surface when the peripheral device is at a first location on the surface, the first image comprising a feature of the first location of the surface; execute a trained machine learning model using data derived from the first image to estimate a displacement of the feature between the first image and a reference image captured at a second location of the surface; and determine a displacement of the peripheral device based on the estimated displacement of the feature.
    Type: Grant
    Filed: August 3, 2018
    Date of Patent: June 22, 2021
    Assignee: Logitech Europe S.A.
    Inventors: Nicolas Chauvin, François Morier, Helmut Grabner
  • Patent number: 10943402
    Abstract: Methods and systems for providing a mixed reality (MR) interaction are provided. In one example, a method comprises: capturing, at a first time and using a camera of a head-mounted display (HMD) of a user, a first image of a physical interaction of the user with a physical object; measuring a movement of the HMD with respect to the physical object between the first time and a second time; processing the first image based on the measurement of the movement of the HMD to generate a second image; generating, based on the second image, a composite image of a virtual interaction involving the user; and displaying, via the HMD and based on the composite image, the virtual interaction in place of the physical interaction to the user at the second time.
    Type: Grant
    Filed: February 4, 2019
    Date of Patent: March 9, 2021
    Assignee: Logitech Europe S.A.
    Inventors: Mario Gutierrez, Thomas Rouvinez, Sidney Bovet, Helmut Grabner, Mathieu Meisser
  • Publication number: 20200043179
    Abstract: Methods and systems for determining a displacement of a peripheral device are provided. In one example, a peripheral device comprises: an image sensor, and a hardware processor configured to: control the image sensor to capture a first image of a surface when the peripheral device is at a first location on the surface, the first image comprising a feature of the first location of the surface; execute a trained machine learning model using data derived from the first image to estimate a displacement of the feature between the first image and a reference image captured at a second location of the surface; and determine a displacement of the peripheral device based on the estimated displacement of the feature.
    Type: Application
    Filed: August 3, 2018
    Publication date: February 6, 2020
    Inventors: Nicolas Chauvin, François Morier, Helmut Grabner
  • Publication number: 20190295323
    Abstract: Methods and systems for providing a mixed reality (MR) interaction are provided. In one example, a method comprises: capturing, at a first time and using a camera of a head-mounted display (HMD) of a user, a first image of a physical interaction of the user with a physical object; measuring a movement of the HMD with respect to the physical object between the first time and a second time; processing the first image based on the measurement of the movement of the HMD to generate a second image; generating, based on the second image, a composite image of a virtual interaction involving the user; and displaying, via the HMD and based on the composite image, the virtual interaction in place of the physical interaction to the user at the second time.
    Type: Application
    Filed: February 4, 2019
    Publication date: September 26, 2019
    Inventors: Mario Gutierrez, Thomas Rouvinez, Sidney Bovet, Helmut Grabner, Mathieu Meisser
  • Publication number: 20190295322
    Abstract: Methods and systems for providing a mixed reality (MR) interaction are provided. In one example, the method comprises: capturing a first image of a user's appendage and a physical peripheral input device, wherein the user's appendage covers at least a part of the physical peripheral input device; performing a segmentation of the first image into a first region of pixels corresponding to the user's appendage, wherein a second region of pixels corresponds to the physical peripheral input device; generating a composite image from the first image based on rendering, in place of the second region of pixels, pixels representing a virtual peripheral input device, such that the virtual peripheral input device appears in the composite image as at a same spatial location as the physical peripheral input device and covered at least partially by the user's appendage; and displaying the composite image to the user.
    Type: Application
    Filed: February 4, 2019
    Publication date: September 26, 2019
    Inventors: Mario Gutierrez, Thomas Rouvinez, Sidney Bovet, Helmut Grabner, Mathieu Meisser
  • Patent number: 10299017
    Abstract: In one embodiment, captured video summaries are tagged with metadata so the videos can be easily searched. The videos are classified into different scenes, depending on the type of action in the video, so searching can be based on the type of scene. In one embodiment, tags are provided for moving objects or people. The type of object that is moving is tagged (car, ball, person, pet, etc.). Video search results are ranked based on the weighting of the video events or video summaries. The video summary weighting focuses on important events, with multiple videos/images over a period of time condensed into a short summary video. This creates a weighted video summary with different time-lapse speeds that focuses on important events.
    Type: Grant
    Filed: September 14, 2015
    Date of Patent: May 21, 2019
    Assignee: Logitech Europe S.A.
    Inventors: Vincent Borel, Aaron Standridge, Fabian Nater, Helmut Grabner
  • Patent number: 9805567
    Abstract: In one embodiment of the present invention, a video camera selectively streams to a remote server. Still images or short video events are intermittently transmitted when there is no significant motion detected. When significant motion is detected, video is streamed to the remote server. The images and video can be higher resolution than the bandwidth used, by locally buffering the images and video, and transmitting it at a lower frame rate that extends to when there is no live streaming. This provides a time-delayed stream, but with more resolution at lower bandwidth.
    Type: Grant
    Filed: September 14, 2015
    Date of Patent: October 31, 2017
    Assignee: Logitech Europe S.A.
    Inventors: Vincent Borel, Aaron Standridge, Fabian Nater, Helmut Grabner
  • Publication number: 20170075508
    Abstract: In one embodiment, the User Interface (UI) provides a live stream from the webcam, with markers on the side indicating the stored, detected important events (such as by using a series of bubbles indicating how long ago an event occurred). The indicators are marked to indicate the relative importance, such as with color coding. Upon selection of an indicator by the user, the time-lapse summary is displayed, along with a time of day indication. Alternately, the user can select to have a time-lapse display of all the events in sequence, using a more condensed time lapse, with less important events having less time or being left out.
    Type: Application
    Filed: March 31, 2016
    Publication date: March 16, 2017
    Inventors: Vincent Borel, Aaron Standridge, Fabian Nater, Helmut Grabner
  • Publication number: 20170076571
    Abstract: In one embodiment of the present invention, a video camera selectively streams to a remote server. Still images or short video events are intermittently transmitted when there is no significant motion detected. When significant motion is detected, video is streamed to the remote server. The images and video can be higher resolution than the bandwidth used, by locally buffering the images and video, and transmitting it at a lower frame rate that extends to when there is no live streaming. This provides a time-delayed stream, but with more resolution at lower bandwidth.
    Type: Application
    Filed: September 14, 2015
    Publication date: March 16, 2017
    Inventors: Vincent Borel, Aaron Standridge, Fabian Nater, Helmut Grabner
  • Publication number: 20170078767
    Abstract: In one embodiment, captured video summaries are tagged with metadata so the videos can be easily searched. The videos are classified into different scenes, depending on the type of action in the video, so searching can be based on the type of scene. In one embodiment, tags are provided for moving objects or people. The type of object that is moving is tagged (car, ball, person, pet, etc.). Video search results are ranked based on the weighting of the video events or video summaries. The video summary weighting focuses on important events, with multiple videos/images over a period of time condensed into a short summary video. This creates a weighted video summary with different time-lapse speeds that focuses on important events.
    Type: Application
    Filed: September 14, 2015
    Publication date: March 16, 2017
    Inventors: Vincent Borel, Aaron Standridge, Fabian Nater, Helmut Grabner
  • Publication number: 20170076156
    Abstract: In one embodiment, image recognition is used to determine the type of location where the camera is mounted, such as indoors or outdoors, in a conference room or in a dining room. A filter for selecting the types of events for a summary has parameters varied depending on the type of location. For example, an indoor location may tag events where humans are detected, and ignore animals (pets). An outdoor location can have the parameters set to detect both human and animal movement.
    Type: Application
    Filed: September 14, 2015
    Publication date: March 16, 2017
    Inventors: Vincent Borel, Aaron Standridge, Fabian Nater, Helmut Grabner
  • Patent number: 9588640
    Abstract: In one embodiment, the User Interface (UI) provides a live stream from the webcam, with markers on the side indicating the stored, detected important events (such as by using a series of bubbles indicating how long ago an event occurred). The indicators are marked to indicate the relative importance, such as with color coding. Upon selection of an indicator by the user, the time-lapse summary is displayed, along with a time of day indication. Alternately, the user can select to have a time-lapse display of all the events in sequence, using a more condensed time lapse, with less important events having less time or being left out.
    Type: Grant
    Filed: March 31, 2016
    Date of Patent: March 7, 2017
    Assignee: Logitech Europe S.A.
    Inventors: Vincent Borel, Aaron Standridge, Fabian Nater, Helmut Grabner
  • Patent number: 9313556
    Abstract: In one embodiment, the User Interface (UI) provides a live stream from the webcam, with markers on the side indicating the stored, detected important events (such as by using a series of bubbles indicating how long ago an event occurred). The indicators are marked to indicate the relative importance, such as with color coding. Upon selection of an indicator by the user, the time-lapse summary is displayed, along with a time of day indication. Alternately, the user can select to have a time-lapse display of all the events in sequence, using a more condensed time lapse, with less important events having less time or being left out.
    Type: Grant
    Filed: September 14, 2015
    Date of Patent: April 12, 2016
    Assignee: Logitech Europe S.A.
    Inventors: Vincent Borel, Aaron Standridge, Fabian Nater, Helmut Grabner