Patents Assigned to Leap Motion, Inc.
  • Publication number: 20190050661
    Abstract: The technology disclosed relates to coordinating motion-capture of a hand by a network of motion-capture sensors having overlapping fields of view. In particular, it relates to designating a first sensor among three or more motion-capture sensors as having a master frame of reference, observing motion of a hand as it passes through overlapping fields of view of the respective motion-capture sensors, synchronizing capture of images of the hand within the overlapping fields of view by pairs of the motion-capture devices, and using the pairs of the hand images captured by the synchronized motion-capture devices to automatically calibrate the motion-capture sensors to the master frame of reference frame.
    Type: Application
    Filed: June 28, 2018
    Publication date: February 14, 2019
    Applicant: Leap Motion, Inc.
    Inventor: David S. HOLZ
  • Publication number: 20190050509
    Abstract: The technology disclosed relates to simplifying updating of a predictive model using clustering observed points. In particular, it relates to observing a set of points in 3D sensory space, determining surface normal directions from the points, clustering the points by their surface normal directions and adjacency, accessing a predictive model of a hand, refining positions of segments of the predictive model, matching the clusters of the points to the segments, and using the matched clusters to refine the positions of the matched segments. It also relates to distinguishing between alternative motions between two observed locations of a control object in a 3D sensory space by accessing first and second positions of a segment of a predictive model of a control object such that motion between the first position and the second position was at least partially occluded from observation in a 3D sensory space.
    Type: Application
    Filed: June 8, 2018
    Publication date: February 14, 2019
    Applicant: Leap Motion, Inc.
    Inventors: David S. HOLZ, Raffi BEDIKIAN, Kevin HOROWITZ, Hua YANG
  • Publication number: 20190042957
    Abstract: The technology disclosed relates to manipulating a virtual object. In particular, it relates to detecting a hand in a three-dimensional (3D) sensory space and generating a predictive model of the hand, and using the predictive model to track motion of the hand. The predictive model includes positions of calculation points of fingers, thumb and palm of the hand. The technology disclosed relates to dynamically selecting at least one manipulation point proximate to a virtual object based on the motion tracked by the predictive model and positions of one or more of the calculation points, and manipulating the virtual object by interaction between at least some of the calculation points of the predictive model and the dynamically selected manipulation point.
    Type: Application
    Filed: June 5, 2018
    Publication date: February 7, 2019
    Applicant: Leap Motion, Inc.
    Inventors: David S. HOLZ, Raffi BEDIKIAN, Adrian GASINSKI, Maxwell SILLS, Hua YANG, Gabriel HARE
  • Publication number: 20190037132
    Abstract: The technology disclosed relates to adjusting the monitored field of view of a camera and/or a view of a virtual scene from a point of view of a virtual camera based on the distance between tracked objects. For example, if the user's hand is being tracked for gestures, the closer the hand gets to another object, the tighter the frame can becomeā€”i.e., the more the camera can zoom in so that the hand and the other object occupy most of the frame. The camera can also be reoriented so that the hand and the other object remain in the center of the field of view. The distance between two objects in a camera's field of view can be determined and a parameter of a motion-capture system adjusted based thereon. In particular, the pan and/or zoom levels of the camera may be adjusted in accordance with the distance.
    Type: Application
    Filed: May 24, 2018
    Publication date: January 31, 2019
    Applicant: Leap Motion, Inc.
    Inventor: David S. HOLZ
  • Publication number: 20190033979
    Abstract: During control of a user interface via free-space motions of a hand or other suitable control object, switching between control modes can be facilitated by tracking the control object's movements relative to, and its penetration of, a virtual control construct (such as a virtual surface construct). The position of the virtual control construct can be updated, continuously or from time to time, based on the control object's location.
    Type: Application
    Filed: August 3, 2018
    Publication date: January 31, 2019
    Applicant: Leap Motion, Inc.
    Inventors: Raffi BEDIKIAN, Jonathan MARSDEN, Keith MERTENS, David HOLZ
  • Publication number: 20190033975
    Abstract: The technology disclosed relates to relates to providing command input to a machine under control. It further relates to gesturally interacting with the machine. The technology disclosed also relates to providing monitoring information about a process under control. The technology disclosed further relates to providing biometric information about an individual. The technology disclosed yet further relates to providing abstract features information (pose, grab strength, pinch strength, confidence, and so forth) about an individual.
    Type: Application
    Filed: May 24, 2018
    Publication date: January 31, 2019
    Applicant: Leap Motion, Inc.
    Inventors: Kevin A. HOROWITZ, Matias PEREZ, Raffi BEDIKIAN, David S. HOLZ, Gabriel A. HARE
  • Publication number: 20190017813
    Abstract: Methods and systems for capturing motion and/or determining the shapes and positions of one or more objects in 3D space utilize cross-sections thereof. In various embodiments, images of the cross-sections are captured using a camera based on reflections therefrom or shadows cast thereby.
    Type: Application
    Filed: April 13, 2018
    Publication date: January 17, 2019
    Applicant: Leap Motion, Inc.
    Inventor: David S. HOLZ
  • Publication number: 20190018141
    Abstract: The technology disclosed relates to determining positional information of an object in a field of view. In particular, it relates to calculating a distance of the object from a reference such as a sensor including scanning the field of view by selectively illuminating directionally oriented light sources and measuring one or more differences in property of returning light emitted from the light sources and reflected from the object. The property can be intensity or phase difference of the light. It also relates to finding an object in a region of space. In particular, it relates to scanning the region of space with directionally controllable illumination, determining a difference in a property of the illumination received for two or more points in the scanning, and determining positional information of the object based in part upon the points in the scanning corresponding to the difference in the property.
    Type: Application
    Filed: March 26, 2018
    Publication date: January 17, 2019
    Applicant: Leap Motion, Inc.
    Inventor: David HOLZ
  • Publication number: 20190018497
    Abstract: Methods and systems for processing input from an image-capture device for gesture-recognition. The method further includes computationally interpreting user gestures in accordance with a first mode of operation; analyzing the path of movement of an object to determine an intent of a user to change modes of operation; and, upon determining an intent of the user to change modes of operation, subsequently interpreting user gestures in accordance with the second mode of operation.
    Type: Application
    Filed: March 22, 2018
    Publication date: January 17, 2019
    Applicant: Leap Motion, Inc.
    Inventor: David S. HOLZ
  • Publication number: 20190019332
    Abstract: Free space machine interface and control can be facilitated by predictive entities useful in interpreting a control object's position and/or motion (including objects having one or more articulating members, i.e., humans and/or animals and/or machines). Predictive entities can be driven using motion information captured using image information or the equivalents. Predictive information can be improved applying techniques for correlating with information from observations.
    Type: Application
    Filed: March 30, 2018
    Publication date: January 17, 2019
    Applicant: Leap Motion, Inc.
    Inventors: Kevin A. HOROWITZ, David S. HOLZ
  • Publication number: 20190018495
    Abstract: Methods and systems for processing an input are disclosed that detect a portion of a hand and/or other detectable object in a region of space monitored by a 3D sensor. The method further includes determining a zone corresponding to the region of space in which the portion of the hand or other detectable object was detected. Also, the method can include determining from the zone a correct way to interpret inputs made by a position, shape or a motion of the portion of the hand or other detectable object.
    Type: Application
    Filed: March 9, 2018
    Publication date: January 17, 2019
    Applicant: LEAP MOTION, INC.
    Inventors: Michael ZAGORSEK, Avinash DABIR, Paul DURDIK, Keith MERTENS
  • Publication number: 20190012564
    Abstract: Enhanced contrast between an object of interest and background surfaces visible in an image is provided using controlled lighting directed at the object. Exploiting the falloff of light intensity with distance, a light source (or multiple light sources), such as an infrared light source, can be positioned near one or more cameras to shine light onto the object while the camera(s) capture images. The captured images can be analyzed to distinguish object pixels from background pixels.
    Type: Application
    Filed: March 27, 2018
    Publication date: January 10, 2019
    Applicant: Leap Motion, Inc.
    Inventors: David S. Holz, Hua Yang
  • Patent number: 10168873
    Abstract: The technology disclosed relates to providing simplified manipulation of virtual objects by detected hand motions. In particular, it relates to a detecting hand motion and positions of the calculation points relative to a virtual object to be manipulated, dynamically selecting at least one manipulation point proximate to the virtual object based on the detected hand motion and positions of one or more of the calculation points, and manipulating the virtual object by interaction between the detected hand motion and positions of one or more of the calculation points and the dynamically selected manipulation point.
    Type: Grant
    Filed: October 29, 2014
    Date of Patent: January 1, 2019
    Assignee: LEAP MOTION, INC.
    Inventors: David S Holz, Raffi Bedikian, Adrian Gasinski, Maxwell Sills, Hua Yang, Gabriel Hare
  • Patent number: 10152824
    Abstract: Methods and systems for capturing motion and/or determining the shapes and positions of one or more objects in 3D space utilize cross-sections thereof. In various embodiments, images of the cross-sections are captured using a camera based on edge points thereof.
    Type: Grant
    Filed: January 4, 2018
    Date of Patent: December 11, 2018
    Assignee: Leap Motion, Inc.
    Inventor: David Holz
  • Patent number: 10152136
    Abstract: The technology disclosed relates to automatically interpreting a gesture of a control object in a three dimensional sensor space by sensing a movement of the control object in the three dimensional sensor space, sensing orientation of the control object, defining a control plane tangential to a surface of the control object and interpreting the gesture based on whether the movement of the control object is more normal to the control plane or more parallel to the control plane.
    Type: Grant
    Filed: October 16, 2014
    Date of Patent: December 11, 2018
    Assignee: LEAP MOTION, INC.
    Inventors: Isaac Cohen, David S Holz, Maxwell Sills
  • Patent number: 10139918
    Abstract: Embodiments of display control based on dynamic user interactions generally include capturing a plurality of temporally sequential images of the user, or a body part or other control object manipulated by the user, and computationally analyzing the images to recognize a gesture performed by the user. In some embodiments, a scale indicative of an actual gesture distance traversed in performance of the gesture is identified, and a movement or action is displayed on the device based, at least in part, on a ratio between the identified scale and the scale of the displayed movement. In some embodiments, a degree of completion of the recognized gesture is determined, and the display contents are modified in accordance therewith. In some embodiments, a dominant gesture is computationally determined from among a plurality of user gestures, and an action displayed on the device is based on the dominant gesture.
    Type: Grant
    Filed: September 28, 2016
    Date of Patent: November 27, 2018
    Assignee: Leap Motion, Inc.
    Inventors: Raffi Bedikian, Jonathan Marsden, Keith Mertens, David Samuel Holz, Maxwell Sills, Matias Perez, Gabriel A. Hare, Ryan Christopher Julian
  • Publication number: 20180329509
    Abstract: The technology disclosed relates to automatically (e.g., programmatically) initializing predictive information for tracking a complex control object (e.g., hand, hand and tool combination, robot end effector) based upon information about characteristics of the object determined from sets of collected observed information. Automated initialization techniques obviate the need for special and often bizarre start-up rituals (place your hands on the screen at the places indicated during a full moon, and so forth) required by conventional techniques. In implementations, systems can refine initial predictive information to reflect an observed condition based on comparison of the observed with an analysis of sets of collected observed information.
    Type: Application
    Filed: February 5, 2018
    Publication date: November 15, 2018
    Applicant: Leap Motion, Inc.
    Inventor: Kevin A. HOROWITZ
  • Patent number: 10126822
    Abstract: The technology disclosed relates to creating user-defined interaction spaces and modalities in a three dimensional (3D) sensor space in response to control gestures. It also relates to controlling virtual cameras in the 3D sensor space using control gestures and manipulating controls of the virtual cameras through the control gestures. In particular, it relates to defining one or more spatial attributes of the interaction spaces and modalities in response to one or more gesture parameters of the control gesture. It also particularly relates to defining one or more visual parameters of a virtual camera in response to one or more gesture parameters of the control gesture.
    Type: Grant
    Filed: December 16, 2014
    Date of Patent: November 13, 2018
    Assignee: Leap Motion, Inc.
    Inventors: Isaac Cohen, Maxwell Sills
  • Patent number: 10095269
    Abstract: A motion control assembly includes a motion control device electrically connected to a battery pack and to a mobile computing device for at least data transmission therebetween. The motion control device can generate inputs, such as inputs corresponding to an attribute of a sensed object, for transmission to the mobile computing device. The drain on a battery of a battery-powered mobile computing device can be reduced when used with a motion control device as follows. A motion control assembly, comprising a motion control device and a battery pack, capable of powering the motion control device, as an integral, one-piece unit, is selected. The motion control device is connected to an electrical connector of a battery-powered mobile computing device. The motion control device is supplied with power from the battery pack during use so the motion control device can be operated using the power from the battery pack.
    Type: Grant
    Filed: August 15, 2017
    Date of Patent: October 9, 2018
    Assignee: Leap Motion, Inc.
    Inventors: Robert Samuel Gordon, Paul Alan Durdik
  • Patent number: 10097754
    Abstract: The technology disclosed provides systems and methods for reducing the overall power consumption of an optical motion-capture system without compromising the quality of motion capture and tracking. In implementations, this is accomplished by operating the motion-detecting cameras and associated image-processing hardware in a low-power mode (e.g., at a low frame rate or in a standby or sleep mode) unless and until touch gestures of an object such as a tap, sequence of taps, or swiping motions are performed with a surface proximate to the cameras. A contact microphone or other appropriate sensor is used for detecting audio signals or other vibrations generated by contact of the object with the surface.
    Type: Grant
    Filed: April 14, 2017
    Date of Patent: October 9, 2018
    Assignee: Leap Motion, Inc.
    Inventor: David S. Holz