Patents by Inventor Leah M. GUM
Leah M. GUM has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240402800Abstract: Various implementations disclosed herein include devices, systems, and methods that interpret user activity as user interactions with user interface (UI) elements positioned within a three-dimensional (3D) space such as an extended reality (XR) environment. Some implementations enable user interactions with virtual elements displayed in 3D environments that utilize alternative input modalities, e.g., XR environments that interpret user activity as either direct interactions or indirect interactions with virtual elements.Type: ApplicationFiled: May 29, 2024Publication date: December 5, 2024Inventors: Julian K. Shutzberg, David J. Meyer, David M. Teitelbaum, Mehmet N. Agaoglu, Ian R. Fasel, Chase B. Lortie, Daniel J. Brewer, Tim H. Cornelissen, Leah M. Gum, Alexander G. Berardino, Lorenzo Soto Doblado, Vinay Chawda, Itay Bar Yosef, Dror Irony, Eslam A. Mostafa, Guy Engelhard, Paul A. Lacey, Ashwin Kumar Asoka Kumar Shenoi, Bhavin Vinodkumar Nayak, Liuhao Ge, Lucas Soffer, Victor Belyaev, Bharat C. Dandu, Matthias M. Schroeder, Yirong Tang
-
Publication number: 20240393876Abstract: Various implementations provide views of 3D environments (e.g., extended reality (XR) environments). Non-eye-based user activity, such as hand gestures, is associated with some types of eye-based activity, such as the user gazing at a particular user interface component displayed within a view of a 3D environment. For example, a user's pinching hand gesture may be associated with the user gazing at a particular user interface component, such as a button, at around the same time as the pinching hand gesture is made. These associated behaviors (e.g., the pinch and gaze at the button) may then be interpreted as user input, e.g., user input selecting or otherwise acting upon that user interface component. In some implementations, non-eye-based user activity is only associated with types of eye-based user activity that are likely to correspond to a user perceiving what they are seeing and/or intentionally looking at something.Type: ApplicationFiled: July 31, 2024Publication date: November 28, 2024Inventors: Vinay Chawda, Mehmet N. Agaoglu, Leah M. Gum, Paul A. Lacey, Julian K. Shutzberg, Tim H. Cornelissen, Alexander G. Berardino
-
Publication number: 20240361901Abstract: In some embodiments, a computer system enables a user to invoke display of transport controls (and/or other controls associated with controlling playback of content) using gaze inputs, gesture inputs, or a combination of these. In some embodiments, in response to detecting a first user input, the computer system displays a first set of controls in a reduced-prominence state (e.g., in a manner that is not unduly distracting to the user), and in response to detecting a second user input, the computer system displays a second set of controls in an increased-prominence state (e.g., in a more visually prominent state). The second set of controls optionally includes more controls than the first set of controls.Type: ApplicationFiled: January 30, 2024Publication date: October 31, 2024Inventors: Jonathan RAVASZ, Angel Suet Yan CHEUNG, Ashwin Kumar ASOKA KUMAR SHENOI, Leah M. GUM, Zoey C. TAYLOR, Evgenii KRIVORUCHKO, Christopher D. MCKENZIE, Matan STAUBER, Yonghyun A. KIM, Gregory T. SCOTT, Lucio MORENO RUFO, Fredric R. VINNA, Brian K. SHIRAISHI, So TANAKA
-
Patent number: 12099653Abstract: Various implementations provide views of 3D environments (e.g., extended reality (XR) environments). Non-eye-based user activity, such as hand gestures, is associated with some types of eye-based activity, such as the user gazing at a particular user interface component displayed within a view of a 3D environment. For example, a user's pinching hand gesture may be associated with the user gazing at a particular user interface component, such as a button, at around the same time as the pinching hand gesture is made. These associated behaviors (e.g., the pinch and gaze at the button) may then be interpreted as user input, e.g., user input selecting or otherwise acting upon that user interface component. In some implementations, non-eye-based user activity is only associated with types of eye-based user activity that are likely to correspond to a user perceiving what they are seeing and/or intentionally looking at something.Type: GrantFiled: September 11, 2023Date of Patent: September 24, 2024Assignee: APPLE INC.Inventors: Vinay Chawda, Mehmet N. Agaoglu, Leah M. Gum, Paul A. Lacey, Julian K. Shutzberg, Tim H. Cornelissen, Alexander G. Berardino
-
Publication number: 20240103634Abstract: Techniques for mapping a user input motion includes detecting an input motion by a user, determining an origin for an input motion in a user-centric spherical coordinate system, determining an arc length for the input motion based on the determined origin, mapping the arc length of the input motion to a 2D plane of a user input component, and presenting a movement of a user input component on the 2D plane in accordance with the mapping.Type: ApplicationFiled: September 22, 2023Publication date: March 28, 2024Inventors: Vinay Chawda, Chase B. Lortie, Daniel J. Brewer, Julian K. Shutzberg, Leah M. Gum, Yirong Tang, Alexander T. Wing
-
Publication number: 20240103705Abstract: Various implementations disclosed herein facilitate interactions with a user interface in 3D environment in which a user interface element is moved based on a user movement in a way that the user interface element appears to lag behind or follow a portion of the user (e.g., the user's fingertip). The user interface element may be moved in a way that it converges with and thus catches up to the portion of the user. Such convergence may be based on the speed of the movement of the portion of the user. No convergence may occur when the portion of the user is not moving or is moving below a threshold speed. When the portion of the user is moving (e.g., above a threshold speed), the user interface component may converge with the portion of the user and the rate of convergence may be increased with faster speeds.Type: ApplicationFiled: September 12, 2023Publication date: March 28, 2024Inventors: Vinay Chawda, Julian K. Shutzberg, Chase B. Lortie, Daniel J. Brewer, David J. Meyer, Leah M. Gum
-
Publication number: 20240103613Abstract: Various implementations provide views of 3D environments (e.g., extended reality (XR) environments). Non-eye-based user activity, such as hand gestures, is associated with some types of eye-based activity, such as the user gazing at a particular user interface component displayed within a view of a 3D environment. For example, a user's pinching hand gesture may be associated with the user gazing at a particular user interface component, such as a button, at around the same time as the pinching hand gesture is made. These associated behaviors (e.g., the pinch and gaze at the button) may then be interpreted as user input, e.g., user input selecting or otherwise acting upon that user interface component. In some implementations, non-eye-based user activity is only associated with types of eye-based user activity that are likely to correspond to a user perceiving what they are seeing and/or intentionally looking at something.Type: ApplicationFiled: September 11, 2023Publication date: March 28, 2024Inventors: Vinay Chawda, Mehmet N. Agaoglu, Leah M. Gum, Paul A. Lacey, Julian K. Shutzberg, Tim H. Cornelissen, Alexander G. Birardino
-
Publication number: 20240103635Abstract: Suppressing a hand gesture upon detecting peripheral events on a peripheral device includes determining a first hand pose for a first hand a second hand pose for a second hand in response to a detected peripheral device peripheral event, determining, based on the first hand pose and the second hand pose, at least one hand of the first hand and the second hand in a peripheral use mode, detecting an input gesture from a hand of the at least one hand determined to be in the peripheral use mode, and rejecting the input gesture by a user input pipeline in accordance with the determination that the hand is in the peripheral use mode. The presence of a peripheral device is confirmed by activating a computer vision system in response to determining that a peripheral use condition is satisfied.Type: ApplicationFiled: September 22, 2023Publication date: March 28, 2024Inventors: Daniel J. Brewer, Ashwin Kumar Asoka Kumar Shenoi, Tian Qiu, Leah M. Gum, David J. Meyer, Julian K. Shutzberg, Yirong Tang
-
Publication number: 20240094825Abstract: Aspects of the subject technology provide improved techniques for gesture recognition. Improved techniques may include detecting and/or classifying an interaction between the body part and another object in a scan of the body part, and then controlling recognition of a gesture based on the interaction. In an aspect, recognition parameters may be selected based on the interaction classification that disable recognition of one or more gestures while not disabling recognition of other gestures.Type: ApplicationFiled: September 15, 2023Publication date: March 21, 2024Inventors: Lailin CHEN, Ashwin Kumar ASOKA KUMAR SHENOI, Daniel J. BREWER, Eslam A. MOSTAFA, Itay BAR YOSEF, Julian K. SHUTZBERG, Leah M. GUM, Martin MELOUN, Minhaeng LEE, Victor BELYAEV
-
Publication number: 20240036699Abstract: While a view of an environment is visible via a display generation component of a computer system, the computer system detects a gaze input directed to a first location, corresponding to a first user interface element, in the environment. In response to detecting the gaze input: if a user's hand is in a predefined configuration during the gaze input, the computer system: provides, to the first user interface element, information about the gaze input; and then, in response to detecting the gaze input moving to a different, second location in the environment while the user's hand is maintained in the predefined configuration, provides, to a second user interface element that corresponds to the second location, information about the gaze input. If the user's hand is not in the predefined configuration during the gaze input, the computer system forgoes providing, to the first user interface element, information about the gaze input.Type: ApplicationFiled: July 25, 2023Publication date: February 1, 2024Inventors: Mark A. Ebbole, Leah M. Gum, Chia-Ling Li, Ashwin Kumar Asoka Kumar Shenoi
-
Publication number: 20230333665Abstract: Techniques for managing an engagement zone include tracking, by a system, a hand of a user and determining that a height of the hand of the user satisfies a first threshold height. In accordance with determining that the height of the hand of the user satisfies the first threshold height, the techniques also include initiating a UI engagement state, wherein the system monitors the user for user input during the UI engagement state, and determining user input into the system based on a user motion detected while the hand is tracked. The threshold height is associated with a boundary of a UI engagement zone and is modifiable based on user activity.Type: ApplicationFiled: April 19, 2023Publication date: October 19, 2023Inventors: Ashwin Kumar Asoka Kumar Shenoi, Julian K. Shutzberg, Leah M. Gum, Daniel J. Brewer, Chia-Ling Li
-
Patent number: 11086368Abstract: An electronic device detects an input sequence that includes an increase in intensity of an input that corresponds to a first input event. If the input sequence includes a second input event, including a decrease in intensity of the input after the first input event, a long press is not detected and a first operation is performed. If the second input event is not detected within a second time period that is longer than the first time period and the input had a characteristic intensity above a respective intensity threshold between when the first input event was detected and when the second time period elapsed, a long press is detected on an accelerated basis, and a second operation is performed. Finally, a long press is detected if the second input event is not detected within a longer, default time period, and the second operation is performed.Type: GrantFiled: August 13, 2020Date of Patent: August 10, 2021Assignee: APPLE INC.Inventors: Nicole M. Wells, Leah M. Gum, Kenneth L. Kocienda, Camille Moussette, Jean-Pierre M. Mouilleseaux, Joshua B. Kopin, Jules K. Fennis
-
Publication number: 20200371630Abstract: An electronic device detects an input sequence that includes an increase in intensity of an input that corresponds to a first input event. If the input sequence includes a second input event, including a decrease in intensity of the input after the first input event, a long press is not detected and a first operation is performed. If the second input event is not detected within a second time period that is longer than the first time period and the input had a characteristic intensity above a respective intensity threshold between when the first input event was detected and when the second time period elapsed, a long press is detected on an accelerated basis, and a second operation is performed. Finally, a long press is detected if the second input event is not detected within a longer, default time period, and the second operation is performed.Type: ApplicationFiled: August 13, 2020Publication date: November 26, 2020Inventors: Nicole M. Wells, Leah M. Gum, Kenneth L. Kocienda, Camille Moussette, Jean-Pierre M. Mouilleseaux, Joshua B. Kopin, Jules K. Fennis
-
Patent number: 10775915Abstract: An electronic device detects a change in intensity of an input on an input element that includes detecting an increase in intensity followed by a decrease in intensity, and determines whether the first decrease in intensity of the input meets up-click detection criteria, which require that the intensity of the input decrease below a first up-click intensity threshold in order for the up-click detection criteria to be met. The first up-click intensity threshold is time varying, in accordance with the detected intensity of the input during the first decrease in intensity of the input. If the first decrease in intensity of the input meets the up-click detection criteria, first feedback is provided, indicating that the first decrease in intensity of the input was recognized as an up-click, and if the first decrease in intensity of the input does not meet the up-click detection criteria, the first feedback is not provided.Type: GrantFiled: January 18, 2018Date of Patent: September 15, 2020Assignee: APPLE INC.Inventors: Nicole M. Wells, Leah M. Gum, Nicholas A. Fahrner, Camille Moussette, Jean-Pierre M. Mouilleseaux, Joshua B. Kopin, Jules K. Fennis
-
Patent number: 10599247Abstract: Thermal compensation can be applied to force measurements of a force-sensitive button. A temperature differential between an object and the force-sensitive button can result in changes in the reconstructed force by the force sensor due to thermal effects rather than actual user force, which in turn can result in degraded performance of the force sensor (e.g., false positive or inconsistent activation force). In some examples, a force-sensitive button can include a force sensor configured to measure an amount of force applied to the force-sensitive button, and a temperature sensor configured to measure a temperature associated with the force sensor. The measured temperature can be used to compensate the amount of force measured by the force sensor based on the temperature associated with the force sensor. In some examples, the thermal compensation can be applied when an object is detected contacting the force-sensitive button (i.e., when rapid temperature differentials can occur).Type: GrantFiled: September 6, 2016Date of Patent: March 24, 2020Assignee: Apple Inc.Inventors: Eric S. Winokur, Robert Leo Sheridan, Leah M. Gum, Sora Kim, Giovanni Gozzini
-
Publication number: 20180143723Abstract: An electronic device detects a change in intensity of an input on an input element that includes detecting an increase in intensity followed by a decrease in intensity, and determines whether the first decrease in intensity of the input meets up-click detection criteria, which require that the intensity of the input decrease below a first up-click intensity threshold in order for the up-click detection criteria to be met. The first up-click intensity threshold is time varying, in accordance with the detected intensity of the input during the first decrease in intensity of the input. If the first decrease in intensity of the input meets the up-click detection criteria, first feedback is provided, indicating that the first decrease in intensity of the input was recognized as an up-click, and if the first decrease in intensity of the input does not meet the up-click detection criteria, the first feedback is not provided.Type: ApplicationFiled: January 18, 2018Publication date: May 24, 2018Inventors: Nicole M. Wells, Leah M. Gum, Nicholas A. Fahrner, Camille Moussette, Jean-Pierre M. Mouilleseaux, Joshua B. Kopin, Jules K. Fennis
-
Publication number: 20180067601Abstract: Thermal compensation can be applied to force measurements of a force-sensitive button. A temperature differential between an object and the force-sensitive button can result in changes in the reconstructed force by the force sensor due to thermal effects rather than actual user force, which in turn can result in degraded performance of the force sensor (e.g., false positive or inconsistent activation force). In some examples, a force-sensitive button can include a force sensor configured to measure an amount of force applied to the force-sensitive button, and a temperature sensor configured to measure a temperature associated with the force sensor. The measured temperature can be used to compensate the amount of force measured by the force sensor based on the temperature associated with the force sensor. In some examples, the thermal compensation can be applied when an object is detected contacting the force-sensitive button (i.e., when rapid temperature differentials can occur).Type: ApplicationFiled: September 6, 2016Publication date: March 8, 2018Inventors: Eric S. WINOKUR, Robert Leo SHERIDAN, Leah M. GUM, Sora KIM, Giovanni GOZZINI
-
Publication number: 20180067596Abstract: An electronic device detects a change in intensity of an input on an input element that includes detecting an increase in intensity followed by a decrease in intensity, and recognizes at least a portion of the change in intensity of the input as a first input event that is associated with a first operation, for example a single click operation. After recognizing the first input event, the device delays performance of the first operation while monitoring subsequent changes in intensity of the input for a second input event, wherein the delay is limited by a default delay time period. If the second event is recognized before default delay time period has elapsed, a second operation is performed and the first operation is not performed. However, if early-confirmation criteria for the first input event are met before the default delay period elapses, the first operation is performed.Type: ApplicationFiled: September 20, 2016Publication date: March 8, 2018Inventors: Nicole M. Wells, Leah M. Gum, Kenneth L. Kocienda, Camille Moussette, Jean-Pierre M. Mouilleseaux, Joshua B. Kopin, Jules K. Fennis
-
Patent number: 9910524Abstract: An electronic device detects a change in intensity of an input on an input element that includes detecting an increase in intensity followed by a decrease in intensity, and recognizes at least a portion of the change in intensity of the input as a first input event that is associated with a first operation, for example a single click operation. After recognizing the first input event, the device delays performance of the first operation while monitoring subsequent changes in intensity of the input for a second input event, wherein the delay is limited by a default delay time period. If the second event is recognized before default delay time period has elapsed, a second operation is performed and the first operation is not performed. However, if early-confirmation criteria for the first input event are met before the default delay period elapses, the first operation is performed.Type: GrantFiled: September 20, 2016Date of Patent: March 6, 2018Assignee: APPLE INC.Inventors: Nicole M. Wells, Leah M. Gum, Kenneth L. Kocienda, Camille Moussette, Jean-Pierre M. Mouilleseaux, Joshua B. Kopin, Jules K. Fennis
-
Publication number: 20170153760Abstract: An electronic device can include gain-based error tracking for improved force sensing performance. The electronic device can comprise a plurality of force sensors (e.g., coupled to a touch sensor panel configured to detect an object touching the touch sensor panel). The plurality of force sensors can be configured to detect an amount of force with which the object touches the touch sensor panel. A processor can be coupled to the plurality of force sensors, and the processor can be configured to: in accordance with a determination that an acceleration characteristic of the electronic device is less than a threshold, determine an error metric for one or more of the plurality of force sensors, and in accordance with a determination that the acceleration characteristic of the electronic device is not less than the threshold, forgo determining the error metric for one or more of the plurality of force sensors.Type: ApplicationFiled: April 1, 2016Publication date: June 1, 2017Inventors: Vinay CHAWDA, Vikrham GOWREESUNKER, Leah M. GUM, Teera SONGATIKAMAS