Patents by Inventor Mark K. Hauenstein

Mark K. Hauenstein has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11947751
    Abstract: An electronic device that is in communication with a display generation component, and sensor(s) to detect location of an input object displays a content selection object within selectable content, wherein the content selection object includes a first edge and a second edge. The device detects an input by the input object, including detecting the input object at a first hover location that corresponds to the first edge of the content selection object. In response to detecting the first portion of the input: in accordance with a determination that the first portion of the input meets first criteria that require the input object meets proximity criteria with respect to the content selection object, the device changes an appearance of the first edge relative to the second edge of the content selection object to indicate that the first edge will be selected for movement when the input object meets second criteria.
    Type: Grant
    Filed: March 30, 2023
    Date of Patent: April 2, 2024
    Assignee: APPLE INC.
    Inventors: Mark K. Hauenstein, Jeffrey T. Bernstein, Julian Missig, Marek A. Bereza
  • Patent number: 11907446
    Abstract: An electronic device with a touch-sensitive display displays an electronic calendar and detects a hand-drawn input on the touch-sensitive display. The hand-drawn input extends over a vertical portion, less than all, of an area corresponding to a first day in the calendar. In response to detecting the hand-drawn input, the device displays, over the vertical portion of the first day, one or more graphics that correspond to the hand-drawn input. While displaying the graphics, the device detects an input that corresponds to a request to create an event in the electronic calendar. In response, the device creates, in the first day in the electronic calendar, an event with a start time and an end time, wherein the start time and the end time of the event are in accordance with the first day that the hand drawn input extends over.
    Type: Grant
    Filed: May 28, 2020
    Date of Patent: February 20, 2024
    Assignee: APPLE INC.
    Inventors: Jeffrey Traer Bernstein, Linda L. Dong, Mark K. Hauenstein, Julian Missig
  • Publication number: 20240029377
    Abstract: A computer system displays a three-dimensional environment and detects, via an input device that includes a first portion and a second portion that can be physically coupled in a first configuration and physically decoupled in a second configuration, a first input. In response to detecting the first input while the first portion is coupled to the second portion of the input device in the first configuration, the computer system performs a first operation in the three-dimensional environment. While the first portion of the input device and the second portion of the input device are decoupled in the second configuration, the computer system detects a sequence of one or more inputs that includes movement of the first portion of the input device relative to the second portion of the input device. In response to detecting the sequence of one or more inputs, the computer system performs one or more second operations.
    Type: Application
    Filed: July 17, 2023
    Publication date: January 25, 2024
    Inventors: Fiona P. O'Leary, Julian K. Missig, Mark K. Hauenstein
  • Patent number: 11818455
    Abstract: A first device sends a request to a second device to initiate a shared annotation session. In response to receiving acceptance of the request, a first prompt to move the first device toward the second device is displayed. In accordance with a determination that connection criteria for the first device and the second device are met, a representation of a field of view of the camera(s) of the first device is displayed in the shared annotation session with the second device. During the shared annotation session, one or more annotations are displayed via the first display generation component and one or more second virtual annotations corresponding to annotation input directed to the respective location in the physical environment by the second device is displayed via the first display generation component, provided that the respective location is included in the field of view of the first set of cameras.
    Type: Grant
    Filed: February 8, 2023
    Date of Patent: November 14, 2023
    Assignee: APPLE INC.
    Inventors: Joseph A. Malia, Mark K. Hauenstein, Praveen Sharma, Matan Stauber, Julian K. Missig, Jeffrey T. Bernstein, Lukas Robert Tom Girling, Matthaeus Krenn
  • Publication number: 20230305674
    Abstract: A computer system displays, in a first viewing mode, a simulated environment that is oriented relative to a physical environment of the computer system. In response to detecting a first change in attitude, the computer system changes an appearance of a first virtual user interface object so as to maintain a fixed spatial relationship between the first virtual user interface object and the physical environment. The computing system detects a gesture. In response to detecting a second change in attitude, in accordance with a determination that the gesture met mode change criteria, the computer system transitions from displaying the simulated environment in the first viewing mode to displaying the simulated environment in a second viewing mode. Displaying the virtual model in the simulated environment in the second viewing mode includes forgoing changing the appearance of the first virtual user interface object to maintain the fixed spatial relationship.
    Type: Application
    Filed: April 27, 2023
    Publication date: September 28, 2023
    Inventors: Mark K. Hauenstein, Joseph A. Malia, Julian K. Missig, Matthaeus Krenn, Jeffrey T. Bernstein
  • Publication number: 20230297180
    Abstract: An electronic device that is in communication with a display generation component, and sensor(s) to detect location of an input object displays a content selection object within selectable content, wherein the content selection object includes a first edge and a second edge. The device detects an input by the input object, including detecting the input object at a first hover location that corresponds to the first edge of the content selection object. In response to detecting the first portion of the input: in accordance with a determination that the first portion of the input meets first criteria that require the input object meets proximity criteria with respect to the content selection object, the device changes an appearance of the first edge relative to the second edge of the content selection object to indicate that the first edge will be selected for movement when the input object meets second criteria.
    Type: Application
    Filed: March 30, 2023
    Publication date: September 21, 2023
    Inventors: Mark K. Hauenstein, Jeffrey T. Bernstein, Julian Missig, Marek A. Bereza
  • Patent number: 11740755
    Abstract: A computer system while displaying an augmented reality environment, concurrently displays: a representation of at least a portion of a field of view of one or more cameras that includes a physical object, and a virtual user interface object at a location in the representation of the field of view, where the location is determined based on the respective physical object in the field of view. While displaying the augmented reality environment, in response to detecting an input that changes a virtual environment setting for the augmented reality environment, the computer system adjusts an appearance of the virtual user interface object in accordance with the change made to the virtual environment setting and applies to at least a portion of the representation of the field of view a filter selected based on the change made to the virtual environment setting.
    Type: Grant
    Filed: September 28, 2021
    Date of Patent: August 29, 2023
    Assignee: APPLE INC.
    Inventors: Mark K. Hauenstein, Joseph A. Malia, Julian K. Missig, Matthaeus Krenn, Jeffrey T. Bernstein
  • Publication number: 20230199296
    Abstract: A first device sends a request to a second device to initiate a shared annotation session. In response to receiving acceptance of the request, a first prompt to move the first device toward the second device is displayed. In accordance with a determination that connection criteria for the first device and the second device are met, a representation of a field of view of the camera(s) of the first device is displayed in the shared annotation session with the second device. During the shared annotation session, one or more annotations are displayed via the first display generation component and one or more second virtual annotations corresponding to annotation input directed to the respective location in the physical environment by the second device is displayed via the first display generation component, provided that the respective location is included in the field of view of the first set of cameras.
    Type: Application
    Filed: February 8, 2023
    Publication date: June 22, 2023
    Inventors: Joseph A. Malia, Mark K. Hauenstein, Praveen Sharma, Matan Stauber, Julian K. Missig, Jeffrey T. Bernstein, Lukas Robert Tom Girling, Matthaeus Krenn
  • Patent number: 11644917
    Abstract: An electronic device, while displaying a first user interface, detects an input for an input object, detects that first hover proximity criteria are met by the input object, and displays first visual feedback. While displaying the first visual feedback, the device detects a change in a current value of a hover proximity parameter of the input object and that second hover proximity criteria are met by the input object after the change. In response to detecting that the second hover proximity criteria are met, the device displays second visual feedback, distinct from the first visual feedback.
    Type: Grant
    Filed: January 19, 2022
    Date of Patent: May 9, 2023
    Assignee: APPLE INC.
    Inventors: Mark K. Hauenstein, Jeffrey T. Bernstein, Julian Missig, Marek A. Bereza
  • Patent number: 11632600
    Abstract: While displaying playback of a first portion of a video in a video playback region, a device receives a request to add a first annotation to the video playback. In response to receiving the request, the device pauses playback of the video at a first position in the video and displays a still image that corresponds to the first, paused position of the video. While displaying the still image, the device receives the first annotation on a first portion of a physical environment captured in the still image. After receiving the first annotation, the device displays, in the video playback region, a second portion of the video that corresponds to a second position in the video, where the first portion of the physical environment is captured in the second portion of the video and the first annotation is displayed in the second portion of the video.
    Type: Grant
    Filed: April 8, 2022
    Date of Patent: April 18, 2023
    Assignee: APPLE INC.
    Inventors: Joseph A. Malia, Mark K. Hauenstein, Praveen Sharma, Matan Stauber, Julian K. Missig, Jeffrey T. Bernstein, Lukas Robert Tom Girling, Matthaeus Krenn
  • Publication number: 20230019329
    Abstract: A device with a display and a touch-sensitive surface displays a user interface including an object. While displaying the user interface, the device detects a first and a second portion of an input, where the first portion includes contact(s) at a location corresponding to the object and the second portion includes movement of the contact(s). In response: upon determining that the second portion was detected shortly after detecting the contact(s): when the first input has a first predefined number of contacts, the device drags the user interface object or a representation; and when the first input has a second predefined number of contacts, the device forgoes the dragging. Further in response, upon determining that the second portion was detected after the contact(s) had been detected at the location for at least the first threshold amount of time, the device drags the user interface object or the representation thereof.
    Type: Application
    Filed: September 16, 2022
    Publication date: January 19, 2023
    Inventors: Mark K. Hauenstein, Raymond S. Sepulveda, Marcos Alonso Ruiz, Julian K. Missig, Jeffrey T. Bernstein, Patrick L. Coffman, Stephen O. Lemay
  • Patent number: 11449222
    Abstract: System and method of moving selected content is disclosed. In response to detecting a first portion of an input that activates a move operation for selected first content, a copy of at least a portion of the first content is displayed in a container object, and the container object is moved from a first placement location to a second placement location relative to a characteristic location of the contact(s) on the touch-screen display. After the movement from the first placement location to the second placement location relative to the characteristic location of the one or more contacts, the container object moves in accordance with movement of the one or more contacts while the container object is maintained at the second placement location relative to the characteristic location of the one or more contacts as the one or more contacts move across the touch-screen display.
    Type: Grant
    Filed: December 30, 2020
    Date of Patent: September 20, 2022
    Assignee: APPLE INC.
    Inventors: Mark K. Hauenstein, Raymond S. Sepulveda, Marcos Alonso Ruiz, Julian K. Missig, Jeffrey T. Bernstein, Patrick L. Coffman, Stephen O. Lemay
  • Publication number: 20220239842
    Abstract: While displaying playback of a first portion of a video in a video playback region, a device receives a request to add a first annotation to the video playback. In response to receiving the request, the device pauses playback of the video at a first position in the video and displays a still image that corresponds to the first, paused position of the video. While displaying the still image, the device receives the first annotation on a first portion of a physical environment captured in the still image. After receiving the first annotation, the device displays, in the video playback region, a second portion of the video that corresponds to a second position in the video, where the first portion of the physical environment is captured in the second portion of the video and the first annotation is displayed in the second portion of the video.
    Type: Application
    Filed: April 8, 2022
    Publication date: July 28, 2022
    Inventors: Joseph A. Malia, Mark K. Hauenstein, Praveen Sharma, Matan Stauber, Julian K. Missig, Jeffrey T. Bernstein, Lukas Robert Tom Girling, Matthaeus Krenn
  • Publication number: 20220147173
    Abstract: An electronic device, while displaying a first user interface, detects an input for an input object, detects that first hover proximity criteria are met by the input object, and displays first visual feedback. While displaying the first visual feedback, the device detects a change in a current value of a hover proximity parameter of the input object and that second hover proximity criteria are met by the input object after the change. In response to detecting that the second hover proximity criteria are met, the device displays second visual feedback, distinct from the first visual feedback.
    Type: Application
    Filed: January 19, 2022
    Publication date: May 12, 2022
    Inventors: Mark K. Hauenstein, Jeffrey T. Bernstein, Julian Missig, Marek A. Bereza
  • Publication number: 20220130118
    Abstract: A computer system displays a representation of a previously-captured media item, which includes or is associated with depth information corresponding to a physical environment in which the media item was captured. While displaying the representation of the media item, the system receives, via one or more input devices, one or more inputs corresponding to a request to display a representation of a measurement corresponding to a respective portion of the physical environment captured in the media item, and, in response, the system displays the representation of the measurement over at least a portion of the representation of the media item that corresponds to the respective portion of the physical environment, based on the depth information, and a label corresponding to the representation of the measurement that describes the measurement based on the depth information.
    Type: Application
    Filed: January 4, 2022
    Publication date: April 28, 2022
    Inventors: Joseph A. Malia, Praveen Sharma, Mark K. Hauenstein
  • Patent number: 11303812
    Abstract: A computer system displays a first previously captured media object including one or more first images, wherein the first previously captured media object was recorded and stored with first depth data corresponding to a first physical environment captured in each of the one or more first images. In response to a first user request to add a first virtual object to the first previously captured media object, the computer system displays the first virtual object over at least a portion of a respective image in the first previously captured media object, wherein the first virtual object is displayed with at least a first position or orientation that is determined using the first depth data that corresponds to the respective image in the first previously captured media object.
    Type: Grant
    Filed: August 19, 2020
    Date of Patent: April 12, 2022
    Assignee: APPLE INC.
    Inventors: Joseph A. Malia, Mark K. Hauenstein, Praveen Sharma, Matan Stauber, Julian K. Missig, Jeffrey T. Bernstein, Lukas Robert Tom Girling, Matthaeus Krenn
  • Patent number: 11243627
    Abstract: An electronic device with a touch-sensitive surface, a display, one or more first sensors to detect proximity of an input object above the touch-sensitive surface and one or more second sensors to detect intensity of contact displays a user interface object at a first location. While displaying the user interface object, the device detects a first portion of an input at the first location of the user interface object while the input object meets hover criteria. In response, the device dynamically changes an appearance of the user interface object in accordance with a current hover proximity parameter of the input object. Afterwards, the device detects a second portion of the input at an initial contact location that corresponds to the first location; and, in response, dynamically changes the appearance of the user interface object in accordance with a current intensity of a contact by the input object.
    Type: Grant
    Filed: August 13, 2020
    Date of Patent: February 8, 2022
    Assignee: APPLE INC.
    Inventors: Mark K. Hauenstein, Jeffrey T. Bernstein, Julian Missig, Marek A. Bereza
  • Publication number: 20220019334
    Abstract: A computer system while displaying an augmented reality environment, concurrently displays: a representation of at least a portion of a field of view of one or more cameras that includes a physical object, and a virtual user interface object at a location in the representation of the field of view, where the location is determined based on the respective physical object in the field of view. While displaying the augmented reality environment, in response to detecting an input that changes a virtual environment setting for the augmented reality environment, the computer system adjusts an appearance of the virtual user interface object in accordance with the change made to the virtual environment setting and applies to at least a portion of the representation of the field of view a filter selected based on the change made to the virtual environment setting.
    Type: Application
    Filed: September 28, 2021
    Publication date: January 20, 2022
    Inventors: Mark K. Hauenstein, Joseph A. Malia, Julian K. Missig, Matthaeus Krenn, Jeffrey T. Bernstein
  • Patent number: 11227446
    Abstract: A computer system captures, via one or more cameras, information indicative of the physical environment, including respective portions of the physical environment that are in a field of view. The respective portions of the physical environment include a plurality of primary features of the physical environment and secondary features of the physical environment. After capturing the information indicative of the physical environment, the system displays a user interface, including concurrently displaying graphical representations of the plurality of primary features that are generated with a first level of fidelity to the corresponding plurality of primary features of the physical environment, and one or more graphical representations of secondary features that are generated with a second level of fidelity to the corresponding one or more secondary features of the physical environment, where the second level of fidelity is lower than the first level of fidelity.
    Type: Grant
    Filed: September 23, 2020
    Date of Patent: January 18, 2022
    Assignee: APPLE INC.
    Inventors: Joseph A. Malia, Praveen Sharma, Mark K. Hauenstein
  • Patent number: 11163417
    Abstract: A computer system concurrently displays, in an augmented reality environment, a representation of at least a portion of a field of view of one or more cameras that includes a respective physical object, which is updated as contents of the field of view change; and a respective virtual user interface object, at a respective location in the virtual user interface determined based on the location of the respective physical object in the field of view. While detecting an input at a location that corresponds to the displayed respective virtual user interface object, in response to detecting movement of the input relative to the respective physical object in the field of view of the one or more cameras, the system adjusts an appearance of the respective virtual user interface object in accordance with a magnitude of movement of the input relative to the respective physical object.
    Type: Grant
    Filed: August 29, 2018
    Date of Patent: November 2, 2021
    Assignee: APPLE INC.
    Inventors: Mark K. Hauenstein, Joseph A. Malia, Julian K. Missig, Matthaeus Krenn, Jeffrey T. Bernstein