Patents by Inventor Paul Armistead Hoover

Paul Armistead Hoover has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 8261212
    Abstract: A computing system for displaying a GUI element on a natural user interface is described herein. The computing system includes a display configured to display a natural user interface of a program executed on the computing system, and a gesture sensor configured to detect a gesture input directed at the natural user interface by a user. The computing system also includes a processor configured to execute a gesture-recognizing module for recognizing a registration phase, an operation phase, and a termination phase of the gesture input, and a gesture assist module configured to first display a GUI element overlaid upon the natural user interface in response to recognition of the registration phase. The GUI element includes a visual or audio operation cue to prompt the user to carry out the operation phase of the gesture input, and a selector manipulatable by the user via the operation phase of the gesture.
    Type: Grant
    Filed: October 20, 2009
    Date of Patent: September 4, 2012
    Assignee: Microsoft Corporation
    Inventors: Daniel J. Wigdor, Paul Armistead Hoover
  • Publication number: 20120133680
    Abstract: User input interpretation mechanisms can pass along user input without delay by initially treating it as freeform movement and passing it along unmodified, rather than waiting to receive a sufficient quantity of user input from which to derive a user's intentions. User input within a defined range from an initial point is freeform movement, passed along unmodified, while user input beyond that range can be confined to a groove if it evidences an intention to move in a precisely linear manner. Grooves can be established to aid in precise vertical movement, with no horizontal deviations, precise horizontal movement, with no vertical deviations, or any other orientation. Before being passed along, user input in groove regions can be modified to center the input within the groove region and eliminate human inaccuracies, such as slight deviations to either side. User input can escape groove regions by moving orthogonally to them.
    Type: Application
    Filed: November 29, 2010
    Publication date: May 31, 2012
    Applicant: MICROSOFT CORPORATION
    Inventors: Paul Armistead Hoover, Michael Patten, Maxim V. Mazeev
  • Publication number: 20120092381
    Abstract: An invention is disclosed for using touch gestures to zoom a video to full-screen. As the user reverse-pinches on a touch-sensitive surface to zoom in on a video, the invention tracks the amount of a zoom. When the user has zoomed to the point where one of the dimensions (height or width) of the video reaches a threshold (such as some percentage of a dimension of the display device—e.g. the width of the video reaches 80% of the width of the display device), the invention determines to display the video in full-screen, and “snaps” the video to full-screen. The invention may do this by way of an animation, such as expanding the video to fill the screen.
    Type: Application
    Filed: October 19, 2010
    Publication date: April 19, 2012
    Applicant: Microsoft Corporation
    Inventors: Paul Armistead Hoover, Vishnu Sivaji, Jarrod Lombardo, Daniel John Wigdor
  • Publication number: 20110157025
    Abstract: A method of controlling a virtual object within a virtual workspace includes recognizing a hand posture of an initial touch gesture directed to a touch-input receptor, and a mode constraint is set based on the hand posture. The mode constraint specifies a constrained parameter of a virtual object that is to be maintained responsive to a subsequent touch gesture. The method further includes recognizing a subsequent touch gesture directed to the touch-input receptor. An unconstrained parameter of the virtual object is modulated responsive to the subsequent touch gesture while the constrained parameter of the virtual object is maintained in accordance with the mode constraint.
    Type: Application
    Filed: December 30, 2009
    Publication date: June 30, 2011
    Inventors: Paul Armistead Hoover, Maxim Oustiogov, Daniel J. Wigdor, Hrvoje Benko, Jarrod Lombardo
  • Publication number: 20110134047
    Abstract: Embodiments are disclosed herein that relate to multi-modal interaction on a computing device comprising a multi-touch display. One disclosed embodiment comprises a method of multi-modal interaction including recognizing a hand posture of a user's first hand directed at the display and displaying a modal region based on the hand posture, wherein the modal region defines an area on the display. The method further includes receiving an input selecting a mode to be applied to the modal region, wherein the mode indicates functionalities to be associated with the modal region and defines a mapping of touch gestures to actions associated with the mode. The method further includes, while the modal region remains displayed, recognizing a touch gesture from a user's second hand directed at the display within the modal region and performing an action on the display based upon a mapping of the touch gesture.
    Type: Application
    Filed: December 4, 2009
    Publication date: June 9, 2011
    Applicant: MICROSOFT CORPORATION
    Inventors: Daniel J. Wigdor, Paul Armistead Hoover, Kay Hofmeester
  • Publication number: 20110115745
    Abstract: An interactive display system with a contact geometry interface is disclosed. The interactive display system may include a multi-touch display, a touch detection system configured to detect a touch input on the multi-touch display and to generate contact geometry for a contact region of the touch input, and an application programming interface executed on a processor of the interactive display system. The application programming interface may be configured to receive the contact geometry and to send the contact geometry to a requesting application program for application-level processing. Further, the application programming interface may be configured to receive from the application program a display command based on the application level-processing. The application programming interface may be configured to send the display command to the multi-touch display to adjust a display of a graphical element on the multi-touch display.
    Type: Application
    Filed: November 13, 2009
    Publication date: May 19, 2011
    Applicant: MICROSOFT CORPORATION
    Inventors: Luis Eduardo Cabrera Cordon, Robert Levy, Sundaram Ramani, Daniel Wigdor, Joyce Wu, Ian Middleton, Paul Armistead Hoover, Sarat Subramaniam, Carlos Pessoa
  • Publication number: 20110093821
    Abstract: A computing system for displaying a GUI element on a natural user interface is described herein. The computing system includes a display configured to display a natural user interface of a program executed on the computing system, and a gesture sensor configured to detect a gesture input directed at the natural user interface by a user. The computing system also includes a processor configured to execute a gesture-recognizing module for recognizing a registration phase, an operation phase, and a termination phase of the gesture input, and a gesture assist module configured to first display a GUI element overlaid upon the natural user interface in response to recognition of the registration phase. The GUI element includes a visual or audio operation cue to prompt the user to carry out the operation phase of the gesture input, and a selector manipulatable by the user via the operation phase of the gesture.
    Type: Application
    Filed: October 20, 2009
    Publication date: April 21, 2011
    Applicant: MICROSOFT CORPORATION
    Inventors: Daniel J. Wigdor, Paul Armistead Hoover
  • Publication number: 20100218249
    Abstract: The claimed subject matter provides a system and/or a method that facilitates authentication of a user in a surface computing environment. A device or authentication object can be carried by a user and employed to retain authentication information. An authentication component can obtain the authentication information from the device and analyze the information to verify an identity of the user. A touch input component can ascertain if a touch input is authentication by associating touch input with the user. In addition, authentication information can be employed to establish a secure communications channel for transfer of user data.
    Type: Application
    Filed: February 25, 2009
    Publication date: August 26, 2010
    Applicant: MICROSOFT CORPORATION
    Inventors: Andrew D. Wilson, Stephen E. Hodges, Peter B. Thompson, Meredith June Morris, Paul Armistead Hoover, William J. Westerinen, Steven N. Bathiche, Ian M. Sands, Shahram Izadi, David Alexander Butler, Matthew B. MacLaurin, Arthur T. Whitten, William Ben Kunz, Shawn R. LeProwse, Hrvoje Benko
  • Publication number: 20100177931
    Abstract: Various embodiments related to the location and adjustment of a virtual object on a display in response to a detected physical object are disclosed. One disclosed embodiment provides a computing device comprising a multi-touch display, a processor and memory comprising instructions executable by the processor to display on the display a virtual object, to detect a change in relative location between the virtual object and a physical object that constrains a viewable area of the display, and to adjust a location of the virtual object on the display in response to detecting the change in relative location between the virtual object and the physical object.
    Type: Application
    Filed: January 15, 2009
    Publication date: July 15, 2010
    Applicant: MICROSOFT CORPORATION
    Inventors: Chris Whytock, Carlos Pessoa, Paul Armistead Hoover
  • Publication number: 20100141589
    Abstract: A method of operating a touch display includes interpreting a touch input on the touch display as a first kind of gesture if a source of the touch input is of a first type and a parameter of the touch input is below a first threshold. The touch input is interpreted as a second kind of gesture if the source is of the first type and the parameter of the touch input is above the first threshold. The touch input is interpreted as the first kind of gesture if the source is of a second type and the parameter of the touch input is below a second threshold, the second threshold being different than the first threshold. The touch input is interpreted as the second kind of gesture if the source is of the second type and the parameter of the touch input is above the second threshold.
    Type: Application
    Filed: December 9, 2008
    Publication date: June 10, 2010
    Applicant: Microsoft Corporation
    Inventor: Paul Armistead Hoover
  • Publication number: 20100146387
    Abstract: A scroll control recognizes a touch input directed to a selectable item while the selectable item is scrolling on a touch display. The scroll control determines if the selectable item is scrolling above a threshold velocity when the touch input directed to the selectable item is recognized. If the selectable item is determined to be scrolling above the threshold velocity, scrolling of the selectable item is stopped. If the selectable item is determined to be scrolling below the threshold velocity, the selectable item is selected.
    Type: Application
    Filed: December 5, 2008
    Publication date: June 10, 2010
    Applicant: Microsoft Corporation
    Inventor: Paul Armistead Hoover
  • Publication number: 20090327886
    Abstract: An interactive media display system and a method of activating a graphical user interface element presented by the interactive media display system are provided. The method includes presenting the graphical user interface element via a touch-sensitive display surface of the interactive media display system; receiving a user input at the touch-sensitive display surface; determining whether one or more secondary factors associated with the user input indicate an intentional contact with the graphical user interface element that is presented via the touch sensitive display surface; activating the graphical user interface element if the one or more secondary factors indicate the intentional contact with the graphical user interface element; and disregarding the user input by not activating the graphical user interface if the one or more secondary factors do not indicate the intentional contact.
    Type: Application
    Filed: October 17, 2008
    Publication date: December 31, 2009
    Applicant: MICROSOFT CORPORATION
    Inventors: Chris Whytock, Peter Vale, Steven Seow, Carlos Pessoa, Paul Armistead Hoover