Patents by Inventor Ian M. Sands

Ian M. Sands has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 10409381
    Abstract: Aspects relate to detecting gestures that relate to a desired action, wherein the detected gestures are common across users and/or devices within a surface computing environment. Inferred intentions and goals based on context, history, affordances, and objects are employed to interpret gestures. Where there is uncertainty in intention of the gestures for a single device or across multiple devices, independent or coordinated communication of uncertainty or engagement of users through signaling and/or information gathering can occur.
    Type: Grant
    Filed: August 10, 2015
    Date of Patent: September 10, 2019
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Meredith June Morris, Eric J. Horvitz, Andrew David Wilson, F. David Jones, Stephen E. Hodges, Kenneth P. Hinckley, David Alexander Butler, Ian M. Sands, V. Kevin Russ, Hrvoje Benko, Shawn R. LeProwse, Shahram Izadi, William Ben Kunz
  • Patent number: 9207806
    Abstract: A virtual mouse input device is created in response to a placement of a card on a touch surface. When the card is placed on the touch surface, the boundaries of the card are captured and a virtual mouse appears around the card. The virtual mouse may be linked with a user through an identifier that is contained on the card. Other controls and actions may be presented in menus that appear with the virtual mouse. For instance, the user may select the type of input (e.g. mouse, keyboard, ink or trackball) driven by the business card. Once created, the virtual mouse is configured to receive user input until the card is removed from the touch surface. The virtual mouse is configured to move a cursor on a display in response to movement of the card on the touch surface.
    Type: Grant
    Filed: May 28, 2009
    Date of Patent: December 8, 2015
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Edwin Russ Burtner, V. Kevin Russ, Ian M. Sands, John A. Snavely
  • Publication number: 20150346837
    Abstract: Aspects relate to detecting gestures that relate to a desired action, wherein the detected gestures are common across users and/or devices within a surface computing environment. Inferred intentions and goals based on context, history, affordances, and objects are employed to interpret gestures. Where there is uncertainty in intention of the gestures for a single device or across multiple devices, independent or coordinated communication of uncertainty or engagement of users through signaling and/or information gathering can occur.
    Type: Application
    Filed: August 10, 2015
    Publication date: December 3, 2015
    Inventors: Meredith June Morris, Eric J. Horvitz, Andrew Daivd Wilson, F. David Jones, Stephen E. Hodges, Kenneth P. Hinckley, David Alexander Butler, Ian M. Sands, V. Kevin Russ, Hrvoje Benko, Shawn R. LeProwse, Shahram Izadi, William Ben Kunz
  • Patent number: 9141284
    Abstract: An input device is created on a touch screen in response to a user's placement of their hand. When a user places their hand on the touch screen, an input device sized for their hand is dynamically created. Alternatively, some other input device may be created. For example, when the user places two hands on the device a split keyboard input device may be dynamically created on the touch screen that is split between the user's hand locations. Once the input device is determined, the user may enter input through the created device on the input screen. The input devices may be configured for each individual user such that the display of the input device changes based on physical characteristics that are associated with the user.
    Type: Grant
    Filed: May 28, 2009
    Date of Patent: September 22, 2015
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Ian M. Sands, John A. Snavely, Edwin Russ Burtner, V. Kevin Russ
  • Patent number: 9134798
    Abstract: Aspects relate to detecting gestures that relate to a desired action, wherein the detected gestures are common across users and/or devices within a surface computing environment. Inferred intentions and goals based on context, history, affordances, and objects are employed to interpret gestures. Where there is uncertainty in intention of the gestures for a single device or across multiple devices, independent or coordinated communication of uncertainty or engagement of users through signaling and/or information gathering can occur.
    Type: Grant
    Filed: December 15, 2008
    Date of Patent: September 15, 2015
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Meredith June Morris, Eric J. Horvitz, Andrew David Wilson, F. David Jones, Stephen E. Hodges, Kenneth P. Hinckley, David Alexander Butler, Ian M. Sands, V. Kevin Russ, Hrvoje Benko, Shawn R. LeProwse, Shahram Izadi, William Ben Kunz
  • Patent number: 8903430
    Abstract: A user of a mobile device is able to display information about objects in the surrounding environment and to optionally interact with those objects. The information may be displayed as a graphical overlay on top of a real-time display of imagery from a camera in the mobile device with the overlay indexed to the real-time display. The graphical overlay may include positional information about an external object and may include navigational information intended to assist the user in moving to the object's location. There may also be a graphical user interface which allows the user to utilize the mobile device to interact with an external object.
    Type: Grant
    Filed: February 21, 2008
    Date of Patent: December 2, 2014
    Assignee: Microsoft Corporation
    Inventors: Ian M. Sands, V. Kevin Russ
  • Publication number: 20140344706
    Abstract: A dual module portable device may be provided. A motion of a first module of the dual module portable device may be detected. Based at least in part on the detected motion, a position of the first module may be determined relative to the second module of the portable device. Once the relative position of the first module has been determined, a portion of a user interface associated with the relative position may be displayed at the first module.
    Type: Application
    Filed: August 5, 2014
    Publication date: November 20, 2014
    Inventors: V. Kevin Russ, John A. Snavely, Edwin R. Burtner, Ian M. Sands
  • Patent number: 8849570
    Abstract: Navigation information may be provided. First, a destination location may be received at a portable device. Next, a current location of the portable device maybe detected. Then, at least one way-point may be calculated based on the current location and the destination location. An orientation and a level of the portable device may be determined and the at least one way-point may then be projected from the portable device.
    Type: Grant
    Filed: March 19, 2009
    Date of Patent: September 30, 2014
    Assignee: Microsoft Corporation
    Inventors: V. Kevin Russ, John A. Snavely, Edwin R. Burtner, Ian M. Sands
  • Patent number: 8798669
    Abstract: A dual module portable device may be provided. A motion of a first module of the dual module portable device may be detected. Based at least in part on the detected motion, a position of the first module may be determined relative to the second module of the portable device. Once the relative position of the first module has been determined, a portion of a user interface associated with the relative position may be displayed at the first module.
    Type: Grant
    Filed: February 15, 2012
    Date of Patent: August 5, 2014
    Assignee: Microsoft Corporation
    Inventors: V. Kevin Russ, John A. Snavely, Edwin R. Burtner, Ian M. Sands
  • Patent number: 8681127
    Abstract: In some implementations, a touch point on a surface of a touchscreen device may be determined. An image of a region of space above the surface and surrounding the touch point may be determined. The image may include a brightness gradient that captures a brightness of objects above the surface. A binary image that includes one or more binary blobs may be created based on a brightness of portions of the image. A determination may be made as to which of the one more binary blobs are connected to each other to form portions of a particular user. A determination may be made that the particular user generated the touch point.
    Type: Grant
    Filed: April 22, 2013
    Date of Patent: March 25, 2014
    Assignee: Microsoft Corporation
    Inventors: Stephen E. Hodges, Hrvoje Benko, Ian M. Sands, David Alexander Butler, Shahram Izadi, William Ben Kunz, Kenneth P. Hinckley
  • Publication number: 20130234992
    Abstract: In some implementations, a touch point on a surface of a touchscreen device may be determined. An image of a region of space above the surface and surrounding the touch point may be determined The image may include a brightness gradient that captures a brightness of objects above the surface. A binary image that includes one or more binary blobs may be created based on a brightness of portions of the image. A determination may be made as to which of the one more binary blobs are connected to each other to form portions of a particular user. A determination may be made that the particular user generated the touch point.
    Type: Application
    Filed: April 22, 2013
    Publication date: September 12, 2013
    Applicant: Microsoft Corporation
    Inventors: Stephen E. Hodges, Hrvoje Benko, Ian M. Sands, David Alexander Butler, Shahram Izadi, William Ben Kunz, Kenneth P. Hinckley
  • Patent number: 8490157
    Abstract: Within a surface computing environment users are provided a seamless and intuitive manner of modifying security levels associated with information. If a modification is to be made the user can perceive the modifications and the result of such modifications, such as on a display. When information is rendered within the surface computing environment and a condition changes, the user can quickly have that information concealed in order to mitigate unauthorized access to the information.
    Type: Grant
    Filed: February 26, 2009
    Date of Patent: July 16, 2013
    Assignee: Microsoft Corporation
    Inventors: Peter B. Thompson, Ian M. Sands, Ali M. Vassigh, Eric I-Chao Chang
  • Patent number: 8432366
    Abstract: The claimed subject matter provides a system and/or a method that facilitates distinguishing input among one or more users in a surface computing environment. A variety of information can be obtained and analyzed to infer an association between a particular input and a particular user. Touch point information can be acquired from a surface wherein the touch point information relates to a touch point. In addition, one or more environmental sensors can monitor the surface computing environment and provide environmental information. The touch point information and the environmental information can be analyzed to determine direction of inputs, location of users, and movement of users and so on. Individual analysis results can be correlated and/or aggregated to generate a inference of association between a touch point and user.
    Type: Grant
    Filed: March 3, 2009
    Date of Patent: April 30, 2013
    Assignee: Microsoft Corporation
    Inventors: Stephen E. Hodges, Hrvoje Benko, Ian M. Sands, David Alexander Butler, Shahram Izadi, William Ben Kunz, Kenneth P. Hinckley
  • Patent number: 8386963
    Abstract: A virtual inking device is created in response to a touch input device detecting a user's inking gesture. For example, when a user places one of their hands in a pen gesture (i.e. by connecting the index finger with the thumb while holding the other fingers near the palm), the user may perform inking operations. When the user changes the pen gesture to an erase gesture (i.e. making a fist) then the virtual pen may become a virtual eraser. Other inking gestures may also be utilized.
    Type: Grant
    Filed: May 28, 2009
    Date of Patent: February 26, 2013
    Assignee: Microsoft Corporation
    Inventors: V. Kevin Russ, Ian M. Sands, John A. Snavely, Edwin Russ Burtner
  • Publication number: 20120139939
    Abstract: A dual module portable device may be provided. A motion of a first module of the dual module portable device may be detected. Based at least in part on the detected motion, a position of the first module may be determined relative to the second module of the portable device. Once the relative position of the first module has been determined, a portion of a user interface associated with the relative position may be displayed at the first module.
    Type: Application
    Filed: February 15, 2012
    Publication date: June 7, 2012
    Applicant: Microsoft Corporation
    Inventors: V. Kevin Russ, John A. Snavely, Edwin R. Burtner, Ian M. Sands
  • Patent number: 8121640
    Abstract: A dual module portable device may be provided. A motion of a first module of the dual module portable device may be detected. Based at least in part on the detected motion, a position of the first module may be determined relative to the second module of the portable device. Once the relative position of the first module has been determined, a portion of a user interface associated with the relative position may be displayed at the first module.
    Type: Grant
    Filed: March 19, 2009
    Date of Patent: February 21, 2012
    Assignee: Microsoft Corporation
    Inventors: V. Kevin Russ, John A. Snavely, Edwin R. Burtner, Ian M. Sands
  • Publication number: 20100313150
    Abstract: The claimed subject matter relates to a display that is physically separable and to an associated architecture that can facilitate data mobility or collaboration in connection with the separable display. In particular, the separable display can be configured as an apparent unitary or singular UI for an associated multi-node computer, yet for which portion of the separable display can be physically decoupled. The multi-node computer can include a set of computing nodes, each of which can potentially operate autonomously, yet also in unison with other nodes to form a collective multiprocessor computing platform. Moreover, each of the computing nodes can be embedded in and distributed throughout the separable display. Accordingly, when a portion of the separable display is decoupled from a remainder of the separable display, both the portion and the remainder can include some subset of the computing nodes, and can therefore maintain the UI.
    Type: Application
    Filed: June 3, 2009
    Publication date: December 9, 2010
    Applicant: MICROSOFT CORPORATION
    Inventors: Meredith J. Morris, Steven N. Bathiche, Stephen Edward Hodges, Ian C. LeGrow, Victor Kevin Russ, Ian M. Sands, William J. Westerinen, John Christopher Whytock, Andrew D. Wilson, David Alexander Butler, Shahram Izadi
  • Publication number: 20100302155
    Abstract: An input device is created on a touch screen in response to a user's placement of their hand. When a user places their hand on the touch screen, an input device sized for their hand is dynamically created. Alternatively, some other input device may be created. For example, when the user places two hands on the device a split keyboard input device may be dynamically created on the touch screen that is split between the user's hand locations. Once the input device is determined, the user may enter input through the created device on the input screen. The input devices may be configured for each individual user such that the display of the input device changes based on physical characteristics that are associated with the user.
    Type: Application
    Filed: May 28, 2009
    Publication date: December 2, 2010
    Applicant: MICROSOFT CORPORATION
    Inventors: Ian M. Sands, John A. Snavely, Edwin Russ Burtner, V. Kevin Russ
  • Publication number: 20100306649
    Abstract: A virtual inking device is created in response to a touch input device detecting a user's inking gesture. For example, when a user places one of their hands in a pen gesture (i.e. by connecting the index finger with the thumb while holding the other fingers near the palm), the user may perform inking operations. When the user changes the pen gesture to an erase gesture (i.e. making a first) then the virtual pen may become a virtual eraser. Other inking gestures may also be utilized.
    Type: Application
    Filed: May 28, 2009
    Publication date: December 2, 2010
    Applicant: MICROSOFT CORPORATION
    Inventors: V. Kevin Russ, Ian M. Sands, John A. Snavely, Edwin Russ Burtner
  • Publication number: 20100302144
    Abstract: A virtual mouse input device is created in response to a placement of a card on a touch surface. When the card is placed on the touch surface, the boundaries of the card are captured and a virtual mouse appears around the card. The virtual mouse may be linked with a user through an identifier that is contained on the card. Other controls and actions may be presented in menus that appear with the virtual mouse. For instance, the user may select the type of input (e.g. mouse, keyboard, ink or trackball) driven by the business card. Once created, the virtual mouse is configured to receive user input until the card is removed from the touch surface. The virtual mouse is configured to move a cursor on a display in response to movement of the card on the touch surface.
    Type: Application
    Filed: May 28, 2009
    Publication date: December 2, 2010
    Applicant: MICROSOFT CORPORATION
    Inventors: Edwin Russ Burtner, V. Kevin Russ, Ian M. Sands, John A. Snavely