Patents by Inventor Mark J. Finocchio

Mark J. Finocchio has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 8866821
    Abstract: Techniques for efficiently tracking points on a depth map using an optical flow are disclosed. In order to optimize the use of optical flow, isolated regions of the depth map may be tracked. The sampling regions may comprise a 3-dimensional box (width, height and depth). Each region may be “colored” as a function of depth information to generate a “zebra” pattern as a function of depth data for each sample. The disclosed techniques may provide for handling optical flow tracking when occlusion occurs by utilizing a weighting process for application of optical flow vs. velocity prediction to stabilize tracking.
    Type: Grant
    Filed: January 30, 2009
    Date of Patent: October 21, 2014
    Assignee: Microsoft Corporation
    Inventor: Mark J. Finocchio
  • Publication number: 20140228123
    Abstract: A system recognizes human beings in their natural environment, without special sensing devices attached to the subjects, uniquely identifies them and tracks them in three dimensional space. The resulting representation is presented directly to applications as a multi-point skeletal model delivered in real-time. The device efficiently tracks humans and their natural movements by understanding the natural mechanics and capabilities of the human muscular-skeletal system. The device also uniquely recognizes individuals in order to allow multiple people to interact with the system via natural movements of their limbs and body as well as voice commands/responses.
    Type: Application
    Filed: April 14, 2014
    Publication date: August 14, 2014
    Applicant: Microsoft Corporation
    Inventors: R. Stephen Polzin, Alex A. Kipman, Mark J. Finocchio, Ryan Michael Geiss, Kathryn Stone Perez, Kudo Tsunoda, Darren Alexander Bennett
  • Patent number: 8762894
    Abstract: Techniques for managing virtual ports are disclosed herein. Each such virtual port may have different associated features such as, for example, privileges, rights or options. When one or more users are in a capture scene of a gesture based system, the system may associate virtual ports with the users and maintain the virtual ports. Also provided are techniques for disassociating virtual ports with users or swapping virtual ports between two or more users.
    Type: Grant
    Filed: February 10, 2012
    Date of Patent: June 24, 2014
    Assignee: Microsoft Corporation
    Inventors: Kathryn Stone-Perez, Jeffrey Margolis, Mark J. Finocchio, Brian E. Keane, Rudy Jacobus Poot, Stephen G. Latta
  • Patent number: 8744121
    Abstract: A system recognizes human beings in their natural environment, without special sensing devices attached to the subjects, uniquely identifies them and tracks them in three dimensional space. The resulting representation is presented directly to applications as a multi-point skeletal model delivered in real-time. The device efficiently tracks humans and their natural movements by understanding the natural mechanics and capabilities of the human muscular-skeletal system. The device also uniquely recognizes individuals in order to allow multiple people to interact with the system via natural movements of their limbs and body as well as voice commands/responses.
    Type: Grant
    Filed: May 29, 2009
    Date of Patent: June 3, 2014
    Assignee: Microsoft Corporation
    Inventors: R. Stephen Polzin, Alex A. Kipman, Mark J. Finocchio, Ryan Michael Geiss, Kathryn Stone Perez, Kudo Tsunoda, Darren Alexander Bennett
  • Publication number: 20140085193
    Abstract: A media feed interface may be provided that may be used to extract a media frame from a media feed. The media feed interface may access a capture device, a file, and/or a network resource. Upon accessing the capture device, file, and/or network resource, the media feed interface may populate buffers with data and then may create a media feed from the buffers. Upon request, the media feed interface may isolate a media frame within the media feed. For example, the media feed interface analyze media frames in the media feed to determine whether a media frame includes information associated with, for example, the request. If the media frame includes the requested information, the media feed interface may isolate the media frame associated with the information and may provide access to the isolated media frame.
    Type: Application
    Filed: November 27, 2013
    Publication date: March 27, 2014
    Inventors: Mark J. Finocchio, Jeffrey Margolis
  • Patent number: 8625837
    Abstract: A media feed interface may be provided that may be used to extract a media frame from a media feed. The media feed interface may access a capture device, a file, and/or a network resource. Upon accessing the capture device, file, and/or network resource, the media feed interface may populate buffers with data and then may create a media feed from the buffers. Upon request, the media feed interface may isolate a media frame within the media feed. For example, the media feed interface analyze media frames in the media feed to determine whether a media frame includes information associated with, for example, the request. If the media frame includes the requested information, the media feed interface may isolate the media frame associated with the information and may provide access to the isolated media frame.
    Type: Grant
    Filed: June 16, 2009
    Date of Patent: January 7, 2014
    Assignee: Microsoft Corporation
    Inventors: Mark J. Finocchio, Jeffrey Margolis
  • Publication number: 20140002607
    Abstract: Technology is described for determining and using invariant features for computer vision. A local orientation may be determined for each depth pixel in a subset of the depth pixels in a depth map. The local orientation may an in-plane orientation, an out-out-plane orientation or both. A local coordinate system is determined for each of the depth pixels in the subset based on the local orientation of the corresponding depth pixel. A feature region is defined relative to the local coordinate system for each of the depth pixels in the subset. The feature region for each of the depth pixels in the subset is transformed from the local coordinate system to an image coordinate system of the depth map. The transformed feature regions are used to process the depth map.
    Type: Application
    Filed: November 28, 2012
    Publication date: January 2, 2014
    Inventors: Jamie D.J. Shotton, Mark J. Finocchio, Richard E. Moore, Alexandru O. Balan, Kyungsuk David Lee
  • Publication number: 20130286004
    Abstract: Technology is described for displaying a collision between objects by an augmented reality display device system. A collision between a real object and a virtual object is identified based on three dimensional space position data of the objects. At least one effect on at least one physical property of the real object is determined based on physical properties of the real object, like a change in surface shape, and physical interaction characteristics of the collision. Simulation image data is generated and displayed simulating the effect on the real object by the augmented reality display. Virtual objects under control of different executing applications can also interact with one another in collisions.
    Type: Application
    Filed: April 27, 2012
    Publication date: October 31, 2013
    Inventors: Daniel J. McCulloch, Stephen G. Latta, Brian J. Mount, Kevin A. Geisner, Roger Sebastian Kevin Sylvan, Arnulfo Zepeda Navratil, Jason Scott, Jonathan T. Steed, Ben J. Sugden, Britta Silke Hummel, Kyungsuk David Lee, Mark J. Finocchio, Alex Aben-Athar Kipman, Jeffrey N. Margolis
  • Publication number: 20120314031
    Abstract: Technology is described for determining and using invariant features for computer vision. A local orientation may be determined for each depth pixel in a subset of the depth pixels in a depth map. The local orientation may an in-plane orientation, an out-out-plane orientation or both. A local coordinate system is determined for each of the depth pixels in the subset based on the local orientation of the corresponding depth pixel. A feature region is defined relative to the local coordinate system for each of the depth pixels in the subset. The feature region for each of the depth pixels in the subset is transformed from the local coordinate system to an image coordinate system of the depth map. The transformed feature regions are used to process the depth map.
    Type: Application
    Filed: June 7, 2011
    Publication date: December 13, 2012
    Applicant: MICROSOFT CORPORATION
    Inventors: Jamie D. J. Shotton, Mark J. Finocchio, Richard E. Moore, Alexandru O. Balan, Kyungsuk David Lee
  • Publication number: 20120280897
    Abstract: Attribute state classification techniques are described. In one or more implementations, one or more pixels of an image are classified by a computing device as having one or several states for one or more attributes that do not identify corresponding body parts of a user. A gesture is recognized by the computing device that is operable to initiate one or more operations of the computing device based at least in part of the state classifications of the one or more pixels of one or more attributes.
    Type: Application
    Filed: May 2, 2011
    Publication date: November 8, 2012
    Applicant: MICROSOFT CORPORATION
    Inventors: Alexandru O. Balan, Richard E. Moore, Mark J. Finocchio
  • Patent number: 8281287
    Abstract: A non-tree representation of a UI control tree is provided by a compact UI binary file that is generated by encoding a UI definition markup file expressing UI controls and behavior in human-readable form. The UI binary file is utilized in a runtime environment on a computing device as a source of a binary instruction stream. The stream can be efficiently processed by an interpreter on the device without needing validation between loading and rendering the UI. The encoding places much of the representation into an object and script section of the UI binary file. The interpreter runs this section without it being entirely resident in the device's memory to minimize the memory footprint. At runtime, operation code (“op-code”) streams contained in this section are used to build UI objects, and implement scriptable behavior for manipulating the UI objects to render the UI on the device with the designed behavior.
    Type: Grant
    Filed: February 26, 2008
    Date of Patent: October 2, 2012
    Inventor: Mark J. Finocchio
  • Publication number: 20120206452
    Abstract: Technology is described for providing realistic occlusion between a virtual object displayed by a head mounted, augmented reality display system and a real object visible to the user's eyes through the display. A spatial occlusion in a user field of view of the display is typically a three dimensional occlusion determined based on a three dimensional space mapping of real and virtual objects. An occlusion interface between a real object and a virtual object can be modeled at a level of detail determined based on criteria such as distance within the field of view, display size or position with respect to a point of gaze. Technology is also described for providing three dimensional audio occlusion based on an occlusion between a real object and a virtual object in the user environment.
    Type: Application
    Filed: April 10, 2012
    Publication date: August 16, 2012
    Inventors: Kevin A. Geisner, Brian J. Mount, Stephen G. Latta, Daniel J. McCulloch, Kyungsuk David Lee, Ben J. Sugden, Jeffrey N. Margolis, Kathryn Stone Perez, Sheridan Martin Small, Mark J. Finocchio, Robert L. Crocco, JR.
  • Publication number: 20120144348
    Abstract: Techniques for managing virtual ports are disclosed herein. Each such virtual port may have different associated features such as, for example, privileges, rights or options. When one or more users are in a capture scene of a gesture based system, the system may associate virtual ports with the users and maintain the virtual ports. Also provided are techniques for disassociating virtual ports with users or swapping virtual ports between two or more users.
    Type: Application
    Filed: February 10, 2012
    Publication date: June 7, 2012
    Applicant: MICROSOFT CORPORATION
    Inventors: Kathryn Stone-Perez, Jeffrey Margolis, Mark J. Finocchio, Brian E. Keane, Rudy Jacobus Poot, Stephen G. Latta
  • Patent number: 8181123
    Abstract: Techniques for managing virtual ports are disclosed herein. Each such virtual port may have different associated features such as, for example, privileges, rights or options. When one or more users are in a capture scene of a gesture based system, the system may associate virtual ports with the users and maintain the virtual ports. Also provided are techniques for disassociating virtual ports with users or swapping virtual ports between two or more users.
    Type: Grant
    Filed: May 1, 2009
    Date of Patent: May 15, 2012
    Assignee: Microsoft Corporation
    Inventors: Kathryn Stone-Perez, Jeffrey Margolis, Mark J. Finocchio, Brian E. Keane, Rudy Jacobus Poot, Stephen G. Latta
  • Patent number: 8133119
    Abstract: Mechanisms for adjusting signals between gaming controllers and gaming consoles are disclosed. In an embodiment, the output signals of a mouse control a gaming console which is normally controlled by an analog thumbstick. The output signals of mouse are adjusted to compensate for the analog thumbstick controller assist techniques employed by the gaming console. The adjusted signals are sent to the gaming console. The result is that the user is able to control the game using the mouse and have the same feel as if the user was using the analog thumbstick controller.
    Type: Grant
    Filed: January 30, 2009
    Date of Patent: March 13, 2012
    Assignee: Microsoft Corporation
    Inventor: Mark J. Finocchio
  • Publication number: 20100304813
    Abstract: A media feed interface may be provided that may be used to extract a media frame from a media feed. The media feed interface may access a capture device, a file, and/or a network resource. Upon accessing the capture device, file, and/or network resource, the media feed interface may populate buffers with data and then may create a media feed from the buffers. Upon request, the media feed interface may isolate a media frame within the media feed. For example, the media feed interface analyze media frames in the media feed to determine whether a media frame includes information associated with, for example, the request. If the media frame includes the requested information, the media feed interface may isolate the media frame associated with the information and may provide access to the isolated media frame.
    Type: Application
    Filed: June 16, 2009
    Publication date: December 2, 2010
    Applicant: Microsoft Corporation
    Inventors: Mark J. Finocchio, Jeffrey Margolis
  • Publication number: 20100303289
    Abstract: A system recognizes human beings in their natural environment, without special sensing devices attached to the subjects, uniquely identifies them and tracks them in three dimensional space. The resulting representation is presented directly to applications as a multi-point skeletal model delivered in real-time. The device efficiently tracks humans and their natural movements by understanding the natural mechanics and capabilities of the human muscular-skeletal system. The device also uniquely recognizes individuals in order to allow multiple people to interact with the system via natural movements of their limbs and body as well as voice commands/responses.
    Type: Application
    Filed: May 29, 2009
    Publication date: December 2, 2010
    Applicant: MICROSOFT CORPORATION
    Inventors: R. Stephen Polzin, Alex A. Kipman, Mark J. Finocchio, Ryan Michael Geiss, Kathryn Stone Perez, Kudo Tsunoda, Darren Alexander Bennett
  • Publication number: 20100303302
    Abstract: A depth image of a scene may be received, observed, or captured by a device. The depth image may include a human target that may have, for example, a portion thereof non-visible or occluded. For example, a user may be turned such that a body part may not be visible to the device, may have one or more body parts partially outside a field of view of the device, may have a body part or a portion of a body part behind another body part or object, or the like such that the human target associated with the user may also have a portion body part or a body part non-visible or occluded in the depth image. A position or location of the non-visible or occluded portion or body part of the human target associated with the user may then be estimated.
    Type: Application
    Filed: June 26, 2009
    Publication date: December 2, 2010
    Applicant: Microsoft Corporation
    Inventors: Alex A. Kipman, Kathryn Stone Perez, Mark J. Finocchio, Ryan Michael Geiss, Kudo Tsunoda
  • Publication number: 20100302365
    Abstract: A depth image of a scene may be received, observed, or captured by a device. The depth image may then be analyzed to determine whether the depth image includes noise. For example, the depth image may include one or more holes having one or more empty pixels or pixels without a depth value. Depth values for the one or more empty pixels may be estimated and a depth image that includes the estimated depth values for the one or empty more pixels may be rendered.
    Type: Application
    Filed: May 29, 2009
    Publication date: December 2, 2010
    Applicant: Microsoft Corporation
    Inventors: Mark J. Finocchio, Ryan Michael Geiss
  • Publication number: 20100281437
    Abstract: Techniques for managing virtual ports are disclosed herein. Each such virtual port may have different associated features such as, for example, privileges, rights or options. When one or more users are in a capture scene of a gesture based system, the system may associate virtual ports with the users and maintain the virtual ports. Also provided are techniques for disassociating virtual ports with users or swapping virtual ports between two or more users.
    Type: Application
    Filed: May 1, 2009
    Publication date: November 4, 2010
    Applicant: Microsoft Corporation
    Inventors: Kathryn Stone-Perez, Jeffrey Margolis, Mark J. Finocchio, Brian E. Keane, Rudy Jacobus Poot, Stephen G. Latta