Patents by Inventor Mark J. Finocchio
Mark J. Finocchio has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 8866821Abstract: Techniques for efficiently tracking points on a depth map using an optical flow are disclosed. In order to optimize the use of optical flow, isolated regions of the depth map may be tracked. The sampling regions may comprise a 3-dimensional box (width, height and depth). Each region may be “colored” as a function of depth information to generate a “zebra” pattern as a function of depth data for each sample. The disclosed techniques may provide for handling optical flow tracking when occlusion occurs by utilizing a weighting process for application of optical flow vs. velocity prediction to stabilize tracking.Type: GrantFiled: January 30, 2009Date of Patent: October 21, 2014Assignee: Microsoft CorporationInventor: Mark J. Finocchio
-
Publication number: 20140228123Abstract: A system recognizes human beings in their natural environment, without special sensing devices attached to the subjects, uniquely identifies them and tracks them in three dimensional space. The resulting representation is presented directly to applications as a multi-point skeletal model delivered in real-time. The device efficiently tracks humans and their natural movements by understanding the natural mechanics and capabilities of the human muscular-skeletal system. The device also uniquely recognizes individuals in order to allow multiple people to interact with the system via natural movements of their limbs and body as well as voice commands/responses.Type: ApplicationFiled: April 14, 2014Publication date: August 14, 2014Applicant: Microsoft CorporationInventors: R. Stephen Polzin, Alex A. Kipman, Mark J. Finocchio, Ryan Michael Geiss, Kathryn Stone Perez, Kudo Tsunoda, Darren Alexander Bennett
-
Patent number: 8762894Abstract: Techniques for managing virtual ports are disclosed herein. Each such virtual port may have different associated features such as, for example, privileges, rights or options. When one or more users are in a capture scene of a gesture based system, the system may associate virtual ports with the users and maintain the virtual ports. Also provided are techniques for disassociating virtual ports with users or swapping virtual ports between two or more users.Type: GrantFiled: February 10, 2012Date of Patent: June 24, 2014Assignee: Microsoft CorporationInventors: Kathryn Stone-Perez, Jeffrey Margolis, Mark J. Finocchio, Brian E. Keane, Rudy Jacobus Poot, Stephen G. Latta
-
Patent number: 8744121Abstract: A system recognizes human beings in their natural environment, without special sensing devices attached to the subjects, uniquely identifies them and tracks them in three dimensional space. The resulting representation is presented directly to applications as a multi-point skeletal model delivered in real-time. The device efficiently tracks humans and their natural movements by understanding the natural mechanics and capabilities of the human muscular-skeletal system. The device also uniquely recognizes individuals in order to allow multiple people to interact with the system via natural movements of their limbs and body as well as voice commands/responses.Type: GrantFiled: May 29, 2009Date of Patent: June 3, 2014Assignee: Microsoft CorporationInventors: R. Stephen Polzin, Alex A. Kipman, Mark J. Finocchio, Ryan Michael Geiss, Kathryn Stone Perez, Kudo Tsunoda, Darren Alexander Bennett
-
Publication number: 20140085193Abstract: A media feed interface may be provided that may be used to extract a media frame from a media feed. The media feed interface may access a capture device, a file, and/or a network resource. Upon accessing the capture device, file, and/or network resource, the media feed interface may populate buffers with data and then may create a media feed from the buffers. Upon request, the media feed interface may isolate a media frame within the media feed. For example, the media feed interface analyze media frames in the media feed to determine whether a media frame includes information associated with, for example, the request. If the media frame includes the requested information, the media feed interface may isolate the media frame associated with the information and may provide access to the isolated media frame.Type: ApplicationFiled: November 27, 2013Publication date: March 27, 2014Inventors: Mark J. Finocchio, Jeffrey Margolis
-
Patent number: 8625837Abstract: A media feed interface may be provided that may be used to extract a media frame from a media feed. The media feed interface may access a capture device, a file, and/or a network resource. Upon accessing the capture device, file, and/or network resource, the media feed interface may populate buffers with data and then may create a media feed from the buffers. Upon request, the media feed interface may isolate a media frame within the media feed. For example, the media feed interface analyze media frames in the media feed to determine whether a media frame includes information associated with, for example, the request. If the media frame includes the requested information, the media feed interface may isolate the media frame associated with the information and may provide access to the isolated media frame.Type: GrantFiled: June 16, 2009Date of Patent: January 7, 2014Assignee: Microsoft CorporationInventors: Mark J. Finocchio, Jeffrey Margolis
-
Publication number: 20140002607Abstract: Technology is described for determining and using invariant features for computer vision. A local orientation may be determined for each depth pixel in a subset of the depth pixels in a depth map. The local orientation may an in-plane orientation, an out-out-plane orientation or both. A local coordinate system is determined for each of the depth pixels in the subset based on the local orientation of the corresponding depth pixel. A feature region is defined relative to the local coordinate system for each of the depth pixels in the subset. The feature region for each of the depth pixels in the subset is transformed from the local coordinate system to an image coordinate system of the depth map. The transformed feature regions are used to process the depth map.Type: ApplicationFiled: November 28, 2012Publication date: January 2, 2014Inventors: Jamie D.J. Shotton, Mark J. Finocchio, Richard E. Moore, Alexandru O. Balan, Kyungsuk David Lee
-
Publication number: 20130286004Abstract: Technology is described for displaying a collision between objects by an augmented reality display device system. A collision between a real object and a virtual object is identified based on three dimensional space position data of the objects. At least one effect on at least one physical property of the real object is determined based on physical properties of the real object, like a change in surface shape, and physical interaction characteristics of the collision. Simulation image data is generated and displayed simulating the effect on the real object by the augmented reality display. Virtual objects under control of different executing applications can also interact with one another in collisions.Type: ApplicationFiled: April 27, 2012Publication date: October 31, 2013Inventors: Daniel J. McCulloch, Stephen G. Latta, Brian J. Mount, Kevin A. Geisner, Roger Sebastian Kevin Sylvan, Arnulfo Zepeda Navratil, Jason Scott, Jonathan T. Steed, Ben J. Sugden, Britta Silke Hummel, Kyungsuk David Lee, Mark J. Finocchio, Alex Aben-Athar Kipman, Jeffrey N. Margolis
-
Publication number: 20120314031Abstract: Technology is described for determining and using invariant features for computer vision. A local orientation may be determined for each depth pixel in a subset of the depth pixels in a depth map. The local orientation may an in-plane orientation, an out-out-plane orientation or both. A local coordinate system is determined for each of the depth pixels in the subset based on the local orientation of the corresponding depth pixel. A feature region is defined relative to the local coordinate system for each of the depth pixels in the subset. The feature region for each of the depth pixels in the subset is transformed from the local coordinate system to an image coordinate system of the depth map. The transformed feature regions are used to process the depth map.Type: ApplicationFiled: June 7, 2011Publication date: December 13, 2012Applicant: MICROSOFT CORPORATIONInventors: Jamie D. J. Shotton, Mark J. Finocchio, Richard E. Moore, Alexandru O. Balan, Kyungsuk David Lee
-
Publication number: 20120280897Abstract: Attribute state classification techniques are described. In one or more implementations, one or more pixels of an image are classified by a computing device as having one or several states for one or more attributes that do not identify corresponding body parts of a user. A gesture is recognized by the computing device that is operable to initiate one or more operations of the computing device based at least in part of the state classifications of the one or more pixels of one or more attributes.Type: ApplicationFiled: May 2, 2011Publication date: November 8, 2012Applicant: MICROSOFT CORPORATIONInventors: Alexandru O. Balan, Richard E. Moore, Mark J. Finocchio
-
Patent number: 8281287Abstract: A non-tree representation of a UI control tree is provided by a compact UI binary file that is generated by encoding a UI definition markup file expressing UI controls and behavior in human-readable form. The UI binary file is utilized in a runtime environment on a computing device as a source of a binary instruction stream. The stream can be efficiently processed by an interpreter on the device without needing validation between loading and rendering the UI. The encoding places much of the representation into an object and script section of the UI binary file. The interpreter runs this section without it being entirely resident in the device's memory to minimize the memory footprint. At runtime, operation code (“op-code”) streams contained in this section are used to build UI objects, and implement scriptable behavior for manipulating the UI objects to render the UI on the device with the designed behavior.Type: GrantFiled: February 26, 2008Date of Patent: October 2, 2012Inventor: Mark J. Finocchio
-
Publication number: 20120206452Abstract: Technology is described for providing realistic occlusion between a virtual object displayed by a head mounted, augmented reality display system and a real object visible to the user's eyes through the display. A spatial occlusion in a user field of view of the display is typically a three dimensional occlusion determined based on a three dimensional space mapping of real and virtual objects. An occlusion interface between a real object and a virtual object can be modeled at a level of detail determined based on criteria such as distance within the field of view, display size or position with respect to a point of gaze. Technology is also described for providing three dimensional audio occlusion based on an occlusion between a real object and a virtual object in the user environment.Type: ApplicationFiled: April 10, 2012Publication date: August 16, 2012Inventors: Kevin A. Geisner, Brian J. Mount, Stephen G. Latta, Daniel J. McCulloch, Kyungsuk David Lee, Ben J. Sugden, Jeffrey N. Margolis, Kathryn Stone Perez, Sheridan Martin Small, Mark J. Finocchio, Robert L. Crocco, JR.
-
Publication number: 20120144348Abstract: Techniques for managing virtual ports are disclosed herein. Each such virtual port may have different associated features such as, for example, privileges, rights or options. When one or more users are in a capture scene of a gesture based system, the system may associate virtual ports with the users and maintain the virtual ports. Also provided are techniques for disassociating virtual ports with users or swapping virtual ports between two or more users.Type: ApplicationFiled: February 10, 2012Publication date: June 7, 2012Applicant: MICROSOFT CORPORATIONInventors: Kathryn Stone-Perez, Jeffrey Margolis, Mark J. Finocchio, Brian E. Keane, Rudy Jacobus Poot, Stephen G. Latta
-
Patent number: 8181123Abstract: Techniques for managing virtual ports are disclosed herein. Each such virtual port may have different associated features such as, for example, privileges, rights or options. When one or more users are in a capture scene of a gesture based system, the system may associate virtual ports with the users and maintain the virtual ports. Also provided are techniques for disassociating virtual ports with users or swapping virtual ports between two or more users.Type: GrantFiled: May 1, 2009Date of Patent: May 15, 2012Assignee: Microsoft CorporationInventors: Kathryn Stone-Perez, Jeffrey Margolis, Mark J. Finocchio, Brian E. Keane, Rudy Jacobus Poot, Stephen G. Latta
-
Patent number: 8133119Abstract: Mechanisms for adjusting signals between gaming controllers and gaming consoles are disclosed. In an embodiment, the output signals of a mouse control a gaming console which is normally controlled by an analog thumbstick. The output signals of mouse are adjusted to compensate for the analog thumbstick controller assist techniques employed by the gaming console. The adjusted signals are sent to the gaming console. The result is that the user is able to control the game using the mouse and have the same feel as if the user was using the analog thumbstick controller.Type: GrantFiled: January 30, 2009Date of Patent: March 13, 2012Assignee: Microsoft CorporationInventor: Mark J. Finocchio
-
Publication number: 20100304813Abstract: A media feed interface may be provided that may be used to extract a media frame from a media feed. The media feed interface may access a capture device, a file, and/or a network resource. Upon accessing the capture device, file, and/or network resource, the media feed interface may populate buffers with data and then may create a media feed from the buffers. Upon request, the media feed interface may isolate a media frame within the media feed. For example, the media feed interface analyze media frames in the media feed to determine whether a media frame includes information associated with, for example, the request. If the media frame includes the requested information, the media feed interface may isolate the media frame associated with the information and may provide access to the isolated media frame.Type: ApplicationFiled: June 16, 2009Publication date: December 2, 2010Applicant: Microsoft CorporationInventors: Mark J. Finocchio, Jeffrey Margolis
-
Publication number: 20100303289Abstract: A system recognizes human beings in their natural environment, without special sensing devices attached to the subjects, uniquely identifies them and tracks them in three dimensional space. The resulting representation is presented directly to applications as a multi-point skeletal model delivered in real-time. The device efficiently tracks humans and their natural movements by understanding the natural mechanics and capabilities of the human muscular-skeletal system. The device also uniquely recognizes individuals in order to allow multiple people to interact with the system via natural movements of their limbs and body as well as voice commands/responses.Type: ApplicationFiled: May 29, 2009Publication date: December 2, 2010Applicant: MICROSOFT CORPORATIONInventors: R. Stephen Polzin, Alex A. Kipman, Mark J. Finocchio, Ryan Michael Geiss, Kathryn Stone Perez, Kudo Tsunoda, Darren Alexander Bennett
-
Publication number: 20100303302Abstract: A depth image of a scene may be received, observed, or captured by a device. The depth image may include a human target that may have, for example, a portion thereof non-visible or occluded. For example, a user may be turned such that a body part may not be visible to the device, may have one or more body parts partially outside a field of view of the device, may have a body part or a portion of a body part behind another body part or object, or the like such that the human target associated with the user may also have a portion body part or a body part non-visible or occluded in the depth image. A position or location of the non-visible or occluded portion or body part of the human target associated with the user may then be estimated.Type: ApplicationFiled: June 26, 2009Publication date: December 2, 2010Applicant: Microsoft CorporationInventors: Alex A. Kipman, Kathryn Stone Perez, Mark J. Finocchio, Ryan Michael Geiss, Kudo Tsunoda
-
Publication number: 20100302365Abstract: A depth image of a scene may be received, observed, or captured by a device. The depth image may then be analyzed to determine whether the depth image includes noise. For example, the depth image may include one or more holes having one or more empty pixels or pixels without a depth value. Depth values for the one or more empty pixels may be estimated and a depth image that includes the estimated depth values for the one or empty more pixels may be rendered.Type: ApplicationFiled: May 29, 2009Publication date: December 2, 2010Applicant: Microsoft CorporationInventors: Mark J. Finocchio, Ryan Michael Geiss
-
Publication number: 20100281437Abstract: Techniques for managing virtual ports are disclosed herein. Each such virtual port may have different associated features such as, for example, privileges, rights or options. When one or more users are in a capture scene of a gesture based system, the system may associate virtual ports with the users and maintain the virtual ports. Also provided are techniques for disassociating virtual ports with users or swapping virtual ports between two or more users.Type: ApplicationFiled: May 1, 2009Publication date: November 4, 2010Applicant: Microsoft CorporationInventors: Kathryn Stone-Perez, Jeffrey Margolis, Mark J. Finocchio, Brian E. Keane, Rudy Jacobus Poot, Stephen G. Latta