Patents by Inventor Zicheng Liu

Zicheng Liu has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 9001118
    Abstract: A method for constructing an avatar of a human subject includes acquiring a depth map of the subject, obtaining a virtual skeleton of the subject based on the depth map, and harvesting from the virtual skeleton a set of characteristic metrics. Such metrics correspond to distances between predetermined points of the virtual skeleton. In this example method, the characteristic metrics are provided as input to an algorithm trained using machine learning. The algorithm may be trained using a human model in a range of poses, and a range of human models in a single pose, to output a virtual body mesh as a function of the characteristic metrics. The method also includes constructing a virtual head mesh distinct from the virtual body mesh, with facial features resembling those of the subject, and connecting the virtual body mesh to the virtual head mesh.
    Type: Grant
    Filed: August 14, 2012
    Date of Patent: April 7, 2015
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: David Molyneaux, Xin Tong, Zicheng Liu, Eric Chang, Fan Yang, Jay Kapur, Emily Yang, Yang Liu, Hsiang-Tao Wu
  • Patent number: 8941710
    Abstract: A system facilitates managing one or more devices utilized for communicating data within a telepresence session. A telepresence session can be initiated within a communication framework that includes a first user and one or more second users. In response to determining a temporary absence of the first user from the telepresence session, a recordation of the telepresence session is initialized to enable a playback of a portion or a summary of the telepresence session that the first user has missed.
    Type: Grant
    Filed: August 13, 2012
    Date of Patent: January 27, 2015
    Assignee: Microsoft Corporation
    Inventors: Christian Huitema, William A. S. Buxton, Jonathan E. Paff, Zicheng Liu, Rajesh Kutpadi Hegde, Zhengyou Zhang, Kori Marie Quinn, Jin Li, Michel Pahud
  • Patent number: 8943526
    Abstract: Technologies described herein relate to estimating engagement of a person with respect to content being presented to the person. A sensor outputs a stream of data relating to the person as the person is consuming the content. At least one feature is extracted from the stream of data, and a level of engagement of the person is estimated based at least in part upon the at least one feature. A computing function is performed based upon the estimated level of engagement of the person.
    Type: Grant
    Filed: April 19, 2013
    Date of Patent: January 27, 2015
    Assignee: Microsoft Corporation
    Inventors: Javier Hernandez Rivera, Zicheng Liu, Geoff Hulten, Michael Conrad, Kyle Krum, David DeBarr, Zhengyou Zhang
  • Patent number: 8929600
    Abstract: A plurality of depth maps corresponding to respective depth measurements determined over a respective plurality of time frames may be obtained. A plurality of skeleton representations respectively corresponding to the respective time frames may be obtained. Each skeleton representation may include joints associated with an observed entity. Local feature descriptors corresponding to the respective time frames may be determined, based on the depth maps and the joints associated with the skeleton representations. An activity recognition associated with the observed entity may be determined, based on the obtained skeleton representations and the determined local feature descriptors.
    Type: Grant
    Filed: December 19, 2012
    Date of Patent: January 6, 2015
    Assignee: Microsoft Corporation
    Inventors: Zicheng Liu, Jiang Wang
  • Publication number: 20140184803
    Abstract: A technique for multi-camera object tracking is disclosed that preserves privacy of imagery from each camera or group of cameras. This technique uses secure multi-party computation to compute a distance metric across data from multiple cameras without revealing any information to operators of the cameras except whether or not an object was observed by both cameras. This is achieved by a distance metric learning technique that reduces the computing complexity of secure computation while maintaining object identification accuracy.
    Type: Application
    Filed: December 31, 2012
    Publication date: July 3, 2014
    Applicant: MICROSOFT CORPORATION
    Inventors: Chun-Te Chu, Jaeyeon Jung, Zicheng Liu, Ratul Mahajan
  • Publication number: 20140169623
    Abstract: A plurality of depth maps corresponding to respective depth measurements determined over a respective plurality of time frames may be obtained. A plurality of skeleton representations respectively corresponding to the respective time frames may be obtained. Each skeleton representation may include joints associated with an observed entity. Local feature descriptors corresponding to the respective time frames may be determined, based on the depth maps and the joints associated with the skeleton representations. An activity recognition associated with the observed entity may be determined, based on the obtained skeleton representations and the determined local feature descriptors.
    Type: Application
    Filed: December 19, 2012
    Publication date: June 19, 2014
    Applicant: MICROSOFT CORPORATION
    Inventors: Zicheng Liu, Jiang Wang
  • Publication number: 20140160123
    Abstract: Described herein are technologies pertaining to generating a relatively accurate virtual three-dimensional model of a head/face of a user. Depth frames are received from a depth sensor and color frames are received from a camera, wherein such frames capture a head of a user. Based upon the depth frames and the color frames, the three-dimensional model of the head of the user is generated.
    Type: Application
    Filed: December 12, 2012
    Publication date: June 12, 2014
    Applicant: MICROSOFT CORPORATION
    Inventors: Fan Yang, Minmin Gong, Zicheng Liu, Xin Tong
  • Publication number: 20140111401
    Abstract: A pendant for use with a mobile terminal, including: a pendant body containing an antenna; and a pendant cord connected with the pendant body, the pendant cord enclosing a cable connected to the antenna for transmitting and receiving radio frequency signals.
    Type: Application
    Filed: December 23, 2013
    Publication date: April 24, 2014
    Applicant: Xiaomi Inc.
    Inventors: Lei Zhang, Zicheng Liu, Zhu Mao
  • Patent number: 8675067
    Abstract: The subject disclosure is directed towards an immersive conference, in which participants in separate locations are brought together into a common virtual environment (scene), such that they appear to each other to be in a common space, with geometry, appearance, and real-time natural interaction (e.g., gestures) preserved. In one aspect, depth data and video data are processed to place remote participants in the common scene from the first person point of view of a local participant. Sound data may be spatially controlled, and parallax computed to provide a realistic experience. The scene may be augmented with various data, videos and other effects/animations.
    Type: Grant
    Filed: May 4, 2011
    Date of Patent: March 18, 2014
    Assignee: Microsoft Corporation
    Inventors: Philip A. Chou, Zhengyou Zhang, Cha Zhang, Dinei A. Florencio, Zicheng Liu, Rajesh K. Hegde, Nirupama Chandrasekaran
  • Patent number: 8639042
    Abstract: Described is a hierarchical filtered motion field technology such as for use in recognizing actions in videos with crowded backgrounds. Interest points are detected, e.g., as 2D Harris corners with recent motion, e.g. locations with high intensities in a motion history image (MHI). A global spatial motion smoothing filter is applied to the gradients of MHI to eliminate low intensity corners that are likely isolated, unreliable or noisy motions. At each remaining interest point, a local motion field filter is applied to the smoothed gradients by computing a structure proximity between sets of pixels in the local region and the interest point. The motion at a pixel/pixel set is enhanced or weakened based on its structure proximity with the interest point (nearer pixels are enhanced).
    Type: Grant
    Filed: June 22, 2010
    Date of Patent: January 28, 2014
    Assignee: Microsoft Corporation
    Inventors: Zicheng Liu, Yingli Tian, Liangliang Cao, Zhengyou Zhang
  • Publication number: 20140009562
    Abstract: Multi-device capture and spatial browsing of conferences is described. In one implementation, a system detects cameras and microphones, such as the webcams on participants' notebook computers, in a conference room, group meeting, or table game, and enlists an ad-hoc array of available devices to capture each participant and the spatial relationships between participants. A video stream composited from the array is browsable by a user to navigate a 3-dimensional representation of the meeting. Each participant may be represented by a video pane, a foreground object, or a 3-D geometric model of the participant's face or body displayed in spatial relation to the other participants in a 3-dimensional arrangement analogous to the spatial arrangement of the meeting. The system may automatically re-orient the 3-dimensional representation as needed to best show a currently interesting event.
    Type: Application
    Filed: July 9, 2013
    Publication date: January 9, 2014
    Inventors: Rajesh K. HEGDE, Zhengyou ZHANG, Philip A. CHOU, Cha ZHANG, Zicheng LIU, Sasa JUNUZOVIC
  • Publication number: 20130342527
    Abstract: A method for constructing an avatar of a human subject includes acquiring a depth map of the subject, obtaining a virtual skeleton of the subject based on the depth map, and harvesting from the virtual skeleton a set of characteristic metrics. Such metrics correspond to distances between predetermined points of the virtual skeleton. In this example method, the characteristic metrics are provided as input to an algorithm trained using machine learning. The algorithm may be trained using a human model in a range of poses, and a range of human models in a single pose, to output a virtual body mesh as a function of the characteristic metrics. The method also includes constructing a virtual head mesh distinct from the virtual body mesh, with facial features resembling those of the subject, and connecting the virtual body mesh to the virtual head mesh.
    Type: Application
    Filed: August 14, 2012
    Publication date: December 26, 2013
    Applicant: MICROSOFT CORPORATION
    Inventors: David Molyneaux, Xin Tong, Zicheng Liu, Eric Chang, Fan Yang, Jay Kapur, Emily Yang, Yang Liu, Hsiang-Tao Wu
  • Patent number: 8537196
    Abstract: Multi-device capture and spatial browsing of conferences is described. In one implementation, a system detects cameras and microphones, such as the webcams on participants' notebook computers, in a conference room, group meeting, or table game, and enlists an ad-hoc array of available devices to capture each participant and the spatial relationships between participants. A video stream composited from the array is browsable by a user to navigate a 3-dimensional representation of the meeting. Each participant may be represented by a video pane, a foreground object, or a 3-D geometric model of the participant's face or body displayed in spatial relation to the other participants in a 3-dimensional arrangement analogous to the spatial arrangement of the meeting.
    Type: Grant
    Filed: October 6, 2008
    Date of Patent: September 17, 2013
    Assignee: Microsoft Corporation
    Inventors: Rajesh K. Hegde, Zhengyou Zhang, Philip A. Chou, Cha Zhang, Zicheng Liu, Sasa Junuzovic
  • Publication number: 20130232515
    Abstract: Technologies described herein relate to estimating engagement of a person with respect to content being presented to the person. A sensor outputs a stream of data relating to the person as the person is consuming the content. At least one feature is extracted from the stream of data, and a level of engagement of the person is estimated based at least in part upon the at least one feature. A computing function is performed based upon the estimated level of engagement of the person.
    Type: Application
    Filed: April 19, 2013
    Publication date: September 5, 2013
    Applicant: Microsoft Corporation
    Inventors: Javier Hernandez Rivera, Zicheng Liu, Geoff Hulten, Michael Conrad, Kyle Krum, David DeBarr, Zhengyou Zhang
  • Publication number: 20130201291
    Abstract: Head pose tracking technique embodiments are presented that use a group of sensors configured so as to be disposed on a user's head. This group of sensors includes a depth sensor apparatus used to identify the three dimensional locations of features within a scene, and at least one other type of sensor. Data output by each sensor in the group of sensors is periodically input, and each time the data is input it is used to compute a transformation matrix that when applied to a previously determined head pose location and orientation established when the first sensor data was input identifies a current head pose location and orientation. This transformation matrix is then applied to the previously determined head pose location and orientation to identify a current head pose location and orientation.
    Type: Application
    Filed: February 8, 2012
    Publication date: August 8, 2013
    Applicant: MICROSOFT CORPORATION
    Inventors: Zicheng Liu, Zhengyou Zhang, Zhenning Li
  • Patent number: 8416302
    Abstract: In some implementations, invisible light is emitted toward a subject being imaged in a low-light environment. A camera having a first color image sensor captures an image of the subject. Image processing is used to correct distortion in the image caused by the invisible light, and an augmented color image is output.
    Type: Grant
    Filed: February 10, 2009
    Date of Patent: April 9, 2013
    Assignee: Microsoft Corporation
    Inventors: Chunhui Zhang, Yasuyuki Matsushita, Yuan Kong, Zicheng Liu
  • Patent number: 8416990
    Abstract: Described is a technology by which video, which may be relatively high-resolution video, is efficiently processed to determine whether the video contains a specified action. The video corresponds to a spatial-temporal volume. The volume is searched with a top-k search that finds a plurality of the most likely sub-volumes simultaneously in a single search round. The score volumes of larger spatial resolution videos may be down-sampled into lower-resolution score volumes prior to searching.
    Type: Grant
    Filed: August 17, 2010
    Date of Patent: April 9, 2013
    Assignee: Microsoft Corporation
    Inventors: Zicheng Liu, Norberto Adrian Goussies
  • Patent number: 8416715
    Abstract: Gaze tracking or other interest indications are used during a video conference to determine one or more audio sources that are of interest to one or more participants to the video conference, such as by determining a conversation from among multiple conversations that a subset of participants are participating in or listening to, for enhancing the audio experience of one or more of the participants.
    Type: Grant
    Filed: June 15, 2009
    Date of Patent: April 9, 2013
    Assignee: Microsoft Corporation
    Inventors: Daniel A. Rosenfeld, Zicheng Liu, Ross G. Cutler, Philip A. Chou, Christian Huitema, Kori Quinn
  • Patent number: 8396247
    Abstract: A system that facilitates automatically determining an action of an animate object is described herein. The system includes a receiver component that receives video data that includes images of an animate object. The system additionally includes a determiner component that accesses a data store that includes an action graph and automatically determines an action undertaken by the animate object in the received video data based at least in part upon the action graph. The action graph comprises a plurality of nodes that are representative of multiple possible postures of the animate object. At least one node in the action graph is shared amongst multiple actions represented in the action graph.
    Type: Grant
    Filed: July 31, 2008
    Date of Patent: March 12, 2013
    Assignee: Microsoft Corporation
    Inventors: Zhengyou Zhang, Wanqing Li, Zicheng Liu
  • Patent number: 8339459
    Abstract: Techniques and technologies for tracking a face with a plurality of cameras wherein a geometry between the cameras is initially unknown. One disclosed method includes detecting a head with two of the cameras and registering a head model with the image of the head (as detected by one of the cameras). The method also includes back projecting the other detected face image to the head model and determining a head pose from the back-projected head image. Furthermore, the determined geometry is used to track the face with at least one of the cameras.
    Type: Grant
    Filed: September 16, 2009
    Date of Patent: December 25, 2012
    Assignee: Microsoft Corporation
    Inventors: Zhengyou Zhang, Aswin Sankaranarayanan, Qing Zhang, Zicheng Liu, Qin Cai