Patents by Inventor Zicheng Liu

Zicheng Liu has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20110063403
    Abstract: Techniques and technologies for tracking a face with a plurality of cameras wherein a geometry between the cameras is initially unknown. One disclosed method includes detecting a head with two of the cameras and registering a head model with the image of the head (as detected by one of the cameras). The method also includes back projecting the other detected face image to the head model and determining a head pose from the back-projected head image. Furthermore, the determined geometry is used to track the face with at least one of the cameras.
    Type: Application
    Filed: September 16, 2009
    Publication date: March 17, 2011
    Applicant: Microsoft Corporation
    Inventors: Zhengyou Zhang, Aswin Sankaranarayanan, Qing Zhang, Zicheng Liu, Qin Cai
  • Patent number: 7892079
    Abstract: Disclosed are a unique DPC (detect point click) based game system and method. The DPC based game system involves generating one or a plurality of DPC images, presenting them to a game participant, and collecting the participant's clicks (that identify which object in the DPC image the participant believes to be the correct DPC object), and determining whether the participant's clicks represent the correct object. DPC images can be created in part by selecting a base image, altering some portion of the base image to create at least one confusion image, mapping these images to a geometric model, and applying one or more distortion filters to at least one of the base or confusing image to obscure the DPC object from clear view. Locating the DPC object nearly hidden in the DPC image can advance the participant in the DPC based game or other game including DPC images as a part thereof.
    Type: Grant
    Filed: August 10, 2004
    Date of Patent: February 22, 2011
    Assignee: Microsoft Corporation
    Inventors: Zicheng Liu, Yong Rui
  • Patent number: 7869705
    Abstract: A subject captured by a camera may be affected by environmental lighting provided by nearby light sources and the sun or moon, which may cause underexposure or overexposure of the image or aesthetically displeasing color tones. Image processing and camera adjustments may mitigate some imaging problems with limited effect and introduce undesirable side effects. A lighting array may be devised to expose the subject to various types of light (e.g., white light comprising full spectrum illumination and red, green, and blue lights comprising partial spectrum illumination) to resolve lighting problems in a more effective manner. Moreover, the lighting array may be responsively controlled to adjust the subject image with respect to one or more target spectra specifying desirable colors for the subject image. The lighting array may be iteratively controlled, e.g. by a gradient descent algorithm, for incrementally adjusting parameters with respect to proximate target spectra for the image.
    Type: Grant
    Filed: January 21, 2008
    Date of Patent: January 11, 2011
    Assignee: Microsoft Corporation
    Inventors: Zicheng Liu, Mingxuan Sun, Jingyu Qiu, Zhengyou Zhang, Michael J. Sinclair
  • Publication number: 20100315482
    Abstract: Gaze tracking or other interest indications are used during a video conference to determine one or more audio sources that are of interest to one or more participants to the video conference, such as by determining a conversation from among multiple conversations that a subset of participants are participating in or listening to, for enhancing the audio experience of one or more of the participants.
    Type: Application
    Filed: June 15, 2009
    Publication date: December 16, 2010
    Applicant: Microsoft Corporation
    Inventors: Daniel A. Rosenfeld, Zicheng Liu, Ross G. Cutler, Philip A. Chou, Christian Huitema, Kori Quinn
  • Publication number: 20100315506
    Abstract: Described is a technology by which video is processed to determine whether the video contains a specified action. The video corresponds to a spatial-temporal volume. The volume is searched to find a sub-volume therein that has a maximum score with respect to whether the video contains the action. Searching for the sub-volume is performed by separating the search space into a spatial subspace and a temporal subspace. The spatial subspace is searched for an optimal spatial window using upper-bounds searching. Also described is discriminative pattern matching.
    Type: Application
    Filed: June 10, 2009
    Publication date: December 16, 2010
    Applicant: Microsoft Corporation
    Inventors: Zicheng Liu, Junsong Yuan
  • Patent number: 7853053
    Abstract: The computer-readable media provides improved procedures to estimate head motion between two images of a face. Locations of a number of distinct facial features are determined in two images. The locations are converted into as a set of physical face parameters based on the symmetry of the identified distinct facial features. An estimation objective function is determined by: (a) estimating each of the set of physical parameters, (b) estimating a first head pose transform corresponding to the first image, and (c) estimating a second head pose transform corresponding to the second image. The motion is estimated between the two images based on the set of physical face parameters by multiplying each term of the estimation objective function by a weighted contribution factor based on the confidence of data corresponding to the estimation objective function.
    Type: Grant
    Filed: March 31, 2010
    Date of Patent: December 14, 2010
    Assignee: Microsoft Corporation
    Inventors: Zicheng Liu, Zhengyou Zhang
  • Publication number: 20100302462
    Abstract: A virtual media device is described for processing one or more input signals from one or more physical media input devices, to thereby generate an output signal for use by a consuming application module. The consuming application module interacts with the virtual media device as if it were a physical media input device. The virtual media device thereby frees the application module and its user from the burden of having to take specific account of the physical media input devices that are connected to a computing environment. The virtual media device can be coupled to one or more microphone devices, one or more video input devices, or a combination of audio and video input devices, etc. The virtual media device can apply any number of processing modules to generate the output signal, each performing a different respective operation.
    Type: Application
    Filed: May 28, 2009
    Publication date: December 2, 2010
    Applicant: Microsoft Corporation
    Inventors: Zicheng Liu, Rajesh K. Hegde, Philip A. Chou
  • Patent number: 7840638
    Abstract: A multimedia conference technique is disclosed that allows physically remote users to participate in an immersive telecollaborative environment by synchronizing multiple data, images and sounds. The multimedia conference implementation provides users with the perception of being in the same room visually as well as acoustically according to an orientation plan which reflects each remote user's position within the multimedia conference environment.
    Type: Grant
    Filed: June 27, 2008
    Date of Patent: November 23, 2010
    Assignee: Microsoft Corporation
    Inventors: Zhengyou Zhang, Xuedong David Huang, Zicheng Liu, Cha Zhang, Philip A. Chou, Christian Huitema
  • Publication number: 20100289904
    Abstract: Systems are disclosed that provide improved transfer speed of video data from a video capture device to a computing device using multiple video feeds respectively comprising different resolutions. A high-resolution image sensor is used to convert light images into a high-resolution video data stream. A down sampler converts the high-resolution video data stream to a low-resolution video data stream, so that both a low-resolution data stream and a high-resolution data stream are available. While the low resolution-data stream can be sent to the computing device, a digital signal processor (DSP) processes the high-resolution video data stream in accordance with an input control signal that is comprised of desired high-resolution video stream parameters derived from the low-resolution video data stream.
    Type: Application
    Filed: May 15, 2009
    Publication date: November 18, 2010
    Applicant: Microsoft Corporation
    Inventors: Cha Zhang, Zhengyou Zhang, Zicheng Liu, Wanghong Yuan, Christian Huitema
  • Patent number: 7813923
    Abstract: A first set of signals from an array of one or more microphones, and a second signal from a reference microphone are used to calibrate a set of filter parameters such that the filter parameters minimize a difference between the second signal and a beamformer output signal that is based on the first set of signals. Once calibrated, the filter parameters are used to form a beamformer output signal that is filtered using a non-linear adaptive filter that is adapted based on portions of a signal that do not contain speech, as determined by a speech detection sensor.
    Type: Grant
    Filed: October 14, 2005
    Date of Patent: October 12, 2010
    Assignee: Microsoft Corporation
    Inventors: Alejandro Acero, Michael L. Seltzer, Zhengyou Zhang, Zicheng Liu
  • Publication number: 20100245536
    Abstract: The claimed subject matter provides a system and/or a method that facilitates managing one or more devices utilized for communicating data within a telepresence session. A telepresence session can be initiated within a communication framework that includes two or more virtually represented users that communicate therein. A device can be utilized by at least one virtually represented user that enables communication within the telepresence session, the device includes at least one of an input to transmit a portion of a communication to the telepresence session or an output to receive a portion of a communication from the telepresence session. A detection component can adjust at least one of the input related to the device or the output related to the device based upon the identification of a cue, the cue is at least one of a movement detected, an event detected, or an ambient variation.
    Type: Application
    Filed: March 30, 2009
    Publication date: September 30, 2010
    Applicant: Microsoft Corporation
    Inventors: Christian Huitema, William A.S. Buxton, Jonathan E. Paff, Zicheng Liu, Rajesh Kutpadi Hegde, Zhengyou Zhang, Kori Marie Quinn, Jin Li, Michel Pahud
  • Publication number: 20100201823
    Abstract: In some implementations, invisible light is emitted toward a subject being imaged in a low-light environment. A camera having a first color image sensor captures an image of the subject. Image processing is used to correct distortion in the image caused by the invisible light, and an augmented color image is output.
    Type: Application
    Filed: February 10, 2009
    Publication date: August 12, 2010
    Applicant: Microsoft Corporation
    Inventors: Chunhui Zhang, Yasuyuki Matsushita, Yuan Kong, Zicheng Liu
  • Patent number: 7770116
    Abstract: A system that captures both whiteboard content and audio signals of a meeting using a digital camera and a microphone. The system can be retrofit to any existing whiteboard. It computes the time stamps of pen strokes on the whiteboard by analyzing the sequence of captured snapshots. It also automatically produces a set of key frames representing all the written content on the whiteboard before each erasure. The whiteboard content serves as a visual index to efficiently browse the audio meeting. The system not only captures the whiteboard content, but also helps the users to view and manage the captured meeting content efficiently and securely.
    Type: Grant
    Filed: November 30, 2006
    Date of Patent: August 3, 2010
    Assignee: Microsoft Corp.
    Inventors: Zhengyou Zhang, Ross Cutler, Zicheng Liu, Anoop Gupta, Li-wei He
  • Publication number: 20100189310
    Abstract: The computer-readable media provides improved procedures to estimate head motion between two images of a face. Locations of a number of distinct facial features are determined in two images. The locations are converted into as a set of physical face parameters based on the symmetry of the identified distinct facial features. An estimation objective function is determined by: (a) estimating each of the set of physical parameters, (b) estimating a first head pose transform corresponding to the first image, and (c) estimating a second head pose transform corresponding to the second image. The motion is estimated between the two images based on the set of physical face parameters by multiplying each term of the estimation objective function by a weighted contribution factor based on the confidence of data corresponding to the estimation objective function.
    Type: Application
    Filed: March 31, 2010
    Publication date: July 29, 2010
    Applicant: Microsoft Corporation
    Inventors: Zicheng Liu, Zhengyou Zhang
  • Patent number: 7725395
    Abstract: A system and method for automatically determining if a remote client is a human or a computer. A set of HIP design guidelines which are important to ensure the security and usability of a HIP system are described. Furthermore, one embodiment of this new HIP system and method is based on human face and facial feature detection. Because human face is the most familiar object to all human users the embodiment of the invention employing a face is possibly the most universal HIP system so far.
    Type: Grant
    Filed: September 19, 2003
    Date of Patent: May 25, 2010
    Assignee: Microsoft Corp.
    Inventors: Yong Rui, Zicheng Liu
  • Patent number: 7706610
    Abstract: An “Image Segmenter” provides a variational energy formulation for segmentation of natural objects from images. In general, the Image Segmenter operates by adopting Gaussian mixture models (GMM) to capture the appearance variation of objects in one or more images. A global image data likelihood potential is then computed and combined with local region potentials to obtain a robust and accurate estimation of pixel foreground and background distributions. Iterative minimization of a “global-local energy function” is then accomplished by evolution of a foreground/background boundary curve by level set, and estimation of a foreground/background model by fixed-point iteration, termed “quasi-semi-supervised EM.” In various embodiments, this process is further improved by providing general object shape information for use in rectifying objects segmented from the image.
    Type: Grant
    Filed: November 29, 2005
    Date of Patent: April 27, 2010
    Assignee: Microsoft Corporation
    Inventors: Zhengyou Zhang, Zicheng Liu, Gang Hua
  • Patent number: 7706575
    Abstract: The system provides improved procedures to estimate head motion between two images of a face. Locations of a number of distinct facial features are identified in two images. The identified locations can correspond to the eye corners, mouth corners and nose tip. The locations are converted into as a set of physical face parameters based on the symmetry of the identified distinct facial features. The set of physical parameters reduces the number of unknowns as compared to the number of equations used to determine the unknowns. An initial head motion estimate is determined by: (a) estimating each of the set of physical parameters, (b) estimating a first head pose transform corresponding to the first image, and (c) estimating a second head pose transform corresponding to the second image. The head motion estimate can be incorporated into a feature matching algorithm to refine the head motion estimation and the physical facial parameters.
    Type: Grant
    Filed: August 4, 2004
    Date of Patent: April 27, 2010
    Assignee: Microsoft Corporation
    Inventors: Zicheng Liu, Zhengyou Zhang
  • Publication number: 20100085416
    Abstract: Multi-device capture and spatial browsing of conferences is described. In one implementation, a system detects cameras and microphones, such as the webcams on participants' notebook computers, in a conference room, group meeting, or table game, and enlists an ad-hoc array of available devices to capture each participant and the spatial relationships between participants. A video stream composited from the array is browsable by a user to navigate a 3-dimensional representation of the meeting. Each participant may be represented by a video pane, a foreground object, or a 3-D geometric model of the participant's face or body displayed in spatial relation to the other participants in a 3-dimensional arrangement analogous to the spatial arrangement of the meeting.
    Type: Application
    Filed: October 6, 2008
    Publication date: April 8, 2010
    Applicant: Microsoft Corporation
    Inventors: Rajesh K. Hegde, Zhengyou Zhang, Philip A. Chou, Cha Zhang, Zicheng Liu, Sasa Junuzovic
  • Patent number: 7694214
    Abstract: A multimodal, multilanguage mobile device which can be employed to enhance note taking and/or annotation of a document, and gaming. Input data types such as optical character recognition (OCR), speech, handwriting, and visual information (e.g., image and/or video), etc., can be fused to generate rich documents with a multidimensional level of data to provide an increased level of context over conventional documents. Such architecture can be utilized by students for homework management, as well as entertainment (e.g., gaming).
    Type: Grant
    Filed: June 29, 2005
    Date of Patent: April 6, 2010
    Assignee: Microsoft Corporation
    Inventors: Zicheng Liu, Zhengyou Zhang, David Kurlander, David W. Williams
  • Patent number: 7680656
    Abstract: A method and apparatus determine a likelihood of a speech state based on an alternative sensor signal and an air conduction microphone signal. The likelihood of the speech state is used, together with the alternative sensor signal and the air conduction microphone signal, to estimate a clean speech value for a clean speech signal.
    Type: Grant
    Filed: June 28, 2005
    Date of Patent: March 16, 2010
    Assignee: Microsoft Corporation
    Inventors: Zhengyou Zhang, Zicheng Liu, Alejandro Acero, Amarnag Subramanya, James G. Droppo