Patents by Inventor Hans Peter Graf

Hans Peter Graf has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20140180977
    Abstract: Systems and methods are disclosed for classifying histological tissues or specimens with two phases. In a first phase, the method includes providing off-line training using a processor during which one or more classifiers are trained based on examples, including: finding a split of features into sets of increasing computational cost, assigning a computational cost to each set; training for each set of features a classifier using training examples; training for each classifier, a utility function that scores a usefulness of extracting the next feature set for a given tissue unit using the training examples.
    Type: Application
    Filed: November 12, 2013
    Publication date: June 26, 2014
    Applicant: NEC Laboratories America, Inc.
    Inventors: Eric Cosatto, Pierre-Francois Laquerre, Christopher Malon, Hans-Peter Graf, Iain Melvin
  • Patent number: 8638846
    Abstract: Systems and methods for encoding/decoding a video stream. Animated talking heads are coded using partial offline encoding, multiple video streams, and multiple reference frames. The content of a face animation video that is known beforehand is encoded offline and the remaining content is encoded online and included in the video stream. To reduce bit rate, a server can stream multiple video sequences to the client and the video sequences are stored in the client's frame store. The server sends instructions to play a particular video sequence instead of streaming the particular video sequence. Multiple video streams can also be streamed to the client. Positional data and blending data are also sent to properly position one video stream relative to another video stream and to blend one video stream into another video stream.
    Type: Grant
    Filed: June 23, 2003
    Date of Patent: January 28, 2014
    Assignee: AT&T Intellectual Property II, L.P.
    Inventors: Eric Cosatto, Hans Peter Graf, Joern Ostermann
  • Publication number: 20140023237
    Abstract: A method of improving the lighting conditions of a real scene or video sequence. Digitally generated light is added to a scene for video conferencing over telecommunication networks. A virtual illumination equation takes into account light attenuation, lambertian and specular reflection. An image of an object is captured, a virtual light source illuminates the object within the image. In addition, the object can be the head of the user. The position of the head of the user is dynamically tracked so that an three-dimensional model is generated which is representative of the head of the user. Synthetic light is applied to a position on the model to form an illuminated model.
    Type: Application
    Filed: September 24, 2013
    Publication date: January 23, 2014
    Applicant: AT&T Intellectual Property II, L.P.
    Inventors: Andrea Basso, Eric Cosatto, David Crawford Gibbon, Hans Peter Graf, Shan Liu
  • Patent number: 8583896
    Abstract: Systems and methods for massively parallel processing on an accelerator that includes a plurality of processing cores. Each processing core includes multiple processing chains configured to perform parallel computations, each of which includes a plurality of interconnected processing elements. The cores further include multiple of smart memory blocks configured to store and process data, each memory block accepting the output of one of the plurality of processing chains. The cores communicate with at least one off-chip memory bank.
    Type: Grant
    Filed: July 26, 2010
    Date of Patent: November 12, 2013
    Assignee: NEC Laboratories America, Inc.
    Inventors: Srihari Cadambi, Abhinandan Majumdar, Michela Becchi, Srimat Chakradhar, Hans Peter Graf
  • Patent number: 8542943
    Abstract: A method of improving the lighting conditions of a real scene or video sequence. Digitally generated light is added to a scene for video conferencing over telecommunication networks. A virtual illumination equation takes into account light attenuation, lambertian and specular reflection. An image of an object is captured, a virtual light source illuminates the object within the image. In addition, the object can be the head of the user. The position of the head of the user is dynamically tracked so that an three-dimensional model is generated which is representative of the head of the user. Synthetic light is applied to a position on the model to form an illuminated model.
    Type: Grant
    Filed: November 23, 2011
    Date of Patent: September 24, 2013
    Assignee: AT&T Intellectual Property II, L.P.
    Inventors: Andrea Basso, Eric Cosatto, David Crawford Gibbon, Hans Peter Graf, Shan Liu
  • Patent number: 8359281
    Abstract: A method system for training an apparatus to recognize a pattern includes providing the apparatus with a host processor executing steps of a machine learning process; providing the apparatus with an accelerator including at least two processors; inputting training pattern data into the host processor; determining coefficient changes in the machine learning process with the host processor using the training pattern data; transferring the training data to the accelerator; determining kernel dot-products with the at least two processors of the accelerator using the training data; and transferring the dot-products back to the host processor.
    Type: Grant
    Filed: June 4, 2009
    Date of Patent: January 22, 2013
    Assignee: NEC Laboratories America, Inc.
    Inventors: Srihari Cadambi, Igor Durdanovic, Venkata Jakkula, Eric Cosatto, Murugan Sankaradass, Hans Peter Graf, Srimat T. Chakradhar
  • Patent number: 8200493
    Abstract: A system and method of controlling the movement of a virtual agent while the agent is listening to a human user during a conversation are disclosed. The method includes receiving speech data from the user to yield received speech data and analyzing the received speech data according to a prosodic characterization of the speech data to yield the speech data analysis. The method further includes controlling listening movement of the agent according to the speech data analysis, wherein the listening movement of the agent occurs while the user is speaking the received speech data to the agent such that the agent appears to respond to the specific received speech data.
    Type: Grant
    Filed: January 25, 2008
    Date of Patent: June 12, 2012
    Assignee: AT&T Intellectual Property II, L.P.
    Inventors: Eric Cosatto, Hans Peter Graf, Thomas M. Isaacson, Volker Franz Strom
  • Publication number: 20120069139
    Abstract: A method of improving the lighting conditions of a real scene or video sequence. Digitally generated light is added to a scene for video conferencing over telecommunication networks. A virtual illumination equation takes into account light attenuation, lambertian and specular reflection. An image of an object is captured, a virtual light source illuminates the object within the image. In addition, the object can be the head of the user. The position of the head of the user is dynamically tracked so that an three-dimensional model is generated which is representative of the head of the user. Synthetic light is applied to a position on the model to form an illuminated model.
    Type: Application
    Filed: November 23, 2011
    Publication date: March 22, 2012
    Applicant: AT&T Intellectual Property II, L.P.
    Inventors: Andrea Basso, Eric Cosatto, David Crawford Goibbon, Hans Peter Graf, Shan Liu
  • Patent number: 8135652
    Abstract: Disclosed is an improved technique for training a support vector machine using a distributed architecture. A training data set is divided into subsets, and the subsets are optimized in a first level of optimizations, with each optimization generating a support vector set. The support vector sets output from the first level optimizations are then combined and used as input to a second level of optimizations. This hierarchical processing continues for multiple levels, with the output of each prior level being fed into the next level of optimizations. In order to guarantee a global optimal solution, a final set of support vectors from a final level of optimization processing may be fed back into the first level of the optimization cascade so that the results may be processed along with each of the training data subsets.
    Type: Grant
    Filed: April 28, 2008
    Date of Patent: March 13, 2012
    Assignee: NEC Laboratories America, Inc.
    Inventors: Hans Peter Graf, Eric Cosatto, Leon Bottou, Vladimir N. Vapnik
  • Patent number: 8135202
    Abstract: An automated method and system for analyzing a digital image of a biopsy to determine whether the biopsy is normal or abnormal, i.e., exhibits some type of disease such as, but not limited to, cancer. In the method and system, a classifier is trained to recognize well formed nuclei outlines from imperfect nuclei outlines in digital biopsy images. The trained classifier may then be used to filter nuclei outlines from one or more digital biopsy images to be analyzed, to obtain the well formed nuclei outlines. The well formed nuclei outlines may then be used to obtain statistics on the size or area of the nuclei for use in determining whether the biopsy is normal or abnormal.
    Type: Grant
    Filed: June 2, 2008
    Date of Patent: March 13, 2012
    Assignee: NEC Laboratories America, Inc.
    Inventors: Eric Cosatto, Hans-Peter Graf, Matthew L. Miller
  • Patent number: 8131551
    Abstract: A system and method of controlling the movement of a virtual agent while the agent is speaking to a human user during a conversation is disclosed. The method comprises receiving speech data to be spoken by the virtual agent, performing a prosodic analysis of the speech data, selecting matching prosody patterns from a speaking database and controlling the virtual agent movement according to the selected prosody patterns.
    Type: Grant
    Filed: July 18, 2006
    Date of Patent: March 6, 2012
    Assignee: AT&T Intellectual Property II, L.P.
    Inventors: Eric Cosatto, Hans Peter Graf, Volker Franz Strom
  • Patent number: 8086464
    Abstract: Methods and apparatus for rendering a talking head on a client device are disclosed. The client device has a client cache capable of storing audio/visual data associated with rendering the talking head. The method comprises storing sentences in a client cache of a client device that relate to bridging delays in a dialog, storing sentence templates to be used in dialogs, generating a talking head response to a user inquiry from the client device, and determining whether sentences or stored templates stored in the client cache relate to the talking head response. If the stored sentences or stored templates relate to the talking head response, the method comprises instructing the client device to use the appropriate stored sentence or template from the client cache to render at least a part of the talking head response and transmitting a portion of the talking head response not stored in the client cache, if any, to the client device to render a complete talking head response.
    Type: Grant
    Filed: November 30, 2009
    Date of Patent: December 27, 2011
    Assignee: AT&T Intellectual Property II, L.P.
    Inventors: Eric Cosatto, Hans Peter Graf, Joern Ostermann
  • Patent number: 8086066
    Abstract: A method of improving the lighting conditions of a real scene or video sequence. Digitally generated light is added to a scene for video conferencing over telecommunication networks. A virtual illumination equation takes into account light attenuation, lambertian and specular reflection. An image of an object is captured, a virtual light source illuminates the object within the image. In addition, the object can be the head of the user. The position of the head of the user is dynamically tracked so that an three-dimensional model is generated which is representative of the head of the user. Synthetic light is applied to a position on the model to form an illuminated model.
    Type: Grant
    Filed: September 8, 2010
    Date of Patent: December 27, 2011
    Assignee: AT&T Intellectual Property II, L.P.
    Inventors: Andrea Basso, Eric Cosatto, David Crawford Gibbon, Hans Peter Graf, Shan Liu
  • Patent number: 8078466
    Abstract: A method for generating animated sequences of talking heads in text-to-speech applications wherein a processor samples a plurality of frames comprising image samples. The processor reads first data comprising one or more parameters associated with noise-producing orifice images of sequences of at least three concatenated phonemes which correspond to an input stimulus. The processor reads, based on the first data, second data comprising images of a noise-producing entity. The processor generates an animated sequence of the noise-producing entity.
    Type: Grant
    Filed: November 30, 2009
    Date of Patent: December 13, 2011
    Assignee: AT&T Intellectual Property II, L.P.
    Inventors: Eric Cosatto, Hans Peter Graf, Juergen Schroeter
  • Publication number: 20110295520
    Abstract: A method for measuring structural entropy of cell nuclei in a histological micrograph of a biopsy tissue sample involves the steps of: obtaining a dye color map from a color image of the biopsy tissue sample; locating cell nuclei in the dye color map; and measuring structural features within small groups (cliques or paths) of cell nuclei to determine their degree of organization (or structural entropy). Also, an apparatus for measuring structural entropy of cell nuclei in a histological micrograph of a biopsy tissue sample includes a processor executing instructions for: obtaining a dye color map of the biopsy tissue sample; locating cell nuclei in the dye color map; and measuring structural features within small groups (cliques or paths) of cell nuclei to determine their degree of organization (or structural entropy).
    Type: Application
    Filed: May 25, 2011
    Publication date: December 1, 2011
    Applicant: NEC LABORATORIES AMERICA, INC.
    Inventors: Eric Cosatto, Christopher D. Malon, Hans Peter Graf
  • Patent number: 7990384
    Abstract: A system and method for generating photo-realistic talking-head animation from a text input utilizes an audio-visual unit selection process. The lip-synchronization is obtained by optimally selecting and concatenating variable-length video units of the mouth area. The unit selection process utilizes the acoustic data to determine the target costs for the candidate images and utilizes the visual data to determine the concatenation costs. The image database is prepared in a hierarchical fashion, including high-level features (such as a full 3D modeling of the head, geometric size and position of elements) and pixel-based, low-level features (such as a PCA-based metric for labeling the various feature bitmaps).
    Type: Grant
    Filed: September 15, 2003
    Date of Patent: August 2, 2011
    Assignee: AT&T Intellectual Property II, L.P.
    Inventors: Eric Cosatto, Hans Peter Graf, Gerasimos Potamianos, Juergen Schroeter
  • Patent number: 7949109
    Abstract: A computing device and computer-readable medium storing instructions for controlling a computing device to customize a voice in a multi-media message created by a sender for a recipient, the multi-media message comprising a text message from the sender to be delivered by an animated entity. The instructions comprise receiving from the sender inserted voice emoticons, which may be repeated, into the text message associated with parameters of a voice used by an animated entity to deliver the text message; and transmitting the text message such that a recipient device can deliver the multi-media message at a variable level associated with a number of times a respective voice emoticon is repeated.
    Type: Grant
    Filed: December 29, 2009
    Date of Patent: May 24, 2011
    Assignee: AT&T Intellectual Property II, L.P.
    Inventors: Joern Ostermann, Mehmat Reha Civanlar, Hans Peter Graf, Thomas M. Isaacson
  • Publication number: 20110119467
    Abstract: Systems and methods for massively parallel processing on an accelerator that includes a plurality of processing cores. Each processing core includes multiple processing chains configured to perform parallel computations, each of which includes a plurality of interconnected processing elements. The cores further include multiple of smart memory blocks configured to store and process data, each memory block accepting the output of one of the plurality of processing chains. The cores communicate with at least one off-chip memory bank.
    Type: Application
    Filed: July 26, 2010
    Publication date: May 19, 2011
    Applicant: NEC Laboratories America, Inc.
    Inventors: Srihari Cadambi, Abhinandan Majumdar, Michela Becchi, Srimat Chakradhar, Hans Peter Graf
  • Patent number: 7933772
    Abstract: A system and method for generating a video sequence having mouth movements synchronized with speech sounds are disclosed. The system utilizes a database of n-phones as the smallest selectable unit, wherein n is larger than 1 and preferably 3. The system calculates a target cost for each candidate n-phone for a target frame using a phonetic distance, coarticulation parameter, and speech rate. For each n-phone in a target sequence, the system searches for candidate n-phones that are visually similar according to the target cost. The system samples each candidate n-phone to get a same number of frames as in the target sequence and builds a video frame lattice of candidate video frames. The system assigns a joint cost to each pair of adjacent frames and searches the video frame lattice to construct the video sequence by finding the optimal path through the lattice according to the minimum of the sum of the target cost and the joint cost over the sequence.
    Type: Grant
    Filed: March 19, 2008
    Date of Patent: April 26, 2011
    Assignee: AT&T Intellectual Property II, L.P.
    Inventors: Eric Cosatto, Hans Peter Graf, Fu Jie Huang
  • Patent number: 7921013
    Abstract: A system and method of providing sender-customization of multi-media messages through the use of emoticons is disclosed. The sender inserts the emoticons into a text message. As an animated face audibly delivers the text, emoticons associated with the message are started a predetermined period of time or number of words prior to the position of the emoticon in the message text and completed a predetermined length of time or number of words following the location of the emoticon. The sender may insert emoticons through the use of emoticon buttons that are icons available for choosing. Upon sender selections of an emoticon, an icon representing the emoticon is inserted into the text at the position of the cursor. Once an emoticon is chosen, the sender may also choose the amplitude for the emoticon and increased or decreased amplitude will be displayed in the icon inserted into the message text.
    Type: Grant
    Filed: August 30, 2005
    Date of Patent: April 5, 2011
    Assignee: AT&T Intellectual Property II, L.P.
    Inventors: Joern Ostermann, Mehmet Reha Civanlar, Eric Cosatto, Hans Peter Graf, Yann Andre LeCun