Patents by Inventor Chalapathy V. Neti

Chalapathy V. Neti has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 10438499
    Abstract: Methods and arrangements for identifying burden comprehension in multimedia content. A contemplated method includes: accepting multimedia input; detecting components of the multimedia input; determining a comprehension burden score of each of the detected components; and thereupon calculating a total comprehension burden score for the multimedia input. Other variants and embodiments are broadly contemplated herein.
    Type: Grant
    Filed: August 1, 2014
    Date of Patent: October 8, 2019
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Danish Contractor, Mukesh Kumar Mohania, Sumit Negi, Chalapathy V. Neti, Nitendra Rajput
  • Publication number: 20160034816
    Abstract: Methods and arrangements for identifying burden comprehension in multimedia content. A contemplated method includes: accepting multimedia input; detecting components of the multimedia input; determining a comprehension burden score of each of the detected components; and thereupon calculating a total comprehension burden score for the multimedia input. Other variants and embodiments are broadly contemplated herein.
    Type: Application
    Filed: August 1, 2014
    Publication date: February 4, 2016
    Inventors: Danish Contractor, Mukesh Kumar Mohania, Sumit Negi, Chalapathy V. Neti, Nitendra Rajput
  • Patent number: 7793212
    Abstract: A manual annotation system of multi-modal characteristics in multimedia files. There is provided an arrangement for selection an observation modality of video with audio, video without audio, audio with video, or audio without video, to be used to annotate multimedia content. While annotating video or audio features is isolation results in less confidence in the identification of features, observing both audio and video simultaneously and annotating that observation results in a higher confidence level.
    Type: Grant
    Filed: December 19, 2003
    Date of Patent: September 7, 2010
    Assignee: International Business Machines Corporation
    Inventors: Hugh W. Adams, Jr., Giridharen Iyengar, Ching-Yung Lin, Chalapathy V. Neti, John R. Smith, Belle L. Tseng
  • Patent number: 7664637
    Abstract: An arrangement for yielding enhanced audio features towards the provision of enhanced audio-visual features for speech recognition. Input is provided in the form of noisy audio-visual features and noisy audio features related to the noisy audio-visual features.
    Type: Grant
    Filed: October 31, 2007
    Date of Patent: February 16, 2010
    Assignee: Nuance Communications, Inc.
    Inventors: Sabine Deligne, Chalapathy V. Neti, Gerasimos Potamianos
  • Publication number: 20080141357
    Abstract: Methods and apparatus for enabling a Pervasive Authentication Domain. A Pervasive Authentication Domain allows many registered Pervasive Devices to obtain authentication credentials from a single Personal Authentication Gateway and to use these credentials on behalf of users to enable additional capabilities for the devices. It provides an arrangement for a user to store credentials in one device (the Personal Authentication Gateway), and then make use of those credentials from many authorized Pervasive Devices without re-entering the credentials. It provides a convenient way for a user to share credentials among many devices, particularly when it is not convenient to enter credentials as in a smart wristwatch environment. It further provides an arrangement for disabling access to credentials to devices that appear to be far from the Personal Authentication Gateway as measured by metrics such as communications signal strengths.
    Type: Application
    Filed: October 31, 2007
    Publication date: June 12, 2008
    Inventors: Sabine Deligne, Chalapathy V. Neti, Gerasimos Potamianos
  • Patent number: 7319955
    Abstract: An arrangement for yielding enhanced audio features towards the provision of enhanced audio-visual features for speech recognition. Input is provided in the form of noisy audio-visual features and noisy audio features related to the noisy audio-visual features.
    Type: Grant
    Filed: November 29, 2002
    Date of Patent: January 15, 2008
    Assignee: International Business Machines Corporation
    Inventors: Sabine Deligne, Chalapathy V. Neti, Gerasimos Potamianos
  • Patent number: 7046300
    Abstract: The use of multiple complementary classes of measure to assess face and speech consistency in video. In an exemplary embodiment, both synchrony measures and plausibility measures are employed.
    Type: Grant
    Filed: November 29, 2002
    Date of Patent: May 16, 2006
    Assignee: International Business Machines Corporation
    Inventors: Giridharan Iyengar, Chalapathy V. Neti, Harriet J. Nock
  • Publication number: 20040123231
    Abstract: A manual annotation system of multi-modal characteristics in multimedia files. There is provided an arrangement for selection an observation modality of video with audio, video without audio, audio with video, or audio without video, to be used to annotate multimedia content. While annotating video or audio features in isolation results in less confidence in the identification of the features, observing both audio and video simultaneously and annotating that observation results in a higher confidence level.
    Type: Application
    Filed: December 20, 2002
    Publication date: June 24, 2004
    Inventors: Hugh W. Adams, Giridharan Iyengar, Ching-Yung Lin, Chalapathy V. Neti, John R. Smith, Belle L. Tseng
  • Patent number: 6754373
    Abstract: A system for activating a microphone based on visual speech cues, in accordance with the invention, includes a feature tracker coupled to an image acquisition device. The feature tracker tracks features in an image of a user. A region of interest extractor is coupled to the feature tracker. The region of interest extractor extracts a region of interest from the image of the user. A visual speech activity detector is coupled to the region of interest extractor and measures changes in the region of interest to determine if a visual speech cue has been generated by the user. A microphone is turned on by the visual speech activity detector when a visual speech cue has been determined by the visual speech activity detector. Methods for activating a microphone based on visual speech cues are also included.
    Type: Grant
    Filed: July 14, 2000
    Date of Patent: June 22, 2004
    Assignee: International Business Machines Corporation
    Inventors: Philippe de Cuetos, Giridharan R. Iyengar, Chalapathy V. Neti, Gerasimos Potamianos
  • Publication number: 20040107098
    Abstract: An arrangement for yielding enhanced audio features towards the provision of enhanced audio-visual features for speech recognition. Input is provided in the form of noisy audio-visual features and noisy audio features related to the noisy audio-visual features.
    Type: Application
    Filed: November 29, 2002
    Publication date: June 3, 2004
    Applicant: IBM Corporation
    Inventors: Sabine Deligne, Chalapathy V. Neti, Gerasimos Potamianos
  • Publication number: 20040107103
    Abstract: The use of multiple complementary classes of measure to assess face and speech consistency in video. In an exemplary embodiment, both synchrony measures and plausibility measures are employed.
    Type: Application
    Filed: November 29, 2002
    Publication date: June 3, 2004
    Applicant: IBM Corporation
    Inventors: Giridharan Iyengar, Chalapathy V. Neti, Harriet J. Nock
  • Patent number: 6377913
    Abstract: In accordance with the invention, a method and system for accessing a dialog system employing a plurality of different clients, includes providing a first client device for accessing a conversational system and presenting a command to the conversational system by converting the command to a form understandable to the conversational system. The command is interpreted by employing a mediator, a dialog manager and a multi-modal history to determine the intent of the command based on a context of the command. A second client device is determined based on a predetermined device preference stored in the conversational system. An application is abstracted to perform the command, and the results of the performance of the command are set to the second client device.
    Type: Grant
    Filed: August 13, 1999
    Date of Patent: April 23, 2002
    Assignee: International Business Machines Corporation
    Inventors: Daniel M. Coffman, Popani Gopalakrishnan, Ganesh N. Ramaswamy, Jan Kleindienst, Chalapathy V. Neti
  • Patent number: 6366885
    Abstract: A method of speech driven lip synthesis which applies viseme based training models to units of visual speech. The audio data is grouped into a smaller number of visually distinct visemes rather than the larger number of phonemes. These visemes then form the basis for a Hidden Markov Model (HMM) state sequence or the output nodes of a neural network. During the training phase, audio and visual features are extracted from input speech, which is then aligned according to the apparent viseme sequence with the corresponding audio features being used to calculate the HMM state output probabilities or the output of the neutral network. During the synthesis phase, the acoustic input is aligned with the most likely viseme HMM sequence (in the case of an HMM based model) or with the nodes of the network (in the case of a neural network based system), which is then used for animation.
    Type: Grant
    Filed: August 27, 1999
    Date of Patent: April 2, 2002
    Assignee: International Business Machines Corporation
    Inventors: Sankar Basu, Tanveer Atzal Faruquie, Chalapathy V. Neti, Nitendra Rajput, Andrew William Senior, L. Venkata Subramaniam, Ashish Verma
  • Patent number: 5768474
    Abstract: A method for noise-robust speech processing with cochlea filters within a computer system is disclosed. This invention provides a method for producing feature vectors from a segment of speech, that is more robust to variations in the environment due to additive noise. A first output is produced by convolving a speech signal input with spatially dependent impulse responses that resemble cochlea filters. The temporal transient and the spatial transient of the first output is then enhanced by taking a time derivative and a spatial derivative, respectively, of the first output to produce a second output. Next, all the negative values of the second output are replaced with zeros. A feature vector is then obtained from each frame of the second output by a multiple resolution extraction.
    Type: Grant
    Filed: December 29, 1995
    Date of Patent: June 16, 1998
    Assignee: International Business Machines Corporation
    Inventor: Chalapathy V. Neti