Patents by Inventor Ananth Nagaraja Iyer

Ananth Nagaraja Iyer has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11694697
    Abstract: A system and method are presented for the correction of packet loss in audio in automatic speech recognition (ASR) systems. Packet loss correction, as presented herein, occurs at the recognition stage without modifying any of the acoustic models generated during training. The behavior of the ASR engine in the absence of packet loss is thus not altered. To accomplish this, the actual input signal may be rectified, the recognition scores may be normalized to account for signal errors, and a best-estimate method using information from previous frames and acoustic models may be used to replace the noisy signal.
    Type: Grant
    Filed: June 29, 2020
    Date of Patent: July 4, 2023
    Inventors: Srinath Cheluvaraja, Ananth Nagaraja Iyer, Aravind Ganapathiraju, Felix Immanuel Wyss
  • Patent number: 11574642
    Abstract: A system and method are presented for the correction of packet loss in audio in automatic speech recognition (ASR) systems. Packet loss correction, as presented herein, occurs at the recognition stage without modifying any of the acoustic models generated during training. The behavior of the ASR engine in the absence of packet loss is thus not altered. To accomplish this, the actual input signal may be rectified, the recognition scores may be normalized to account for signal errors, and a best-estimate method using information from previous frames and acoustic models may be used to replace the noisy signal.
    Type: Grant
    Filed: June 29, 2020
    Date of Patent: February 7, 2023
    Inventors: Srinath Cheluvaraja, Ananth Nagaraja Iyer, Aravind Ganapathiraju, Felix Immanuel Wyss
  • Patent number: 11294955
    Abstract: A system and method are presented for optimization of audio fingerprint search. In an embodiment, the audio fingerprints are organized into a recursive tree with different branches containing fingerprint sets that are dissimilar to each other. The tree is constructed using a clustering algorithm based on a similarity measure. The similarity measure may comprise a Hamming distance for a binary fingerprint or a Euclidean distance for continuous valued fingerprints. In another embodiment, each fingerprint is stored at a plurality of resolutions and clustering is performed hierarchically. The recognition of an incoming fingerprint begins from the root of the tree and proceeds down its branches until a match or mismatch is declared. In yet another embodiment, a fingerprint definition is generalized to include more detailed audio information than in the previous definition.
    Type: Grant
    Filed: April 8, 2019
    Date of Patent: April 5, 2022
    Inventors: Srinath Cheluvaraja, Ananth Nagaraja Iyer, Felix Immanuel Wyss
  • Publication number: 20200335110
    Abstract: A system and method are presented for the correction of packet loss in audio in automatic speech recognition (ASR) systems. Packet loss correction, as presented herein, occurs at the recognition stage without modifying any of the acoustic models generated during training. The behavior of the ASR engine in the absence of packet loss is thus not altered. To accomplish this, the actual input signal may be rectified, the recognition scores may be normalized to account for signal errors, and a best-estimate method using information from previous frames and acoustic models may be used to replace the noisy signal.
    Type: Application
    Filed: June 29, 2020
    Publication date: October 22, 2020
    Applicant: GENESYS TELECOMMUNICATIONS LABORATORIES, INC.
    Inventors: SRINATH CHELUVARAJA, ANANTH NAGARAJA IYER, ARAVIND GANAPATHIRAJU, FELIX IMMANUEL WYSS
  • Patent number: 10789962
    Abstract: A system and method are presented for the correction of packet loss in audio in automatic speech recognition (ASR) systems. Packet loss correction, as presented herein, occurs at the recognition stage without modifying any of the acoustic models generated during training. The behavior of the ASR engine in the absence of packet loss is thus not altered. To accomplish this, the actual input signal may be rectified, the recognition scores may be normalized to account for signal errors, and a best-estimate method using information from previous frames and acoustic models may be used to replace the noisy signal.
    Type: Grant
    Filed: November 12, 2018
    Date of Patent: September 29, 2020
    Inventors: Srinath Cheluvaraja, Ananth Nagaraja Iyer, Aravind Ganapathiraju, Felix Immanuel Wyss
  • Patent number: 10614814
    Abstract: Technologies for authenticating a speaker in a voice authentication system using voice biometrics include a speech collection computing device and a speech authentication computing device. The speech collection computing device is configured to collect a speech signal from a speaker and transmit the speech signal to the speech authentication computing device. The speech authentication computing device is configured to compute a speech signal feature vector for the received speech signal, retrieve a speech signal classifier associated with the speaker, and feed the speech signal feature vector to the retrieved speech signal classifier. Additionally, the speech authentication computing device is configured to determine whether the speaker is an authorized speaker based on an output of the retrieved speech signal classifier. Additional embodiments are described herein.
    Type: Grant
    Filed: June 2, 2017
    Date of Patent: April 7, 2020
    Inventors: Rajesh Dachiraju, Aravind Ganapathiraju, Ananth Nagaraja Iyer, Felix Immanuel Wyss
  • Patent number: 10535000
    Abstract: A method for training a neural network of a neural network based speaker classifier for use in speaker change detection. The method comprises: a) preprocessing input speech data; b) extracting a plurality of feature frames from the preprocessed input speech data; c) normalizing the extracted feature frames of each speaker within the preprocessed input speech data with each speaker's mean and variance; d) concatenating the normalized feature frames to form overlapped longer frames having a frame length and a hop size; e) inputting the overlapped longer frames to the neural network based speaker classifier; and f) training the neural network through forward-backward propagation.
    Type: Grant
    Filed: October 6, 2017
    Date of Patent: January 14, 2020
    Inventors: Zhenhao Ge, Ananth Nagaraja Iyer, Srinath Cheluvaraja, Aravind Ganapathiraju
  • Publication number: 20190236101
    Abstract: A system and method are presented for optimization of audio fingerprint search. In an embodiment, the audio fingerprints are organized into a recursive tree with different branches containing fingerprint sets that are dissimilar to each other. The tree is constructed using a clustering algorithm based on a similarity measure. The similarity measure may comprise a Hamming distance for a binary fingerprint or a Euclidean distance for continuous valued fingerprints. In another embodiment, each fingerprint is stored at a plurality of resolutions and clustering is performed hierarchically. The recognition of an incoming fingerprint begins from the root of the tree and proceeds down its branches until a match or mismatch is declared. In yet another embodiment, a fingerprint definition is generalized to include more detailed audio information than in the previous definition.
    Type: Application
    Filed: April 8, 2019
    Publication date: August 1, 2019
    Inventors: Srinath Cheluvaraja, Ananth Nagaraja Iyer, Felix Immanuel Wyss
  • Patent number: 10303800
    Abstract: A system and method are presented for optimization of audio fingerprint search. In an embodiment, the audio fingerprints are organized into a recursive tree with different branches containing fingerprint sets that are dissimilar to each other. The tree is constructed using a clustering algorithm based on a similarity measure. The similarity measure may comprise a Hamming distance for a binary fingerprint or a Euclidean distance for continuous valued fingerprints. In another embodiment, each fingerprint is stored at a plurality of resolutions and clustering is performed hierarchically. The recognition of an incoming fingerprint begins from the root of the tree and proceeds down its branches until a match or mismatch is declared. In yet another embodiment, a fingerprint definition is generalized to include more detailed audio information than in the previous definition.
    Type: Grant
    Filed: March 3, 2015
    Date of Patent: May 28, 2019
    Inventors: Srinath Cheluvaraja, Ananth Nagaraja Iyer, Felix Immanuel Wyss
  • Patent number: 10283112
    Abstract: A system and method are presented for neural network based feature extraction for acoustic model development. A neural network may be used to extract acoustic features from raw MFCCs or the spectrum, which are then used for training acoustic models for speech recognition systems. Feature extraction may be performed by optimizing a cost function used in linear discriminant analysis. General non-linear functions generated by the neural network are used for feature extraction. The transformation may be performed using a cost function from linear discriminant analysis methods which perform linear operations on the MFCCs and generate lower dimensional features for speech recognition. The extracted acoustic features may then be used for training acoustic models for speech recognition systems.
    Type: Grant
    Filed: February 26, 2018
    Date of Patent: May 7, 2019
    Inventors: Srinath Cheluvaraja, Ananth Nagaraja Iyer
  • Publication number: 20190080701
    Abstract: A system and method are presented for the correction of packet loss in audio in automatic speech recognition (ASR) systems. Packet loss correction, as presented herein, occurs at the recognition stage without modifying any of the acoustic models generated during training. The behavior of the ASR engine in the absence of packet loss is thus not altered. To accomplish this, the actual input signal may be rectified, the recognition scores may be normalized to account for signal errors, and a best-estimate method using information from previous frames and acoustic models may be used to replace the noisy signal.
    Type: Application
    Filed: November 12, 2018
    Publication date: March 14, 2019
    Inventors: Srinath Cheluvaraja, Ananth Nagaraja Iyer, Aravind Ganapathiraju, Felix Immanuel Wyss
  • Patent number: 10157620
    Abstract: A system and method are presented for the correction of packet loss in audio in automatic speech recognition (ASR) systems. Packet loss correction, as presented herein, occurs at the recognition stage without modifying any of the acoustic models generated during training. The behavior of the ASR engine in the absence of packet loss is thus not altered. To accomplish this, the actual input signal may be rectified, the recognition scores may be normalized to account for signal errors, and a best-estimate method using information from previous frames and acoustic models may be used to replace the noisy signal.
    Type: Grant
    Filed: March 4, 2015
    Date of Patent: December 18, 2018
    Inventors: Srinath Cheluvaraja, Ananth Nagaraja Iyer, Aravind Ganapathiraju, Felix Immanuel Wyss
  • Publication number: 20180190267
    Abstract: A system and method are presented for neural network based feature extraction for acoustic model development. A neural network may be used to extract acoustic features from raw MFCCs or the spectrum, which are then used for training acoustic models for speech recognition systems. Feature extraction may be performed by optimizing a cost function used in linear discriminant analysis. General non-linear functions generated by the neural network are used for feature extraction. The transformation may be performed using a cost function from linear discriminant analysis methods which perform linear operations on the MFCCs and generate lower dimensional features for speech recognition. The extracted acoustic features may then be used for training acoustic models for speech recognition systems.
    Type: Application
    Filed: February 26, 2018
    Publication date: July 5, 2018
    Inventors: Srinath Cheluvaraja, Ananth Nagaraja Iyer
  • Patent number: 9972310
    Abstract: A system and method are presented for neural network based feature extraction for acoustic model development. A neural network may be used to extract acoustic features from raw MFCCs or the spectrum, which are then used for training acoustic models for speech recognition systems. Feature extraction may be performed by optimizing a cost function used in linear discriminant analysis. General non-linear functions generated by the neural network are used for feature extraction. The transformation may be performed using a cost function from linear discriminant analysis methods which perform linear operations on the MFCCs and generate lower dimensional features for speech recognition. The extracted acoustic features may then be used for training acoustic models for speech recognition systems.
    Type: Grant
    Filed: December 31, 2015
    Date of Patent: May 15, 2018
    Assignee: Interactive Intelligence Group, Inc.
    Inventors: Srinath Cheluvaraja, Ananth Nagaraja Iyer
  • Publication number: 20180039888
    Abstract: A method for training a neural network of a neural network based speaker classifier for use in speaker change detection. The method comprises: a) preprocessing input speech data; b) extracting a plurality of feature frames from the preprocessed input speech data; c) normalizing the extracted feature frames of each speaker within the preprocessed input speech data with each speaker's mean and variance; d) concatenating the normalized feature frames to form overlapped longer frames having a frame length and a hop size; e) inputting the overlapped longer frames to the neural network based speaker classifier; and f) training the neural network through forward-backward propagation.
    Type: Application
    Filed: October 6, 2017
    Publication date: February 8, 2018
    Inventors: ZHENHAO GE, ANANTH NAGARAJA IYER, SRINATH CHELUVARAJA, ARAVIND GANAPATHIRAJU
  • Publication number: 20170352353
    Abstract: Technologies for authenticating a speaker in a voice authentication system using voice biometrics include a speech collection computing device and a speech authentication computing device. The speech collection computing device is configured to collect a speech signal from a speaker and transmit the speech signal to the speech authentication computing device. The speech authentication computing device is configured to compute a speech signal feature vector for the received speech signal, retrieve a speech signal classifier associated with the speaker, and feed the speech signal feature vector to the retrieved speech signal classifier. Additionally, the speech authentication computing device is configured to determine whether the speaker is an authorized speaker based on an output of the retrieved speech signal classifier. Additional embodiments are described herein.
    Type: Application
    Filed: June 2, 2017
    Publication date: December 7, 2017
    Inventors: Rajesh Dachiraju, Aravind Ganapathiraju, Ananth Nagaraja Iyer, Felix Immanuel Wyss
  • Patent number: 9767792
    Abstract: A system and method for learning alternate pronunciations for speech recognition is disclosed. Alternative name pronunciations may be covered, through pronunciation learning, that have not been previously covered in a general pronunciation dictionary. In an embodiment, the detection of phone-level and syllable-level mispronunciations in words and sentences may be based on acoustic models trained by Hidden Markov Models. Mispronunciations may be detected by comparing the likelihood of the potential state of the targeting pronunciation unit with a pre-determined threshold through a series of tests. It is also within the scope of an embodiment to detect accents.
    Type: Grant
    Filed: October 12, 2016
    Date of Patent: September 19, 2017
    Assignee: Interactive Intelligence Group, Inc.
    Inventors: Zhenhao Ge, Vivek Tyagi, Aravind Ganapathiraju, Ananth Nagaraja Iyer, Scott Allen Randal, Felix Immanuel Wyss
  • Publication number: 20170193988
    Abstract: A system and method are presented for neural network based feature extraction for acoustic model development. A neural network may be used to extract acoustic features from raw MFCCs or the spectrum, which are then used for training acoustic models for speech recognition systems. Feature extraction may be performed by optimizing a cost function used in linear discriminant analysis. General non-linear functions generated by the neural network are used for feature extraction. The transformation may be performed using a cost function from linear discriminant analysis methods which perform linear operations on the MFCCs and generate lower dimensional features for speech recognition. The extracted acoustic features may then be used for training acoustic models for speech recognition systems.
    Type: Application
    Filed: December 31, 2015
    Publication date: July 6, 2017
    Inventors: Srinath Cheluvaraja, Ananth Nagaraja Iyer
  • Patent number: 9672815
    Abstract: A system and method are presented for real-time speech analytics in the speech analytics field. Real time audio is fed along with a keyword model, into a recognition engine. The recognition engine computes the probability of the audio stream data matching keywords in the keyword model. The probability is compared to a threshold where the system determines if the probability is indicative of whether or not the keyword has been spotted. Empirical metrics are computed and any false alarms are identified and rejected. The keyword may be reported as found when it is deemed not to be a false alarm and passes the threshold for detection.
    Type: Grant
    Filed: July 20, 2012
    Date of Patent: June 6, 2017
    Assignee: Interactive Intelligence Group, Inc.
    Inventors: Aravind Ganapathiraju, Ananth Nagaraja Iyer
  • Patent number: 9628141
    Abstract: A system and method are presented for acoustic echo cancellation. The echo canceller performs reduction of acoustic and hybrid echoes which may arise in a situation such as a long-distance conference call with multiple speakers in varying environments, for example. Echo cancellation, in at least one embodiment, may be based on similarity measurement, statistical determination of echo cancellation parameters from historical values, frequency domain operation, double talk detection, packet loss detection, signal detection, and noise subtraction.
    Type: Grant
    Filed: October 22, 2013
    Date of Patent: April 18, 2017
    Assignee: Interactive Intelligence Group, Inc.
    Inventors: Felix Immanuel Wyss, Rivarol Vergin, Ananth Nagaraja Iyer, Aravind Ganapathiraju, Kevin Charles Vlack, Srinath Cheluvaraja