Patents by Inventor Sarangarajan Parthasarathy
Sarangarajan Parthasarathy has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 8082147Abstract: A system and method of updating automatic speech recognition parameters on a mobile device are disclosed. The method comprises storing user account-specific adaptation data associated with ASR on a computing device associated with a wireless network, generating new ASR adaptation parameters based on transmitted information from the mobile device when a communication channel between the computing device and the mobile device becomes available and transmitting the new ASR adaptation data to the mobile device when a communication channel between the computing device and the mobile device becomes available. The new ASR adaptation data on the mobile device more accurately recognizes user utterances.Type: GrantFiled: October 8, 2010Date of Patent: December 20, 2011Assignee: AT&T Intellectual Property II, L.P.Inventors: Sarangarajan Parthasarathy, Richard Cameron Rose
-
Publication number: 20110218808Abstract: A system and method for non-speech input or keypad-aided word and spelling recognition is disclosed. The method includes generating an unweighted grammar, selecting a database of words, generating a weighted grammar using the unweighted grammar and a statistical letter model trained on the database of words, receiving speech from a user after receiving the non-speech input and after generating the weighted grammar, and performing automatic speech recognition on the speech and non-speech input using the weighted grammar If a confidence is below a predetermined level, then the method includes receiving non-speech input from the user, disambiguating possible spellings by generating a letter lattice based on a user input modality, and constraining the letter lattice and generating a new letter string of possible word spellings until a letter string is correctly recognized.Type: ApplicationFiled: May 17, 2011Publication date: September 8, 2011Applicant: AT&T Intellectual Property II, LPInventor: Sarangarajan PARTHASARATHY
-
Patent number: 7949528Abstract: A system and method for non-speech input or keypad-aided word and spelling recognition is disclosed. The method includes generating an unweighted grammar, selecting a database of words, generating a weighted grammar using the unweighted grammar and a statistical letter model trained on the database of words, receiving speech from a user after receiving the non-speech input and after generating the weighted grammar, and performing automatic speech recognition on the speech and non-speech input using the weighted grammar. If a confidence is below a predetermined level, then the method includes receiving non-speech input from the user, disambiguating possible spellings by generating a letter lattice based on a user input modality, and constraining the letter lattice and generating a new letter string of possible word spellings until a letter string is correctly recognized.Type: GrantFiled: July 22, 2009Date of Patent: May 24, 2011Assignee: AT&T Intellectual Property II, L.P.Inventor: Sarangarajan Parthasarathy
-
Patent number: 7930179Abstract: Systems and methods for unsupervised segmentation of multi-speaker speech or audio data by speaker. A front-end analysis is applied to input speech data to obtain feature vectors. The speech data is initially segmented and then clustered into groups of segments that correspond to different speakers. The clusters are iteratively modeled and resegmented to obtain stable speaker segmentations. The overlap between segmentation sets is checked to ensure successful speaker segmentation. Overlapping segments are combined and remodeled and resegmented. Optionally, the speech data is processed to produce a segmentation lattice to maximize the overall segmentation likelihood.Type: GrantFiled: October 2, 2007Date of Patent: April 19, 2011Assignee: AT&T Intellectual Property II, L.P.Inventors: Allen Louis Gorin, Zhu Liu, Sarangarajan Parthasarathy, Aaron Edward Rosenberg
-
Publication number: 20110029307Abstract: A system and method of updating automatic speech recognition parameters on a mobile device are disclosed. The method comprises storing user account-specific adaptation data associated with ASR on a computing device associated with a wireless network, generating new ASR adaptation parameters based on transmitted information from the mobile device when a communication channel between the computing device and the mobile device becomes available and transmitting the new ASR adaptation data to the mobile device when a communication channel between the computing device and the mobile device becomes available. The new ASR adaptation data on the mobile device more accurately recognizes user utterances.Type: ApplicationFiled: October 8, 2010Publication date: February 3, 2011Applicant: AT&T Intellectual Property II, L.P. via transfer form AT&T Corp.Inventors: Sarangarajan Parthasarathy, Richard Cameron Rose
-
Patent number: 7822603Abstract: A system and method of updating automatic speech recognition parameters on a mobile device are disclosed. The method comprises storing user account-specific adaptation data associated with ASR on a computing device associated with a wireless network, generating new ASR adaptation parameters based on transmitted information from the mobile device when a communication channel between the computing device and the mobile device becomes available and transmitting the new ASR adaptation data to the mobile device when a communication channel between the computing device and the mobile device becomes available. The new ASR adaptation data on the mobile device more accurately recognizes user utterances.Type: GrantFiled: May 5, 2008Date of Patent: October 26, 2010Assignee: AT&T Intellectual Property II, L.P.Inventors: Sarangarajan Parthasarathy, Richard Cameron Rose
-
Publication number: 20100211396Abstract: A digital speech enabled middleware module is disclosed that facilitates interaction between a large number of client devices and network-based automatic speech recognition (ASR) resources. The module buffers feature vectors associated with speech received from the client devices when the number of client devices is greater than the available ASR resources. When an ASR decoder becomes available, the module transmits the feature vectors to the ASR decoder and a recognition result is returned.Type: ApplicationFiled: May 3, 2010Publication date: August 19, 2010Applicant: AT&T Intellectual Property II, LP via transfer from AT&T Corp.Inventors: Iker Arizmendi, Sarangarajan Parthasarathy, Richard Cameron Rose
-
Publication number: 20100166157Abstract: The invention provides a system and method for indexing and organizing voice mail message by the speaker of the message. One or more speaker models are created from voice mail messages received. As additional messages are left, each of the new messages are compared with existing speaker models to determine the identity of the callers of each of the new messages. The voice mail messages are organized within a user's mailbox by caller. Unknown callers may be identified and tagged by the user and then used to create new speaker models and/or update existing speaker models.Type: ApplicationFiled: December 29, 2009Publication date: July 1, 2010Applicant: AT&T Corp.Inventors: Julia Hirschberg, Sarangarajan Parthasarathy, Aaron Edward Rosenberg, Stephen Whittaker
-
Patent number: 7711568Abstract: A digital speech enabled middleware module is disclosed that facilitates interaction between a large number of client devices and network-based automatic speech recognition (ASR) resources. The module buffers feature vectors associated with speech received from the client devices when the number of client devices is greater than the available ASR resources. When an ASR decoder becomes available, the module transmits the feature vectors to the ASR decoder and a recognition result is returned.Type: GrantFiled: April 3, 2003Date of Patent: May 4, 2010Assignee: AT&T Intellectual Property II, LPInventors: Iker Arizmendi, Sarangarajan Parthasarathy, Richard Cameron Rose
-
Patent number: 7664636Abstract: The invention provides a system and method for indexing and organizing voice mail message by the speaker of the message. One or more speaker models are created from voice mail messages received. As additional messages are left, each of the new messages are compared with existing speaker models to determine the identity of the callers of each of the new messages. The voice mail messages are organized within a user's mailbox by caller. Unknown callers may be identified and tagged by the user and then used to create new speaker models and/or update existing speaker models.Type: GrantFiled: April 17, 2000Date of Patent: February 16, 2010Assignee: AT&T Intellectual Property II, L.P.Inventors: Julia Hirschberg, Sarangarajan Parthasarathy, Aaron Edward Rosenberg, Stephen Whittaker
-
Publication number: 20090281806Abstract: A system and method for non-speech input or keypad-aided word and spelling recognition is disclosed. The method includes generating an unweighted grammar, selecting a database of words, generating a weighted grammar using the unweighted grammar and a statistical letter model trained on the database of words, receiving speech from a user after receiving the non-speech input and after generating the weighted grammar, and performing automatic speech recognition on the speech and non-speech input using the weighted grammar. If a confidence is below a predetermined level, then the method includes receiving non-speech input from the user, disambiguating possible spellings by generating a letter lattice based on a user input modality, and constraining the letter lattice and generating a new letter string of possible word spellings until a letter string is correctly recognized.Type: ApplicationFiled: July 22, 2009Publication date: November 12, 2009Applicant: AT&T Corp.Inventor: Sarangarajan Parthasarathy
-
Patent number: 7574356Abstract: A system and method for non-speech input or keypad-aided word and spelling recognition is disclosed. The method comprises performing spelling recognition via automatic speech recognition (ASR) on received speech from a user, the ASR being performed using a statistical letter model trained on domain data and producing a letter lattice RLN. If an ASR confidence is below a predetermined level, then the method comprises receiving non-speech input from the user, generating a keypad constraint grammar K and generating a letter string based on a composition of finite state transducers RLN and K. Other variations of the invention include recognizing input by first receiving non-speech input, dynamically generating an unweighted grammar, generating a weighted grammar using domain data, and then performing speech, and thus spelling, recognition on input speech using the weighted grammar.Type: GrantFiled: July 19, 2004Date of Patent: August 11, 2009Assignee: AT&T Intellectual Property II, L.P.Inventor: Sarangarajan Parthasarathy
-
Publication number: 20090006088Abstract: Speech recognition models are dynamically re-configurable based on user information, application information, background information such as background noise and transducer information such as transducer response characteristics to provide users with alternate input modes to keyboard text entry. Word recognition lattices are generated for each data field of an application and dynamically concatenated into a single word recognition lattice. A language model is applied to the concatenated word recognition lattice to determine the relationships between the word recognition lattices and repeated until the generated word recognition lattices are acceptable or differ from a predetermined value only by a threshold amount. These techniques of dynamic re-configurable speech recognition provide for deployment of speech recognition on small devices such as mobile phones and personal digital assistants as well environments such as office, home or vehicle while maintaining the accuracy of the speech recognition.Type: ApplicationFiled: September 9, 2008Publication date: January 1, 2009Applicant: AT&T Corp.Inventors: Bojana Gajic, Shrikanth Sambasivan Narayanan, Sarangarajan Parthasarathy, Richard Cameron Rose, Aaron Edward Rosenberg
-
Patent number: 7451081Abstract: Speech recognition models are dynamically re-configurable based on user information, application information, background information such as background noise and transducer information such as transducer response characteristics to provide users with alternate input modes to keyboard text entry. Word recognition lattices are generated for each data field of an application and dynamically concatenated into a single word recognition lattice. A language model is applied to the concatenated word recognition lattice to determine the relationships between the word recognition lattices and repeated until the generated word recognition lattices are acceptable or differ from a predetermined value only by a threshold amount. These techniques of dynamic re-configurable speech recognition provide for deployment of speech recognition on small devices such as mobile phones and personal digital assistants as well environments such as office, home or vehicle while maintaining the accuracy of the speech recognition.Type: GrantFiled: March 13, 2007Date of Patent: November 11, 2008Assignee: AT&T Corp.Inventors: Bojana Gajic, Shrikanth Sambasivan Narayanan, Sarangarajan Parthasarathy, Richard Cameron Rose, Aaron Edward Rosenberg
-
Patent number: 7386443Abstract: A system and method of updating automatic speech recognition parameters on a mobile device are disclosed. The method comprises storing user account-specific adaptation data associated with ASR on a computing device associated with a wireless network, generating new ASR adaptation parameters based on transmitted information from the mobile device when a communication channel between the computing device and the mobile device becomes available and transmitting the new ASR adaptation data to the mobile device when a communication channel between the computing device and the mobile device becomes available. The new ASR adaptation data on the mobile device more accurately recognizes user utterances.Type: GrantFiled: January 9, 2004Date of Patent: June 10, 2008Assignee: AT&T Corp.Inventors: Sarangarajan Parthasarathy, Richard Cameron Rose
-
Publication number: 20080015848Abstract: A digital speech enabled middleware module is disclosed that facilitates interaction between a large number of client devices and network-based automatic speech recognition (ASR) resources. The module buffers feature vectors associated with speech received from the client devices when the number of client devices is greater than the available ASR resources. When an ASR decoder becomes available, the module transmits the feature vectors to the ASR decoder and a recognition result is returned.Type: ApplicationFiled: July 18, 2007Publication date: January 17, 2008Applicant: AT&T Corp.Inventors: Iker Arizmendi, Sarangarajan Parthasarathy, Richard Rose
-
Patent number: 7295970Abstract: Systems and methods for unsupervised segmentation of multi-speaker speech or audio data by speaker. A front-end analysis is applied to input speech data to obtain feature vectors. The speech data is initially segmented and then clustered into groups of segments that correspond to different speakers. The clusters are iteratively modeled and resegmented to obtain stable speaker segmentations. The overlap between segmentation sets is checked to ensure successful speaker segmentation. Overlapping segments are combined and remodeled and resegmented. Optionally, the speech data is processed to produce a segmentation lattice to maximize the overall segmentation likelihood.Type: GrantFiled: January 24, 2003Date of Patent: November 13, 2007Assignee: AT&T CorpInventors: Allen Louis Gorin, Zhu Liu, Sarangarajan Parthasarathy, Aaron Edward Rosenberg
-
Patent number: 7209880Abstract: Speech recognition models are dynamically re-configurable based on user information, application information, background information such as background noise and transducer information such as transducer response characteristics to provide users with alternate input modes to keyboard text entry. Word recognition lattices are generated for each data field of an application and dynamically concatenated into a single word recognition lattice. A language model is applied to the concatenated word recognition lattice to determine the relationships between the word recognition lattices and repeated until the generated word recognition lattices are acceptable or differ from a predetermined value only by a threshold amount. These techniques of dynamic re-configurable speech recognition provide for deployment of speech recognition on small devices such as mobile phones and personal digital assistants as well environments such as office, home or vehicle while maintaining the accuracy of the speech recognition.Type: GrantFiled: March 6, 2002Date of Patent: April 24, 2007Assignee: AT&T Corp.Inventors: Bojana Gajic, Shrikanth Sambasivan Narayanan, Sarangarajan Parthasarathy, Richard Cameron Rose, Aaron Edward Rosenberg
-
Publication number: 20060015336Abstract: A system and method for non-speech input or keypad-aided word and spelling recognition is disclosed. The method comprises performing spelling recognition via automatic speech recognition (ASR) on received speech from a user, the ASR being performed using a statistical letter model trained on domain data and producing a letter lattice RLN. If an ASR confidence is below a predetermined level, then the method comprises receiving non-speech input from the user, generating a keypad constraint grammar K and generating a letter string based on a composition of finite state transducers RLN and K. Other variations of the invention include recognizing input by first receiving non-speech input, dynamically generating an unweighted grammar, generating a weighted grammar using domain data, and then performing speech, and thus spelling, recognition on input speech using the weighted grammar.Type: ApplicationFiled: July 19, 2004Publication date: January 19, 2006Inventor: Sarangarajan Parthasarathy
-
Publication number: 20040199393Abstract: A digital speech enabled middleware module is disclosed that facilitates interaction between a large number of client devices and network-based automatic speech recognition (ASR) resources. The module buffers feature vectors associated with speech received from the client devices when the number of client devices is greater than the available ASR resources. When an ASR decoder becomes available, the module transmits the feature vectors to the ASR decoder and a recognition result is returned.Type: ApplicationFiled: April 3, 2003Publication date: October 7, 2004Inventors: Iker Arizmendi, Sarangarajan Parthasarathy, Richard Cameron Rose