Patents by Inventor Srinivas Bangalore

Srinivas Bangalore has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 10114818
    Abstract: Disclosed herein are systems, methods, and non-transitory computer-readable storage media for bootstrapping a language translation system. A system configured to practice the method performs a bidirectional web crawl to identify a bilingual website. The system analyzes data on the bilingual website to make a classification decision about whether the root of the bilingual website is an entry point for the bilingual website. The bilingual site can contain pairs of parallel pages. Each pair can include a first website in a first language and a second website in a second language, and a first portion of the first web page corresponds to a second portion of the second web page. Then the system analyzes the first and second web pages to identify corresponding information pairs in the first and second languages, and extracts the corresponding information pairs from the first and second web pages for use in a language translation model.
    Type: Grant
    Filed: October 17, 2016
    Date of Patent: October 30, 2018
    Assignee: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: Luciano De Andrade Barbosa, Srinivas Bangalore, Vivek Kumar Rangarajan Sridhar
  • Publication number: 20180301145
    Abstract: Disclosed herein are systems, methods, and non-transitory computer-readable storage media for approximating relevant responses to a user query with voice-enabled search. A system practicing the method receives a word lattice generated by an automatic speech recognizer based on a user speech and a prosodic analysis of the user speech, generates a reweighted word lattice based on the word lattice and the prosodic analysis, approximates based on the reweighted word lattice one or more relevant responses to the query, and presents to a user the responses to the query. The prosodic analysis examines metalinguistic information of the user speech and can identify the most salient subject matter of the speech, assess how confident a speaker is in the content of his or her speech, and identify the attitude, mood, emotion, sentiment, etc. of the speaker. Other information not described in the content of the speech can also be used.
    Type: Application
    Filed: June 18, 2018
    Publication date: October 18, 2018
    Inventors: Srinivas BANGALORE, Junlan FENG, Michael JOHNSTON, Taniya MISHRA
  • Publication number: 20180268810
    Abstract: Disclosed herein are systems, methods, and non-transitory computer-readable storage media for generating domain-specific speech recognition models for a domain of interest by combining and tuning existing speech recognition models when a speech recognizer does not have access to a speech recognition model for that domain of interest and when available domain-specific data is below a minimum desired threshold to create a new domain-specific speech recognition model. A system configured to practice the method identifies a speech recognition domain and combines a set of speech recognition models, each speech recognition model of the set of speech recognition models being from a respective speech recognition domain. The system receives an amount of data specific to the speech recognition domain, wherein the amount of data is less than a minimum threshold to create a new domain-specific model, and tunes the combined speech recognition model for the speech recognition domain based on the data.
    Type: Application
    Filed: May 21, 2018
    Publication date: September 20, 2018
    Inventors: Srinivas BANGALORE, Robert BELL, Diamantino Antonio CASEIRO, Mazin GILBERT, Patrick HAFFNER
  • Publication number: 20180261206
    Abstract: Disclosed herein are systems, computer-implemented methods, and computer-readable media for dialog modeling. The method includes receiving spoken dialogs annotated to indicate dialog acts and task/subtask information, parsing the spoken dialogs with a hierarchical, parse-based dialog model which operates incrementally from left to right and which only analyzes a preceding dialog context to generate parsed spoken dialogs, and constructing a functional task structure of the parsed spoken dialogs. The method can further either interpret user utterances with the functional task structure of the parsed spoken dialogs or plan system responses to user utterances with the functional task structure of the parsed spoken dialogs. The parse-based dialog model can be a shift-reduce model, a start-complete model, or a connection path model.
    Type: Application
    Filed: May 15, 2018
    Publication date: September 13, 2018
    Inventors: Amanda STENT, Srinivas BANGALORE
  • Patent number: 10042877
    Abstract: Information is aggregated and made available to users. A system monitors over the internet a first set of external information sources for a first user based on instructions from a first user profile that specifies information to aggregate for the first user. The system detects, based on the monitoring, new data at one of the first set of information sources. The system obtains the new data at the one of the first set of information sources, independent of preferences of the one of the first set of information sources. The system updates aggregated information for the first user with the new data from the one of the first set of information sources. The updated aggregated information for the first user is made available to the first user.
    Type: Grant
    Filed: June 5, 2015
    Date of Patent: August 7, 2018
    Assignee: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: Junlan Feng, Srinivas Bangalore, Michael James Robert Johnston, Taniya Mishra
  • Publication number: 20180197566
    Abstract: Disclosed herein are systems, methods, and computer-readable storage media for improving speech recognition accuracy using textual context. The method includes retrieving a recorded utterance, capturing text from a device display associated with the spoken dialog and viewed by one party to the recorded utterance, and identifying words in the captured text that are relevant to the recorded utterance. The method further includes adding the identified words to a dynamic language model, and recognizing the recorded utterance using the dynamic language model. The recorded utterance can be a spoken dialog. A time stamp can be assigned to each identified word. The method can include adding identified words to and/or removing identified words from the dynamic language model based on their respective time stamps. A screen scraper can capture text from the device display associated with the recorded utterance. The device display can contain customer service data.
    Type: Application
    Filed: March 5, 2018
    Publication date: July 12, 2018
    Inventors: Dan MELAMED, Srinivas Bangalore, Michael JOHNSTON
  • Publication number: 20180174578
    Abstract: A natural language processing system has a hierarchy of user intents related to a domain of interest, the hierarchy having specific intents corresponding to leaf nodes of the hierarchy, and more general intents corresponding to ancestor nodes of the leaf nodes. The system also has a trained understanding model that can classify natural language utterances according to user intent. When the understanding model cannot determine with sufficient confidence that a natural language utterance corresponds to one of the specific intents, the natural language processing system traverses the hierarchy of intents to find a more general user intent that is related to the most applicable specific intent of the utterance and for which there is sufficient confidence. The general intent can then be used to prompt the user with questions applicable to the general intent to obtain the missing information needed for a specific intent.
    Type: Application
    Filed: December 19, 2016
    Publication date: June 21, 2018
    Inventors: SRINIVAS BANGALORE, John Chen
  • Patent number: 10002608
    Abstract: Disclosed herein are systems, methods, and non-transitory computer-readable storage media for approximating relevant responses to a user query with voice-enabled search. A system practicing the method receives a word lattice generated by an automatic speech recognizer based on a user speech and a prosodic analysis of the user speech, generates a reweighted word lattice based on the word lattice and the prosodic analysis, approximates based on the reweighted word lattice one or more relevant responses to the query, and presents to a user the responses to the query. The prosodic analysis examines metalinguistic information of the user speech and can identify the most salient subject matter of the speech, assess how confident a speaker is in the content of his or her speech, and identify the attitude, mood, emotion, sentiment, etc. of the speaker. Other information not described in the content of the speech can also be used.
    Type: Grant
    Filed: September 17, 2010
    Date of Patent: June 19, 2018
    Assignee: NUANCE COMMUNICATIONS, INC.
    Inventors: Srinivas Bangalore, Junlan Feng, Michael Johnston, Taniya Mishra
  • Patent number: 9978363
    Abstract: Disclosed herein are systems, methods, and non-transitory computer-readable storage media for generating domain-specific speech recognition models for a domain of interest by combining and tuning existing speech recognition models when a speech recognizer does not have access to a speech recognition model for that domain of interest and when available domain-specific data is below a minimum desired threshold to create a new domain-specific speech recognition model. A system configured to practice the method identifies a speech recognition domain and combines a set of speech recognition models, each speech recognition model of the set of speech recognition models being from a respective speech recognition domain. The system receives an amount of data specific to the speech recognition domain, wherein the amount of data is less than a minimum threshold to create a new domain-specific model, and tunes the combined speech recognition model for the speech recognition domain based on the data.
    Type: Grant
    Filed: June 12, 2017
    Date of Patent: May 22, 2018
    Assignee: NUANCE COMMUNICATIONS, INC.
    Inventors: Srinivas Bangalore, Robert Bell, Diamantino Antonio Caseiro, Mazin Gilbert, Patrick Haffner
  • Patent number: 9972307
    Abstract: Disclosed herein are systems, computer-implemented methods, and computer-readable media for dialog modeling. The method includes receiving spoken dialogs annotated to indicate dialog acts and task/subtask information, parsing the spoken dialogs with a hierarchical, parse-based dialog model which operates incrementally from left to right and which only analyzes a preceding dialog context to generate parsed spoken dialogs, and constructing a functional task structure of the parsed spoken dialogs. The method can further either interpret user utterances with the functional task structure of the parsed spoken dialogs or plan system responses to user utterances with the functional task structure of the parsed spoken dialogs. The parse-based dialog model can be a shift-reduce model, a start-complete model, or a connection path model.
    Type: Grant
    Filed: September 4, 2015
    Date of Patent: May 15, 2018
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Amanda Stent, Srinivas Bangalore
  • Patent number: 9911437
    Abstract: Disclosed herein are systems, methods, and computer-readable storage media for improving speech recognition accuracy using textual context. The method includes retrieving a recorded utterance, capturing text from a device display associated with the spoken dialog and viewed by one party to the recorded utterance, and identifying words in the captured text that are relevant to the recorded utterance. The method further includes adding the identified words to a dynamic language model, and recognizing the recorded utterance using the dynamic language model. The recorded utterance can be a spoken dialog. A time stamp can be assigned to each identified word. The method can include adding identified words to and/or removing identified words from the dynamic language model based on their respective time stamps. A screen scraper can capture text from the device display associated with the recorded utterance. The device display can contain customer service data.
    Type: Grant
    Filed: May 4, 2016
    Date of Patent: March 6, 2018
    Assignee: Nuance Communications, Inc.
    Inventors: Dan Melamed, Srinivas Bangalore, Michael Johnston
  • Publication number: 20180046617
    Abstract: In an embodiment, a method of providing an on demand translation service is provided. A subscriber may be charged a reduced fee or no fee for use of the on demand translation service in exchange for displaying commercial messages to the subscriber, the commercial messages being selected based on subscriber information. A multimedia signal including information in a source language may be received. The information may be obtained as text in the source language from the multimedia signal. The text may be translated from the source language to a target language. Translated information, based on the translated text, may be transmitted to a processing device for presentation to the subscriber. The received multimedia signal may be sent to a multimedia device for viewing.
    Type: Application
    Filed: October 30, 2017
    Publication date: February 15, 2018
    Inventors: Srinivas BANGALORE, David Crawford GIBBON, Mazin GILBERT, Patrick Guy HAFFNER, Zhu LIU, Behzad SHAHRARAY
  • Publication number: 20180001482
    Abstract: A system, method and computer-readable storage devices are for processing natural language commands, such as commands to a robotic arm, using a Tag & Parse approach to semantic parsing. The system first assigns semantic tags to each word in a sentence and then parses the tag sequence into a semantic tree. The system can use statistical approach for tagging, parsing, and reference resolution. Each stage can produce multiple hypotheses, which are re-ranked using spatial validation. Then the system selects a most likely hypothesis after spatial validation, and generates or outputs a command. In the case of a robotic arm, the command is output in Robot Control Language (RCL).
    Type: Application
    Filed: September 15, 2017
    Publication date: January 4, 2018
    Inventors: Svetlana STOYANCHEV, Srinivas BANGALORE, John CHEN, Hyuckchul JUNG
  • Patent number: 9858345
    Abstract: A method and apparatus for using a classifier for processing a query are disclosed. For example, the method receives a query from a user, and processes the query to locate one or more documents in accordance with a search engine having a discriminative classifier, wherein the discriminative classifier is trained with a plurality of artificial query examples. The method then presents a result of the processing to the user.
    Type: Grant
    Filed: September 19, 2016
    Date of Patent: January 2, 2018
    Assignee: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: Ilija Zeljkovic, Srinivas Bangalore, Patrick Haffner, Jay Wilpon
  • Publication number: 20170372693
    Abstract: A system, method and computer-readable storage device which balance latency and accuracy of machine translations by segmenting the speech upon locating a conjunction. The system, upon receiving speech, will buffer speech until a conjunction is detected. Upon detecting a conjunction, the speech received until that point is segmented. The system then continues performing speech recognition on the segment, searching for the next conjunction, while simultaneously initiating translation of the segment. Upon translating the segment, the system converts the translation to a speech output, allowing a user to hear an audible translation of the speech originally heard.
    Type: Application
    Filed: August 14, 2017
    Publication date: December 28, 2017
    Inventors: Vivek Kumar RANGARAJAN SRIDHAR, Srinivas BANGALORE, John CHEN
  • Publication number: 20170345418
    Abstract: Disclosed herein are systems, methods, and non-transitory computer-readable storage media for generating domain-specific speech recognition models for a domain of interest by combining and tuning existing speech recognition models when a speech recognizer does not have access to a speech recognition model for that domain of interest and when available domain-specific data is below a minimum desired threshold to create a new domain-specific speech recognition model. A system configured to practice the method identifies a speech recognition domain and combines a set of speech recognition models, each speech recognition model of the set of speech recognition models being from a respective speech recognition domain. The system receives an amount of data specific to the speech recognition domain, wherein the amount of data is less than a minimum threshold to create a new domain-specific model, and tunes the combined speech recognition model for the speech recognition domain based on the data.
    Type: Application
    Filed: June 12, 2017
    Publication date: November 30, 2017
    Inventors: Srinivas BANGALORE, Robert BELL, Diamantino Antonio CASEIRO, Mazin GILBERT, Patrick HAFFNER
  • Publication number: 20170344665
    Abstract: Extracting, from user activity data, quantitative attributes and qualitative attributes collected for users having user profiles. The quantitative attributes and the qualitative attributes are extracted during a specified time period determined before the user activity data is collected. Values for the quantitative attributes and the qualitative attributes are plotted, and subsets of the user profiles are clustered into separate group of users based on the plotted values. Delivering a product related content to the groups of users based on the clustering.
    Type: Application
    Filed: August 21, 2017
    Publication date: November 30, 2017
    Applicant: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: Srinivas BANGALORE, Junlan FENG, Michael J. JOHNSTON, Taniya MISHRA
  • Publication number: 20170345416
    Abstract: Disclosed herein are systems, methods, and non-transitory computer-readable storage media for processing speech. A system configured to practice the method monitors user utterances to generate a conversation context. Then the system receives a current user utterance independent of non-natural language input intended to trigger speech processing. The system compares the current user utterance to the conversation context to generate a context similarity score, and if the context similarity score is above a threshold, incorporates the current user utterance into the conversation context. If the context similarity score is below the threshold, the system discards the current user utterance. The system can compare the current user utterance to the conversation context based on an n-gram distribution, a perplexity score, and a perplexity threshold. Alternately, the system can use a task model to compare the current user utterance to the conversation context.
    Type: Application
    Filed: August 21, 2017
    Publication date: November 30, 2017
    Inventor: Srinivas BANGALORE
  • Publication number: 20170337185
    Abstract: Disclosed herein are systems, methods, and non-transitory computer-readable storage media for collecting web data in order to create diverse language models. A system configured to practice the method first crawls, such as via a crawler operating on a computing device, a set of documents in a network of interconnected devices according to a visitation policy, wherein the visitation policy is configured to focus on novelty regions for a current language model built from previous crawling cycles by crawling documents whose vocabulary considered likely to fill gaps in the current language model. A language model from a previous cycle can be used to guide the creation of a language model in the following cycle. The novelty regions can include documents with high perplexity values over the current language model.
    Type: Application
    Filed: August 7, 2017
    Publication date: November 23, 2017
    Inventors: Luciano De Andrade BARBOSA, Srinivas BANGALORE
  • Publication number: 20170330554
    Abstract: A system and method are disclosed for generating customized text-to-speech voices for a particular application. The method comprises generating a custom text-to-speech voice by selecting a voice for generating a custom text-to-speech voice associated with a domain, collecting text data associated with the domain from a pre-existing text data source and using the collected text data, generating an in-domain inventory of synthesis speech units by selecting speech units appropriate to the domain via a search of a pre-existing inventory of synthesis speech units, or by recording the minimal inventory for a selected level of synthesis quality. The text-to-speech custom voice for the domain is generated utilizing the in-domain inventory of synthesis speech units. Active learning techniques may also be employed to identify problem phrases wherein only a few minutes of recorded data is necessary to deliver a high quality TTS custom voice.
    Type: Application
    Filed: July 31, 2017
    Publication date: November 16, 2017
    Inventors: Srinivas BANGALORE, Junlan FENG, Mazin GILBERT, Juergen SCHROETER, Ann K. SYRDAL, David SCHULZ