Patents by Inventor Laurence S. Gillick

Laurence S. Gillick has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20180174585
    Abstract: An interaction assistant conducts multiple turn interaction dialogs with a user in which context is maintained between turns, and the system manages the dialog to achieve an inferred goal for the user. The system includes a linguistic interface to a user and a parser for processing linguistic events from the user. A dialog manager of the system is configured to receive alternative outputs from the parser, and selecting an action and causing the action to be performed based on the received alternative outputs. The system further includes a dialog state for an interaction with the user, and the alternative outputs represent alternative transitions from a current dialog state to a next dialog state. The system further includes a storage for a plurality of templates, and wherein each dialog state is defined in terms of an interrelationship of one or more instances of the templates.
    Type: Application
    Filed: February 14, 2018
    Publication date: June 21, 2018
    Inventors: Jacob Andreas, Taylor D. Berg-Kirkpatrick, Pengyu Chen, Jordan R. Cohen, Laurence S. Gillick, David Leo Wright Hall, Daniel Klein, Michael Newman, Adam D. Pauls, Daniel L. Roth, Jesse Daniel Eskes Rusak, Andrew R. Volpe, Steven A. Wegmann
  • Publication number: 20180108346
    Abstract: Systems and processes are disclosed for discovering trending terms in automatic speech recognition. Candidate terms (e.g., words, phrases, etc.) not yet found in a speech recognizer vocabulary or having low language model probability can be identified based on trending usage in a variety of electronic data sources (e.g., social network feeds, news sources, search queries, etc.). When candidate terms are identified, archives of live or recent speech traffic can be searched to determine whether users are uttering the candidate terms in dictation or speech requests. Such searching can be done using open vocabulary spoken term detection to find phonetic matches in the audio archives. As the candidate terms are found in the speech traffic, notifications can be generated that identify the candidate terms, provide relevant usage statistics, identify the context in which the terms are used, and the like.
    Type: Application
    Filed: November 3, 2017
    Publication date: April 19, 2018
    Inventors: Matthias PAULIK, Gunnar EVERMANN, Laurence S. GILLICK
  • Patent number: 9818400
    Abstract: Systems and processes are disclosed for discovering trending terms in automatic speech recognition. Candidate terms (e.g., words, phrases, etc.) not yet found in a speech recognizer vocabulary or having low language model probability can be identified based on trending usage in a variety of electronic data sources (e.g., social network feeds, news sources, search queries, etc.). When candidate terms are identified, archives of live or recent speech traffic can be searched to determine whether users are uttering the candidate terms in dictation or speech requests. Such searching can be done using open vocabulary spoken term detection to find phonetic matches in the audio archives. As the candidate terms are found in the speech traffic, notifications can be generated that identify the candidate terms, provide relevant usage statistics, identify the context in which the terms are used, and the like.
    Type: Grant
    Filed: August 28, 2015
    Date of Patent: November 14, 2017
    Assignee: Apple Inc.
    Inventors: Matthias Paulik, Gunnar Evermann, Laurence S. Gillick
  • Publication number: 20170147554
    Abstract: A method for configuring an automated dialogue system uses traces of interactions via a graphical user interface (GUI) for an application. Each trace includes interactions in the context of a plurality of presentations of the GUI. Elements of one or more presentations of the GUI are identified, and templates are associated with portions of the trace. Each template has one or more defined inputs and a defined output. For each template of the plurality of templates, the portions of the traces are processed to automatically configure the template by specifying a procedure for providing values of inputs to the template via the GUI and obtaining a value of an output. The automated dialogue system is configured with the configured templates, thereby avoiding manual configuration of the dialogue system.
    Type: Application
    Filed: November 22, 2016
    Publication date: May 25, 2017
    Inventors: Pengyu Chen, Jordan R. Cohen, Laurence S. Gillick, David Leo Wright Hall, Daniel Klein, Adam D. Pauls, Daniel L. Roth, Jesse Daniel Eskes Rusak
  • Publication number: 20170118344
    Abstract: An approach to providing communication assistance to an operator of a vehicle makes use software having a first component executing on a personal device of the operator as well as a second component executing on a server in communication with the personal device. In some implementations, handling a call involves establishing a first two-way audio link between the server and the calling device is established, and a second two-way audio link between a server and the user device. The server passes some of the audio from the calling device to the user device, and monitors a user's voice input, of lack thereof, to determine how to handle the call.
    Type: Application
    Filed: October 20, 2016
    Publication date: April 27, 2017
    Inventors: Jordan R. Cohen, Daniel L. Roth, David Leo Wright Hall, Jesse Daniel Eskes Rusak, Andrew Robert Volpe, Sean Daniel True, Damon R. Pender, Laurence S. Gillick, Yan Virin
  • Publication number: 20160078860
    Abstract: Systems and processes are disclosed for discovering trending terms in automatic speech recognition. Candidate terms (e.g., words, phrases, etc.) not yet found in a speech recognizer vocabulary or having low language model probability can be identified based on trending usage in a variety of electronic data sources (e.g., social network feeds, news sources, search queries, etc.). When candidate terms are identified, archives of live or recent speech traffic can be searched to determine whether users are uttering the candidate terms in dictation or speech requests. Such searching can be done using open vocabulary spoken term detection to find phonetic matches in the audio archives. As the candidate terms are found in the speech traffic, notifications can be generated that identify the candidate terms, provide relevant usage statistics, identify the context in which the terms are used, and the like.
    Type: Application
    Filed: August 28, 2015
    Publication date: March 17, 2016
    Inventors: Matthias PAULIK, Gunnar EVERMANN, Laurence S. GILLICK
  • Publication number: 20140032216
    Abstract: A method for a portable device includes receiving a spoken utterance of a word or phrase, generating a plurality of alternative pronunciations of the spoken utterance, scoring one or more pronunciations of the plurality of alternative pronunciations using the spoken utterance, and updating a lexicon with at least one scored pronunciation.
    Type: Application
    Filed: September 30, 2013
    Publication date: January 30, 2014
    Applicant: Nuance Communications, Inc.
    Inventors: Daniel L. Roth, Laurence S. Gillick, Michael L. Shire
  • Patent number: 8577681
    Abstract: A method of generating an alternative pronunciation for a word or phrase, given an initial pronunciation and a spoken example of the word or phrase, includes providing the initial pronunciation of the word or phrase, and generating the alternative pronunciation by searching a neighborhood of pronunciations about the initial pronunciation via a constrained hypothesis, wherein the neighborhood includes pronunciations that differ from the initial pronunciation by at most one phoneme. The method further includes selecting a highest scoring pronunciation within the neighborhood of pronunciations.
    Type: Grant
    Filed: September 13, 2004
    Date of Patent: November 5, 2013
    Assignee: Nuance Communications, Inc.
    Inventors: Daniel L. Roth, Laurence S. Gillick, Mike Shire
  • Patent number: 8065144
    Abstract: A method for speech recognition. The method uses a single pronunciation estimator to train acoustic phoneme models and recognize utterances from multiple languages. The method includes accepting text spellings of training words in a plurality of sets of training words, each set corresponding to a different one of a plurality of languages. The method also includes, for each of the sets of training words in the plurality, receiving pronunciations for the training words in the set, the pronunciations being characteristic of native speakers of the language of the set, the pronunciations also being in terms of subword units at least some of which are common to two or more of the languages. The method also includes training a single pronunciation estimator using data comprising the text spellings and the pronunciations of the training words.
    Type: Grant
    Filed: February 3, 2010
    Date of Patent: November 22, 2011
    Assignee: Voice Signal Technologies, Inc.
    Inventors: Laurence S. Gillick, Thomas E. Lynch, Michael J. Newman, Daniel L. Roth, Steven A. Wegmann, Jonathan P. Yamron
  • Publication number: 20100298010
    Abstract: A method of operating a mobile communication device having a set of one or more applications, each with its own associated user-configurable customization, the method comprising detecting whether the user-configurable customization of any of the applications has changed since an earlier time, and for all applications for which the user-configurable customization has changed since said earlier time, wirelessly transmitting those changes to a remote server. The method further comprises maintaining a set of flags indicating whether changes have occurred to the user-configurable customization, wherein detecting whether the user-configurable customization of any of the applications has changed since said earlier time includes reading the set of flags. The remote server is one of a carrier server and a third party provider server.
    Type: Application
    Filed: December 22, 2009
    Publication date: November 25, 2010
    Applicant: NUANCE COMMUNICATIONS, INC.
    Inventors: Daniel L. ROTH, Laurence S. GILLICK, Jordan COHEN
  • Patent number: 7797157
    Abstract: Channel normalization for automatic speech recognition is provided. Statistics are measured from an initial portion of a speech utterance. Feature normalization parameters are estimated based on the measured statistics and a statistically derived mapping relating measured statistics and feature normalization parameters. In some examples, the measured statistics comprise measures of an energy from the initial portion of the speech utterance. In some examples, measures of the energy comprise extreme values of the energy.
    Type: Grant
    Filed: January 10, 2005
    Date of Patent: September 14, 2010
    Assignee: Voice Signal Technologies, Inc.
    Inventors: Igor Zlokarnik, Laurence S. Gillick, Jordan Cohen
  • Patent number: 7716050
    Abstract: A method for speech recognition. The method uses a single pronunciation estimator to train acoustic phoneme models and recognize utterances from multiple languages. The method includes accepting text spellings of training words in a plurality of sets of training words, each set corresponding to a different one of a plurality of languages. The method also includes, for each of the sets of training words in the plurality, receiving pronunciations for the training words in the set, the pronunciations being characteristic of native speakers of the language of the set, the pronunciations also being in terms of subword units at least some of which are common to two or more of the languages. The method also includes training a single pronunciation estimator using data comprising the text spellings and the pronunciations of the training words.
    Type: Grant
    Filed: November 17, 2003
    Date of Patent: May 11, 2010
    Assignee: Voice Signal Technologies, Inc.
    Inventors: Laurence S. Gillick, Thomas E. Lynch, Michael J. Newman, Daniel L. Roth, Steven A. Wegmann, Jonathan P. Yamron
  • Patent number: 7676367
    Abstract: A method of constructing a list of alternate transcripts from a recognized transcript includes generating a list of close call records, matching partial sub-histories from the recognized transcript with one of the history pairs stored in each of the records, and substituting the other of the history pairs for the partial sub-history of the recognized transcript. A close call record is generated each time a pair of partial hypotheses attempt to seed a common word. Each close call record includes history information and scoring information associated with a particular pair of partial hypotheses seeding a common word. Alternate transcripts are constructed by substituting close call histories for partial histories of the recognized transcripts, and also by substituting close call histories for partial histories of other alternate transcript.
    Type: Grant
    Filed: February 20, 2004
    Date of Patent: March 9, 2010
    Assignee: Voice Signal Technologies, Inc.
    Inventors: Robert Roth, Arkady Khasin, Laurence S. Gillick
  • Patent number: 7467087
    Abstract: The error rate of a pronunciation guesser that guesses the phonetic spelling of words used in speech recognition is improved by causing its training to weigh letter-to-phoneme mappings used as data in such training as a function of the frequency of the words in which such mappings occur. Preferably the ratio of the weight to word frequency increases as word frequencies decreases. Acoustic phoneme models for use in speech recognition with phonetic spellings generated by a pronunciation guesser that makes errors are trained against word models whose phonetic spellings have been generated by a pronunciation guesser that makes similar errors. As a result, the acoustic models represent blends of phoneme sounds that reflect the spelling errors made by the pronunciation guessers. Speech recognition enabled systems are made by storing in them both a pronunciation guesser and a corresponding set of such blended acoustic models.
    Type: Grant
    Filed: October 10, 2003
    Date of Patent: December 16, 2008
    Inventors: Laurence S. Gillick, Steven A. Wegmann, Jonathan P. Yamron
  • Publication number: 20080154612
    Abstract: A method implemented on a mobile device that includes speech recognition functionality involves: receiving an utterance from a user of the mobile device, the utterance including a spoken search request; recognizing that the utterance includes a spoken search request; sending a representation of the spoken search request to a remote server over a wireless data connection; receiving search results over the wireless data connection that are responsive to the search request; storing the results on the mobile device; receiving a subsequent search request; performing a subsequent search responsive to the subsequent search request to generate subsequent search results, the subsequent search including searching the stored search results; and presenting the subsequent results on the mobile device.
    Type: Application
    Filed: February 12, 2007
    Publication date: June 26, 2008
    Applicant: Voice Signal Technologies, Inc.
    Inventors: Gunnar Evermann, Daniel L. Roth, Laurence S. Gillick, James Coughlin
  • Publication number: 20080153465
    Abstract: Methods and devices for providing a user of a mobile communications device with mobile voice-mediated search capability. The methods and devices involve receiving an utterance from a user of the mobile device, the utterance including a search request; using the speech recognition functionality to recognize that the utterance includes a search request; as a result of recognizing that the utterance includes a search request, establishing a wireless data connection to a remote server; sending a representation of the search request to the remote server over the wireless data connection; receiving search results that are responsive to the search request; and presenting the search results on the mobile device.
    Type: Application
    Filed: February 9, 2007
    Publication date: June 26, 2008
    Applicant: Voice Signal Technologies, Inc.
    Inventors: Gunnar Evermann, Daniel L. Roth, Laurence S. Gillick, James Coughlin
  • Publication number: 20080154611
    Abstract: A method implemented on a mobile device that includes speech recognition functionality involves presenting to a user of the mobile device a voice-control interface that supports two types of commands at a common level of the interface, the two types of commands including a first type and a second type, the first type being command and control commands and the second type being search request commands. The method further involves: receiving an utterance from the user that corresponds to a command of either of the first type or the second type; recognizing the utterance; if the received utterance is a command of the first type, performing a corresponding command and control function; and if the received utterance is a command of the second type, generating a representation of a corresponding search request and then using the representation to request a search that is responsive to the search request.
    Type: Application
    Filed: February 12, 2007
    Publication date: June 26, 2008
    Applicant: VOICE SIGNAL TECHNOLOGIES, INC.
    Inventors: Gunnar Evermann, Daniel L. ROTH, Laurence S. GILLICK, James COUGHLIN
  • Publication number: 20080154870
    Abstract: Methods and systems for providing voice-mediated search capability to a mobile communications device involve receiving a signal from the mobile device that includes a representation of a spoken search request from a user of the mobile device, using speech recognition software to convert the search request into a text search request, extracting side information contained implicitly within the received signal, using the extracted side information to assign the user to a category, sending the text search request and the user category to content providers, receiving from the content providers content that is responsive to the text search request and the user category, and sending to the mobile device search results that are based on content from content providers. The methods and systems further involve sending searches and user categories to advertising providers, and sending advertisements returned by the advertising providers to the mobile device along with the search results.
    Type: Application
    Filed: February 12, 2007
    Publication date: June 26, 2008
    Applicant: VOICE SIGNAL TECHNOLOGIES, INC.
    Inventors: Gunnar Evermann, Daniel L. ROTH, Laurence S. GILLICK, James COUGHLIN
  • Publication number: 20080154608
    Abstract: A method implemented on a mobile device that includes speech recognition functionality involves: receiving an utterance that includes a search request from a user of the device; recognizing that the utterance includes a search request; sending a representation of the search request to a remote server over a wireless data connection; receiving information over the wireless data connection that is responsive to the search request; presenting the information on the mobile device; receiving an input from the user selecting an item present in the received information, the item identifying a remote resource; using the selected item to connect to the remote resource, the connection to the remote resource not involving the remote server; and sending to the remote server an indication that a connection was made to the resource identified by the selected item. The method further involves storing a log of the user's connection to remote resources and sending the log to the server.
    Type: Application
    Filed: February 12, 2007
    Publication date: June 26, 2008
    Inventors: Gunnar Evermann, Daniel L. ROTH, Laurence S. GILLICK, James Coughlin
  • Patent number: 7133827
    Abstract: A new word model is trained from synthetic word samples derived by Monte Carlo techniques from one or more prior word models. The prior word model can be a phonetic word model and the new word model can be a non-phonetic, whole-word, word model. The prior word model can be trained from data that has undergone a first channel normalization and the synthesized word samples from which the new word model is trained can undergo a different channel normalization similar to that to be used in a given speech recognition context. The prior word model can have a first model structure and the new word model can have a second, different, model structure. These differences in model structure can include, for example, differences of model topology; differences of model complexity; and differences in the type of basis function used in a description of such probability distributions.
    Type: Grant
    Filed: February 6, 2003
    Date of Patent: November 7, 2006
    Assignee: Voice Signal Technologies, Inc.
    Inventors: Laurence S. Gillick, Donald R. McAllaster, Daniel L. Roth