Patents Examined by Abdelali Serrou
  • Patent number: 9002717
    Abstract: A system that incorporates teachings of the present disclosure may include, for example, a controller configured to obtain information associated with media content, to generate a first group of tones representative of the information associated with the media content, and to generate a media stream comprising the media content and the first group of tones; and a communication interface configured to transmit the media stream to a media device whereby the media device presents the media content and a sequence of tones, where the sequence of tones is generated based at least in part on the first group of tones, where the first group of tones comprises high frequency tones and low frequency tones, and where one of the high and low frequency tones represents a binary one and the other of the high and low frequency tones represents a binary zero. Other embodiments are disclosed.
    Type: Grant
    Filed: December 3, 2010
    Date of Patent: April 7, 2015
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Ke Yu, Ashwini Sule
  • Patent number: 8990086
    Abstract: A recognition confidence measurement method, medium and system which can more accurately determine whether an input speech signal is an in-vocabulary, by extracting an optimum number of candidates that match a phone string extracted from the input speech signal and estimating a lexical distance between the extracted candidates is provided. A recognition confidence measurement method includes: extracting a phoneme string from a feature vector of an input speech signal; extracting candidates by matching the extracted phoneme string and phoneme strings of vocabularies registered in a predetermined dictionary and; estimating a lexical distance between the extracted candidates; and determining whether the input speech signal is an in-vocabulary, based on the lexical distance.
    Type: Grant
    Filed: July 31, 2006
    Date of Patent: March 24, 2015
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Sang-Bae Jeong, Nam Hoon Kim, Ick Sang Han, In Jeong Choi, Gil Jin Jang, Jae-Hoon Jeong
  • Patent number: 8990088
    Abstract: A runtime framework and authoring tool are provided for enabling linguistic experts to author text normalization maps and grammar libraries without requiring high level of technical or programming skills. Authors define or select terminals, map the terminals, and define rules for the mapping. The tool enables an author to validate their work, by executing the map in the same way the recognition engine does, causing consistency in results from authoring to user operations. The runtime is used by the speech engines and by the tools to provide consistent normalization for supported scenarios.
    Type: Grant
    Filed: January 28, 2009
    Date of Patent: March 24, 2015
    Assignee: Microsoft Corporation
    Inventors: Rachel I. Morton, Nicholas J. Gedge, Heiko W. Rahmel
  • Patent number: 8965772
    Abstract: Methods, systems, and products are disclosed for displaying speech command input state information in a multimodal browser including displaying an icon representing a speech command type and displaying an icon representing the input state of the speech command. In typical embodiments, the icon representing a speech command type and the icon representing the input state of the speech command also includes attributes of a single icon. Typical embodiments include accepting from a user a speech command of the speech command type, changing the input state of the speech command, and displaying another icon representing the changed input state of the speech command. Typical embodiments also include displaying the text of the speech command in association with the icon representing the speech command type.
    Type: Grant
    Filed: March 20, 2014
    Date of Patent: February 24, 2015
    Assignee: Nuance Communications, Inc.
    Inventors: Charles W. Cross, Jr., Michael C. Hollinger, Igor R. Jablokov, Benjamin D. Lewis, Hilary A. Pike, Daniel M. Smith, David W. Wintermute, Michael A. Zaitzeff
  • Patent number: 8954334
    Abstract: A voice-activated pulser can trigger an oscilloscope or a meter, upon a simple voice command, thereby enabling hands-free signal measurements. The pulser can also be used to control the circuit under test, activating it or changing parameters, all under voice control. The pulser includes numerous switch-selectable output modes that allow users to generate complex, tightly-controlled diagnostic sequences, all activated upon a voice command and hands-free. The invention includes a fast, robust command-interpretation protocol that completely eliminates the expense and complexity of word recognition. Visual indicators display the device status and various operating modes, and also confirm each output pulse. The device receives voice commands directly through an internal microphone, or through a detachable headset, and confirms each command with an acoustical signal in the headset.
    Type: Grant
    Filed: October 15, 2011
    Date of Patent: February 10, 2015
    Assignee: Zanavox
    Inventor: David Edward Newman
  • Patent number: 8954332
    Abstract: A computer-implemented system and method for masking special data is provided. Speakers of a call recording are identified. The call recording is separated into strands corresponding to each of the speakers. A prompt list of elements that prompt the speaker of the other strand to utter special information is applied to one of the strands. At least one of the elements of the prompt list is identified in the one strand. A special information candidate is identified in the other strand and is located after a location in time where the element was found in the voice recording of the one strand. A confidence score is assigned to the element located in the one strand and to the special information candidate in the other strand. The confidence scores are combined and a threshold is applied. The special information candidate is rendered unintelligible when the combined confidence scores satisfy the threshold.
    Type: Grant
    Filed: November 4, 2013
    Date of Patent: February 10, 2015
    Assignee: Intellisist, Inc.
    Inventors: Howard M. Lee, Steven Lutz, Gilad Odinak
  • Patent number: 8942974
    Abstract: A device searches for an available network. The device automatically sends a request message to an identified available network. After the device receives a response message from the identified available network, the device selects a language and/or other device setting based on contents of the response message.
    Type: Grant
    Filed: March 4, 2011
    Date of Patent: January 27, 2015
    Assignee: Amazon Technologies, Inc.
    Inventor: Jonathan A. Leblang
  • Patent number: 8909528
    Abstract: A method (and system) of determining confusable list items and resolving this confusion in a spoken dialog system includes receiving user input, processing the user input and determining if a list of items needs to be played back to the user, retrieving the list to be played back to the user, identifying acoustic confusions between items on the list, changing the items on the list as necessary to remove the acoustic confusions, and playing unambiguous list items back to the user.
    Type: Grant
    Filed: May 9, 2007
    Date of Patent: December 9, 2014
    Assignee: Nuance Communications, Inc.
    Inventors: Ellen Marie Eide, Vaibhava Goel, Ramesh Gopinath, Osamuyimen T. Stewart
  • Patent number: 8849665
    Abstract: A machine translation method, system for using the method, and computer readable media are disclosed. The method includes the steps of receiving a source language sentence, selecting a set of target language n-grams using a lexical classifier and based on the source language sentence. When selecting the set of target language n-grams, in at least one n-gram, n is greater than 1. The method continues by combining the selected set of target language n-grams as a finite state acceptor (FSA), weighting the FSA with data from the lexical classifier, and generating an n-best list of target sentences from the FSA. As an alternate to using the FSA, N strings may be generated from the n-grams and ranked using a language model. The N strings may be represented by an FSA for efficiency but it is not necessary.
    Type: Grant
    Filed: January 30, 2008
    Date of Patent: September 30, 2014
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Srinivas Bangalore, Emil Ettelaie
  • Patent number: 8805696
    Abstract: An audio encoder implements multi-channel coding decision, band truncation, multi-channel rematrixing, and header reduction techniques to improve quality and coding efficiency. In the multi-channel coding decision technique, the audio encoder dynamically selects between joint and independent coding of a multi-channel audio signal via an open-loop decision based upon (a) energy separation between the coding channels, and (b) the disparity between excitation patterns of the separate input channels. In the band truncation technique, the audio encoder performs open-loop band truncation at a cut-off frequency based on a target perceptual quality measure. In multi-channel rematrixing technique, the audio encoder suppresses certain coefficients of a difference channel by scaling according to a scale factor, which is based on current average levels of perceptual quality, current rate control buffer fullness, coding mode, and the amount of channel separation in the source.
    Type: Grant
    Filed: October 7, 2013
    Date of Patent: August 12, 2014
    Assignee: Microsoft Corporation
    Inventors: Wei-Ge Chen, Naveen Thumpudi, Ming-Chieh Lee
  • Patent number: 8793132
    Abstract: An apparatus, method and program for dividing a conversational dialog into utterance. The apparatus includes: a computer processor; a word database for storing spellings and pronunciations of words; a grammar database for storing syntactic rules on words; a pause detecting section which detects a pause location in a channel making a main speech among conversational dialogs inputted in at least two channels; an acknowledgement detecting section which detects an acknowledgement location in a channel not making the main speech; a boundary-candidate extracting section which extracts boundary candidates in the main speech, by extracting pauses existing within a predetermined range before and after a base point that is the acknowledgement location; and a recognizing unit which outputs a word string of the main speech segmented by one of the extracted boundary candidates after dividing the segmented speech into optimal utterance in reference to the word database and grammar database.
    Type: Grant
    Filed: December 26, 2007
    Date of Patent: July 29, 2014
    Assignee: Nuance Communications, Inc.
    Inventors: Nobuyasu Itoh, Gakuto Kurata
  • Patent number: 8731938
    Abstract: A computer-implemented system and method for identifying and masking special information within recorded speech is provided. A field for entry of special information is identified. Movement of a pointer device along a trajectory towards the field is also identified. A correlation of the pointer device movement and entry of the special information is determined based on a location of the trajectory in relation to the field. A threshold is applied to the correlation. The special information is received as verbal speech. A recording of the special information is rendered unintelligible when the threshold is satisfied.
    Type: Grant
    Filed: April 26, 2013
    Date of Patent: May 20, 2014
    Assignee: Intellisist, Inc.
    Inventor: G. Kevin Doren
  • Patent number: 8725505
    Abstract: A computer implemented method and system for speech recognition are provided. The method and system generally maintain a set of verbs for speech recognition commands. Upon recognizing utterance of a verb of the set in combination with an invalid object or objects for the verb, the method and system generate an indication relative to the verb and invalid object. The indication can include informing the user that the system is unsure how to execute the command associated with the verb with the invalid object. The method and system can then receive a user input to specify how the verb and invalid object should be treated.
    Type: Grant
    Filed: October 22, 2004
    Date of Patent: May 13, 2014
    Assignee: Microsoft Corporation
    Inventors: David Mowatt, Robert L. Chambers
  • Patent number: 8719034
    Abstract: Methods, systems, and products are disclosed for displaying speech command input state information in a multimodal browser including displaying an icon representing a speech command type and displaying an icon representing the input state of the speech command. In typical embodiments, the icon representing a speech command type and the icon representing the input state of the speech command also includes attributes of a single icon. Typical embodiments include accepting from a user a speech command of the speech command type, changing the input state of the speech command, and displaying another icon representing the changed input state of the speech command. Typical embodiments also include displaying the text of the speech command in association with the icon representing the speech command type.
    Type: Grant
    Filed: September 13, 2005
    Date of Patent: May 6, 2014
    Assignee: Nuance Communications, Inc.
    Inventors: Charles W. Cross, Jr., Michael Charles Hollinger, Igor R. Jablokov, Benjamin D. Lewis, Hilary A. Pike, Daniel M. Smith, David W. Wintermute, Michael A. Zaitzeff
  • Patent number: 8694316
    Abstract: An automatic speech recognition (ASR) system includes a speech-responsive application and a recognition engine. The ASR system generates user prompts to elicit certain spoken inputs, and the speech-responsive application performs operations when the spoken inputs are recognized. The recognition engine compares sounds within an input audio signal with phones within an acoustic model, to identify candidate matching phones. A recognition confidence score is calculated for each candidate matching phone, and the confidence scores are used to help identify one or more likely sequences of matching phones that appear to match a word within the grammar of the speech-responsive application. The per-phone confidence scores are evaluated against predefined confidence score criteria (for example, identifying scores below a ‘low confidence’ threshold) and the results of the evaluation are used to influence subsequent selection of user prompts.
    Type: Grant
    Filed: October 20, 2005
    Date of Patent: April 8, 2014
    Assignee: Nuance Communications, Inc.
    Inventors: John Brian Pickering, Timothy David Poultney, Benjamin Terrick Staniford, Matthew Whitbourne
  • Patent number: 8694322
    Abstract: A voice-activated user interface for an application is described. The voice activated user interface invokes a speech recognition component to recognize audio input from a user. If the audio input is a command, then a validation component is invoked to determine whether to validate the command prior to execution.
    Type: Grant
    Filed: October 21, 2005
    Date of Patent: April 8, 2014
    Assignee: Microsoft Corporation
    Inventors: Alex G. Snitkovskiy, David Mowatt, Felix G. T. I. Andrew, Robert Edward Dewar, Oliver Scholz
  • Patent number: 8666750
    Abstract: A voice control system allows a user to control a device through voice commands. The voice control system includes a speech recognition unit that receives a control signal from a mobile device and a speech signal from a user. The speech recognition unit configures speech recognition settings in response to the control signal to improve speech recognition.
    Type: Grant
    Filed: January 31, 2008
    Date of Patent: March 4, 2014
    Assignee: Nuance Communications, Inc.
    Inventors: Markus Buck, Lars König, Gerhard Uwe Schmidt
  • Patent number: 8655660
    Abstract: The present invention is a system and method for generating a personal voice font including, monitoring voice segments automatically from phone conversations of a user by a voice learning processor to generate a personalized voice font and delivering the personalized voice font (PVF) to the a server.
    Type: Grant
    Filed: February 10, 2009
    Date of Patent: February 18, 2014
    Assignee: International Business Machines Corporation
    Inventors: Zsolt Szalai, Philippe Bazot, Bernard Pucci, Joel Vitale
  • Patent number: 8655646
    Abstract: An apparatus and method for detecting a named-entity. The apparatus includes a candidate-named-entity extraction module that detects a candidate-named-entity based on an initial learning example and feature information regarding morphemes constituting an inputted sentence, the candidate-named-entity extraction module providing a tagged sentence including the detected candidate-named-entity; a storage module that stores information regarding a named-entity dictionary and a rule; and a learning-example-regeneration module for finally determining whether the candidate-named-entity included in the provided sentence is a valid named-entity, based on the named-entity dictionary and the rule, the learning-example-regeneration module providing the sentence as a learning example, based on a determination result, so that a probability of candidate-named-entity detection is gradually updated.
    Type: Grant
    Filed: August 3, 2006
    Date of Patent: February 18, 2014
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Jae-won Lee, In-ho Kang, Jeong-su Kim, Hae-chang Rim
  • Patent number: 8645138
    Abstract: Disclosed are apparatus and methods for processing spoken speech. Input speech can be received at a computing system. During a first pass of speech recognition, a plurality of language model outputs can be determined by: providing the input speech to each of a plurality of language models and responsively receiving a language model output from each language model. A language model of the plurality of language models can be selected using a classifier operating on the plurality of language model outputs. During a second pass of speech recognition, a revised language model output can be determined by: providing the input speech and the language model output from the selected language model to the selected language model and responsively receiving the revised language model output from the selected language model. The computing system can generate a result based on the revised language model output.
    Type: Grant
    Filed: December 20, 2012
    Date of Patent: February 4, 2014
    Assignee: Google Inc.
    Inventors: Eugene Weinstein, Austin Waters