Patents Examined by Jesse Pullias
  • Patent number: 9549068
    Abstract: A method for adaptive voice interaction includes monitoring voice communications between a service recipient and a service representative, measuring a set of voice communication features based upon the voice communications between the service recipient and the service representative, analyzing the set of voice communication features to generate emotion metric values, and generating a response based on the analysis of the set of voice communication features.
    Type: Grant
    Filed: January 28, 2015
    Date of Patent: January 17, 2017
    Assignee: Simple Emotion, Inc.
    Inventors: Akash Krishnan, Matthew Fernandez
  • Patent number: 9547645
    Abstract: A machine translation apparatus configured to translate an input sentence and output a translated sentence in a target language, the machine translation apparatus includes a rule acquirer configured to acquire a difference between an input example sentence and a replaced example sentence which is obtained by replacing the input example sentence, and acquire a replacement rule based on the difference and each of meaning representations which indicate each relationship of words in the input example sentence and the replaced example sentence; and a translator configured to apply the replacement rule acquired by the rule acquirer to the input sentence and output the translated sentence based on the meaning representations and a meaning representation indicating of relationship of words in the input sentence.
    Type: Grant
    Filed: January 15, 2015
    Date of Patent: January 17, 2017
    Assignee: FUJITSU LIMITED
    Inventors: Masataka Kadotani, Tomoki Nagase
  • Patent number: 9548051
    Abstract: A system and method are disclosed that improve automatic speech recognition in a spoken dialog system. The method comprises partitioning speech recognizer output into self-contained clauses, identifying a dialog act in each of the self-contained clauses, qualifying dialog acts by identifying a current domain object and/or a current domain action, and determining whether further qualification is possible for the current domain object and/or current domain action. If further qualification is possible, then the method comprises identifying another domain action and/or another domain object associated with the current domain object and/or current domain action, reassigning the another domain action and/or another domain object as the current domain action and/or current domain object and then recursively qualifying the new current domain action and/or current object. This process continues until nothing is left to qualify.
    Type: Grant
    Filed: February 3, 2016
    Date of Patent: January 17, 2017
    Assignee: AT&T Intellectual Property II, L.P.
    Inventors: Srinivas Bangalore, Narendra K. Gupta, Mazin G. Rahim
  • Patent number: 9542948
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for speaker verification. The methods, systems, and apparatus include actions of inputting speech data that corresponds to a particular utterance to a first neural network and determining an evaluation vector based on output at a hidden layer of the first neural network. Additional actions include obtaining a reference vector that corresponds to a past utterance of a particular speaker. Further actions include inputting the evaluation vector and the reference vector to a second neural network that is trained on a set of labeled pairs of feature vectors to identify whether speakers associated with the labeled pairs of feature vectors are the same speaker. More actions include determining, based on an output of the second neural network, whether the particular utterance was likely spoken by the particular speaker.
    Type: Grant
    Filed: February 3, 2015
    Date of Patent: January 10, 2017
    Assignee: Google Inc.
    Inventors: Dominik Roblek, Matthew Sharifi, Raziel Alvarez Guevara
  • Patent number: 9542954
    Abstract: Audio watermarking is the process of embedding watermark information items into an audio signal in an in-audible manner. In a first embodiment, in case the original audio signal has parts of low signal energy, an alternative signal having a level or strength given by the psycho-acoustic model is combined with the original audio signal. The combined signal is watermarked with watermark data to be embedded. In a second embodiment, in case the original audio signal has parts of low signal energy, an alternative signal having a level or strength given by the psycho-acoustic model is watermarked with watermark data to be embedded, and the audio signal is watermarked with the watermark data to be embedded. The watermarked alternative signal is combined with the watermarked audio signal.
    Type: Grant
    Filed: February 4, 2015
    Date of Patent: January 10, 2017
    Assignee: THOMSON LICENSING
    Inventors: Peter Georg Baum, Xiaoming Chen, Michael Arnold, Ulrich Gries
  • Patent number: 9542931
    Abstract: On a computing device a speech utterance is received from a user. The speech utterance is a section of a speech dialog that includes a plurality of speech utterances. One or more features from the speech utterance are identified. Each identified feature from the speech utterance is a specific characteristic of the speech utterance. One or more features from the speech dialog are identified. Each identified feature from the speech dialog is associated with one or more events in the speech dialog. The one or more events occur prior to the speech utterance. One or more identified features from the speech utterance and one or more identified features from the speech dialog are used to calculate a confidence score for the speech utterance.
    Type: Grant
    Filed: October 23, 2014
    Date of Patent: January 10, 2017
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Michael Levit, Bruce Melvin Buntschuh
  • Patent number: 9524720
    Abstract: Systems and methods of performing blind bandwidth extension are disclosed. In an embodiment, a method includes determining, based on a set of low-band parameters of an audio signal, a first set of high-band parameters and a second set of high-band parameters. The method further includes generating a predicted set of high-band parameters based on a weighted combination of the first set of high-band parameters and the second set of high-band parameters.
    Type: Grant
    Filed: July 18, 2014
    Date of Patent: December 20, 2016
    Assignee: QUALCOMM Incorporated
    Inventors: Sen Li, Stephane Pierre Villette, Daniel J. Sinder, Pravin Kumar Ramadas
  • Patent number: 9520126
    Abstract: A voice recognition system that divides a search space for voice recognition into a general domain search space and a specific domain search space. A mobile terminal receives a voice recognition target word from a user, and a voice recognition server divides a search space for voice recognition into a general domain search space and a specific domain search space and stores the spaces and performs voice recognition for the voice recognition target word through linkage of the general domain search space and the specific domain search space.
    Type: Grant
    Filed: January 22, 2015
    Date of Patent: December 13, 2016
    Assignee: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE
    Inventors: Seung Hi Kim, Sang Hun Kim, Ki Hyun Kim, Sang Kyu Park, Soo Jong Lee
  • Patent number: 9514121
    Abstract: A custom dictionary is generated for an e-book. A dictionary management system receives a custom dictionary request from a user client operated by a user, the custom dictionary request identifying the e-book and including dictionary management information describing the user. The dictionary management system chooses a group reader profile that has an associated group reading score for the user based on the dictionary management information and candidate words are identified in the identified e-book for inclusion in the custom dictionary. The dictionary management system selects words for inclusion in the custom dictionary from among the candidate words responsive to the associated group reading score for the chosen group reading profile. The dictionary management system generates the custom dictionary using the selected words, and provides the generated custom dictionary to the user client.
    Type: Grant
    Filed: May 6, 2016
    Date of Patent: December 6, 2016
    Assignee: Google Inc.
    Inventor: Diego Puppin
  • Patent number: 9514762
    Abstract: The present invention relates to an audio signal coding method and apparatus. The method includes: categorizing audio signals into high-frequency audio signals and low-frequency audio signals; coding the low-frequency audio signals by using a corresponding low-frequency coding manner according to characteristics of low-frequency audio signals; and selecting a bandwidth extension mode to code the high-frequency audio signals according to the low-frequency coding manner and/or characteristics of the audio signals.
    Type: Grant
    Filed: February 1, 2016
    Date of Patent: December 6, 2016
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Lei Miao, Zexin Liu
  • Patent number: 9508339
    Abstract: A method for updating language understanding classifier models includes receiving via one or more microphones of a computing device, a digital voice input from a user of the computing device. Natural language processing using the digital voice input is used to determine a user voice request. Upon determining the user voice request does not match at least one of a plurality of pre-defined voice commands in a schema definition of a digital personal assistant, a GUI of an end-user labeling tool is used to receive a user selection of at least one of the following: at least one intent of a plurality of available intents and/or at least one slot for the at least one intent. A labeled data set is generated by pairing the user voice request and the user selection, and is used to update a language understanding classifier.
    Type: Grant
    Filed: January 30, 2015
    Date of Patent: November 29, 2016
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Vishwac Sena Kannan, Aleksandar Uzelac, Daniel J. Hwang
  • Patent number: 9502049
    Abstract: An audio encoder has a window function controller, a windower, a time warper with a final quality check functionality, a time/frequency converter, a TNS stage or a quantizer encoder, the window function controller, the time warper, the TNS stage or an additional noise filling analyzer are controlled by signal analysis results obtained by a time warp analyzer or a signal classifier. Furthermore, a decoder applies a noise filling operation using a manipulated noise filling estimate depending on a harmonic or speech characteristic of the audio signal.
    Type: Grant
    Filed: November 11, 2014
    Date of Patent: November 22, 2016
    Assignee: Fraunhofer-Gesellschaft zur Foerderung der angewandten Forschung e.V.
    Inventors: Stefan Bayer, Sascha Disch, Ralf Geiger, Guillaume Fuchs, Max Neuendorf, Gerald Schuller, Bernd Edler
  • Patent number: 9502045
    Abstract: In general, techniques are described for coding an ambient higher order ambisonic coefficient. An audio decoding device comprising a memory and a processor may perform the techniques. The memory may store a first frame of a bitstream and a second frame of the bitstream. The processor may obtain, from the first frame, one or more bits indicative of whether the first frame is an independent frame that includes additional reference information to enable the first frame to be decoded without reference to the second frame. The processor may further obtain, in response to the one or more bits indicating that the first frame is not an independent frame, prediction information for first channel side information data of a transport channel. The prediction information may be used to decode the first channel side information data of the transport channel with reference to second channel side information data of the transport channel.
    Type: Grant
    Filed: January 29, 2015
    Date of Patent: November 22, 2016
    Assignee: QUALCOMM Incorporated
    Inventors: Nils Günther Peters, Dipanjan Sen
  • Patent number: 9495972
    Abstract: In an embodiment, bitstream elements of sub-frames are encoded differentially to a global gain value so that a change of the global gain value results in an adjustment of an output level of the decoded representation of the audio content. Concurrently, the differential coding saves bits. Even further, the differential coding enables the lowering of the burden of globally adjusting the gain of an encoded bitstream. In another embodiment, a global gain control across CELP coded frames and transform coded frames is achieved by co-controlling the gain of the codebook excitation of the CELP codec, along with a level of the transform or inverse transform of the transform coded frames. In another embodiment, the gain value determination in CELP coding is performed in the weighted domain of the excitation signal.
    Type: Grant
    Filed: May 27, 2014
    Date of Patent: November 15, 2016
    Assignee: Fraunhofer-Gesellschaft zur Foerderung der angewandten Forschung e.V.
    Inventors: Ralf Geiger, Guillaume Fuchs, Markus Multrus, Bernhard Grill
  • Patent number: 9489450
    Abstract: Disclosed is a method and apparatus for responding to an inquiry from a client via a network. The method and apparatus receive the inquiry from a client via a network. Based on the inquiry, question-answer pairs retrieved from the network are analyzed to determine a response to the inquiry. The QA pairs are not predefined. As a result, the QA pairs have to be analyzed in order to determine whether they are responsive to a particular inquiry. Questions of the QA pairs may be repetitive and, without more, will not be useful in determining whether their corresponding answer responds to an inquiry.
    Type: Grant
    Filed: November 10, 2015
    Date of Patent: November 8, 2016
    Inventors: Junlan Feng, Mazin Gilbert, Dilek Hakkani-Tur, Gokhan Tur
  • Patent number: 9489955
    Abstract: In general, techniques are described for indicating frame parameter reusability for decoding vectors. A device comprising a processor and a memory may perform the techniques. The processor may be configured to obtain a bitstream comprising a vector representative of an orthogonal spatial axis in a spherical harmonics domain. The bitstream may further comprise an indicator for whether to reuse, from a previous frame, at least one syntax element indicative of information used when compressing the vector. The memory may be configured to store the bitstream.
    Type: Grant
    Filed: January 29, 2015
    Date of Patent: November 8, 2016
    Assignee: QUALCOMM Incorporated
    Inventors: Nils Günther Peters, Dipanjan Sen
  • Patent number: 9489171
    Abstract: A computing system is configured to listen to user speech and translate the user speech into voice commands that control operation of the computing system. The identity of a user interacting with the computing system is determined, and a voice command is selected from a set of voice commands based on the user identity. A voice-command suggestion corresponding to the voice command is selected and presented via a display. If the user speaks the voice-command suggestion, the computing system executes the voice command corresponding to the voice-command suggestion.
    Type: Grant
    Filed: January 20, 2015
    Date of Patent: November 8, 2016
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Christian Klein, Gregg Wygonik
  • Patent number: 9484012
    Abstract: According to an embodiment, a speech synthesis dictionary generation apparatus includes an analyzer, a speaker adapter, a level designation unit, and a determination unit. The analyzer analyzes speech data and generates a speech database containing characteristics of utterance by an object speaker. The speaker adapter generates the model of the object speaker by speaker adaptation of converting a base model to be closer to characteristics of the object speaker based on the database. The level designation unit accepts designation of a target speaker level representing a speaker's utterance skill and/or a speaker's native level in a language of the speech synthesis dictionary. The determination determines a parameter related to fidelity of reproduction of speaker properties in the speaker adaptation, in accordance with a relationship between the target speaker level and a speaker level of the object speaker.
    Type: Grant
    Filed: January 27, 2015
    Date of Patent: November 1, 2016
    Assignee: KABUSHIKI KAISHA TOSHIBA
    Inventor: Masahiro Morita
  • Patent number: 9478219
    Abstract: Disclosed are techniques and systems to provide a narration of a text. In some aspects, the techniques and systems described herein include generating a timing file that includes elapsed time information for expected portions of text that provides an elapsed time period from a reference time in an audio recording to each portion of text in recognized portions of text.
    Type: Grant
    Filed: December 1, 2014
    Date of Patent: October 25, 2016
    Assignee: K-NFB Reading Technology, Inc.
    Inventors: Raymond C. Kurzweil, Paul Albrecht, Peter Chapman, Lucy Gibson
  • Patent number: 9472186
    Abstract: An automated system to build a user audio profile for a natural or continuous language speech to text dictation/transcription system is provided. The system uses previously recorded audio files that may have been already transcribed. The previously recorded audio file is split into a plurality of smaller audio files of about 15 seconds in length. The plurality of smaller audio files are matched to the transcribed text (e.g., small text files) or the smaller audio files are transcribed. All, some, or a selection of the small audio files and the small text files are linked as a training pair. The training pair may be edited in certain embodiments herein, both the text and the audio. The training pairs are submitted to the server to build the initial user audio profile.
    Type: Grant
    Filed: January 20, 2015
    Date of Patent: October 18, 2016
    Assignee: NVOQ INCORPORATED
    Inventor: Michael Clark