Patents Examined by Abdelali Serrou
  • Patent number: 9454960
    Abstract: The present invention addresses the deficiencies in the prior art by providing an improved dialog for disambiguating a user utterance containing more than one intent. The invention comprises methods, computer-readable media, and systems for engaging in a dialog. The method embodiment of the invention relates to a method of disambiguating a user utterance containing at least two user intents. The method comprises establishing a confidence threshold for spoken language understanding to encourage that multiple intents are returned, determining whether a received utterance comprises a first intent and a second intent and, if the received utterance contains the first intent and the second intent, disambiguating the first intent and the second intent by presenting a disambiguation sub-dialog wherein the user is offered a choice of which intent to process first, wherein the user is first presented with the intent of the first or second intents having the lowest confidence score.
    Type: Grant
    Filed: April 13, 2015
    Date of Patent: September 27, 2016
    Assignee: AT&T Intellectual Property II, L.P.
    Inventor: Osamuyimen Thompson Stewart
  • Patent number: 9443525
    Abstract: An audio encoder implements multi-channel coding decision, band truncation, multi-channel rematrixing, and header reduction techniques to improve quality and coding efficiency. In the multi-channel coding decision technique, the audio encoder dynamically selects between joint and independent coding of a multi-channel audio signal via an open-loop decision based upon (a) energy separation between the coding channels, and (b) the disparity between excitation patterns of the separate input channels. In the band truncation technique, the audio encoder performs open-loop band truncation at a cut-off frequency based on a target perceptual quality measure. In multi-channel rematrixing technique, the audio encoder suppresses certain coefficients of a difference channel by scaling according to a scale factor, which is based on current average levels of perceptual quality, current rate control buffer fullness, coding mode, and the amount of channel separation in the source.
    Type: Grant
    Filed: June 30, 2014
    Date of Patent: September 13, 2016
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Wei-Ge Chen, Naveen Thumpudi, Ming-Chieh Lee
  • Patent number: 9431020
    Abstract: The present invention proposes a new method and a new apparatus for enhancement of audio source coding systems utilizing high frequency reconstruction (HFR). It utilizes a detection mechanism on the encoder side to assess what parts of the spectrum will not be correctly reproduced by the HFR method in the decoder. Information on this is efficiently coded and sent to the decoder, where it is combined with the output of the HFR unit.
    Type: Grant
    Filed: April 18, 2013
    Date of Patent: August 30, 2016
    Assignee: DOLBY INTERNATIONAL AB
    Inventors: Kristofer Kjoerling, Per Ekstrand, Holger Hoerich
  • Patent number: 9396728
    Abstract: Remote controllers and systems thereof are disclosed. The remote controller remotely operates a receiving host, in which the receiving host provides voice input and speech recognition functions. The remote controller comprises a first input unit and a second input unit for generating a voice input request and a speech recognition request. The generated voice input and speech recognition requests are then sent to the receiving host, thereby forcing the receiving host to perform the voice input and speech recognition functions.
    Type: Grant
    Filed: July 22, 2015
    Date of Patent: July 19, 2016
    Assignee: ASUSTEK COMPUTER INC.
    Inventors: Chia-Chen Liu, Yun-Jung Wu, Liang-Yi Huang, Yi-Hsiu Lee
  • Patent number: 9368112
    Abstract: The disclosure provides a method and an apparatus for detecting a voice activity in an input audio signal composed of frames. A noise characteristic of the input signal is determined based on a received frame of the input audio signal. A voice activity detection (VAD) parameter is derived based on the noise characteristic of the input audio signal using an adaptive function. The derived VAD parameter is compared with a threshold value to provide a voice activity detection decision. The input audio signal is processed according to the voice activity detection decision.
    Type: Grant
    Filed: May 10, 2013
    Date of Patent: June 14, 2016
    Assignee: HUAWEI TECHNOLOGIES CO., LTD
    Inventor: Zhe Wang
  • Patent number: 9317595
    Abstract: Techniques are described herein for automatic generation of a title or summary from a long body of text. A grammatical tree representing one or more sentences of the long body of text is generated. One or more nodes from the grammatical tree are selected to be removed. According to one embodiment, a particular node is selected to be removed based on its position in the grammatical tree and its node-type, where the node type represents a grammatical element of the sentence. Once the particular node is selected, a branch of the tree is cut at the node. After branch has been cut, one or more sub-sentences are generated from the remaining nodes in the grammatical tree. The one or more sub-sentences may be returned as a title or summary.
    Type: Grant
    Filed: December 6, 2010
    Date of Patent: April 19, 2016
    Assignee: Yahoo! Inc.
    Inventors: Xin Li, Hongjian Zhao
  • Patent number: 9317501
    Abstract: A method, computer system, and computer program product for translating information. The computer system receives the information for a translation. The computer system identifies portions of the information based on a set of rules for security for the information in response to receiving the information. The computer system sends the portions of the information to a plurality of translation systems. In response to receiving translation results from the plurality of translation systems for respective portions of the information, the computer system combines the translation results for the respective portions to form a consolidated translation of the information.
    Type: Grant
    Filed: March 12, 2015
    Date of Patent: April 19, 2016
    Assignee: International Business Machines Corporation
    Inventors: Carl J. Kraenzel, David M. Lubensky, Baiju Dhirajlal Mandalia, Cheng Wu
  • Patent number: 9304990
    Abstract: Methods and systems for translating a text into multiple languages performed by at least one software component executed by at least one processor, comprise: maintaining a translation repository having a plurality of entries associating different types of content with user-specified languages; monitoring the text received by a program to identify one or more types of content and a source language of the text; retrieving the user-specified languages from the translation repository associated with the identified types of content; and for each of the identified types of content, translating the content thereof from the source language to the corresponding user-specified language when the source language is different from the corresponding user-specified language.
    Type: Grant
    Filed: August 20, 2012
    Date of Patent: April 5, 2016
    Assignee: International Business Machines Corporation
    Inventors: Judith H. Bank, Liam Harpur, Ruthie D. Lyle, Patrick J. O'Sullivan, Lin Sun
  • Patent number: 9305550
    Abstract: An apparatus and method for tracking dialogue and other sound signals in film, television or other systems with multiple channel sound is described. One or more audio channels which is expected to carry the speech of persons appearing in the program or other particular types of sounds is inspected to determine if that channel's audio includes particular sounds such as MUEVs, including phonemes corresponding to human speech patterns. If an improper number of particular sounds such as phonemes are found in the channel(s) an action such as a report, an alarm, a correction, or other action is taken. The inspection of the audio channel(s) may be made in conjunction with the appearance of corresponding images associated with the sound, such as visemes in the video signal, to improve the determination of types of sounds such as phonemes.
    Type: Grant
    Filed: December 7, 2010
    Date of Patent: April 5, 2016
    Inventors: J. Carl Cooper, Mirko Vojnovic, Christopher Smith
  • Patent number: 9245526
    Abstract: A speech recognition method includes receiving a nametag utterance, decoding the nametag utterance to recognize constituent subwords of the nametag utterance, determining the number of subwords in the nametag utterance, and associating the nametag utterance with one or more of a plurality of different nametag clusters based on the number of subwords in the nametag utterance. According to preferred aspects of the method, a confusability check is performed on the nametag utterance within the cluster(s) associated with the nametag utterance, stored nametags are received from memory by decoding the nametag utterance within the cluster(s) associated with the nametag utterance, and the stored nametags are played back by cluster.
    Type: Grant
    Filed: April 25, 2006
    Date of Patent: January 26, 2016
    Assignee: General Motors LLC
    Inventor: Rathinavelu Chengalvarayan
  • Patent number: 9236048
    Abstract: The present invention discloses a method and device for voice control, which are used to solve the problem of low success rate of voice control in the prior art. The method includes: classifying stored recognition information used for voice recognizing to obtain a syntax packet corresponding to each type of recognition information (10); receiving an inputted voice signal, and performing a voice recognition processing respectively on the received voice signal by using each obtained syntax packet in turn (20), and performing a corresponding control processing based on a voice recognition result of the voice signal according to each syntax packet (30).
    Type: Grant
    Filed: January 12, 2011
    Date of Patent: January 12, 2016
    Assignee: ZTE Corporation
    Inventors: Manhai Li, Kaili Xiao, Jingping Wang, Xin Liao
  • Patent number: 9223901
    Abstract: A method for controlling a program by natural language allows a user to efficiently operate a computer-implemented target program through intuitive natural language commands. A list of natural language commands related to the target program is compiled. Each natural language command is stored as an element in an electronic list. Natural language commands generally consist of short sentences comprising at least a predicate (a verb) and an object (a noun). A user can filter the list of natural language commands by entering the initials of a natural language command. The user enters the first character of the first word to be filtered, followed by the first character of the second word to be filtered, and so forth. Filtering by initials very rapidly reduces the number of choices presented to a user and minimizes the number of keystrokes required to select a particular list element.
    Type: Grant
    Filed: November 9, 2011
    Date of Patent: December 29, 2015
    Inventor: Michael Rabben
  • Patent number: 9218336
    Abstract: Constructing an automaton for automated analysis of agglutinative languages comprises: constructing an affix automaton for each of a plurality of affix types of an agglutinative language, where each of the affix types is associated with one or more affixes associated with a morphological concept; combining any of the affix automatons to form a plurality of template automatons, where each of the template automatons is patterned after any of a plurality of agglutination templates of any of the affix types for the language; and combining the template automatons into a master automaton.
    Type: Grant
    Filed: March 28, 2007
    Date of Patent: December 22, 2015
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Daniel Cohen, Yigal Shai Dayan, Josemina Marcella Magdalen, Victoria Mazel
  • Patent number: 9201870
    Abstract: A method, system and computer-readable storage medium are provided for generating translated dynamic Web page content comprising a phrasebook storing source phrases and translated phrases, a Web server, and an application server that generates dynamic Web page content in response to Web page requests and includes a translation engine that replaces source phrases in the dynamic Web page with a corresponding translated phrase stored in the phrasebook. In another embodiment the phrasebook also stores translations in a second dialect so that source phrases in a dynamic Web page can be translated into two or more dialects.
    Type: Grant
    Filed: January 25, 2008
    Date of Patent: December 1, 2015
    Assignee: First Data Corporation
    Inventors: James Edward Jurach, Jr., David Steven Blumenthal
  • Patent number: 9201876
    Abstract: Methods and apparatus related to contextual weighting of words. Methods are provided for determining co-occurrence relationships between words in a corpus of word groupings and for contextually weighting words in a word grouping as a function of which other words are present in the word grouping.
    Type: Grant
    Filed: May 29, 2012
    Date of Patent: December 1, 2015
    Assignee: GOOGLE INC.
    Inventors: Shailesh Kumar, Mahesh Chhaparia, Shweta Pathak
  • Patent number: 9177551
    Abstract: Disclosed are systems, methods and computer-readable media for enabling speech processing in a user interface of a device. The method includes receiving an indication of a field and a user interface of a device, the indication also signaling that speech will follow, receiving the speech from the user at the device, the speech being associated with the field, transmitting the speech as a request to public, common network node that receives and processes speech, processing the transmitted speech and returning text associated with the speech to the device and inserting the text into the field. Upon a second indication from the user, the system processes the text in the field as programmed by the user interface. The present disclosure provides a speech mash up application for a user interface of a mobile or desktop device that does not require expensive speech processing technologies.
    Type: Grant
    Filed: May 28, 2008
    Date of Patent: November 3, 2015
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Jay Wilpon, Giuseppe Di Fabbrizio, Benjamin J. Stern
  • Patent number: 9159316
    Abstract: A method for generating a speech recognition model includes accessing a baseline speech recognition model, obtaining information related to recent language usage from search queries, and modifying the speech recognition model to revise probabilities of a portion of a sound occurrence based on the information. The portion of a sound may include a word. Also, a method for generating a speech recognition model, includes receiving at a search engine from a remote device an audio recording and a transcript that substantially represents at least a portion of the audio recording, synchronizing the transcript with the audio recording, extracting one or more letters from the transcript and extracting the associated pronunciation of the one or more letters from the audio recording, and generating a dictionary entry in a pronunciation dictionary.
    Type: Grant
    Filed: May 7, 2013
    Date of Patent: October 13, 2015
    Assignee: Google Inc.
    Inventors: Michael H. Cohen, Shumeet Baluja, Pedro J. Mengibar
  • Patent number: 9153244
    Abstract: A voice analyzer includes an apparatus body, a strap that is connected to the apparatus body and is used to hang the apparatus body from a neck of a user, a first voice acquisition unit provided in the strap or the apparatus body, a second voice acquisition unit provided at a position where a distance of a sound wave propagation path from a mouth of the user is smaller than a distance of a sound wave propagation path from the mouth of the user to the first voice acquisition unit, and an identification unit that identifies a sound, in which first sound pressure acquired by the first voice acquisition unit is larger by a predetermined value or more than second sound pressure acquired by the second voice acquisition unit, on the basis of a result of comparison between the first sound pressure and the second sound pressure.
    Type: Grant
    Filed: May 18, 2012
    Date of Patent: October 6, 2015
    Assignee: FUJI XEROX CO., LTD.
    Inventors: Kei Shimotani, Yohei Nishino, Hirohito Yoneyama, Kiyoshi Iida, Akira Fujii, Haruo Harada
  • Patent number: 9153245
    Abstract: A pitch detection method and apparatus are disclosed. The method includes: performing pitch detection on an input signal in a signal domain, and obtaining a candidate pitch; performing linear prediction (LP) on the input signal, and obtaining an LP residual signal; setting a candidate pitch range that includes the candidate pitch; searching the candidate pitch range for the LP residual signal, and obtaining a selected pitch.
    Type: Grant
    Filed: April 9, 2010
    Date of Patent: October 6, 2015
    Assignee: Huawei Technologies Co., Ltd.
    Inventors: Fengyan Qi, Dejun Zhang, Lei Miao, Jianfeng Xu, Qing Zhang, Yang Gao
  • Patent number: 9147392
    Abstract: A speech synthesis device includes: a mouth-opening-degree generation unit which generates, for each of phonemes generated from input text, a mouth-opening-degree corresponding to oral-cavity volume, using information generated from the text and indicating the type and position of the phoneme within the text, such that the generated mouth-opening-degree is larger for a phoneme at the beginning of a sentence in the text than for a phoneme at the end of the sentence; a segment selection unit which selects, for each of the generated phonemes, segment information corresponding to the phoneme from among pieces of segment information stored in a segment storage unit and including phoneme type, mouth-opening-degree, and speech segment data, based on the type of the phoneme and the generated mouth-opening-degree; and a synthesis unit which generates synthetic speech of the text, using the selected pieces of segment information and pieces of prosody information generated from the text.
    Type: Grant
    Filed: May 28, 2013
    Date of Patent: September 29, 2015
    Assignee: PANASONIC INTELLECTUAL PROPERTY MANAGEMENT CO., LTD.
    Inventors: Yoshifumi Hirose, Takahiro Kamai