Patents Examined by Jesse Pullias
  • Patent number: 9361291
    Abstract: A custom dictionary is generated for an e-book. A dictionary management system receives a custom dictionary request from a user client operated by a user, the custom dictionary request identifying the e-book and including dictionary management information describing the user. The dictionary management system chooses a group reader profile that has an associated group reading score for the user based on the dictionary management information and candidate words are identified in the identified e-book for inclusion in the custom dictionary. The dictionary management system selects words for inclusion in the custom dictionary from among the candidate words responsive to the associated group reading score for the chosen group reading profile. The dictionary management system generates the custom dictionary using the selected words, and provides the generated custom dictionary to the user client.
    Type: Grant
    Filed: March 20, 2015
    Date of Patent: June 7, 2016
    Assignee: GOOGLE INC.
    Inventor: Diego Puppin
  • Patent number: 9361898
    Abstract: A method for encoding three dimensional audio by a wireless communication device is disclosed. The wireless communication device detects an indication of a plurality of localizable audio sources. The wireless communication device also records a plurality of audio signals associated with the plurality of localizable audio sources. The wireless communication device also encodes the plurality of audio signals.
    Type: Grant
    Filed: September 10, 2015
    Date of Patent: June 7, 2016
    Assignee: QUALCOMM Incorporated
    Inventors: Erik Visser, Lae-Hoon Kim, Pei Xiang
  • Patent number: 9361880
    Abstract: Disclosed herein are systems, computer-implemented methods, and computer-readable media for recognizing speech. The method includes receiving speech from a user, perceiving at least one speech dialect in the received speech, selecting at least one grammar from a plurality of optimized dialect grammars based on at least one score associated with the perceived speech dialect and the perceived at least one speech dialect, and recognizing the received speech with the selected at least one grammar. Selecting at least one grammar can be further based on a user profile. Multiple grammars can be blended. Predefined parameters can include pronunciation differences, vocabulary, and sentence structure. Optimized dialect grammars can be domain specific. The method can further include recognizing initial received speech with a generic grammar until an optimized dialect grammar is selected. Selecting at least one grammar from a plurality of optimized dialect grammars can be based on a certainty threshold.
    Type: Grant
    Filed: June 9, 2015
    Date of Patent: June 7, 2016
    Assignee: Interactions LLC
    Inventors: Gregory Pulz, Harry E. Blanchard, Steven H. Lewis, Lan Zhang
  • Patent number: 9356731
    Abstract: Certain aspects of the present disclosure relate to techniques for low-complexity encoding (compression) of broad class of signals, which are typically not well modeled as sparse signals in either time-domain or frequency-domain. First, the signal can be split in time-segments that may be either sparse in time domain or sparse in frequency domain, for example by using absolute second order differential operator on the input signal. Next, different encoding strategies can be applied for each of these time-segments depending in which domain the sparsity is present.
    Type: Grant
    Filed: February 27, 2015
    Date of Patent: May 31, 2016
    Assignee: QUALCOMM Incorporated
    Inventors: Pawan Kumar Baheti, Harinath Garudadri, Yuejie Chi
  • Patent number: 9355650
    Abstract: Devices, systems, methods, media, and programs for detecting an emotional state change in an audio signal are provided. A plurality of segments of the audio signal is received, with the plurality of segments being sequential. Each segment of the plurality of segments is analyzed, and, for each segment, an emotional state and a confidence score of the emotional state are determined. The emotional state and the confidence score of each segment are sequentially analyzed, and a current emotional state of the audio signal is tracked throughout each of the plurality of segments. For each segment, it is determined whether the current emotional state of the audio signal changes to another emotional state based on the emotional state and the confidence score of the segment.
    Type: Grant
    Filed: May 4, 2015
    Date of Patent: May 31, 2016
    Assignee: AT&T INTELLECTUAL PROPERTY I, L.P.
    Inventors: Dimitrios Dimitriadis, Mazin E. Gilbert, Taniya Mishra, Horst J. Schroeter
  • Patent number: 9357321
    Abstract: Described herein are systems, methods, and apparatus for determining audio context between an audio source and an audio sink and selecting signal profiles based at least in part on that audio context. The signal profiles may include noise cancellation which is configured to facilitate operation within the audio context. Audio context may include user-to-user and user-to-device communications.
    Type: Grant
    Filed: September 30, 2014
    Date of Patent: May 31, 2016
    Assignee: Amazon Technologies, Inc.
    Inventors: Stephen M. Polansky, Matthew P. Bell, Yuzo Watanabe
  • Patent number: 9349371
    Abstract: A speech recognition terminal device communicable with a speech recognition server includes a request device for requesting the speech recognition server to carry out the speech recognition of an acquired speech command, a prediction device for predicting a present delay time until a result of the requested speech recognition is obtained from the speech recognition server, a determination device for determining a filler word with a time length in accordance with the predicted present delay time, a filler speaking device for outputting the determined filler word during a waiting time until the result of the requested speech recognition is obtained from the speech recognition server, and a response device for responding to the user when the result of the speech recognition is acquired from the speech recognition server.
    Type: Grant
    Filed: January 13, 2015
    Date of Patent: May 24, 2016
    Assignee: DENSO CORPORATION
    Inventors: Yuki Fujisawa, Toru Nada
  • Patent number: 9349370
    Abstract: A speech recognition terminal device communicable with a speech recognition device includes a request device for requesting the speech recognition device to carry out speech recognition of a acquired speech command, a determination device for determining a filler word filling a waiting time until a result of the requested speech recognition is obtained from the speech recognition device, a filler speaking device for outputting the determined filler word during the waiting time, and a response device for responding to the user based on the acquired result of the speech recognition when the result of the speech recognition is acquired from the speech recognition device.
    Type: Grant
    Filed: January 13, 2015
    Date of Patent: May 24, 2016
    Assignee: DENSO CORPORATION
    Inventors: Yuki Fujisawa, Toru Nada
  • Patent number: 9350860
    Abstract: Systems and method are provided for rendering different speech-based services to a plurality of users. A service-providing system may be accessed via a plurality of connectivity ports. Each of the connectivity ports may be associated with at least one of a plurality of different speech-related services. The connectivity ports may be associated with the different speech-related services may be performed before receiving user service requests. The service-providing system may comprise a plurality of processing components, each of which may be configurable to provide one or more of a plurality of different speech-related services. The service-providing system may further comprise a connection component, which may be operable to establish a connection between the respective connectivity port and a processing component having a configuration of suitable for performing a service requested through the respective connectivity port.
    Type: Grant
    Filed: October 8, 2014
    Date of Patent: May 24, 2016
    Assignee: SWISSCOM AG
    Inventors: Roger Lagadec, Patrik Estermann, Luciano Butera
  • Patent number: 9348805
    Abstract: Systems, methods and apparatus for learning parsing rules and argument identification from crowdsourcing of proposed command inputs are disclosed. Crowdsourcing techniques are used to generate rules for parsing input sentences. A parse is used to determine whether the input sentence invokes a specific action, and if so, what arguments are to be passed to the invocation of the action.
    Type: Grant
    Filed: July 22, 2015
    Date of Patent: May 24, 2016
    Assignee: Google Inc.
    Inventors: Jakob D. Uszkoreit, Percy Liang
  • Patent number: 9342268
    Abstract: Methods, apparatus, and computer-readable media are described herein related to a user interface (UI) that can be implemented on a head-mountable device (HMD). The UI can include a voice-navigable UI. The voice-navigable UI can include a voice navigable menu that includes one or more menu items. The voice-navigable UI can also present a first visible menu that includes at least a portion of the voice navigable menu. In response to a first utterance comprising one of the one or more menu items, the voice-navigable UI can modify the first visible menu to display one or more commands associated with the first menu item. In response to a second utterance comprising a first command, the voice-navigable UI can invoke the first command. In some embodiments, the voice-navigable UI can display a second visible menu, where the first command can be displayed above other menu items in the second visible menu.
    Type: Grant
    Filed: October 15, 2015
    Date of Patent: May 17, 2016
    Assignee: Google Inc.
    Inventors: Michael J. LeBeau, Clifford Ivar Nass
  • Patent number: 9335965
    Abstract: An apparatus includes at least one input device configured to receive a speech input, a display configured to present predetermined content acquired by the apparatus from which excerpts may be extracted, and a processor configured to execute computer readable program code. The computer readable program code is configured to collect a speech recognition vocabulary set that corresponds to content visible on the display such that at least one vocabulary word in the speech recognition vocabulary set is the same as a word presented on the display, designate a segment of the content to be excerpted based on the speech input, and create a link to a source of the excerpted content and display the link with the excerpted content.
    Type: Grant
    Filed: August 25, 2014
    Date of Patent: May 10, 2016
    Assignee: Core Wireless Licensing S.a.r.l.
    Inventors: Huanglingzi Liu, Yue Zhong Tang, Yu Zhang
  • Patent number: 9330089
    Abstract: Automated user-machine interaction is gaining attraction in many applications and services. However, implementing and offering smart automated user-machine interaction services still present technical challenges. According to at least one example embodiment, a dialogue manager is configured to handle multiple dialogue applications independent of the language, the input modalities, or output modalities used. The dialogue manager employs generic semantic representation of user-input data. At a step of a dialogue, the dialogue manager determines whether the user-input data is indicative of a new request or a refinement request based on the generic semantic representation and at least one of a maintained state of the dialogue, general knowledge data representing one or more concepts, and data representing history of the dialogue. The dialogue manager then responds to determined user-request with multi-facet output data to a client dialogue application indicating action(s) to be performed.
    Type: Grant
    Filed: November 6, 2015
    Date of Patent: May 3, 2016
    Assignee: Nuance Communications, Inc.
    Inventors: Simona Gandrabur, Eric Buist, Andrei Dragoi, Alireza Salimi
  • Patent number: 9330669
    Abstract: A system and method is presented for performing dual mode speech recognition, employing a local recognition module on a mobile device and a remote recognition engine on a server device. The system accepts a spoken query from a user, and both the local recognition module and the remote recognition engine perform speech recognition operations on the query, returning a transcription and confidence score, subject to a latency cutoff time. If both sources successfully transcribe the query, then the system accepts the result having the higher confidence score. If only one source succeeds, then that result is accepted. In either case, if the remote recognition engine does succeed in transcribing the query, then a client vocabulary is updated if the remote system result includes information not present in the client vocabulary.
    Type: Grant
    Filed: February 12, 2015
    Date of Patent: May 3, 2016
    Assignee: SoundHound, Inc.
    Inventors: Timothy P. Stonehocker, Keyvan Mohajer, Bernard Mont-Reynaud
  • Patent number: 9324317
    Abstract: Disclosed herein are systems, methods, and computer readable-media for providing an automatic synthetically generated voice describing media content, the method comprising receiving one or more pieces of metadata for a primary media content, selecting at least one piece of metadata for output, and outputting the at least one piece of metadata as synthetically generated speech with the primary media content. Other aspects of the invention involve alternative output, output speech simultaneously with the primary media content, output speech during gaps in the primary media content, translate metadata in foreign language, tailor voice, accent, and language to match the metadata and/or primary media content. A user may control output via a user interface or output may be customized based on preferences in a user profile.
    Type: Grant
    Filed: September 9, 2014
    Date of Patent: April 26, 2016
    Assignee: AT&T Intellectual Property I, L.P.
    Inventors: Linda Roberts, Hong Thi Nguyen, Horst J. Schroeter
  • Patent number: 9317498
    Abstract: Systems and methods for summarizing online articles for consumption on a user device are disclosed herein. The system extracts the main body of an article's text from the HTML code of an online article. The system may then classify the extracted article into one of several different categories and removes duplicate articles. The system breaks down the article into its component sentences, and each sentence is classified into one of three categories: (1) potential candidate sentences that may be included in the generated summary; (2) weakly rejected sentences that will not be included in the summary but may be used to generate the summary; and (3) strongly rejected sentences that are not included in the summary. Finally, the system applies a document summarizer to generate quickly readable article summaries, for viewing on the user device, using relevant sentences from the article while maintaining the coherence of the article.
    Type: Grant
    Filed: April 8, 2015
    Date of Patent: April 19, 2016
    Assignee: CODEQ LLC
    Inventors: Douglas Dane Baker, Paulo Malvar Fernández, Brian Fernandes, Rodrigo Alarcón Martinez
  • Patent number: 9318120
    Abstract: An exemplary noise reduction system and method processes a speech signal that is delivered in a noisy channel or with ambient noise. Some exemplary embodiments of the system and method use filters to extract speech information, and focus on a subset of harmonics that are least corrupted by noise. Some exemplary embodiments disregard signal harmonics with low signal-to-noise ratio(s), and disregard amplitude modulations that are inconsistent with speech. An exemplary system and method processes a signal that focuses on a subset of harmonics that are least corrupted by noise, disregards the signal harmonics with low signal-to-noise ratio(s), and disregards amplitude modulations that are inconsistent with speech.
    Type: Grant
    Filed: February 9, 2015
    Date of Patent: April 19, 2016
    Assignee: SCTI HOLDINGS, INC.
    Inventor: Mark Pinson
  • Patent number: 9318105
    Abstract: In one embodiment, the invention provides a method for a speech recognition system to select a return value corresponding to a spoken input. The method comprises generating a dictionary comprising return values associated with data provisioned in the speech recognition system; generating a grammar for each return value in the dictionary; analyzing the grammar to determine a subset of return values from the dictionary that are likely alternatives for each return value in the dictionary, based on the grammar; selecting a return value corresponding to the spoken input based on the grammar; and if the selected return value is not confirmed by a user, then presenting the likely alternative for the selected return value to the user.
    Type: Grant
    Filed: June 26, 2013
    Date of Patent: April 19, 2016
    Assignee: West Interactive Corporation II
    Inventor: Ashok M. Khosla
  • Patent number: 9311291
    Abstract: Methods and a system for calculating N-gram probabilities in a language model. A method includes counting N-grams in each page of a plurality of pages or in each document of a plurality of documents to obtain respective N-gram counts therefor. The method further includes applying weights to the respective N-gram counts based on at least one of view counts and rankings to obtain weighted respective N-gram counts. The view counts and the rankings are determined with respect to the plurality of pages or the plurality of documents. The method also includes merging the weighted respective N-gram counts to obtain merged weighted respective N-gram counts for the plurality of pages or the plurality of documents. The method additionally includes calculating a respective probability for each of the N-grams based on the merged weighted respective N-gram counts.
    Type: Grant
    Filed: September 9, 2013
    Date of Patent: April 12, 2016
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Nathan M. Bodenstab, Nobuyasu Itoh, Gakuto Kurata, Masafumi Nishimura, Paul J. Vozila
  • Patent number: 9311929
    Abstract: A speech analysis system uses one or more digital processors to reconstruct a speech signal by accurately extracting speech formants from a digitized version of the speech signal. The system extracts the formants by determining an estimated instantaneous frequency and an estimated instantaneous bandwidth of speech resonances of the digital version of the speech signal in real time. The system digitally filters the digital speech signal using a plurality of complex digital filters in parallel having overlapping bandwidths to ensure that substantially all of the bandwidth of the speech signal is covered. This virtual chain of overlapping complex digital filters produces a corresponding plurality of complex filtered signals. A first estimated frequency and a first estimated bandwidth is generated for each of the filtered signals, and speech resonances of the input speech signal are identified therefrom.
    Type: Grant
    Filed: October 31, 2012
    Date of Patent: April 12, 2016
    Assignee: Eliza Corporation
    Inventors: John P. Kroeker, Janet Slifka, Richard S. McGowan