Patents Examined by Daniel Abebe
  • Patent number: 10553204
    Abstract: A computer-implemented conversational agent engages in a natural language conversation with a user, interpreting the natural language conversation by parsing and tokenizing utterances in the natural language conversation. Based on interpreting, a set of utterances in the natural language conversation to be recorded as a macro is determined. The macro is stored in a database with an associated macro identifier. Replaying of the macro executes a function specified in the set of utterances.
    Type: Grant
    Filed: December 21, 2017
    Date of Patent: February 4, 2020
    Assignee: International Business Machines Corporation
    Inventors: Martin Hirzel, Louis Mandel, Avraham E. Shinnar, Jerome Simeon, Mandana Vaziri
  • Patent number: 10546580
    Abstract: Methods, systems, and vehicle components for providing a corrected pronunciation suggestion to a user are disclosed. A method includes receiving, by a microphone communicatively coupled to a processing device, a voice input from the user, the voice input including a particularly pronounced word. The method further includes comparing, by the processing device, the particularly pronounced word to one or more reference words in a reference table, determining, by the processing device, that the particularly pronounced word has been potentially mispronounced by the user based on the one or more reference words in the reference table, determining, by the processing device, a corrected pronunciation suggestion from the one or more reference words, and providing, via a user interface, the corrected pronunciation suggestion to the user.
    Type: Grant
    Filed: December 5, 2017
    Date of Patent: January 28, 2020
    Assignee: TOYOTA MOTOR ENGINEERING & MANUFACUTURING NORTH AMERICA, INC.
    Inventors: Scott A. Friedman, Prince R. Remegio, Tim Uwe Falkenmayer, Roger Akira Kyle, Ryoma Kakimi, Luke D. Heide, Nishikant Narayan Puranik
  • Patent number: 10546590
    Abstract: Audio signal processing enhances audio watermark embedding and detecting processes. Audio signal processes include audio classification and adapting watermark embedding and detecting based on classification. Advances in audio watermark design include adaptive watermark signal structure data protocols, perceptual models, and insertion methods. Perceptual and robustness evaluation is integrated into audio watermark embedding to optimize audio quality relative the original signal, and to optimize robustness or data capacity. These methods are applied to audio segments in audio embedder and detector configurations to support real time operation. Feature extraction and matching are also used to adapt audio watermark embedding and detecting.
    Type: Grant
    Filed: December 20, 2017
    Date of Patent: January 28, 2020
    Assignee: Digimarc Corporation
    Inventors: Ravi K. Sharma, Brett A. Bradley, Yang Bai, Shankar Thagadur Shivappa, Aparna Gurijala
  • Patent number: 10535358
    Abstract: An apparatus and a method to encode and decode a speech signal using an encoding mode are provided. An encoding apparatus may select an encoding mode of a frame included in an input speech signal, and encode a frame having an unvoiced mode for an unvoiced speech as the selected encoding mode.
    Type: Grant
    Filed: February 8, 2018
    Date of Patent: January 14, 2020
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Ho Sang Sung, Ki Hyun Choo, Jung Hoe Kim, Eun Mi Oh
  • Patent number: 10529336
    Abstract: Technology is described for removing sensitive information. An audio block that represents a portion of a conversation may be identified. A text representation for the audio block may be obtained using a speech-to-text process. The text representation for the audio block may be compared to pattern rules to mark sensitive information in the audio block. A portion of audio data from the audio block marked as sensitive information may be removed in the audio block.
    Type: Grant
    Filed: September 13, 2017
    Date of Patent: January 7, 2020
    Assignee: Amazon Technologies, Inc.
    Inventors: Nicholas Channing Matthews, Jeddel Yeras
  • Patent number: 10521190
    Abstract: A user speech interface for interactive media guidance applications, such as television program guides, guides for audio services, guides for video-on-demand (VOD) services, guides for personal video recorders (PVRs), or other suitable guidance applications is provided. Voice commands may be received from a user and guidance activities may be performed in response to the voice commands.
    Type: Grant
    Filed: June 25, 2018
    Date of Patent: December 31, 2019
    Assignee: Rovi Guides, Inc.
    Inventors: M. Scott Reichardt, David M. Berezowski, Michael D. Ellis, Toby DeWeese
  • Patent number: 10521476
    Abstract: An offline semantic processor of a resource-constrained voice-enabled device such as a mobile device utilizes an offline grammar model with reduced resource requirements to parse voice-based queries received by the device. The offline grammar model may be generated from a larger and more comprehensive grammar model used by an online voice-based query processor, and the generation of the offline grammar model may be based upon query usage data collected from one or more users to enable a subset of more popular voice-based queries from the online grammar model to be incorporated into the offline grammar model. In addition, such a device may collect query usage data and upload such data to an online service to enable an updated offline grammar model to be generated and downloaded back to the device and thereby enable a dynamic update of the offline grammar model to be performed.
    Type: Grant
    Filed: February 4, 2018
    Date of Patent: December 31, 2019
    Assignee: GOOGLE LLC
    Inventors: Sangsoo Sung, Yuli Gao, Prathab Murugesan
  • Patent number: 10522135
    Abstract: A system and method for segmenting an audio file. The method includes analyzing an audio file, wherein the analyzing includes identifying speech recognition features within the audio file; generating metadata based on the audio file, wherein the metadata includes transcription characteristics of the audio file; and determining a segmenting interval for the audio file based on the speech recognition features and the metadata.
    Type: Grant
    Filed: December 31, 2017
    Date of Patent: December 31, 2019
    Assignee: Verbit Software Ltd.
    Inventors: Tom Livne, Kobi Ben Tzvi, Eric Shellef
  • Patent number: 10514815
    Abstract: A computer readable medium containing a set of instructions that causes a computer to perform a process comprised of receiving one or more media files. The one or more media files having one or more scenes and each scene including a starting time point and ending time point. The set of instructions may include changing the starting time point and/or the ending time point of a scene from the one or more scenes in response to an input command. The set of instructions may create a new scene and save the new scene based on the new starting time point and/or ending time point of the scene.
    Type: Grant
    Filed: December 20, 2017
    Date of Patent: December 24, 2019
    Assignee: Thomas Majchrowski & Associates, Inc.
    Inventor: Keri DeWitt
  • Patent number: 10515626
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for neural network adaptive beamforming for multichannel speech recognition are disclosed. In one aspect, a method includes the actions of receiving a first channel of audio data corresponding to an utterance and a second channel of audio data corresponding to the utterance. The actions further include generating a first set of filter parameters for a first filter based on the first channel of audio data and the second channel of audio data and a second set of filter parameters for a second filter based on the first channel of audio data and the second channel of audio data. The actions further include generating a single combined channel of audio data. The actions further include inputting the audio data to a neural network. The actions further include providing a transcription for the utterance.
    Type: Grant
    Filed: December 20, 2017
    Date of Patent: December 24, 2019
    Assignee: Google LLC
    Inventors: Bo Li, Ron J. Weiss, Michiel A. U. Bacchiani, Tara N. Sainath, Kevin William Wilson
  • Patent number: 10490196
    Abstract: A voice encoder/decoder (vocoder) may provide receiving a voice sample and generating zero crossings of the voice sample in response to voice excitation in a first formant and creating a corresponding output signal. Additional operations may include dividing the output signal by two, and sampling the output signal at a predefined frequency such that a resulting combination uses half of a bit rate for an excitation and a remainder for short term spectrum analysis.
    Type: Grant
    Filed: February 6, 2018
    Date of Patent: November 26, 2019
    Assignee: OPEN INVENTION NETWORK LLC
    Inventor: Clyde Holmes
  • Patent number: 10490188
    Abstract: A method and system for language selection and synchronization in a vehicle are provided. The method includes receiving an audio representative of sounds captured within a vehicle, recognizing a language category for propagating information to a user of the vehicle according to the received audio, selecting the language category of the vehicle system according to the recognized language category in response to receiving a user acknowledgment, synchronizing the language category among a plurality of vehicle systems, and propagating information to the user of the vehicle using the synchronized language category.
    Type: Grant
    Filed: September 12, 2017
    Date of Patent: November 26, 2019
    Assignee: TOYOTA MOTOR ENGINEERING & MANUFACTURING NORTH AMERICA, INC.
    Inventors: Ming Michael Meng, Krishna Buddharaju
  • Patent number: 10482893
    Abstract: A sound processing method includes a step of applying a nonlinear filter to a temporal sequence of spectral envelope of an acoustic signal, wherein the nonlinear filter smooths a fine temporal perturbation of the spectral envelope without smoothing out a large temporal change. A sound processing apparatus includes a smoothing processor configured to apply a nonlinear filter to a temporal sequence of spectral envelope of an acoustic signal, wherein the nonlinear filter smooths a fine temporal perturbation of the spectral envelope without smoothing out a large temporal change.
    Type: Grant
    Filed: November 1, 2017
    Date of Patent: November 19, 2019
    Assignee: YAMAHA CORPORATION
    Inventors: Ryunosuke Daido, Hiraku Kayama
  • Patent number: 10482159
    Abstract: Aspects create a multimedia presentation wherein processors are configured to calculate a time it would take to narrate a plurality of words in a document at a specified speech speed in response to determining that the time it would take to narrate the plurality of words in the document at the specified speech speed exceeds a specified maximum time, generate a long summary of the document as a subset of the plurality of words, generate audio content for a first portion of the plurality of words of the long summary by applying a text-to-speech processing mechanism to the portion of the long summary at the desired speech speed, and create a multimedia slide of a multimedia presentation by adding the generated audio content to a presentation of text from a remainder portion of the plurality of words of the long summary.
    Type: Grant
    Filed: November 2, 2017
    Date of Patent: November 19, 2019
    Assignee: International Business Machines Corporation
    Inventors: Nicolas Bainer, Dario Alejando Falasca, Federico Tomas Gimenez Molinelli, Nicolas O. Nappe, Gaston Alejo Rius, Nicolas Tcherechansky, Facundo J. Tomaselli
  • Patent number: 10482904
    Abstract: This disclosure describes, in part, context-driven device arbitration techniques to select a speech interface device from multiple speech interface devices to provide a response to a command included in a speech utterance of a user. In some examples, the context-driven arbitration techniques may include executing multiple pipeline instances to analyze audio signals and device metadata received from each of the multiple speech interface devices which detected the speech utterance. A remote speech processing service may execute the multiple pipeline instances and analyze the audio signals and/or metadata, at various stages of the pipeline instances, to determine which speech interface device is to respond to the speech utterance.
    Type: Grant
    Filed: August 15, 2017
    Date of Patent: November 19, 2019
    Assignee: Amazon Technologies, Inc.
    Inventors: Tony Roy Hardie, Brian Alexander Oliver, Vikram Kumar Gundeti
  • Patent number: 10482892
    Abstract: System and method embodiments are provided for very short pitch detection and coding for speech or audio signals. The system and method include detecting whether there is a very short pitch lag in a speech or audio signal that is shorter than a conventional minimum pitch limitation using a combination of time domain and frequency domain pitch detection techniques. The pitch detection techniques include using pitch correlations in time domain and detecting a lack of low frequency energy in the speech or audio signal in frequency domain. The detected very short pitch lag is coded using a pitch range from a predetermined minimum very short pitch limitation that is smaller than the conventional minimum pitch limitation.
    Type: Grant
    Filed: July 28, 2017
    Date of Patent: November 19, 2019
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Yang Gao, Fengyan Qi
  • Patent number: 10460215
    Abstract: A method for natural language interaction includes recording speech provided by a human user. The recorded speech is translated into a machine-readable natural language input relating to an interaction topic. An interaction timer is maintained that tracks a length of time since a last machine-readable natural language input referring to the interaction topic was translated. Based on a current value of the interaction timer being greater than an interaction engagement threshold, a message relating to the interaction topic is delivered with a first natural language phrasing that includes an interaction topic reminder. Based on the current value of the interaction timer being less than the interaction engagement threshold, the message relating to the interaction topic is delivered with a second natural language phrasing that lacks the interaction topic reminder.
    Type: Grant
    Filed: July 21, 2017
    Date of Patent: October 29, 2019
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Keith Coleman Herold, Oz Solomon
  • Patent number: 10460738
    Abstract: Disclosed is an apparatus for processing an input signal, having a perceptual weighter and a quantizer. The perceptual weighter has a model provider and a model applicator. The model provider provides a perceptual weighted model based on the input signal. The model applicator provides a perceptually weighted spectrum by applying the perceptual weighted model to a spectrum based on the input signal. The quantizer is configured to quantize the perceptually weighted spectrum and for providing a bitstream. The quantizer has a random matrix applicator and a sign function calculator. The random matrix applicator is configured for applying a random matrix to the perceptually weighted spectrum in order to provide a transformed spectrum. The sign function calculator is configured for calculating a sign function of components of the transformed spectrum in order to provide the bitstream. The invention further refers to an apparatus for processing an encoded signal and to corresponding methods.
    Type: Grant
    Filed: March 13, 2017
    Date of Patent: October 29, 2019
    Assignee: Fraunhofer-Gesellschaft zur Foerderung der angewandten Forschung e.V.
    Inventors: Tom Baeckstroem, Florin Ghido, Johannes Fischer
  • Patent number: 10460035
    Abstract: Technologies are provided for determining deficiencies in narrative textual data that may impact decision-making in a decisional context. A candidate text document and a reference corpus of text may be utilized to generate one or more topic models and document-term matrices, and then to determine a corresponding statistical perplexity and probabilistic coherence. The reference corpus may be known or normatively deemed to be epistemically persuasive and adequate with respect to a particular context. Statistical determinations of a degree to which the candidate deviates from the reference normative corpus are determined, in terms of the statistical perplexity and probabilistic coherence of the candidate as compared to the reference.
    Type: Grant
    Filed: December 26, 2017
    Date of Patent: October 29, 2019
    Assignee: Cerner Innovation, Inc.
    Inventor: Douglas S. McNair
  • Patent number: 10453447
    Abstract: Disclosed herein is a system, method, and computer program product for filtering data in an audio stream. Incoming audio is converted into text and analyzed for information of a first type. Based on the detection of information of the first type from a source, a recipient is blocked from receiving the audible information.
    Type: Grant
    Filed: November 28, 2017
    Date of Patent: October 22, 2019
    Assignee: International Business Machines Corporation
    Inventors: Jeffrey A. Schmidt, Adam D. Braham, Jason A. Nikolai, John M. Santosuosso