Patents Examined by Michael N. Opsasnick
  • Patent number: 10559302
    Abstract: A domestic appliance includes a user interface for a user to input commands, a camera for taking an image of an operating area from which the user interface can be operated by the user, a speech recognition device for detecting a speech command, and a control device configured to determine a level of security depending on the image that was taken by the camera and to execute the speech command detected by the speech recognition device depending on the level of security that has been determined.
    Type: Grant
    Filed: February 5, 2016
    Date of Patent: February 11, 2020
    Assignee: BSH Hausgeräte GmbH
    Inventors: Wolfgang Beifuss, Uwe Has
  • Patent number: 10559309
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for collaboration between multiple voice controlled devices are disclosed. In one aspect, a method includes the actions of identifying, by a first computing device, a second computing device that is configured to respond to a particular, predefined hotword; receiving audio data that corresponds to an utterance; receiving a transcription of additional audio data outputted by the second computing device in response to the utterance; based on the transcription of the additional audio data and based on the utterance, generating a transcription that corresponds to a response to the additional audio data; and providing, for output, the transcription that corresponds to the response.
    Type: Grant
    Filed: December 22, 2016
    Date of Patent: February 11, 2020
    Assignee: Google LLC
    Inventors: Victor Carbune, Pedro Gonnet Anders, Thomas Deselaers, Sandro Feuz
  • Patent number: 10559314
    Abstract: In accordance with an example embodiment of the present invention, disclosed is a method and an apparatus thereof for controlling a concealment method for a lost audio frame of a received audio signal. A method for a decoder of concealing a lost audio frame comprises detecting in a property of the previously received and reconstructed audio signal, or in a statistical property of observed frame losses, a condition for which the substitution of a lost frame provides relatively reduced quality. In case such a condition is detected, the concealment method is modified by selectively adjusting a phase or a spectrum magnitude of a substitution frame spectrum.
    Type: Grant
    Filed: May 9, 2019
    Date of Patent: February 11, 2020
    Assignee: Telefonaktiebolaget L M Ericsson (publ)
    Inventors: Stefan Bruhn, Jonas Svedberg
  • Patent number: 10555077
    Abstract: Methods and apparatus for identifying a music service based on a user command. A content type is identified from a received user command and a music service is selected that supports the content type. A selected music service can then transmit audio content associated with the content type for playback.
    Type: Grant
    Filed: October 8, 2018
    Date of Patent: February 4, 2020
    Assignee: Sonos, Inc.
    Inventors: Simon Jarvis, Mark Plagge, Christopher Butts
  • Patent number: 10540994
    Abstract: Aspects relate to computer implemented methods and systems for monitoring a user's hearing and comprehension. The methods and systems include receiving, by an audio capture device, a first audio input, receiving, by the audio capture device, a second audio input, converting the first and second audio inputs into respective first and second audio signals, transcribing the first and second audio signals into respective first and second transcriptions, analyzing, by a processor of the remote resource, the first and second transcriptions to determine if a content of the second transcription is related to a content of the first transcription to determine degradation of hearing of the user.
    Type: Grant
    Filed: April 15, 2019
    Date of Patent: January 21, 2020
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Eli M. Dow, Thomas D. Fitzsimmons, Maurice M. Materise, Jessie Yu
  • Patent number: 10528320
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for enabling screen-specific user interfacing with elements of viewable screens presented by an electronic device are disclosed. In one aspect, a method includes the actions of identifying a character sequence representing a first input that is received while displaying a viewable screen having at least one selectable viewable element. The actions further include accessing an electronic file that provides a text representation of one or more of the at least one selectable viewable element. The actions further include comparing the character sequence to the text representation. The actions further include selecting, within the viewable screen, a selectable viewable element whose text representation matches the character sequence. The actions further include triggering any action linked to the selecting the selectable viewable element.
    Type: Grant
    Filed: February 3, 2017
    Date of Patent: January 7, 2020
    Assignee: Google Technology Holdings LLC
    Inventors: Sanjeev Kumar P. V., Amit K. Agrawal, Satyabrata Rout, Vishal S. Patil
  • Patent number: 10522150
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for collaboration between multiple voice controlled devices are disclosed. In one aspect, a method includes the actions of identifying, by a first computing device, a second computing device that is configured to respond to a particular, predefined hotword; receiving audio data that corresponds to an utterance; receiving a transcription of additional audio data outputted by the second computing device in response to the utterance; based on the transcription of the additional audio data and based on the utterance, generating a transcription that corresponds to a response to the additional audio data; and providing, for output, the transcription that corresponds to the response.
    Type: Grant
    Filed: December 22, 2016
    Date of Patent: December 31, 2019
    Assignee: Google LLC
    Inventors: Victor Carbune, Pedro Gonnet Anders, Thomas Deselaers, Sandro Feuz
  • Patent number: 10522168
    Abstract: An audio signal synthesizer generates a synthesis audio signal having a first frequency band and a second synthesized frequency band derived from the first frequency band and comprises a patch generator, a spectral converter, a raw signal processor and a combiner. The patch generator performs at least two different patching algorithms, each patching algorithm generating a raw signal. The patch generator is adapted to select one of the at least two different patching algorithms in response to a control information. The spectral converter converts the raw signal into a raw signal spectral representation. The raw signal processor processes the raw signal spectral representation in response to spectral domain spectral band replication parameters to obtain an adjusted raw signal spectral representation.
    Type: Grant
    Filed: June 6, 2018
    Date of Patent: December 31, 2019
    Assignee: Fraunhofer-Gesellschaft zur Foerderung der angewandten Forschung e.V.
    Inventors: Frederik Nagel, Sascha Disch, Nikolaus Rettelbach, Max Neuendorf, Bernhard Grill, Ulrich Kraemer, Stefan Wabnik
  • Patent number: 10522152
    Abstract: Systems and methods of diarization using linguistic labeling include receiving a set of diarized textual transcripts. A least one heuristic is automatedly applied to the diarized textual transcripts to select transcripts likely to be associated with an identified group of speakers. The selected transcripts are analyzed to create at least one linguistic model. The linguistic model is applied to transcripted audio data to label a portion of the transcripted audio data as having been spoken by the identified group of speakers. Still further embodiments of diarization using linguistic labeling may serve to label agent speech and customer speech in a recorded and transcripted customer service interaction.
    Type: Grant
    Filed: October 25, 2018
    Date of Patent: December 31, 2019
    Assignee: Verint Systems Ltd.
    Inventors: Omer Ziv, Ran Achituv, Ido Shapira, Jeremie Dreyfuss
  • Patent number: 10522153
    Abstract: Systems and methods of diarization using linguistic labeling include receiving a set of diarized textual transcripts. A least one heuristic is automatedly applied to the diarized textual transcripts to select transcripts likely to be associated with an identified group of speakers. The selected transcripts are analyzed to create at least one linguistic model. The linguistic model is applied to transcripted audio data to label a portion of the transcripted audio data as having been spoken by the identified group of speakers. Still further embodiments of diarization using linguistic labeling may serve to label agent speech and customer speech in a recorded and transcripted customer service interaction.
    Type: Grant
    Filed: October 25, 2018
    Date of Patent: December 31, 2019
    Assignee: Verint Systems Ltd.
    Inventors: Omer Ziv, Ran Achituv, Ido Shapira, Jeremie Dreyfuss
  • Patent number: 10515647
    Abstract: Systems, methods, and storage media for audio processing adapted to voice encoding/decoding are disclosed. In some implementations, a system is adapted to determine prediction coefficients based on coefficient data included within a bitstream to determine quantized prediction coefficients, the coefficient data including one or more model parameters indicating at least one attribute of a signal model. The system inversely quantizes the quantized prediction coefficients to determine dequantized prediction coefficients. The system then determines multiple spectral energy values for multiple corresponding frequency bands based on the dequantized prediction coefficients.
    Type: Grant
    Filed: July 11, 2018
    Date of Patent: December 24, 2019
    Assignee: Dolby International AB
    Inventors: Lars Villemoes, Janusz Klejsa, Per Hedelin
  • Patent number: 10499168
    Abstract: Embodiments of the present invention comprises a system, method, and apparatus that provides for the utilization of a relatively real-time or near real-time interpretation or translation that may be utilized preferably for a relatively short duration of time on a network. A preferred embodiment of the present invention provides online, real-time, short-duration interpreting services in a network-based format. In preferred embodiments, the interpreting system comprises at least one provider computer, such as a server, wherein the provider computer is capable of communicating with user computers via a network. In one preferred embodiment, the provider computer provides a series of web pages that allow access to the interpreting system, including, but not limited to, a request for service page, wherein a user can access the system and input a request for interpreting services. Interpreting services are then provided to a user and a third party desiring to communicate with the user via the network.
    Type: Grant
    Filed: February 8, 2010
    Date of Patent: December 3, 2019
    Assignee: KP Innovations, LLC
    Inventors: Robert Glen Klinefelter, Gregory A. Piccionelli
  • Patent number: 10497375
    Abstract: An apparatus for adapting input audio information, encoding one or more audio objects, to obtain adapted audio information is provided. The input audio information includes two or more input audio downmix channels and further includes input parametric side information. The adapted audio information includes one or more adapted audio downmix channels and further includes adapted parametric side information. The apparatus includes a downmix signal modifier for adapting, depending on adaptation information, the two or more input audio downmix channels to obtain the one or more adapted audio downmix channels. Moreover, the apparatus includes a parametric side information adapter for adapting, depending on the adaptation information, the input parametric side information to obtain the adapted parametric side information.
    Type: Grant
    Filed: February 6, 2015
    Date of Patent: December 3, 2019
    Assignee: Fraunhofer—Gesellschaft zur Foerderung der angewandten Forschung e.V.
    Inventors: Thorsten Kastner, Juergen Herre, Leon Terentiv, Oliver Hellmuth, Jouni Paulus, Falko Ridderbusch
  • Patent number: 10482876
    Abstract: A speech interpretation module interprets the audio of user utterances as sequences of words. To do so, the speech interpretation module parameterizes a literal corpus of expressions by identifying portions of the expressions that correspond to known concepts, and generates a parameterized statistical model from the resulting parameterized corpus. When speech is received the speech interpretation module uses a hierarchical speech recognition decoder that uses both the parameterized statistical model and language sub-models that specify how to recognize a sequence of words. The separation of the language sub-models from the statistical model beneficially reduces the size of the literal corpus needed for training, reduces the size of the resulting model, provides more fine-grained interpretation of concepts, and improves computational efficiency by allowing run-time incorporation of the language sub-models.
    Type: Grant
    Filed: October 1, 2018
    Date of Patent: November 19, 2019
    Assignee: Interactions LLC
    Inventors: Ethan Selfridge, Michael Johnston
  • Patent number: 10460740
    Abstract: A method performed in an audio decoder for decoding M encoded audio channels representing N audio channels is disclosed. The method includes receiving a bitstream containing the M encoded audio channels and a set of spatial parameters, decoding the M encoded audio channels, and extracting the set of spatial parameters from the bitstream. The method also includes analyzing the M audio channels to detect a location of a transient, decorrelating the M audio channels, and deriving N audio channels from the M audio channels and the set of spatial parameters. A first decorrelation technique is applied to a first subset of each audio channel and a second decorrelation technique is applied to a second subset of each audio channel. The first decorrelation technique represents a first mode of operation of a decorrelator, and the second decorrelation technique represents a second mode of operation of the decorrelator.
    Type: Grant
    Filed: December 19, 2018
    Date of Patent: October 29, 2019
    Assignee: Dolby Laboratories Licensing Corporation
    Inventor: Mark F. Davis
  • Patent number: 10453466
    Abstract: A method and apparatus for performing coding and decoding for high-frequency bandwidth extension. The coding apparatus may classify a coding mode of a low-frequency signal of an input signal based on characteristics of the low-frequency signal of an input signal, perform code excited linear prediction coding or audio coding on the LPC excitation signal of the low-frequency signal of an input signal, and perform time-domain (TD) extension coding or frequency-domain (FD) extension coding on a high-frequency signal of an input signal. When the FD extension coding is performed, the coding apparatus may generate a base excitation signal for a high band using an input spectrum, obtain an energy control factor of a sub-band in a frame using the base excitation signal and the input spectrum, generate an energy signal based on the input spectrum and the energy control factor, for the sub-band in the frame, and quantize the energy signal.
    Type: Grant
    Filed: December 10, 2018
    Date of Patent: October 22, 2019
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Ki-hyun Choo, Eun-mi Oh, Ho-sang Sung
  • Patent number: 10453442
    Abstract: A computer-implemented method for automatically analyzing, predicting, and/or modifying acoustic units of prosodic human speech utterances for use in speech synthesis or speech recognition. Possible steps include: initiating analysis of acoustic wave data representing the human speech utterances, via the phase state of the acoustic wave data; using one or more phase state defined acoustic wave metrics as common elements for analyzing, and optionally modifying, pitch, amplitude, duration, and other measurable acoustic parameters of the acoustic wave data, at predetermined time intervals; analyzing acoustic wave data representing a selected acoustic unit to determine the phase state of the acoustic unit; and analyzing the acoustic wave data representing the selected acoustic unit to determine at least one acoustic parameter of the acoustic unit with reference to the determined phase state of the selected acoustic unit. Also included are systems for implementing the described and related methods.
    Type: Grant
    Filed: September 26, 2016
    Date of Patent: October 22, 2019
    Assignee: LESSAC TECHNOLOGIES, INC.
    Inventors: Nishant Chandra, Reiner Wilhelms-Tricarico, Rattima Nitisaroj, Brian Mottershead, Gary A. Marple, John B. Reichenbach
  • Patent number: 10453108
    Abstract: In an example embodiment, input is received from a first user of a computer system. A text object relating to a first item from the input is created, and translated from a first language to a second language. A plurality of text objects, in the second language, having text similar to the translated text object, are located in a database, each text object comprising textual information pertaining to the first item. The plurality of text objects having text similar to the translated text are then ranked based on a comparison of the contextual information about the first item and the contextual information stored in the database for the plurality of text objects having text similar to the translated text object. At least one of the ranked text objects is translated to the first language.
    Type: Grant
    Filed: September 22, 2017
    Date of Patent: October 22, 2019
    Assignee: eBay Inc.
    Inventor: Yan Chelly
  • Patent number: 10446156
    Abstract: Systems and methods of diarization using linguistic labeling include receiving a set of diarized textual transcripts. A least one heuristic is automatedly applied to the diarized textual transcripts to select transcripts likely to be associated with an identified group of speakers. The selected transcripts are analyzed to create at least one linguistic model. The linguistic model is applied to transcripted audio data to label a portion of the transcripted audio data as having been spoken by the identified group of speakers. Still further embodiments of diarization using linguistic labeling may serve to label agent speech and customer speech in a recorded and transcripted customer service interaction.
    Type: Grant
    Filed: October 25, 2018
    Date of Patent: October 15, 2019
    Assignee: Verint Systems Ltd.
    Inventors: Omer Ziv, Ran Achituv, Ido Shapira, Jeremie Dreyfuss
  • Patent number: 10446167
    Abstract: Systems, methods, and devices for user-specific noise suppression are provided. For example, when a voice-related feature of an electronic device is in use, the electronic device may receive an audio signal that includes a user voice. Since noise, such as ambient sounds, also may be received by the electronic device at this time, the electronic device may suppress such noise in the audio signal. In particular, the electronic device may suppress the noise in the audio signal while substantially preserving the user voice via user-specific noise suppression parameters. These user-specific noise suppression parameters may be based at least in part on a user noise suppression preference or a user voice profile, or a combination thereof.
    Type: Grant
    Filed: January 27, 2014
    Date of Patent: October 15, 2019
    Assignee: Apple Inc.
    Inventors: Aram Lindahl, Baptiste Pierre Paquier