Patents Examined by Michael N. Opsasnick
  • Patent number: 10446167
    Abstract: Systems, methods, and devices for user-specific noise suppression are provided. For example, when a voice-related feature of an electronic device is in use, the electronic device may receive an audio signal that includes a user voice. Since noise, such as ambient sounds, also may be received by the electronic device at this time, the electronic device may suppress such noise in the audio signal. In particular, the electronic device may suppress the noise in the audio signal while substantially preserving the user voice via user-specific noise suppression parameters. These user-specific noise suppression parameters may be based at least in part on a user noise suppression preference or a user voice profile, or a combination thereof.
    Type: Grant
    Filed: January 27, 2014
    Date of Patent: October 15, 2019
    Assignee: Apple Inc.
    Inventors: Aram Lindahl, Baptiste Pierre Paquier
  • Patent number: 10446156
    Abstract: Systems and methods of diarization using linguistic labeling include receiving a set of diarized textual transcripts. A least one heuristic is automatedly applied to the diarized textual transcripts to select transcripts likely to be associated with an identified group of speakers. The selected transcripts are analyzed to create at least one linguistic model. The linguistic model is applied to transcripted audio data to label a portion of the transcripted audio data as having been spoken by the identified group of speakers. Still further embodiments of diarization using linguistic labeling may serve to label agent speech and customer speech in a recorded and transcripted customer service interaction.
    Type: Grant
    Filed: October 25, 2018
    Date of Patent: October 15, 2019
    Assignee: Verint Systems Ltd.
    Inventors: Omer Ziv, Ran Achituv, Ido Shapira, Jeremie Dreyfuss
  • Patent number: 10438595
    Abstract: Systems and processes for generating a speaker profile for use in performing speaker identification for a virtual assistant are provided. One example process can include receiving an audio input including user speech and determining whether a speaker of the user speech is a predetermined user based on a speaker profile for the predetermined user. In response to determining that the speaker of the user speech is the predetermined user, the user speech can be added to the speaker profile and operation of the virtual assistant can be triggered. In response to determining that the speaker of the user speech is not the predetermined user, the user speech can be added to an alternate speaker profile and operation of the virtual assistant may not be triggered. In some examples, contextual information can be used to verify results produced by the speaker identification process.
    Type: Grant
    Filed: October 9, 2018
    Date of Patent: October 8, 2019
    Assignee: Apple Inc.
    Inventors: Yoon Kim, Sachin S. Kajarekar
  • Patent number: 10438592
    Abstract: Systems and method of diarization of audio files use an acoustic voiceprint model. A plurality of audio files are analyzed to arrive at an acoustic voiceprint model associated to an identified speaker. Metadata associate with an audio file is used to select an acoustic voiceprint model. The selected acoustic voiceprint model is applied in a diarization to identify audio data of the identified speaker.
    Type: Grant
    Filed: October 25, 2018
    Date of Patent: October 8, 2019
    Assignee: Verint Systems Ltd.
    Inventors: Omer Ziv, Ran Achituv, Ido Shapira, Jeremie Dreyfuss
  • Patent number: 10431211
    Abstract: An apparatus includes multiple microphones to generate audio signals based on sound of a far-field acoustic environment. The apparatus also includes a signal processing system to process the audio signals to generate at least one processed audio signal. The signal processing system is configured to update one or more processing parameters while operating in a first operational mode and is configured to use a static version of the one or more processing parameters while operating in the second operational mode. The apparatus further includes a keyword detection system to perform keyword detection based on the at least one processed audio signal to determine whether the sound includes an utterance corresponding to a keyword and, based on a result of the keyword detection, to send a control signal to the signal processing system to change an operational mode of the signal processing system.
    Type: Grant
    Filed: December 21, 2016
    Date of Patent: October 1, 2019
    Assignee: Qualcomm Incorporated
    Inventors: Lae-Hoon Kim, Erik Visser, Asif Mohammad, Ian Ernan Liu, Ye Jiang
  • Patent number: 10430024
    Abstract: A storage machine holds instructions executable by a logic machine to receive a digital representation of a spoken command. The digital representation is provided to a speech recognizer trained with a user-specific grammar library. The logic machine then receives from the speech recognizer a confidence rating for each of a plurality of different media items. The confidence rating indicates the likelihood that the media item is named in the spoken command. The logic machine then automatically plays back the media item with a greatest confidence rating.
    Type: Grant
    Filed: November 13, 2013
    Date of Patent: October 1, 2019
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Vignesh Sachidanandam, Joe Downing, Cory Cirrincione, Catherine Pham
  • Patent number: 10431229
    Abstract: A signal processing device, method, and program that may obtain audio at a higher audio quality when decoding an audio signal. An envelope information generating unit generates envelope information representing an envelope form of high frequency components of an audio signal to be encoded. A sine wave information generating unit extracts a sine wave signal from the high frequency components of the audio signal, and generates a sine wave information representing an emergence start position of the sine wave signal. An encoding stream generating unit multiplexes the envelope information, the sine wave information, and low frequency components of the audio signal that have been encoded, and outputs an encoding stream obtained as the result. The high frequency components included in the sine wave signal may be predicted at a higher accuracy from the envelope information and the sine wave information at the receiving side of the encoding stream.
    Type: Grant
    Filed: February 3, 2017
    Date of Patent: October 1, 2019
    Assignee: Sony Corporation
    Inventors: Mitsuyuki Hatanaka, Toru Chinen
  • Patent number: 10424292
    Abstract: An audio controlled assistant captures environmental noise and converts the environmental noise into audio signals. The audio signals are provided to a system which analyzes the audio signals for a plurality of audio prompts, which have been customized for the acoustic environment surrounding the audio controlled assistant by an acoustic modeling system. The system configured to detect the presence of an audio prompt in the audio signals and transmit instructions associated with the detected audio prompt to at least one of the audio controlled assistant or one or more cloud based services, in response.
    Type: Grant
    Filed: March 14, 2013
    Date of Patent: September 24, 2019
    Assignee: Amazon Technologies, Inc.
    Inventors: John Daniel Thimsen, Gregory Michael Hart, Ryan Paul Thomas
  • Patent number: 10418035
    Abstract: A method for electronically utilizing content in a communication between a customer and a customer representative is provided. An audible conversation between a customer and a service representative is captured. At least a portion of the audible conversation is converted into computer searchable data. The computer searchable data is analyzed during the audible conversation to identify relevant meta tags previously stored in a data repository or generated during the audible conversation. Each meta tag is associated with the customer. Each meta tag provides a contextual item determined from at least a portion of one of a current or previous conversation with the customer. A meta tag determined to be relevant to the current conversation between the service representative and the customer is displayed in real time to the service representative currently conversing with the customer.
    Type: Grant
    Filed: October 18, 2018
    Date of Patent: September 17, 2019
    Assignee: United Services Automobile Association
    Inventors: Zakery L. Johnson, Jonathan E. Neuse
  • Patent number: 10402489
    Abstract: Embodiments are disclosed for transliterating text entries across different script systems. A method according to some embodiments includes steps of: receiving an input string in a first script system input using a keyboard; segmenting, using a probabilistic model, the input string into phonemes that correspond to characters or sets of characters in a second script system; converting the phonemes in the first script system into the characters or sets of characters in the second script system, the characters or sets of characters forming a word or a word prefix in the second script system; and outputting the word or the word prefix in the second script system.
    Type: Grant
    Filed: December 21, 2016
    Date of Patent: September 3, 2019
    Assignee: FACEBOOK, INC.
    Inventors: Juan Miguel Pino, Stanislav Funiak, Mridul Malpani, Gaurav Lochan
  • Patent number: 10403297
    Abstract: A method performed in an audio decoder for decoding M encoded audio channels representing N audio channels is disclosed. The method includes receiving a bitstream containing the M encoded audio channels and a set of spatial parameters, decoding the M encoded audio channels, and extracting the set of spatial parameters from the bitstream. The method also includes analyzing the M audio channels to detect a location of a transient, decorrelating the M audio channels, and deriving N audio channels from the M audio channels and the set of spatial parameters. A first decorrelation technique is applied to a first subset of each audio channel and a second decorrelation technique is applied to a second subset of each audio channel. The first decorrelation technique represents a first mode of operation of a decorrelator, and the second decorrelation technique represents a second mode of operation of the decorrelator.
    Type: Grant
    Filed: December 19, 2018
    Date of Patent: September 3, 2019
    Assignee: Dolby Laboratories Licensing Corporation
    Inventor: Mark F. Davis
  • Patent number: 10394960
    Abstract: Embodiments are disclosed for transliteration decoding using a tree structure. A method according to some embodiments includes steps of: generating a tree structure for an input string in a first script system, the tree structure including nodes representing segments of the input string; identifying segmentation candidates for the input string based on paths of the tree structure, the segmentation candidates segmenting the input string into character groups; selecting a segmentation candidate based on probabilities of the segmentation candidates predicted by a probabilistic model; segmenting the input string into character groups that correspond to characters in a second script system; decoding the character groups in the first script system into the characters in the second script system, the characters forming a word or a word prefix in the second script system; and outputting the word or the word prefix in the second script system.
    Type: Grant
    Filed: December 21, 2016
    Date of Patent: August 27, 2019
    Assignee: FACEBOOK, INC.
    Inventors: Juan Miguel Pino, Stanislav Funiak, Mridul Malpani, Gaurav Lochan
  • Patent number: 10367928
    Abstract: A method and system on an electronic device which uses speech recognition to initiate a communication from a mobile device having access to contact information for a number of contacts. In one example, the method comprises receiving through an audio input interface a voice input for initiating a communication, extracting from the voice input a type of communication and at least part of a contact name, and outputting, to an output interface, a selectable list of all contacts from the contact information which have the part of the contact name and which have a contact address associated with the type of communication. The mobile device may also be configured to access remote contact information from a remote server.
    Type: Grant
    Filed: July 28, 2017
    Date of Patent: July 30, 2019
    Assignees: 2236008 Ontario Inc., BlackBerry Limited
    Inventors: Stephen Lau, Darrin Kenneth John Fry, Jianqiang Shi
  • Patent number: 10362269
    Abstract: The present disclosure relates to systems and methods for automatically determining the active users in an audio or video conference session including a plurality of users. In one implementation, the system may include a memory storing instructions and a processor configured to execute the instructions to receive packets from a device associated with each of a plurality of users, update a loudest level associated with each user if the received packet from the user reflects a communication louder than all of the previously received packets from the user, calculate a loudness ratio associated with each user based on the loudest level associated with each user and a corresponding time-decaying average, sort the plurality of users by a loudness score associated with each user, and mute audio or hide video received from a subset of the plurality of users in the conference session.
    Type: Grant
    Filed: January 11, 2017
    Date of Patent: July 23, 2019
    Assignee: RingCentral, Inc.
    Inventors: Dmitriy Solovey, Aleksei Petrov
  • Patent number: 10354650
    Abstract: In one aspect, a method comprises accessing audio data generated by a computing device based on audio input from a user, the audio data encoding one or more user utterances. The method further comprises generating a first transcription of the utterances by performing speech recognition on the audio data using a first speech recognizer that employs a language model based on user-specific data. The method further comprises generating a second transcription of the utterances by performing speech recognition on the audio data using a second speech recognizer that employs a language model independent of user-specific data. The method further comprises determining that the second transcription of the utterances includes a term from a predefined set of one or more terms. The method further comprises, based on determining that the second transcription of the utterance includes the term, providing an output of the first transcription of the utterance.
    Type: Grant
    Filed: March 15, 2013
    Date of Patent: July 16, 2019
    Assignee: Google LLC
    Inventors: Alexander H. Gruenstein, Petar Aleksic
  • Patent number: 10347274
    Abstract: An apparatus for decoding an encoded audio signal, includes: a spectral domain audio decoder for generating a first decoded representation of a first set of first spectral portions being spectral prediction residual values; a frequency regenerator for generating a reconstructed second spectral portion using a first spectral portion of the first set of first spectral portions, wherein the reconstructed second spectral portion additionally includes spectral prediction residual values; and an inverse prediction filter for performing an inverse prediction over frequency using the spectral residual values for the first set of first spectral portions and the reconstructed second spectral portion using prediction filter information included in the encoded audio signal.
    Type: Grant
    Filed: February 13, 2017
    Date of Patent: July 9, 2019
    Assignee: Fraunhofer-Gesellschaft zur Foerderung der angewandten Forschung e.V.
    Inventors: Sascha Disch, Frederik Nagel, Ralf Geiger, Balaji Nagendran Thoshkahna, Konstantin Schmidt, Stefan Bayer, Christian Neukam, Bernd Edler, Christian Helmrich
  • Patent number: 10339919
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for generating responses using task-independent conversational systems.
    Type: Grant
    Filed: April 20, 2018
    Date of Patent: July 2, 2019
    Assignee: botbotbotbot Inc.
    Inventor: Antoine Raux
  • Patent number: 10339960
    Abstract: Aspects relate to computer implemented methods and systems for monitoring a user's hearing and comprehension. The methods and systems include receiving, by an audio capture device, a first audio input, receiving, by the audio capture device, a second audio input, converting the first and second audio inputs into respective first and second audio signals, transmitting said first and second audio signals to a remote resource, transcribing the first and second audio signals into respective first and second transcriptions, analyzing, by a processor of the remote resource, the first and second transcriptions to determine if a content of the second transcription is related to a content of the first transcription, and recording the analysis to track a hearing and comprehension ability of the user.
    Type: Grant
    Filed: October 13, 2016
    Date of Patent: July 2, 2019
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Eli M. Dow, Thomas D. Fitzsimmons, Maurice M. Materise, Jessie Yu
  • Patent number: 10339216
    Abstract: Selecting a grammar for use in a machine question-answering system, such as a Natural Language Understanding System, can be difficult for non-experts in such grammars. A tool, according to an example embodiment, can compare annotations of sample sentences, performed correctly by a human, the annotations having intents and mentions, against annotations performed by multiple grammars. Each grammar can be scored, and the system can select the best scored grammar for the user. In one embodiment, a method of selecting a grammar includes comparing manually-generated annotations against machine-generated annotations as a function of a given grammar among multiple grammars. The method can further include applying scores to the machine-generated annotations that are a function of weightings of the intents and mentions. The method can additionally include recommending whether to employ the given grammar based on the scores.
    Type: Grant
    Filed: July 26, 2013
    Date of Patent: July 2, 2019
    Assignee: Nuance Communications, Inc.
    Inventor: Jeffrey N. Marcus
  • Patent number: 10339452
    Abstract: Systems and methods of automated ontology development include a corpus of communication data. The corpus of communication data includes communication data from a plurality of interactions and is processed. A plurality of terms are extracted from the corpus. Each term of the plurality is a plurality of words that identify a single concept within the corpus. An ontology is automatedly generated from the extracted terms.
    Type: Grant
    Filed: February 5, 2014
    Date of Patent: July 2, 2019
    Assignee: VERINT SYSTEMS LTD.
    Inventors: Roni Romano, Yair Horesh, Jeremie Dreyfuss