Patents Examined by Oluwadamilola M. Ogunbiyi
  • Patent number: 10468036
    Abstract: A method for mixing, processing and enhancing signals using signal decomposition is presented. A method for improving sorting of decomposed signal parts using cross-component similarity is also provided.
    Type: Grant
    Filed: April 30, 2014
    Date of Patent: November 5, 2019
    Assignee: ACCUSONUS, INC.
    Inventors: Alexandros Tsilfidis, Elias Kokkinis
  • Patent number: 10468022
    Abstract: A voice assistant (VA) can switch between a voice input mode, in which the VA produces audible responses to voice queries, and a gesture input mode that can be triggered by a predetermined gesture, in which the VA produces visual responses to gesture-based queries.
    Type: Grant
    Filed: April 3, 2017
    Date of Patent: November 5, 2019
    Assignee: Motorola Mobility LLC
    Inventors: Jun-ki Min, Mir Farooq Ali, Navin Tulsibhai Dabhi
  • Patent number: 10438581
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for speech recognition using neural networks. A feature vector that models audio characteristics of a portion of an utterance is received. Data indicative of latent variables of multivariate factor analysis is received. The feature vector and the data indicative of the latent variables is provided as input to a neural network. A candidate transcription for the utterance is determined based on at least an output of the neural network.
    Type: Grant
    Filed: July 31, 2013
    Date of Patent: October 8, 2019
    Assignee: Google LLC
    Inventors: Andrew W. Senior, Ignacio L. Moreno
  • Patent number: 10431214
    Abstract: The disclosure relates to methods, systems and other embodiments directed to determining an information domain match for a natural language (NL) input (e.g., a spoken utterance), and confirming whether the NL input is correctly matched to the information domain. For example, after receiving an NL input, a first information domain to which the NL input belongs and a feature value set may be determined based on a semantic pattern matching technique. Further, a second information domain to which the NL input belongs, and a corresponding confidence score related to the second information domain may be determined. The second information domain may be determined based on a first statistical classification technique. Based on the determined feature value set and the confidence score related to the second information domain, it may be confirmed whether the NL input correctly belongs to the first information domain, e.g., based on a second statistical classification technique.
    Type: Grant
    Filed: November 24, 2015
    Date of Patent: October 1, 2019
    Assignee: Voicebox Technologies Corporation
    Inventors: Yikun Guo, Safiyyah Saleem, Jiayuan Huang
  • Patent number: 10424289
    Abstract: A speech recognition system includes a phone classifier and a boundary classifier. The phone classifier generates combined boundary posteriors from a combination of auditory attention features and phone posteriors by feeding phone posteriors of neighboring frames of an audio signal into a machine learning algorithm to classify phone posterior context information. The boundary classifier estimates boundaries in speech contained in the audio signal from the combined boundary posteriors.
    Type: Grant
    Filed: August 14, 2018
    Date of Patent: September 24, 2019
    Assignee: SONY INTERACTIVE ENTERTAINMENT INC.
    Inventor: Ozlem Kalinli-Akbacak
  • Patent number: 10424305
    Abstract: An error-concealing audio decoding method comprises: receiving a packet comprising a set of MDCT coefficients encoding a frame of time-domain samples of an audio signal; identifying the received packet as erroneous; generating estimated MDCT coefficients to replace the set of MDCT coefficients of the erroneous packet, based on corresponding MDCT coefficients associated with a received packet directly preceding the erroneous packet; assigning signs of a first subset of MDCT coefficients of the estimated MDCT coefficients, wherein the first subset comprises such MDCT coefficients that are associated with tonal-like spectral bins, to coincide with signs of corresponding MDCT coefficients of said preceding packet; randomly assigning signs of a second subset of MDCT coefficients of the estimated MDCT coefficients, wherein the second subset comprises MDCT coefficients associated with noise-like spectral bins; replacing the erroneous packet by a concealment packet containing the estimated MDCT coefficients and the s
    Type: Grant
    Filed: December 8, 2015
    Date of Patent: September 24, 2019
    Assignee: Dolby International AB
    Inventors: Arijit Biswas, Tobias Friedrich, Klaus Peichl
  • Patent number: 10417345
    Abstract: A system in which a customer service agent (CSA) is able to assist a customer with obtaining a desired response from a speech-controlled appliance while protecting customer data. The customer service agent submits queries to a natural language understanding (NLU) processor that performs entity resolution using personalized library information stored in an entity library based on the customer identity information and/or an device identifier. The CSA is shielded from the entity library itself, as well as data stored on the speech-controlled appliance. The CSA can instruct the NLU processor to deliver results to multiple endpoints, including both the customer's appliance and the CSA agent's console.
    Type: Grant
    Filed: December 22, 2014
    Date of Patent: September 17, 2019
    Assignee: Amazon Technologies, Inc.
    Inventors: Ty Loren Carlson, Hsuan-Cheng Lai
  • Patent number: 10403289
    Abstract: A voice processing device includes: a processor; and a memory which stores a plurality of instructions, which when executed by the processor, cause the processor to execute: detecting a first utterance section included in a first voice and a second utterance section included in a second voice; specifying an overlapping section within which the first utterance section and the second utterance section overlap with each other; calculating a first utterance continuation section from a start point of the overlapping section to an end point of the first utterance section; and evaluating an impression regarding the first voice at least on the basis of information relating to a length of the first utterance continuation section.
    Type: Grant
    Filed: November 25, 2015
    Date of Patent: September 3, 2019
    Assignee: FUJITSU LIMITED
    Inventors: Taro Togawa, Chisato Shioda, Sayuri Kohmura, Takeshi Otani
  • Patent number: 10395650
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for suppressing hotword triggers when detecting a hotword in recorded media are disclosed. In one aspect, a method includes the actions of receiving, by a computing device, audio corresponding to playback of an item of media content. The actions further include determining, by the computing device, that the audio includes an utterance of a predefined hotword and that the audio includes an audio watermark. The actions further include analyzing, by the computing device, the audio watermark. The actions further include based on analyzing the audio watermark, determining, by the computing device, whether to perform speech recognition on a portion of the audio following the predefined hotword.
    Type: Grant
    Filed: June 5, 2017
    Date of Patent: August 27, 2019
    Assignee: Google LLC
    Inventor: Ricardo Antonio Garcia
  • Patent number: 10388282
    Abstract: Embodiments cover a voice command device and a server computing device that communicates with the voice command device. In one embodiment, a voice command device comprises a speaker, a microphone, a wireless communication module, and a processing device. The processing device is to scan for wireless advertising packets from a plurality of medical devices at an interval and detect a wireless advertising packet from a medical device of the plurality of medical devices as a result of the scanning. The processing device is further to receive medical data for a living entity from the medical device and send the medical data to a server computing device, wherein the server computing device is to generate a message associated with the medical data. The processing device is to receive the message and output the message via the speaker.
    Type: Grant
    Filed: January 25, 2017
    Date of Patent: August 20, 2019
    Assignee: CliniCloud Inc.
    Inventors: Hon Weng Chong, An Lin
  • Patent number: 10360914
    Abstract: Using many speech recognition engines, one can select which one is best at any given iteration of sending a command to a device to be interpreted and carried out. Depending on the context, a different result of many results received from speech recognition engines is chosen. The context is determined based on window history, including rendered webpages represented by URLs previously displayed on the device or windows resulting from executed code on the computing device. In this manner, the operation of the computer is improved as a more accurate result of receiving audio and processing it to text many times is used.
    Type: Grant
    Filed: January 26, 2017
    Date of Patent: July 23, 2019
    Assignee: ESSENCE, INC
    Inventors: Holly R Corcoran, Barry Klein, Llewellyn Q Morake
  • Patent number: 10360903
    Abstract: According to one embodiment, an apparatus includes a storage unit, a first acquisition unit, a second acquisition unit, an analyzer, and a recognition unit. The storage unit stores first situation information about a situation assumed in advance, a first representation representing a meaning of a sentence assumed, intention information representing an intention to be estimated, and a first value representing a degree of application of the first representation to the first situation information and the intention information. The first acquisition unit acquires a natural sentence. The second acquisition unit acquires second situation information about a situation when acquiring the natural sentence. The analyzer analyzes the natural sentence and generates a second representation representing a meaning of the natural sentence. The recognition unit obtains an estimated value based on the first value associated with the first situation information and the first representation.
    Type: Grant
    Filed: February 15, 2017
    Date of Patent: July 23, 2019
    Assignee: Kabushiki Kaisha Toshiba
    Inventors: Hiromi Wakaki, Kenji Iwata, Masayuki Okamoto
  • Patent number: 10339214
    Abstract: A method, system and computer program product for recognizing terms in a specified corpus. In one embodiment, the method comprises providing a set of known terms t?T, each of the known terms t belonging to a set of types ? (t)={?1, . . . }, wherein each of the terms is comprised of a list of words, t=w1, w2, . . . , wn, and the union of all the words for all the terms is a word set W. The method further comprises using the set of terms T and the set of types to determine a set of pattern-to-type mappings p??; and using the set of pattern-to-type mappings to recognize terms in the specified corpus and, for each of the recognized terms in the specified corpus, to recognize one or more of the types ? for said each recognized term.
    Type: Grant
    Filed: November 2, 2012
    Date of Patent: July 2, 2019
    Assignee: International Business Machines Corporation
    Inventors: Michael R. Glass, Alfio M. Gliozzo
  • Patent number: 10304447
    Abstract: A method for controlling a plurality of environmental factors that trigger a negative emotional state is provided. The method may include analyzing a plurality of user data when a user experiences a plurality of various environmental factors. The method may also include determining an emotional state experienced by the user when each of the plurality of various environmental factors is present based on the plurality of user data. The method may include receiving a plurality of calendar information associated with a user account. The method may also include identifying an upcoming event based on the plurality of calendar information. The method may include identifying an environmental factor within the plurality of various environmental factors is present at the upcoming event. The method may also include, in response to determining the environmental factor causes the user to experience a negative emotional state, executing an accommodation method based on the environmental factor.
    Type: Grant
    Filed: January 25, 2017
    Date of Patent: May 28, 2019
    Assignee: International Business Machines Corporation
    Inventors: Paul R. Bastide, Matthew E. Broomhall, Robert E. Loredo, Fang Lu
  • Patent number: 10303762
    Abstract: A method is disclosed for evaluating a chat message sent between users of an online environment. The method may include associating each word in the chat message with metadata. The metadata identifies a word type and usage for each word in the chat message. This method may also include identifying one or more safety rules associated with the metadata. Each safety rule identifies an ordered sequence of one or more sets of words. This method may also include applying the safety rule to the chat message to determine whether a sequence of words in the chat message present in the ordered sequence of sets of words. Upon determining a word, from each set of words in the ordered sequence of sets of words, matches a respective one of the words in the chat message, the chat message is blocked from being sent to a message recipient.
    Type: Grant
    Filed: April 30, 2013
    Date of Patent: May 28, 2019
    Assignee: Disney Enterprises, Inc.
    Inventors: Vita Markman, Sean O'Dell, Drew Beechum, Paul Pak
  • Patent number: 10297263
    Abstract: A method includes extracting a voicing classification parameter of an audio signal and determining a filter coefficient of a low pass filter based on the voicing classification parameter. The method also includes filtering a low-band portion of the audio signal to generate a low-band audio signal and controlling an amplitude of a temporal envelope of the low-band audio signal based on the filter coefficient. The method also includes modulating a white noise signal based on the amplitude of the temporal envelope to generate a modulated white noise signal and scaling the modulated white noise signal based on a noise gain to generate a scaled modulated white noise signal. The method also includes mixing a scaled version of the low-band audio signal with the scaled modulated white noise signal to generate a high-band excitation signal that is used to generate a decoded version of the audio signal.
    Type: Grant
    Filed: June 1, 2017
    Date of Patent: May 21, 2019
    Assignee: Qualcomm Incorporated
    Inventors: Pravin Kumar Ramadas, Daniel J. Sinder, Stephane Pierre Villette, Vivek Rajendran
  • Patent number: 10269374
    Abstract: In an approach to determining speech effectiveness, one or more computer processors receive speech input. The one or more computer processors determine, based, at least in part, on the received speech input, a first speaking mode. The one or more computer processors detect, based, at least in part, on the first speaking mode, at least one aspect of the received speech input. The one or more computer processors determine, based, at least in part, on the first speaking mode and the detected at least one aspect of the received speech input, whether a rate of occurrence of the detected at least one aspect of the speech input exceeds a pre-defined threshold. Responsive to determining a rate of occurrence of the detected at least one aspect of the speech input exceeds a pre-defined threshold, the one or more computer processors notify a user.
    Type: Grant
    Filed: May 18, 2016
    Date of Patent: April 23, 2019
    Assignee: International Business Machines Corporation
    Inventors: Patrick A. Spizzo, Sara H. Waggoner, Kaleb D. Walton, Aaron T. Wodrich
  • Patent number: 10258295
    Abstract: Disclosed is a method, a device, a system and/or a manufacture of voice controlled assistance for monitoring adverse events of a user and/or coordinating emergency actions such as caregiver communication. In one embodiment, a system includes an assistance device coupled to a voice controlled assistance server over a network. The assistance device includes a memory including computer-executable instructions that when executed, upon an event such as sensing the presence of the user, generates a verbal status query to the user usable to determine an event of concern and/or an adverse event related to the user, such as a negative response or non-response event that may require initiation of an emergency action.
    Type: Grant
    Filed: May 9, 2017
    Date of Patent: April 16, 2019
    Assignee: LifePod Solutions, Inc.
    Inventor: Dennis Fountaine
  • Patent number: 10249314
    Abstract: A voice conversion system for generating realistic, natural-sounding target speech is disclosed. The voice conversion system preferably comprises a neural network for converting the source speech data to estimated target speech data; a global variance correction module; a modulation spectrum correction module; and a waveform generator. The global variance correction module is configured to scale and shift (or normalize and de-normalize) the estimated target speech based on (i) a mean and standard deviation of the source speech data, and further based on (ii) a mean and standard deviation of the estimated target speech data. The modulation spectrum correction module is configured to apply a plurality of filters to the estimated target speech data after it has been scaled and shifted by the global variance correction module. Each filter is designed to correct the trajectory representing the curve of one MCEP coefficient over time.
    Type: Grant
    Filed: July 21, 2017
    Date of Patent: April 2, 2019
    Assignee: OBEN, INC.
    Inventor: Sandesh Aryal
  • Patent number: 10236007
    Abstract: An audio encoder for encoding an audio signal, includes: a first encoding processor for encoding a first audio signal portion in a frequency domain, wherein the first encoding processor includes: a time frequency converter for converting the first audio signal portion into a frequency domain representation having spectral lines up to a maximum frequency of the first audio signal portion; a spectral encoder for encoding the frequency domain representation; a second encoding processor for encoding a second different audio signal portion in the time domain; a cross-processor for calculating, from the encoded spectral representation of the first audio signal portion, initialization data of the second encoding processor, so that the second encoding processing is initialized to encode the second audio signal portion immediately following the first audio signal portion in time in the audio signal.
    Type: Grant
    Filed: January 24, 2017
    Date of Patent: March 19, 2019
    Assignee: Fraunhofer-Gesellschaft zur Foerderung der angewandten Forschung e.V.
    Inventors: Sascha Disch, Martin Dietz, Markus Multrus, Guillaume Fuchs, Emmanuel Ravelli, Matthias Neusinger, Markus Schnell, Benjamin Schubert, Bernhard Grill