Patents Examined by Vijay B. Chawan
  • Patent number: 10319378
    Abstract: According to one embodiment, an interaction apparatus includes an interaction apparatus includes a storage, a first extractor, a retriever, a generator, a second extractor and a register. The storage stores a problem and at least one solution for solving the problem. The first extractor extracts a target problem which is an expression regarded as the problem, from a first speech. The generator generates a first speech-prompting sentence prompting the user to make a speech including the target solution if the storage stores no target solution or if the user rejects the target solution. The second extractor extracts the target solution from a second speech which is a response of the user relating to the first speech-prompting sentence. The register registers, on the storage, the target problem and the target solution.
    Type: Grant
    Filed: September 6, 2016
    Date of Patent: June 11, 2019
    Assignee: Kabushiki Kaisha Toshiba
    Inventor: Yumi Ichimura
  • Patent number: 10319250
    Abstract: Speech synthesis chooses pronunciations of words with multiple acceptable pronunciations based on an indication of a personal, class-based, or global preference or an intended non-preferred pronunciation. A speaker's words can be parroted back on personal devices using preferred pronunciations for accent training. Degrees of pronunciation error are computed and indicated to the user in a visual transcription or audibly as word emphasis in parroted speech. Systems can use sets of phonemes extended beyond those generally recognized for a language. Speakers are classified in order to choose specific phonetic dictionaries or adapt global ones. User profiles maintain lists of which pronunciations are preferred among ones acceptable for words with multiple recognized pronunciations. Systems use multiple correlations of word preferences across users to predict use preferences of unlisted words.
    Type: Grant
    Filed: February 22, 2017
    Date of Patent: June 11, 2019
    Assignee: SOUNDHOUND, INC.
    Inventors: Kiran Garaga Lokeswarappa, Jonah Probell
  • Patent number: 10311860
    Abstract: Methods, systems, and apparatus for receiving audio data corresponding to a user utterance and context data, identifying an initial set of one or more n-grams from the context data, generating an expanded set of one or more n-grams based on the initial set of n-grams, adjusting a language model based at least on the expanded set of n-grams, determining one or more speech recognition candidates for at least a portion of the user utterance using the adjusted language model, adjusting a score for a particular speech recognition candidate determined to be included in the expanded set of n-grams, determining a transcription of user utterance that includes at least one of the one or more speech recognition candidates, and providing the transcription of the user utterance for output.
    Type: Grant
    Filed: February 14, 2017
    Date of Patent: June 4, 2019
    Assignee: Google LLC
    Inventors: Petar Aleksic, Pedro J. Moreno Mengibar
  • Patent number: 10311149
    Abstract: Natural language translation device contains a bus, an input interface connecting to the bus for receiving a source sentence in a first natural language to be translated to a target sentence in second natural language one word at a time in sequential order. A two-dimensional (2-D) symbol containing a super-character characterizing the i-th word of the target sentence based on the received source sentence is formed in accordance with a set of 2-D symbol creation rules. The i-th word of the target sentence is obtained by classifying the 2-D symbol via a deep learning model that contains multiple ordered convolution layers in a Cellular Neural Networks or Cellular Nonlinear Networks (CNN) based integrated circuit.
    Type: Grant
    Filed: August 8, 2018
    Date of Patent: June 4, 2019
    Assignee: Gyrfalcon Technology Inc.
    Inventors: Lin Yang, Patrick Z. Dong, Catherine Chi, Charles Jin Young, Jason Z Dong, Baohua Sun
  • Patent number: 10303428
    Abstract: An electronic device with a function of smart voice service is disclosed. The electronic device is capable of identifying a user who is using the function of smart voice service and of adjusting the frequency of output sound to answer according to the hearing state of the user.
    Type: Grant
    Filed: November 27, 2017
    Date of Patent: May 28, 2019
    Assignee: UNLIMITER MFA CO., LTD.
    Inventors: Kuo-Ping Yang, Kuan-Li Chao, Ho-Hsin Liao
  • Patent number: 10304474
    Abstract: A method of enhancing speech quality includes: generating a high-frequency signal by using a low-frequency signal in a time domain; combining the low-frequency signal with the high-frequency signal; transforming the combined signal into a spectrum in a frequency domain; determining a class of a decoded speech signal; predicting an envelope from a low-frequency spectrum obtained in the transforming; and generating a final high-frequency spectrum by applying the predicted envelope to a high-frequency spectrum obtained in the transforming.
    Type: Grant
    Filed: August 17, 2015
    Date of Patent: May 28, 2019
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Ki-hyun Choo, Anton Viktorovich Porov, Konstantin Sergeevich Osipov, Eun-mi Oh, Woo-jung Park
  • Patent number: 10284706
    Abstract: Provided are a system and method of providing a voice-message call service. A mobile device that performs a call with an external mobile device comprises a control unit configured to obtain text, the text converted from voice data that is exchanged between the mobile device and the external mobile device, during the call between the mobile device and the external mobile device, and obtain input text input to the mobile device and provided text that is received from the external mobile device; and a display unit configured to arrange the text, the input text, and the provided text and display the arranged text, input text, and provided text on a screen of the device, during the call between the mobile device and the external mobile device.
    Type: Grant
    Filed: March 9, 2018
    Date of Patent: May 7, 2019
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Hong-chul Kim, Seon-ae Kim, Hyun-jae Shin
  • Patent number: 10276164
    Abstract: The present invention relates to a multi-speaker speech recognition correction system for determining a speaker of an utterance with a simple method and easily correcting speech-recognized text during speech recognition for a plurality of speakers. According to the present invention, when speech signals are input to a multi-speaker speech recognition system from a plurality of microphones which are each provided to a corresponding one of a plurality of speakers, the multi-speaker speech recognition correction system may detect a speech session from a time point at which input of each of the speech signals is started to a time point at which the input of the speech signal is stopped, and a speech recognizer may convert only the detected speech sessions into text so that a speaker of an utterance can be identified by a simple method and speech recognition can be carried out at a low cost.
    Type: Grant
    Filed: November 28, 2017
    Date of Patent: April 30, 2019
    Assignee: SORIZAVA CO., LTD.
    Inventor: Munhak An
  • Patent number: 10276171
    Abstract: A noise filling method is provided that includes detecting a frequency band including a part encoded to 0 from a spectrum obtained by decoding a bitstream; generating a noise component for the detected frequency band; and adjusting energy of the frequency band in which the noise component is generated and filled by using energy of the noise component and energy of the frequency band including the part encoded to 0.
    Type: Grant
    Filed: July 17, 2017
    Date of Patent: April 30, 2019
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Mi-young Kim, Eun-mi Oh
  • Patent number: 10269028
    Abstract: The present invention provides a system and method for optimizing a message. Components of a starting message are identified, and at least one rule is applied for modifying at least one message component to create at least one variation of the starting message. Message variants are tested by sending each variant to a sample of people and measuring a response rate for each sent message variant. The measured response rates are used to create an optimal version of the message. In one embodiment, message variants may be created and tested in multiple rounds.
    Type: Grant
    Filed: July 19, 2017
    Date of Patent: April 23, 2019
    Assignee: PERSADO INTELLECTUAL PROPERTY LIMITED
    Inventors: Alexios Vratskides, Ori Einstein, Guy Stephane Krief, Assaf Baciu
  • Patent number: 10268779
    Abstract: A method, computer system, and computer program product for determining a server conversational state in an interactive dialog between a server and an administrator is provided. The embodiment may include receiving a query statement from a user. The embodiment may also include updating a context map to reflect a context of the received query statement. The embodiment may further include transmitting a request to one or more cognitive engines capable of resolving the received query statement. The embodiment may also include processing the transmitted request in each of the one or more cognitive engines. The embodiment may further include updating the context map with a resolution operation performed by the one or more cognitive engines. The embodiment may also include displaying a response to the user.
    Type: Grant
    Filed: August 9, 2017
    Date of Patent: April 23, 2019
    Assignee: International Business Machines Corporation
    Inventors: Anu Jalan, Naveen Kumar T, Priya Kannan, Shashank M S
  • Patent number: 10269357
    Abstract: The present invention disclose a speech/audio bitstream decoding method including: acquiring a speech/audio decoding parameter of a current speech/audio frame, where the foregoing current speech/audio frame is a redundant decoded frame or a speech/audio frame previous to the foregoing current speech/audio frame is a redundant decoded frame; performing post processing on the acquired speech/audio decoding parameter according to speech/audio parameters of X speech/audio frames, where the foregoing X speech/audio frames include M speech/audio frames previous to the foregoing current speech/audio frame and/or N speech/audio frames next to the foregoing current speech/audio frame; and recovering a speech/audio signal by using the post-processed speech/audio decoding parameter of the foregoing current speech/audio frame. The technical solutions of the present invention help improve quality of an output speech/audio signal.
    Type: Grant
    Filed: September 2, 2016
    Date of Patent: April 23, 2019
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Xingtao Zhang, Zexin Liu, Lei Miao
  • Patent number: 10261991
    Abstract: One variation of a system for imposing a dynamic sentiment vector to an electronic message includes: a processor; an electronic computing device communicatively coupled to the processor and associated with a particular user; and a sentiment vector generator comprising: a parsing module; a dynamic sentiment value spectrum associated with the particular user; and a program executable by the processor and configured to: receive a text input comprising message content from the electronic computing device; parse, at the parsing module, the message content comprised in the text input for emotionally charged language; based on the emotionally charged language, generate a sentiment value from the dynamic sentiment value spectrum for the text input and, based on the sentiment value, impose a sentiment vector, corresponding to the assigned sentiment value, to the text input, the imposed sentiment vector rendering a sensory effect on the message content designed to convey a corresponding sentiment.
    Type: Grant
    Filed: September 12, 2017
    Date of Patent: April 16, 2019
    Assignee: Aebeze labs
    Inventors: Michael Phillips Moskowitz, Matthew Jordan, Martin Kay, Ray Sidney, Barbara McGillivray, Scott Tong, Bradley Artziniega
  • Patent number: 10255920
    Abstract: Provided are methods, systems, and apparatuses for detecting, processing, and responding to audio signals, including speech signals, within a designated area or space. A platform for multiple media devices connected via a network is configured to process speech, such as voice commands, detected at the media devices, and respond to the detected speech by causing the media devices to simultaneously perform one or more requested actions. The platform is capable of scoring the quality of a speech request, handling speech requests from multiple end points of the platform using a centralized processing approach, a de-centralized processing approach, or a combination thereof, and also manipulating partial processing of speech requests from multiple end points into a coherent whole when necessary.
    Type: Grant
    Filed: June 16, 2017
    Date of Patent: April 9, 2019
    Assignee: Google LLC
    Inventors: Jay Pierre Civelli, Mikhal Shemer, Turaj Zakizadeh Shabestary, David Tapuska
  • Patent number: 10249303
    Abstract: Provided are methods, systems, and apparatuses for detecting, processing, and responding to audio signals, including speech signals, within a designated area or space. A platform for multiple media devices connected via a network is configured to process speech, such as voice commands, detected at the media devices, and respond to the detected speech by causing the media devices to simultaneously perform one or more requested actions. The platform is capable of scoring the quality of a speech request, handling speech requests from multiple end points of the platform using a centralized processing approach, a de-centralized processing approach, or a combination thereof, and also manipulating partial processing of speech requests from multiple end points into a coherent whole when necessary.
    Type: Grant
    Filed: June 14, 2017
    Date of Patent: April 2, 2019
    Assignee: Google LLC
    Inventors: Jay Pierre Civelli, Mikhal Shemer, Turaj Zakizadeh Shabestary, David Tapuska
  • Patent number: 10248649
    Abstract: A natural language processing apparatus is provided with: a dialog processing unit which, every time an analyzable unit of a part of a natural language sentence is inputted, performs analysis processing with respect to the input unit in each of a plurality of analysis processing units in an incremental and parallel manner; and an output unit which obtains an output, such as a dialog response sentence, based on an analysis result from each analysis processing unit in the dialog processing unit. Each of processing units prepared in the dialog processing unit acquires an immediately preceding or an earlier past analysis result from the processing unit, and an immediately preceding or an earlier past analysis results from the other processing units, and obtains one or a limited number of analysis results while making an estimation with reference to the acquired analysis results.
    Type: Grant
    Filed: December 5, 2016
    Date of Patent: April 2, 2019
    Assignee: KAI INC.
    Inventor: Yoshinobu Kano
  • Patent number: 10229696
    Abstract: An audio encoder for providing an output signal using an input audio signal includes a patch generator, a comparator and an output interface. The patch generator generates at least one bandwidth extension high-frequency signal, wherein a bandwidth extension high-frequency signal includes a high-frequency band. The high-frequency band of the bandwidth extension high-frequency signal is based on a low frequency band of the input audio signal. A comparator calculates a plurality of comparison parameters. A comparison parameter is calculated based on a comparison of the input audio signal and a generated bandwidth extension high-frequency signal. Each comparison parameter of the plurality of comparison parameters is calculated based on a different offset frequency between the input audio signal and a generated bandwidth extension high-frequency signal.
    Type: Grant
    Filed: May 12, 2015
    Date of Patent: March 12, 2019
    Assignee: FRAUNHOFER-GESELLSCHAFT ZUR FOERDERUNG DER ANGEWANDTEN FORSCHUNG E.V.
    Inventors: Frederik Nagel, Sascha Disch, Guillaume Fuchs, Juergen Herre, Christian Griebel
  • Patent number: 10224024
    Abstract: Implementations relate to techniques for providing context-dependent search results. A computer-implemented method includes receiving an audio stream at a computing device during a time interval, the audio stream comprising user speech data and background audio, separating the audio stream into a first substream that includes the user speech data and a second substream that includes the background audio, identifying concepts related to the background audio, generating a set of terms related to the identified concepts, influencing a speech recognizer based on at least one of the terms related to the background audio, and obtaining a recognized version of the user speech data using the speech recognizer.
    Type: Grant
    Filed: June 14, 2017
    Date of Patent: March 5, 2019
    Assignee: Google LLC
    Inventors: Jason Sanders, Gabriel Taubman, John J. Lee
  • Patent number: 10216832
    Abstract: A natural language processing system has a hierarchy of user intents related to a domain of interest, the hierarchy having specific intents corresponding to leaf nodes of the hierarchy, and more general intents corresponding to ancestor nodes of the leaf nodes. The system also has a trained understanding model that can classify natural language utterances according to user intent. When the understanding model cannot determine with sufficient confidence that a natural language utterance corresponds to one of the specific intents, the natural language processing system traverses the hierarchy of intents to find a more general user intent that is related to the most applicable specific intent of the utterance and for which there is sufficient confidence. The general intent can then be used to prompt the user with questions applicable to the general intent to obtain the missing information needed for a specific intent.
    Type: Grant
    Filed: December 19, 2016
    Date of Patent: February 26, 2019
    Assignee: Interactions LLC
    Inventors: Srinivas Bangalore, John Chen
  • Patent number: 10192550
    Abstract: Voice input is received from a user. An ASR system generates in memory a set of words it has identified in the voice input, and update the set each time it identifies a new word in the voice input to add the new word to the set. A condition indicative of speech inactivity in the voice input is detected. A response for outputting to the user is generated based on the set of identified words, in response to the detection of the speech inactivity condition. The generated response is outputted to the user after an interval of time—commencing with the detection of the speech inactivity condition—has ended and only if no more words have been identified in the voice input by the ASR system in that interval of time.
    Type: Grant
    Filed: March 1, 2016
    Date of Patent: January 29, 2019
    Assignee: Microsoft Technology Licensing, LLC
    Inventor: Raymond J. Froelich