Probability Patents (Class 704/240)
  • Patent number: 11442932
    Abstract: Systems and methods for mapping natural language to queries using a query grammar are described. For example, methods may include generating, based on a string, a set of tokens of a database syntax; generating a query graph for the set of tokens using a finite state machine representing a query grammar, wherein nodes of the finite state machine represent token types, directed edges of the finite state machine represent valid transitions between token types in the query grammar, vertices of the query graph correspond to respective tokens of the set of tokens, and directed edges of the query graph represent a transition between two tokens in a sequencing of the tokens; determining, based on the query graph, a sequence of the tokens in the set of tokens, forming a database query; and invoking a search of a database using a query based on the database query to obtain search results.
    Type: Grant
    Filed: July 16, 2019
    Date of Patent: September 13, 2022
    Assignee: ThoughtSpot, Inc.
    Inventors: Nikhil Yadav, Ravi Tandon
  • Patent number: 11425252
    Abstract: Exemplary aspects involve a data-communications apparatus or system communicate over a broadband network with a plurality of remotely-located data-communications circuits respectively associated with a plurality of remotely-situated client entities. The system includes a unified-communications and call center (UC-CC) platform that processes incoming data-communication interactions including different types of digitally-represented communications among which are incoming call, and that is integrated with a memory circuit including a database of information sets. Each of the information sets includes experience data corresponding to past incoming data-communication interactions processed by the platform, and with aggregated and organized data based on data collected in previous incoming interactions.
    Type: Grant
    Filed: July 21, 2021
    Date of Patent: August 23, 2022
    Assignee: 8x8, Inc.
    Inventors: Bryan R. Martin, Matt Taylor, Manu Mukerji
  • Patent number: 11410658
    Abstract: Audio data saved at the end of client interactions are sampled, analyzed for pauses in speech, and sliced into stretches of acoustic data containing human speech between those pauses. The acoustic data are accompanied by machine transcripts made by VoiceAI. A suitable distribution of data useful for training and testing are stipulated during data sampling by applying certain filtering criteria. The resulting datasets are sent for transcription by a human transcriber team. The human transcripts are retrieved, some post-transcription processing and cleaning are performed, and the results are added to datastores for training and testing an acoustic model.
    Type: Grant
    Filed: October 29, 2019
    Date of Patent: August 9, 2022
    Assignee: Dialpad, Inc.
    Inventors: Eddie Yee Tak Ma, James Palmer, Kevin James, Etienne Manderscheid
  • Patent number: 11393471
    Abstract: A system is provided for modifying how an output is presented via a multi-device synchronous configuration based on detecting a speech characteristic in the user input. For example, if the user whispers a request, then the system may temporarily modify how the responsive output is presented to the user via multiple devices. In one example, the system may lower the volume on all devices presented the output. In another example, the system may present the output via a single device rather than multiple devices. The system may also determine to operate in a alternate output mode based on certain non-audio data.
    Type: Grant
    Filed: March 30, 2020
    Date of Patent: July 19, 2022
    Assignee: Amazon Technologies, Inc.
    Inventor: Ezekiel Wade Sanborn de Asis
  • Patent number: 11388480
    Abstract: An information processing apparatus according to the present technology includes a reception unit, a first generation unit, a collection unit, and a second generation unit. The reception unit receives a content. The first generation unit analyzes the received content and generates one or more pieces of analysis information related to the content. The collection unit collects content information related to the content on a network on the basis of the one or more pieces of generated analysis information. The second generation unit generates an utterance sentence on the basis of at least one of the one or more pieces of analysis information and the collected content information.
    Type: Grant
    Filed: September 28, 2016
    Date of Patent: July 12, 2022
    Inventor: Hideo Nagasaka
  • Patent number: 11367438
    Abstract: An embodiment of the present invention provides an artificial intelligence (AI) apparatus for recognizing a speech of a user, the artificial intelligence apparatus includes a memory to store a speech recognition model and a processor to obtain a speech signal for a user speech, to convert the speech signal into a text using the speech recognition model, to measure a confidence level for the conversion, to perform a control operation corresponding to the converted text if the measured confidence level is greater than or equal to a reference value, and to provide feedback for the conversion if the measured confidence level is less than the reference value.
    Type: Grant
    Filed: May 16, 2019
    Date of Patent: June 21, 2022
    Inventors: Jaehong Kim, Hyoeun Kim, Hangil Jeong, Heeyeon Choi
  • Patent number: 11355113
    Abstract: A method, apparatus, device, and computer readable storage medium for recognizing and decoding a voice based on a streaming attention model are provided. The method may include generating a plurality of acoustic paths for decoding the voice using the streaming attention model, and then merging acoustic paths with identical last syllables of the plurality of acoustic paths to obtain a plurality of merged acoustic paths. The method may further include selecting a preset number of acoustic paths from the plurality of merged acoustic paths as retained candidate acoustic paths. Embodiments of the present disclosure present a concept that acoustic score calculating of a current voice fragment is only affected by its last voice fragment and has nothing to do with earlier voice history, and merge acoustic paths with the identical last syllables of the plurality of candidate acoustic paths.
    Type: Grant
    Filed: March 9, 2020
    Date of Patent: June 7, 2022
    Assignee: Baidu Online Network Technology (Beijing) Co., Ltd.
    Inventors: Junyao Shao, Sheng Qian, Lei Jia
  • Patent number: 11341988
    Abstract: A hybrid machine learning-based and DSP statistical post-processing technique is disclosed for voice activity detection. The hybrid technique may use a DNN model with a small context window to estimate the probability of speech by frames. The DSP statistical post-processing stage operates on the frame-based speech probabilities from the DNN model to smooth the probabilities and to reduce transitions between speech and non-speech states. The hybrid technique may estimate the soft decision on detected speech in each frame based on the smoothed probabilities, generate a hard decision using a threshold, detect a complete utterance that may include brief pauses, and estimate the end point of the utterance. The hybrid voice activity detection technique may incorporate a target directional probability estimator to estimate the direction of the speech source. The DSP statistical post-processing module may use the direction of the speech source to inform the estimates of the voice activity.
    Type: Grant
    Filed: September 23, 2019
    Date of Patent: May 24, 2022
    Assignee: APPLE INC.
    Inventors: Ramin Pishehvar, Feiping Li, Ante Jukic, Mehrez Souden, Joshua D. Atkins
  • Patent number: 11341340
    Abstract: Adapters for neural machine translation systems. A method includes determining a set of similar n-grams that are similar to a source n-gram, and each similar n-gram and the source n-gram is in a first language; determining, for each n-gram in the set of similar n-grams, a target n-gram is a translation of the similar n-gram in the first language to the target n-gram in the second language; generating a source encoding of the source n-gram, and, for each target n-gram determined from the set of similar n-grams determined for the source n-gram, a target encoding of the target n-gram and a conditional source target memory that is an encoding of each of the target encodings; providing, as input to a first prediction model, the source encoding and the condition source target memory; and generating a predicted translation of the source n-gram from the first language to the second language.
    Type: Grant
    Filed: October 1, 2019
    Date of Patent: May 24, 2022
    Assignee: Google LLC
    Inventors: Ankur Bapna, Ye Tian, Orhan Firat
  • Patent number: 11334182
    Abstract: In some implementations, data indicating a touch received on a proximity-sensitive display is received while the proximity-sensitive display is presenting one or more items. In one aspect, the techniques describe may involve a process for disambiguating touch selections of hypothesized items, such as text or graphical objects that have been generated based on input data, on a proximity-sensitive display. This process may allow a user to more easily select hypothesized items that the user may wish to correct, by determining whether a touch received through the proximity-sensitive display represents a selection of each hypothesized item based at least on a level of confidence that the hypothesized item accurately represents the input data.
    Type: Grant
    Filed: December 10, 2019
    Date of Patent: May 17, 2022
    Assignee: Google LLC
    Inventors: Jakob Nicolaus Foerster, Diego Melendo Casado, Glen Shires
  • Patent number: 11328731
    Abstract: System and methods for identifying a text word from a spoken utterance are provided. An ensemble BPE system that includes a phone BPE system and a character BPE system receives a spoken utterance. Both BPE systems include a multi-level language model (LM) and an acoustic model. The phone BPE system identifies first words from the spoken utterance and determine a first score for each first word. The first words are converted into character sequences. The character BPE model converts the character sequences into second words and determines a second score for each second word. For each word from the first words that matches a word in the second words the first and second scores are combined. The text word is the word with a highest score.
    Type: Grant
    Filed: June 17, 2020
    Date of Patent: May 10, 2022
    Assignee:, inc.
    Inventors: Weiran Wang, Yingbo Zhou, Caiming Xiong
  • Patent number: 11295732
    Abstract: In order to improve the accuracy of ASR, an utterance is transcribed using a plurality of language models, such as for example, an N-gram language model and a neural language model. The language models are trained separately. They each output a probability score or other figure of merit for a partial transcription hypothesis. Model scores are interpolated to determine a hybrid score. While recognizing an utterance, interpolation weights are chosen or updated dynamically, in the specific context of processing. The weights are based on dynamic variables associated with the utterance, the partial transcription hypothesis, or other aspects of context.
    Type: Grant
    Filed: August 1, 2019
    Date of Patent: April 5, 2022
    Assignee: SoundHound, Inc.
    Inventors: Steffen Holm, Terry Kong, Kiran Garaga Lokeswarappa
  • Patent number: 11289099
    Abstract: An information processing device including processing circuitry is provided. The processing circuitry configured to receive voice information regarding a voice of a user collected by a specific microphone of a plurality of microphones. The processing circuitry is configured to determine the user identified on a basis of the voice information regarding the voice of the user collected by the specific microphone among the plurality of microphones to be a specific type of user that has performed speech a predefined number of times or more within at least a certain period of time. Further, the processing circuitry is configured to control a message to be output to the user via a speaker corresponding to the specific microphone based on the user being determined to be the specific type of user.
    Type: Grant
    Filed: August 4, 2017
    Date of Patent: March 29, 2022
    Inventor: Keigo Ihara
  • Patent number: 11276413
    Abstract: Disclosed are an audio signal encoding method and audio signal decoding method, and an encoder and decoder performing the same. The audio signal encoding method includes applying an audio signal to a training model including N autoencoders provided in a cascade structure, encoding an output result derived through the training model, and generating a bitstream with respect to the audio signal based on the encoded output result.
    Type: Grant
    Filed: August 16, 2019
    Date of Patent: March 15, 2022
    Assignees: Electronics and Telecommunications Research Institute, THE TRUSTEES OF INDIANA UNIVERSITY
    Inventors: Mi Suk Lee, Jongmo Sung, Minje Kim, Kai Zhen
  • Patent number: 11243810
    Abstract: The system uses the non-repudiatory persistence of blockchain technology to store all task statuses and results across the distributed computer network in an immutable blockchain database. Coupled with the resiliency of the stored data, the system may determine a sequence of processing tasks for a given processing request and use the sequence to detect and/or predict failures. Accordingly, in the event of a detected system failure, the system may recover the results prior to the failure, minimizing disruptions to processing the request and improving hardware resiliency.
    Type: Grant
    Filed: January 22, 2021
    Date of Patent: February 8, 2022
    Assignee: The Bank of New York Mellon
    Inventors: Sanjay Kumar Stribady, Saket Sharma, Gursel Taskale
  • Patent number: 11217252
    Abstract: A method of zoning a transcription of audio data includes separating the transcription of audio data into a plurality of utterances. A that each word in an utterances is a meaning unit boundary is calculated. The utterance is split into two new utterances at a work with a maximum calculated probability. At least one of the two new utterances that is shorter than a maximum utterance threshold is identified as a meaning unit.
    Type: Grant
    Filed: August 28, 2019
    Date of Patent: January 4, 2022
    Inventors: Roni Romano, Yair Horesh, Jeremie Dreyfuss
  • Patent number: 11204685
    Abstract: User interfaces may enable users to initiate voice-communications with voice-controlled devices via a Wi-Fi network or other network via an Internet Protocol (IP) address. The user interfaces may include controls to enable users to initiate voice communications, such as Voice over Internet Protocol (VoIP) calls, with devices that do not have connectivity with traditional mobile telephone networks, such as traditional circuit transmissions of a Public Switched Telephone Network (PSTN). For example, the user interface may enable initiating a voice communication with a voice-controlled device that includes network connectivity via a home Wi-Fi network. The user interfaces may indicate availability of devices and/or contacts for voice communications and/or recent activity of devices or contact.
    Type: Grant
    Filed: February 21, 2020
    Date of Patent: December 21, 2021
    Assignee: Amazon Technologies, Inc.
    Inventors: Blair Harold Beebe, Katherine Ann Baker, David Michael Rowell, Peter Chin
  • Patent number: 11178082
    Abstract: Methods, systems, and computer programs are presented for a smart communications assistant with an audio interface. One method includes an operation for getting messages addressed to a user. The messages are from one or more message sources and each message comprising message data that includes text. The method further includes operations for analyzing the message data to determine a meaning of each message, for generating a score for each message based on the respective message data and the meaning of the message, and for generating a textual summary for the messages based on the message scores and the meaning of the messages. A speech summary is created based on the textual summary and the speech summary is then sent to a speaker associated with the user. The audio interface further allows the user to verbally request actions for the messages.
    Type: Grant
    Filed: November 15, 2019
    Date of Patent: November 16, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Nikrouz Ghotbi, August Niehaus, Sachin Venugopalan, Aleksandar Antonijevic, Tvrtko Tadic, Vashutosh Agrawal, Lisa Stifelman
  • Patent number: 11164564
    Abstract: According to certain embodiments, a system comprises interface circuitry and processing circuitry. The processing circuitry receives an input via the interface circuitry. The input is based on an utterance of a user, and the processing circuitry uses a probabilistic engine to determine one or more candidate intents associated with the utterance. The processing circuitry determines a number of the one or more candidate intents that exceed a threshold. If the number of candidate intents that exceed the threshold does not equal one, the processing circuitry uses a deterministic engine to compare the input to a set of regular expression patterns. If the input matches one of the regular expression patterns, the processing circuitry uses the matching regular expression pattern to determine the intent of the utterance. The interface circuitry communicates the intent of the utterance as an output.
    Type: Grant
    Filed: June 12, 2020
    Date of Patent: November 2, 2021
    Assignee: Bank of America Corporation
    Inventors: Donatus Asumu, Bhargav Aditya Ayyagari
  • Patent number: 11164584
    Abstract: Systems and methods are provided for application awakening and speech recognition. Such system may comprise a microphone configured to record an audio in an audio queue. The system may further comprise a processor configured to monitor the audio queue for an awakening phrase, in response to detecting the awakening phrase, obtain an audio segment from the audio queue, and transmit the obtained audio segment to a server. The recording of the audio may be continuous from a beginning of the awakening phrase to an end of the audio segment.
    Type: Grant
    Filed: September 9, 2019
    Date of Patent: November 2, 2021
    Assignee: Beijing DiDi Infinity Technology and Development Co., Ltd.
    Inventors: Liting Guo, Gangtao Hu
  • Patent number: 11151332
    Abstract: Embodiments provide for dialog based speech recognition by clustering a plurality of nodes comprising a dialog tree into at least a first cluster and a second cluster; creating a first dataset of natural language sentences for the first cluster and a second dataset of natural language sentences for the second cluster; generating a first specialized language model (LM) associated with the first cluster based on the first dataset; and generating a second specialized LM associated with the second cluster based on the second dataset, wherein the first specialized LM is different from the second specialized LM.
    Type: Grant
    Filed: March 7, 2019
    Date of Patent: October 19, 2021
    Assignee: International Business Machines Business
    Inventors: Julio Nogima, Marcelo C. Grave, Claudio S. Pinhanez
  • Patent number: 11115463
    Abstract: The description relates to predicting terms based on text inputted by a user. One example includes a computing device comprising a processor configured to send, over a communications network, the text to a remote prediction engine. The processor is configured to send the text to a local prediction engine stored at the computing device, and to monitor for a local predicted term from the local prediction engine and a remote predicted term from the remote prediction engine, in response to the sent text. The computing device includes a user interface configured to present a final predicted term to the user such that the user is able to select the final term. The processor is configured to form the final predicted term using either the remote predicted term or the local predicted term on the basis of a time interval running from the time at which the user input the text.
    Type: Grant
    Filed: November 22, 2016
    Date of Patent: September 7, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Adam John Cudworth, Alexander Gautam Primavesi, Piotr Jerzy Holc, Joseph Charles Woodward
  • Patent number: 11094316
    Abstract: A device includes a memory configured to store category labels associated with categories of a natural language processing library. A processor is configured to analyze input audio data to generate a text string and to perform natural language processing on at least the text string to generate an output text string including an action associated with a first device, a speaker, a location, or a combination thereof. The processor is configured to compare the input audio data to audio data of the categories to determine whether the input audio data matches any of the categories and, in response to determining that the input audio data does not match any of the categories: create a new category label, associate the new category label with at least a portion of the output text string, update the categories with the new category label, and generate a notification indicating the new category label.
    Type: Grant
    Filed: May 4, 2018
    Date of Patent: August 17, 2021
    Assignee: QUALCOMM Incorporated
    Inventors: Erik Visser, Fatemeh Saki, Yinyi Guo, Sunkuk Moon, Lae-Hoon Kim, Ravi Choudhary
  • Patent number: 11087764
    Abstract: A speech recognition apparatus includes a speech detection unit configured to detect a speech input by a user, an information providing unit configured to perform information provision to the user, using either first speech recognition information based on a recognition result of the speech by a first speech recognition unit or second speech recognition information based on a recognition result of the speech by a second speech recognition unit different from the first speech recognition unit, and a selection unit configured to select either the first speech recognition information or the second speech recognition information as speech recognition information to be used by the information providing unit on the basis of an elapsed time from the input of the speech, and change a method of the information provision by the information providing unit.
    Type: Grant
    Filed: November 14, 2017
    Date of Patent: August 10, 2021
    Assignee: Clarion Co., Ltd.
    Inventors: Takeshi Homma, Rui Zhang, Takuya Matsumoto, Hiroaki Kokubo
  • Patent number: 11074908
    Abstract: A method, computer program product, and computer system for identifying, by a computing device, at least one language model component of a plurality of language model components in at least one application associated with automatic speech recognition (ASR) and natural language understanding (NLU) usage. A contribution bias may be received for the at least one language model component. The ASR and NLU may be aligned between the plurality of language model components based upon, at least in part, the contribution bias.
    Type: Grant
    Filed: June 14, 2019
    Date of Patent: July 27, 2021
    Assignee: Nuance Communications, Inc.
    Inventors: Nathan Bodenstab, Matt Hohensee, Dermot Connolly, Kenneth Smith, Vittorio Manzone
  • Patent number: 11056118
    Abstract: A method of speaker identification comprises receiving a speech signal and dividing the speech signal into segments. Following each segment, a plurality of features are extracted from a most recently received segment, and scoring information is derived from the extracted features of the most recently received segment. The scoring information derived from the extracted features of the most recently received segment is combined with previously stored scoring information derived from the extracted features of any previously received segment. The new combined scoring information is stored, and an identification score is calculated using the combined scoring information.
    Type: Grant
    Filed: June 28, 2018
    Date of Patent: July 6, 2021
    Assignee: Cirrus Logic, Inc.
    Inventors: David Martínez González, Carlos Vaquero Avilés-Casco
  • Patent number: 11056104
    Abstract: In an approach for acoustic modeling with a language model, a computer isolates an audio stream. The computer identifies one or more language models based at least in part on the isolated audio stream. The computer selects a language model from the identified one or more language models. The computer creates a text based on the selected language model and the isolated audio stream. The computer creates an acoustic model based on the created text. The computer generates a confidence level associated with the created acoustic model. The computer selects a highest ranked language model based at least in part on the generated confidence level.
    Type: Grant
    Filed: May 26, 2017
    Date of Patent: July 6, 2021
    Assignee: International Business Machines Corporation
    Inventors: Aaron K. Baughman, Stephen C. Hammer, Mauro Marzorati
  • Patent number: 11049045
    Abstract: A classification apparatus includes: a calculation unit that outputs, as a classification result, results of classification by each of a plurality of classifiers with respect to learning data formed of data of at least two classes at a learning time and calculates a combination result value obtained by linear combination, using a combination coefficient, of results of classification by each of the plurality of classifiers with respect to the learning data to output the calculated combination result value as the classification result at a classification time; an extraction unit that extracts a correct solution class and an incorrect solution class for each of the classifiers from the classification result; a difference calculation unit that calculates a difference between the correct solution class and the incorrect solution class for each of the classifiers; a conversion unit that calculates a feature vector using the calculated difference for each of the classifiers; and a combination coefficient setting uni
    Type: Grant
    Filed: November 16, 2016
    Date of Patent: June 29, 2021
    Assignee: HONDA MOTOR CO., LTD.
    Inventors: Kotaro Funakoshi, Naoto Iwahashi
  • Patent number: 11037551
    Abstract: Methods, systems, and apparatus for receiving audio data corresponding to a user utterance and context data, identifying an initial set of one or more n-grams from the context data, generating an expanded set of one or more n-grams based on the initial set of n-grams, adjusting a language model based at least on the expanded set of n-grams, determining one or more speech recognition candidates for at least a portion of the user utterance using the adjusted language model, adjusting a score for a particular speech recognition candidate determined to be included in the expanded set of n-grams, determining a transcription of user utterance that includes at least one of the one or more speech recognition candidates, and providing the transcription of the user utterance for output.
    Type: Grant
    Filed: May 21, 2019
    Date of Patent: June 15, 2021
    Inventors: Petar Aleksic, Pedro J. Moreno Mengibar
  • Patent number: 11018885
    Abstract: In general, the disclosure describes techniques for automatically generating summaries of meetings. A computing system obtains a transcript of a meeting and may produce, based on the transcript of the meeting, a data structure that comprises utterance features. Furthermore, the computing system may determine, based on the transcript of the meeting, temporal bounds of a plurality of activity episodes within the meeting. For each respective activity episode of a plurality of activity episodes, the computing system may determine, based on the utterance features associated with the respective activity episode, a conversational activity type associated with the respective activity episode. Additionally, the computing system may produce an episode summary for the respective activity episode that is dependent on the determined conversational activity type associated with the respective activity episode.
    Type: Grant
    Filed: April 17, 2019
    Date of Patent: May 25, 2021
    Assignee: SRI International
    Inventor: John Niekrasz
  • Patent number: 10991363
    Abstract: An apparatus, method, and computer program product for adapting an acoustic model to a specific environment are defined. An adapted model obtained by adapting an original model to the specific environment using adaptation data, the original model being trained using training data and being used to calculate probabilities of context-dependent phones given an acoustic feature. Adapted probabilities obtained by adapting original probabilities using the training data and the adaptation data, the original probabilities being trained using the training data and being prior probabilities of context-dependent phones. An adapted acoustic model obtained from the adapted model and the adapted probabilities.
    Type: Grant
    Filed: November 6, 2017
    Date of Patent: April 27, 2021
    Assignee: International Business Machines Corporation
    Inventors: Gakuto Kurata, Bhuvana Ramabhadran, Masayuki Suzuki
  • Patent number: 10957308
    Abstract: Provided is a method and device to personalize a speech recognition model, the device that personalizes a speech recognition model by identifying a language group corresponding to a user, and generating a personalized speech recognition model by applying a group scale matrix corresponding to the identified language group to at least a layer of a speech recognition model.
    Type: Grant
    Filed: August 31, 2018
    Date of Patent: March 23, 2021
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Ki Soo Kwon, Inchul Song, YoungSang Choi
  • Patent number: 10957129
    Abstract: Methods, systems, and apparatus for monitoring a sound are described. An audio signal is obtained and the audio signal is analyzed to generate an audio signature. An object type is identified based on the audio signature and an action corresponding to the object type is identified.
    Type: Grant
    Filed: December 12, 2018
    Date of Patent: March 23, 2021
    Assignee: eBay Inc.
    Inventor: Sergio Pinzon Gonzales, Jr.
  • Patent number: 10943583
    Abstract: A system to perform automatic speech recognition (ASR) using a dynamic language model. Portions of the language model can include a group of probabilities rather than a single probability. At runtime individual probabilities of the group are weighted and combined to create an adjusted probability for the portion of the language model. The adjusted probability can be used for ASR processing. The weights can be determined based on a characteristic of the utterance, for example an associated speechlet/application, the specific user speaking, or other characteristic. By applying the weights at runtime the system can use a single language model to dynamically adjust to different utterance conditions.
    Type: Grant
    Filed: March 23, 2018
    Date of Patent: March 9, 2021
    Assignee: Amazon Technologies, Inc.
    Inventors: Ankur Gandhe, Ariya Rastrow, Shaswat Pratap Shah
  • Patent number: 10943143
    Abstract: Techniques are disclosed relating to scoring partial matches between words. In certain embodiments, a method may include receiving a request to determine a similarity between an input text data and a stored text data. The method also includes determining, based on comparing one or more words included in the input text data with one or more words included in the stored text data, a set of word pairs and a set of unpaired words. Further, in response to determining that the set of unpaired words passes elimination criteria, the method includes calculating a base similarity score between the input text data and the stored text data based on the set of word pairs. The method also includes determining a scoring penalty based on the set of unpaired words and generating a final similarity score between the input text data and the stored text data by modifying the base similarity score based on the scoring penalty.
    Type: Grant
    Filed: December 28, 2018
    Date of Patent: March 9, 2021
    Assignee: PAYPAL, INC.
    Inventors: Rushik Upadhyay, Dhamodharan Lakshmipathy, Nandhini Ramesh, Aditya Kaulagi
  • Patent number: 10937415
    Abstract: There is provided an information processing device to further improve the operability of user interfaces that use a voice as an input, the information processing device including: an acquisition unit configured to acquire context information in a period for collection of a voice; and a control unit configured to cause a predetermined output unit to present a candidate for character information obtained by converting the voice in a mode in accordance with the context information.
    Type: Grant
    Filed: March 15, 2017
    Date of Patent: March 2, 2021
    Inventors: Ayumi Kato, Shinichi Kawano, Yuhei Taki, Yusuke Nakagawa
  • Patent number: 10922990
    Abstract: A display apparatus and a method for questions and answers includes a display unit includes an input unit configured to receive user's speech voice; a communication unit configured to perform data communication with an answer server; and a processor configured to create and display one or more question sentences using the speech voice in response to the speech voice being a word speech, create a question language corresponding to the question sentence selected from among the displayed one or more question sentences, transmit the created question language to the answer server via the communication unit, and, in response to one or more answer results related to the question language being received from the answer server, display the received one or more answer results. Accordingly, the display apparatus may provide an answer result appropriate to a user's question intention although a non-sentence speech is input.
    Type: Grant
    Filed: May 23, 2019
    Date of Patent: February 16, 2021
    Inventor: Eun-sang Bak
  • Patent number: 10896681
    Abstract: This document describes, among other things, a computer-implemented method for transcribing an utterance. The method can include receiving, at a computing system, speech data that characterizes an utterance of a user. A first set of candidate transcriptions of the utterance can be generated using a static class-based language model that includes a plurality of classes that are each populated with class-based terms selected independently of the utterance or the user. The computing system can then determine whether the first set of candidate transcriptions includes class-based terms. Based on whether the first set of candidate transcriptions includes class-based terms, the computing system can determine whether to generate a dynamic class-based language model that includes at least one class that is populated with class-based terms selected based on a context associated with at least one of the utterance and the user.
    Type: Grant
    Filed: December 29, 2015
    Date of Patent: January 19, 2021
    Assignee: Google LLC
    Inventors: Petar Aleksic, Pedro J. Moreno Mengibar
  • Patent number: 10896293
    Abstract: Provided is an information processing apparatus including a processing unit configured to determine, on a basis of a word of a predetermined unit selected in a text string indicated by text string information, another word connected to the selected word and included in the text string and to set a delimitation in the text string with regard to the selected word.
    Type: Grant
    Filed: April 19, 2017
    Date of Patent: January 19, 2021
    Inventors: Yuhei Taki, Shinichi Kawano
  • Patent number: 10885909
    Abstract: A speech recognition method to be performed by a computer, the method including: detecting a first keyword uttered by a user from an audio signal representing voice of the user; detecting a term indicating a request of the user from sections that follow the first keyword in the audio signal; and determining a type of speech recognition processing applied to the following sections in accordance with the detected term indicating the request of the user.
    Type: Grant
    Filed: February 6, 2018
    Date of Patent: January 5, 2021
    Inventors: Chikako Matsumoto, Naoshi Matsuo
  • Patent number: 10872613
    Abstract: A method includes generating a synthesized non-reference high-band channel based on a non-reference high-band excitation corresponding to a non-reference target channel. The method further includes estimating one or more spectral mapping parameters based on the synthesized non-reference high-band channel and a high-band portion of the non-reference target channel. The method also includes applying the one or more spectral mapping parameters to the synthesized non-reference high-band channel to generate a spectrally shaped synthesized non-reference high-band channel. The method further includes generating an encoded bitstream based on the one or more spectral mapping parameters and the spectrally shaped synthesized non-reference high-band channel.
    Type: Grant
    Filed: November 4, 2019
    Date of Patent: December 22, 2020
    Assignee: QUALCOMM Incorporated
    Inventors: Venkata Subrahmanyam Chandra Sekhar Chebiyyam, Venkatraman Atti
  • Patent number: 10867598
    Abstract: A semantic analysis method, semantic analysis and non-transitory computer-readable medium are provided in this disclosure.
    Type: Grant
    Filed: December 10, 2018
    Date of Patent: December 15, 2020
    Inventors: Yu-Shian Chiu, Wei-Jen Yang
  • Patent number: 10847147
    Abstract: Automatic speech recognition systems can benefit from cues in user voice such as hyperarticulation. Traditional approaches typically attempt to define and detect an absolute state of hyperarticulation, which is very difficult, especially on short voice queries. This disclosure provides for an approach for hyperarticulation detection using pair-wise comparisons and on a real-world speech recognition system. The disclosed approach uses delta features extracted from a pair of repetitive user utterances. The improvements provided by the disclosed systems and methods include improvements in word error rate by using hyperarticulation information as a feature in a second pass N-best hypotheses rescoring setup.
    Type: Grant
    Filed: May 24, 2019
    Date of Patent: November 24, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Ranjitha Gurunath Kulkarni, Ahmed Moustafa El Kholy, Ziad Al Bawab, Noha Alon, Imed Zitouni
  • Patent number: 10847137
    Abstract: An approach to speech recognition, and in particular trigger word detection, implements fixed feature extraction form waveform samples with a neural network (NN). For example, rather than computing Log Frequency Band Energies (LFBEs), a convolutional neural network is used. In some implementations, this NN waveform processing is combined with a trained secondary classification that makes use of phonetic segmentation of a possible trigger word occurrence.
    Type: Grant
    Filed: December 12, 2017
    Date of Patent: November 24, 2020
    Assignee: Amazon Technologies, Inc.
    Inventors: Arindam Mandal, Nikko Strom, Kenichi Kumatani, Sankaran Panchapagesan
  • Patent number: 10841411
    Abstract: Systems, methods, and devices for establishing communications sessions with contacts are disclosed. In some embodiments, a first request may be received from a first device. The first request may be to communicate with a contact name. A user account associated with the first device may then be identified, and a contact list associated with the user account may be accessed to determine contacts associated with the contact name. Based on the contact list, a first contact and a second contact associated with the contact name may be identified. It may be determined, from memory, that the first contact is a first preferred contact. However, based on an intervening event, the second contact, rather than the preferred contact, may be selected for communicating with the contact.
    Type: Grant
    Filed: November 9, 2017
    Date of Patent: November 17, 2020
    Assignee: Amazon Technologies, Inc.
    Inventor: Aparna Nandyal
  • Patent number: 10839796
    Abstract: Multi-turn conversation systems that are personalized to a user based on insights derived from big data are described. A method includes: receiving, by a computer device, input from a user; obtaining, by the computer device, insights about the user; generating, by the computer device, a response based on the insights and the input; and outputting, by the computer device, the response.
    Type: Grant
    Filed: December 15, 2017
    Date of Patent: November 17, 2020
    Inventors: Faried Abrahams, Lalit Agarwalla, Gandhi Sivakumar
  • Patent number: 10832664
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for language models using domain-specific model components. In some implementations, context data for an utterance is obtained. A domain-specific model component is selected from among multiple domain-specific model components of a language model based on the non-linguistic context of the utterance. A score for a candidate transcription for the utterance is generated using the selected domain-specific model component and a baseline model component of the language model that is domain-independent. A transcription for the utterance is determined using the score the transcription is provided as output of an automated speech recognition system.
    Type: Grant
    Filed: August 21, 2017
    Date of Patent: November 10, 2020
    Assignee: Google LLC
    Inventors: Fadi Biadsy, Diamantino Antionio Caseiro
  • Patent number: 10832658
    Abstract: A method, program product and computer system to predict utterances in a dialog system includes receiving a set of utterances associated with a dialog between a client device and a dialog system, mapping the utterances to vector representations of the utterances, and identifying at least one cluster to which the utterances belong from among a plurality of possible clusters. A next cluster is predicted based upon a conditional probability of the next cluster following a set of a predetermined number of previous clusters using a language model. A next utterance is predicted from among a plurality of possible utterances within the predicted next cluster.
    Type: Grant
    Filed: March 8, 2018
    Date of Patent: November 10, 2020
    Inventors: Chulaka Gunasekara, David Nahamoo, Lazaros Polymenakos, Kshitij Fadnis, David Echeverria Ciaurri, Jatin Ganhotra
  • Patent number: 10810472
    Abstract: Techniques are provided for performing sentiment analysis on words in a first data set. An example embodiment includes generating a word embedding model including a first plurality of features. A value indicating sentiment for the words in the first data set can be determined using a convolutional neural network (CNN). A second plurality of features are generated based on bigrams identified in the data set. The bigrams can be generated using a co-occurrence graph. The model is updated to include the second plurality of features, and sentiment analysis can be performed on a second data set using the updated model.
    Type: Grant
    Filed: May 10, 2018
    Date of Patent: October 20, 2020
    Inventors: Michael Malak, Mark L. Kreider
  • Patent number: 10789946
    Abstract: Systems and methods are provided for speech recognition. An example method may be implementable by a server. The method may comprise adding a key phrase into a dictionary comprising a plurality of dictionary phrases, and for each one or more of the dictionary phrases, obtaining a first probability that the dictionary phrase is after the key phrase in a phrase sequence. The key phrase and the dictionary phrase may each comprise one or more words. The first probability may be independent of the key phrase.
    Type: Grant
    Filed: December 27, 2019
    Date of Patent: September 29, 2020
    Inventor: Chen Huang