Probability Patents (Class 704/240)
  • Patent number: 11115463
    Abstract: The description relates to predicting terms based on text inputted by a user. One example includes a computing device comprising a processor configured to send, over a communications network, the text to a remote prediction engine. The processor is configured to send the text to a local prediction engine stored at the computing device, and to monitor for a local predicted term from the local prediction engine and a remote predicted term from the remote prediction engine, in response to the sent text. The computing device includes a user interface configured to present a final predicted term to the user such that the user is able to select the final term. The processor is configured to form the final predicted term using either the remote predicted term or the local predicted term on the basis of a time interval running from the time at which the user input the text.
    Type: Grant
    Filed: November 22, 2016
    Date of Patent: September 7, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Adam John Cudworth, Alexander Gautam Primavesi, Piotr Jerzy Holc, Joseph Charles Woodward
  • Patent number: 11094316
    Abstract: A device includes a memory configured to store category labels associated with categories of a natural language processing library. A processor is configured to analyze input audio data to generate a text string and to perform natural language processing on at least the text string to generate an output text string including an action associated with a first device, a speaker, a location, or a combination thereof. The processor is configured to compare the input audio data to audio data of the categories to determine whether the input audio data matches any of the categories and, in response to determining that the input audio data does not match any of the categories: create a new category label, associate the new category label with at least a portion of the output text string, update the categories with the new category label, and generate a notification indicating the new category label.
    Type: Grant
    Filed: May 4, 2018
    Date of Patent: August 17, 2021
    Assignee: QUALCOMM Incorporated
    Inventors: Erik Visser, Fatemeh Saki, Yinyi Guo, Sunkuk Moon, Lae-Hoon Kim, Ravi Choudhary
  • Patent number: 11087764
    Abstract: A speech recognition apparatus includes a speech detection unit configured to detect a speech input by a user, an information providing unit configured to perform information provision to the user, using either first speech recognition information based on a recognition result of the speech by a first speech recognition unit or second speech recognition information based on a recognition result of the speech by a second speech recognition unit different from the first speech recognition unit, and a selection unit configured to select either the first speech recognition information or the second speech recognition information as speech recognition information to be used by the information providing unit on the basis of an elapsed time from the input of the speech, and change a method of the information provision by the information providing unit.
    Type: Grant
    Filed: November 14, 2017
    Date of Patent: August 10, 2021
    Assignee: Clarion Co., Ltd.
    Inventors: Takeshi Homma, Rui Zhang, Takuya Matsumoto, Hiroaki Kokubo
  • Patent number: 11074908
    Abstract: A method, computer program product, and computer system for identifying, by a computing device, at least one language model component of a plurality of language model components in at least one application associated with automatic speech recognition (ASR) and natural language understanding (NLU) usage. A contribution bias may be received for the at least one language model component. The ASR and NLU may be aligned between the plurality of language model components based upon, at least in part, the contribution bias.
    Type: Grant
    Filed: June 14, 2019
    Date of Patent: July 27, 2021
    Assignee: Nuance Communications, Inc.
    Inventors: Nathan Bodenstab, Matt Hohensee, Dermot Connolly, Kenneth Smith, Vittorio Manzone
  • Patent number: 11056104
    Abstract: In an approach for acoustic modeling with a language model, a computer isolates an audio stream. The computer identifies one or more language models based at least in part on the isolated audio stream. The computer selects a language model from the identified one or more language models. The computer creates a text based on the selected language model and the isolated audio stream. The computer creates an acoustic model based on the created text. The computer generates a confidence level associated with the created acoustic model. The computer selects a highest ranked language model based at least in part on the generated confidence level.
    Type: Grant
    Filed: May 26, 2017
    Date of Patent: July 6, 2021
    Assignee: International Business Machines Corporation
    Inventors: Aaron K. Baughman, Stephen C. Hammer, Mauro Marzorati
  • Patent number: 11056118
    Abstract: A method of speaker identification comprises receiving a speech signal and dividing the speech signal into segments. Following each segment, a plurality of features are extracted from a most recently received segment, and scoring information is derived from the extracted features of the most recently received segment. The scoring information derived from the extracted features of the most recently received segment is combined with previously stored scoring information derived from the extracted features of any previously received segment. The new combined scoring information is stored, and an identification score is calculated using the combined scoring information.
    Type: Grant
    Filed: June 28, 2018
    Date of Patent: July 6, 2021
    Assignee: Cirrus Logic, Inc.
    Inventors: David Martínez González, Carlos Vaquero Avilés-Casco
  • Patent number: 11049045
    Abstract: A classification apparatus includes: a calculation unit that outputs, as a classification result, results of classification by each of a plurality of classifiers with respect to learning data formed of data of at least two classes at a learning time and calculates a combination result value obtained by linear combination, using a combination coefficient, of results of classification by each of the plurality of classifiers with respect to the learning data to output the calculated combination result value as the classification result at a classification time; an extraction unit that extracts a correct solution class and an incorrect solution class for each of the classifiers from the classification result; a difference calculation unit that calculates a difference between the correct solution class and the incorrect solution class for each of the classifiers; a conversion unit that calculates a feature vector using the calculated difference for each of the classifiers; and a combination coefficient setting uni
    Type: Grant
    Filed: November 16, 2016
    Date of Patent: June 29, 2021
    Assignee: HONDA MOTOR CO., LTD.
    Inventors: Kotaro Funakoshi, Naoto Iwahashi
  • Patent number: 11037551
    Abstract: Methods, systems, and apparatus for receiving audio data corresponding to a user utterance and context data, identifying an initial set of one or more n-grams from the context data, generating an expanded set of one or more n-grams based on the initial set of n-grams, adjusting a language model based at least on the expanded set of n-grams, determining one or more speech recognition candidates for at least a portion of the user utterance using the adjusted language model, adjusting a score for a particular speech recognition candidate determined to be included in the expanded set of n-grams, determining a transcription of user utterance that includes at least one of the one or more speech recognition candidates, and providing the transcription of the user utterance for output.
    Type: Grant
    Filed: May 21, 2019
    Date of Patent: June 15, 2021
    Inventors: Petar Aleksic, Pedro J. Moreno Mengibar
  • Patent number: 11018885
    Abstract: In general, the disclosure describes techniques for automatically generating summaries of meetings. A computing system obtains a transcript of a meeting and may produce, based on the transcript of the meeting, a data structure that comprises utterance features. Furthermore, the computing system may determine, based on the transcript of the meeting, temporal bounds of a plurality of activity episodes within the meeting. For each respective activity episode of a plurality of activity episodes, the computing system may determine, based on the utterance features associated with the respective activity episode, a conversational activity type associated with the respective activity episode. Additionally, the computing system may produce an episode summary for the respective activity episode that is dependent on the determined conversational activity type associated with the respective activity episode.
    Type: Grant
    Filed: April 17, 2019
    Date of Patent: May 25, 2021
    Assignee: SRI International
    Inventor: John Niekrasz
  • Patent number: 10991363
    Abstract: An apparatus, method, and computer program product for adapting an acoustic model to a specific environment are defined. An adapted model obtained by adapting an original model to the specific environment using adaptation data, the original model being trained using training data and being used to calculate probabilities of context-dependent phones given an acoustic feature. Adapted probabilities obtained by adapting original probabilities using the training data and the adaptation data, the original probabilities being trained using the training data and being prior probabilities of context-dependent phones. An adapted acoustic model obtained from the adapted model and the adapted probabilities.
    Type: Grant
    Filed: November 6, 2017
    Date of Patent: April 27, 2021
    Assignee: International Business Machines Corporation
    Inventors: Gakuto Kurata, Bhuvana Ramabhadran, Masayuki Suzuki
  • Patent number: 10957308
    Abstract: Provided is a method and device to personalize a speech recognition model, the device that personalizes a speech recognition model by identifying a language group corresponding to a user, and generating a personalized speech recognition model by applying a group scale matrix corresponding to the identified language group to at least a layer of a speech recognition model.
    Type: Grant
    Filed: August 31, 2018
    Date of Patent: March 23, 2021
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Ki Soo Kwon, Inchul Song, YoungSang Choi
  • Patent number: 10957129
    Abstract: Methods, systems, and apparatus for monitoring a sound are described. An audio signal is obtained and the audio signal is analyzed to generate an audio signature. An object type is identified based on the audio signature and an action corresponding to the object type is identified.
    Type: Grant
    Filed: December 12, 2018
    Date of Patent: March 23, 2021
    Assignee: eBay Inc.
    Inventor: Sergio Pinzon Gonzales, Jr.
  • Patent number: 10943583
    Abstract: A system to perform automatic speech recognition (ASR) using a dynamic language model. Portions of the language model can include a group of probabilities rather than a single probability. At runtime individual probabilities of the group are weighted and combined to create an adjusted probability for the portion of the language model. The adjusted probability can be used for ASR processing. The weights can be determined based on a characteristic of the utterance, for example an associated speechlet/application, the specific user speaking, or other characteristic. By applying the weights at runtime the system can use a single language model to dynamically adjust to different utterance conditions.
    Type: Grant
    Filed: March 23, 2018
    Date of Patent: March 9, 2021
    Assignee: Amazon Technologies, Inc.
    Inventors: Ankur Gandhe, Ariya Rastrow, Shaswat Pratap Shah
  • Patent number: 10943143
    Abstract: Techniques are disclosed relating to scoring partial matches between words. In certain embodiments, a method may include receiving a request to determine a similarity between an input text data and a stored text data. The method also includes determining, based on comparing one or more words included in the input text data with one or more words included in the stored text data, a set of word pairs and a set of unpaired words. Further, in response to determining that the set of unpaired words passes elimination criteria, the method includes calculating a base similarity score between the input text data and the stored text data based on the set of word pairs. The method also includes determining a scoring penalty based on the set of unpaired words and generating a final similarity score between the input text data and the stored text data by modifying the base similarity score based on the scoring penalty.
    Type: Grant
    Filed: December 28, 2018
    Date of Patent: March 9, 2021
    Assignee: PAYPAL, INC.
    Inventors: Rushik Upadhyay, Dhamodharan Lakshmipathy, Nandhini Ramesh, Aditya Kaulagi
  • Patent number: 10937415
    Abstract: There is provided an information processing device to further improve the operability of user interfaces that use a voice as an input, the information processing device including: an acquisition unit configured to acquire context information in a period for collection of a voice; and a control unit configured to cause a predetermined output unit to present a candidate for character information obtained by converting the voice in a mode in accordance with the context information.
    Type: Grant
    Filed: March 15, 2017
    Date of Patent: March 2, 2021
    Assignee: SONY CORPORATION
    Inventors: Ayumi Kato, Shinichi Kawano, Yuhei Taki, Yusuke Nakagawa
  • Patent number: 10922990
    Abstract: A display apparatus and a method for questions and answers includes a display unit includes an input unit configured to receive user's speech voice; a communication unit configured to perform data communication with an answer server; and a processor configured to create and display one or more question sentences using the speech voice in response to the speech voice being a word speech, create a question language corresponding to the question sentence selected from among the displayed one or more question sentences, transmit the created question language to the answer server via the communication unit, and, in response to one or more answer results related to the question language being received from the answer server, display the received one or more answer results. Accordingly, the display apparatus may provide an answer result appropriate to a user's question intention although a non-sentence speech is input.
    Type: Grant
    Filed: May 23, 2019
    Date of Patent: February 16, 2021
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventor: Eun-sang Bak
  • Patent number: 10896293
    Abstract: Provided is an information processing apparatus including a processing unit configured to determine, on a basis of a word of a predetermined unit selected in a text string indicated by text string information, another word connected to the selected word and included in the text string and to set a delimitation in the text string with regard to the selected word.
    Type: Grant
    Filed: April 19, 2017
    Date of Patent: January 19, 2021
    Assignee: SONY CORPORATION
    Inventors: Yuhei Taki, Shinichi Kawano
  • Patent number: 10896681
    Abstract: This document describes, among other things, a computer-implemented method for transcribing an utterance. The method can include receiving, at a computing system, speech data that characterizes an utterance of a user. A first set of candidate transcriptions of the utterance can be generated using a static class-based language model that includes a plurality of classes that are each populated with class-based terms selected independently of the utterance or the user. The computing system can then determine whether the first set of candidate transcriptions includes class-based terms. Based on whether the first set of candidate transcriptions includes class-based terms, the computing system can determine whether to generate a dynamic class-based language model that includes at least one class that is populated with class-based terms selected based on a context associated with at least one of the utterance and the user.
    Type: Grant
    Filed: December 29, 2015
    Date of Patent: January 19, 2021
    Assignee: Google LLC
    Inventors: Petar Aleksic, Pedro J. Moreno Mengibar
  • Patent number: 10885909
    Abstract: A speech recognition method to be performed by a computer, the method including: detecting a first keyword uttered by a user from an audio signal representing voice of the user; detecting a term indicating a request of the user from sections that follow the first keyword in the audio signal; and determining a type of speech recognition processing applied to the following sections in accordance with the detected term indicating the request of the user.
    Type: Grant
    Filed: February 6, 2018
    Date of Patent: January 5, 2021
    Assignee: FUJITSU LIMITED
    Inventors: Chikako Matsumoto, Naoshi Matsuo
  • Patent number: 10872613
    Abstract: A method includes generating a synthesized non-reference high-band channel based on a non-reference high-band excitation corresponding to a non-reference target channel. The method further includes estimating one or more spectral mapping parameters based on the synthesized non-reference high-band channel and a high-band portion of the non-reference target channel. The method also includes applying the one or more spectral mapping parameters to the synthesized non-reference high-band channel to generate a spectrally shaped synthesized non-reference high-band channel. The method further includes generating an encoded bitstream based on the one or more spectral mapping parameters and the spectrally shaped synthesized non-reference high-band channel.
    Type: Grant
    Filed: November 4, 2019
    Date of Patent: December 22, 2020
    Assignee: QUALCOMM Incorporated
    Inventors: Venkata Subrahmanyam Chandra Sekhar Chebiyyam, Venkatraman Atti
  • Patent number: 10867598
    Abstract: A semantic analysis method, semantic analysis and non-transitory computer-readable medium are provided in this disclosure.
    Type: Grant
    Filed: December 10, 2018
    Date of Patent: December 15, 2020
    Assignee: INSTITUTE FOR INFORMATION INDUSTRY
    Inventors: Yu-Shian Chiu, Wei-Jen Yang
  • Patent number: 10847147
    Abstract: Automatic speech recognition systems can benefit from cues in user voice such as hyperarticulation. Traditional approaches typically attempt to define and detect an absolute state of hyperarticulation, which is very difficult, especially on short voice queries. This disclosure provides for an approach for hyperarticulation detection using pair-wise comparisons and on a real-world speech recognition system. The disclosed approach uses delta features extracted from a pair of repetitive user utterances. The improvements provided by the disclosed systems and methods include improvements in word error rate by using hyperarticulation information as a feature in a second pass N-best hypotheses rescoring setup.
    Type: Grant
    Filed: May 24, 2019
    Date of Patent: November 24, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Ranjitha Gurunath Kulkarni, Ahmed Moustafa El Kholy, Ziad Al Bawab, Noha Alon, Imed Zitouni
  • Patent number: 10847137
    Abstract: An approach to speech recognition, and in particular trigger word detection, implements fixed feature extraction form waveform samples with a neural network (NN). For example, rather than computing Log Frequency Band Energies (LFBEs), a convolutional neural network is used. In some implementations, this NN waveform processing is combined with a trained secondary classification that makes use of phonetic segmentation of a possible trigger word occurrence.
    Type: Grant
    Filed: December 12, 2017
    Date of Patent: November 24, 2020
    Assignee: Amazon Technologies, Inc.
    Inventors: Arindam Mandal, Nikko Strom, Kenichi Kumatani, Sankaran Panchapagesan
  • Patent number: 10841411
    Abstract: Systems, methods, and devices for establishing communications sessions with contacts are disclosed. In some embodiments, a first request may be received from a first device. The first request may be to communicate with a contact name. A user account associated with the first device may then be identified, and a contact list associated with the user account may be accessed to determine contacts associated with the contact name. Based on the contact list, a first contact and a second contact associated with the contact name may be identified. It may be determined, from memory, that the first contact is a first preferred contact. However, based on an intervening event, the second contact, rather than the preferred contact, may be selected for communicating with the contact.
    Type: Grant
    Filed: November 9, 2017
    Date of Patent: November 17, 2020
    Assignee: Amazon Technologies, Inc.
    Inventor: Aparna Nandyal
  • Patent number: 10839796
    Abstract: Multi-turn conversation systems that are personalized to a user based on insights derived from big data are described. A method includes: receiving, by a computer device, input from a user; obtaining, by the computer device, insights about the user; generating, by the computer device, a response based on the insights and the input; and outputting, by the computer device, the response.
    Type: Grant
    Filed: December 15, 2017
    Date of Patent: November 17, 2020
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Faried Abrahams, Lalit Agarwalla, Gandhi Sivakumar
  • Patent number: 10832664
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for language models using domain-specific model components. In some implementations, context data for an utterance is obtained. A domain-specific model component is selected from among multiple domain-specific model components of a language model based on the non-linguistic context of the utterance. A score for a candidate transcription for the utterance is generated using the selected domain-specific model component and a baseline model component of the language model that is domain-independent. A transcription for the utterance is determined using the score the transcription is provided as output of an automated speech recognition system.
    Type: Grant
    Filed: August 21, 2017
    Date of Patent: November 10, 2020
    Assignee: Google LLC
    Inventors: Fadi Biadsy, Diamantino Antionio Caseiro
  • Patent number: 10832658
    Abstract: A method, program product and computer system to predict utterances in a dialog system includes receiving a set of utterances associated with a dialog between a client device and a dialog system, mapping the utterances to vector representations of the utterances, and identifying at least one cluster to which the utterances belong from among a plurality of possible clusters. A next cluster is predicted based upon a conditional probability of the next cluster following a set of a predetermined number of previous clusters using a language model. A next utterance is predicted from among a plurality of possible utterances within the predicted next cluster.
    Type: Grant
    Filed: March 8, 2018
    Date of Patent: November 10, 2020
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Chulaka Gunasekara, David Nahamoo, Lazaros Polymenakos, Kshitij Fadnis, David Echeverria Ciaurri, Jatin Ganhotra
  • Patent number: 10810472
    Abstract: Techniques are provided for performing sentiment analysis on words in a first data set. An example embodiment includes generating a word embedding model including a first plurality of features. A value indicating sentiment for the words in the first data set can be determined using a convolutional neural network (CNN). A second plurality of features are generated based on bigrams identified in the data set. The bigrams can be generated using a co-occurrence graph. The model is updated to include the second plurality of features, and sentiment analysis can be performed on a second data set using the updated model.
    Type: Grant
    Filed: May 10, 2018
    Date of Patent: October 20, 2020
    Assignee: ORACLE INTERNATIONAL CORPORATION
    Inventors: Michael Malak, Mark L. Kreider
  • Patent number: 10789946
    Abstract: Systems and methods are provided for speech recognition. An example method may be implementable by a server. The method may comprise adding a key phrase into a dictionary comprising a plurality of dictionary phrases, and for each one or more of the dictionary phrases, obtaining a first probability that the dictionary phrase is after the key phrase in a phrase sequence. The key phrase and the dictionary phrase may each comprise one or more words. The first probability may be independent of the key phrase.
    Type: Grant
    Filed: December 27, 2019
    Date of Patent: September 29, 2020
    Assignee: BEIJING DIDI INFINITY TECHNOLOGY AND DEVELOPMENT CO., LTD.
    Inventor: Chen Huang
  • Patent number: 10714080
    Abstract: A weighted finite-state transducer (WFST) decoding system is provided. The WFST decoding system includes a memory that stores WFST data and a WFST decoder including a data fetch logic. The WFST data has a structure including states, and arcs connecting the states with directivity. The WFST data is compressed in the memory. The WFST data includes body data, and header data including state information for each states that is aligned discontinuously. The body data includes arc information of the arcs that is aligned continuously. The state information includes an arc index of the arcs, a number of the arcs, and compression information of the arcs, and the data fetch logic de-compresses the WFST data using the compression information, and retrieves the WFST data from the memory.
    Type: Grant
    Filed: September 8, 2017
    Date of Patent: July 14, 2020
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Jae Sung Yoon, Jun Seok Park
  • Patent number: 10713519
    Abstract: The present invention is directed towards providing automated workflows for the identification of a reading order from text segments extracted from a document. Ordering the text segments is based on trained natural language models. In some embodiments, the workflows are enabled to perform a method for identifying a sequence associated with a portable document. The methods includes iteratively generating a probabilistic language model, receiving the portable document, and selectively extracting features (such as but not limited to text segments) from the document. The method may generate pairs of features (or feature pair from the extracted features). The method may further generate a score for each of the pairs based on the probabilistic language model and determine an order to features based on the scores. The method may provide the extracted features in the determined order.
    Type: Grant
    Filed: June 22, 2017
    Date of Patent: July 14, 2020
    Assignee: ADOBE INC.
    Inventors: Trung Huu Bui, Hung Hai Bui, Shawn Alan Gaither, Walter Wei-Tuh Chang, Michael Frank Kraley, Pranjal Daga
  • Patent number: 10714122
    Abstract: Speech or non-speech detection techniques are discussed and include updating a speech pattern model using probability scores from an acoustic model to generate a score for each state of the speech pattern model, such that the speech pattern model includes a first non-speech state having multiple self loops each associated with a non-speech probability score of the probability scores, a plurality of speech states following the first non-speech state, and a second non-speech state following the speech states, and detecting speech based on a comparison of a score of the first non-speech state and a score of the last speech state of the multiple speech states.
    Type: Grant
    Filed: June 6, 2018
    Date of Patent: July 14, 2020
    Assignee: Intel Corporation
    Inventors: Maciej Muchlinski, Tobias Bocklet
  • Patent number: 10713289
    Abstract: Systems, methods, and devices for performing interactive question answering using data source credibility and conversation entropy are disclosed. A speech-controlled device captures audio including a spoken question, and sends audio data corresponding thereto to a server(s). The server(s) performs speech processing on the audio data, and determines various stored data that can be used to determine an answer to the question. The server(s) determines which stored data to use based on the credibility of the source from which the stored data was received. The server(s) may also determine a number of user interactions needed to obtain data in order to fully answer the question and may select a question for a dialog soliciting further data based on the number of user interactions.
    Type: Grant
    Filed: March 31, 2017
    Date of Patent: July 14, 2020
    Assignee: Amazon Technologies, Inc.
    Inventors: Nina Mishra, Yonatan Naamad
  • Patent number: 10706852
    Abstract: The described technology provides arbitration between speech recognition results generated by different automatic speech recognition (ASR) engines, such as ASR engines trained according to different language or acoustic models. The system includes an arbitrator that selects between a first speech recognition result representing an acoustic utterance as transcribed by a first ASR engine and a second speech recognition result representing the acoustic utterance as transcribed by a second ASR engine. This selection is based on a set of confidence features that is initially used by the first ASR engine or the second ASR engine to generate the first and second speech recognition results.
    Type: Grant
    Filed: November 13, 2015
    Date of Patent: July 7, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Kshitiz Kumar, Hosam Khalil, Yifan Gong, Ziad Al-Bawab, Chaojun Liu
  • Patent number: 10699712
    Abstract: An information processing method and an electronic device are provided. The method includes an electronic device obtaining an input information through a second collection manner when the electronic device is in a speech collection state for obtaining speech information through a first collection manner, and determining a logic boundary position in relation to a first speech information in accordance with the input information, the first speech information is obtained by the electronic device through the first collection manner which is different from the second collection manner. An electronic device corresponding thereto is also disclosed.
    Type: Grant
    Filed: March 4, 2015
    Date of Patent: June 30, 2020
    Assignee: LENOVO (BEIJING) CO., LTD.
    Inventors: Haisheng Dai, Zhepeng Wang
  • Patent number: 10699696
    Abstract: The present disclosure provides a method and apparatus for correcting a speech recognition error based on artificial intelligence, and a storage medium, wherein the method comprises: obtaining a second speech recognition result of a second speech query input by the user; performing error-correcting intention recognition according to the second speech recognition result; extracting error-correcting information from the second speech recognition result when it is determined that the user has an error-correcting intention; screening error-correcting resources according to the error-correcting information, and using a selected best-matched error-correcting resource to perform error correction for the first speech recognition result, the first speech recognition result being a speech recognition result of a first speech query which is input before the second speech query.
    Type: Grant
    Filed: May 22, 2018
    Date of Patent: June 30, 2020
    Assignee: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY CO., LTD.
    Inventors: Hehan Li, Wensong He
  • Patent number: 10649850
    Abstract: Techniques and systems for storing and retrieving data storage devices of a data storage system are disclosed. In some embodiments, inventory holders are used to store data storage devices used by a data storage system. When data is to be transacted with the data storage devices, mobile drive units locate appropriate inventory holders and transport them to a device reading station, where an appropriate device retrieval unit transacts the data. In some embodiments, each inventory holder includes a heterogenous mix of data storage device types, the layout of which may be calculated according to the specific mix allocated to a given inventory holder. After the data has been transacted, the data storage devices are returned to the appropriate inventory holders, and the inventory holders are placed by the mobile drive units in locations where they may be accessed in response to further data transactions.
    Type: Grant
    Filed: June 29, 2015
    Date of Patent: May 12, 2020
    Assignee: Amazon Technologies, Inc.
    Inventors: James Raymond Allard, Paul David Franklin, Samuel Rubin Barrett, Jeremiah Brazeau, Jeffrey Allen Dzado, James Caleb Kirschner, David Levy, Brent James Lutz, Andrew Brendan Tinka, Colin Laird Lazier
  • Patent number: 10652592
    Abstract: Methods and systems are disclosed for enriching a viewing experience of a user watching video content on a screen of a client terminal by increasing the relevance of additional media content proposed or provided to the user. Disambiguation of named entities detected in a video content item being played is performed by identifying and accessing an information source directly associated with the video content item, and/or by analyzing visual content of a segment of the video content item. Selecting, proposing and/or providing an additional media content item is based on the information source and/or on the analyzing.
    Type: Grant
    Filed: March 25, 2018
    Date of Patent: May 12, 2020
    Assignee: Comigo Ltd.
    Inventors: Guy Geva, Menahem Lasser
  • Patent number: 10643621
    Abstract: An electronic device is provided. The electronic device includes a processor configured to perform automatic speech recognition (ASR) on a speech input by using a speech recognition model that is stored in a memory and a communication module configured to provide the speech input to a server and receive a speech instruction, which corresponds to the speech input, from the server. The electronic device may perform different operations according to a confidence score of a result of the ASR. Besides, it may be permissible to prepare other various embodiments speculated through the specification.
    Type: Grant
    Filed: September 11, 2018
    Date of Patent: May 5, 2020
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Seok Yeong Jung, Kyung Tae Kim
  • Patent number: 10622009
    Abstract: A system configured to improve double-talk detection. The system detects when double-talk is present in a voice conversation using two or more speaker models. The system extracts feature data from microphone audio data and compares the feature data to each speaker model. For example, the system may generate a first distance score indicating a likelihood that the feature data corresponds to a far-end speaker model and a second distance score indicating a likelihood that the feature data corresponds to a universal speaker model. The system may determine current system conditions based on the distance scores and may change settings to improve speech quality during the voice conversation. For example, during far-end single-talk the system may aggressively reduce an echo signal, whereas during near-end single-talk and double-talk the system may apply minimal echo cancellation to improve a quality of the local speech.
    Type: Grant
    Filed: September 10, 2018
    Date of Patent: April 14, 2020
    Assignee: Amazon Technologies, Inc.
    Inventors: Xianxian Zhang, Philip Ryan Hilmes, Trausti Thor Kristjansson
  • Patent number: 10621282
    Abstract: A computer-implemented method for providing agent assisted transcriptions of user utterances. A user utterance is received in response to a prompt provided to the user at a remote client device. An automatic transcription is generated from the utterance using a language model based upon an application or context, and presented to a human agent. The agent reviews the transcription and may replace at least a portion of the transcription with a corrected transcription. As the agent inputs the corrected transcription, accelerants are presented to the user comprising suggested texted to be inputted. The accelerants may be determined based upon an agent input, an application or context of the transcription, the portion of the transcription being replaced, or any combination thereof. In some cases, the user provides textual input, to which the agent transcribes an intent associated with the input with the aid of one or more accelerants.
    Type: Grant
    Filed: April 26, 2018
    Date of Patent: April 14, 2020
    Assignee: Interactions LLC
    Inventors: Ethan Selfridge, Michael Johnston, Robert Lifgren, James Dreher, John Leonard
  • Patent number: 10614121
    Abstract: Content from multiple different stations can be divided into segments based on time. Matched segments associated with each station can be identified by comparing content included in a first segment associated with a first station, to content included in a second segment associated with a second station. Syndicated content can be identified and tagged based, at least in part, on a relationship between sequences of matched segments on different stations. Various embodiments also include identifying main sequences associated with each station under consideration, removing some of the main sequences, and consolidating remaining main sequences based on various threshold criteria.
    Type: Grant
    Filed: May 26, 2015
    Date of Patent: April 7, 2020
    Assignee: IHEARTMEDIA MANAGEMENT SERVICES, INC.
    Inventors: Periklis Beltas, Philippe Generali, David C. Jellison, Jr.
  • Patent number: 10593329
    Abstract: A system of multi-modal transmission of packetized data in a voice activated data packet based computer network environment is provided. A natural language processor component can parse an input audio signal to identify a request and a trigger keyword. Based on the input audio signal, a direct action application programming interface can generate a first action data structure, and a content selector component can select a content item. An interface management component can identify first and second candidate interfaces, and respective resource utilization values. The interface management component can select, based on the resource utilization values, the first candidate interface to present the content item. The interface management component can provide the first action data structure to the client computing device for rendering as audio output, and can transmit the content item converted for a first modality to deliver the content item for rendering from the selected interface.
    Type: Grant
    Filed: July 18, 2018
    Date of Patent: March 17, 2020
    Assignee: Google LLC
    Inventors: Gaurav Bhaya, Robert Stets, Umesh Patil
  • Patent number: 10573300
    Abstract: The invention provides a method of automatic speech recognition. The method includes receiving a speech signal, dividing the speech signal into time windows, for each time window determining acoustic parameters of the speech signal within that window, and identifying phonological features from the acoustic parameters, such that a sequence of phonological features are generated for the speech signal, separating the sequence of phonological features into a sequence of zones, and comparing the sequences of zones to a lexical entry comprising a sequence of phonological segments to a stored lexicon to identify one or more words in the speech signal.
    Type: Grant
    Filed: August 22, 2018
    Date of Patent: February 25, 2020
    Assignee: Oxford University Innovation Limited
    Inventors: Aditi Lahiri, Henning Reetz, Philip Roberts
  • Patent number: 10535342
    Abstract: Techniques and systems are disclosed for context-dependent speech recognition. The techniques and systems described enable accurate recognition of speech by accessing sub-libraries associated with the context of the speech to be recognized. These techniques translate audible input into audio data at a smart device and determine context for the speech, such as location-based, temporal-based, recipient-based, and application based context. The smart device then accesses a context-dependent library to compare the audio data with phrase-associated translation data in one or more sub-libraries of the context-dependent library to determine a match. In this way, the techniques allow access to a large quantity of phrases while reducing incorrect matching of the audio data to translation data caused by organizing the phrases into context-dependent sub-libraries.
    Type: Grant
    Filed: April 10, 2017
    Date of Patent: January 14, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventor: Christian Liensberger
  • Patent number: 10535348
    Abstract: A system of multi-modal transmission of packetized data in a voice activated data packet based computer network environment is provided. A natural language processor component can parse an input audio signal to identify a request and a trigger keyword. Based on the input audio signal, a direct action application programming interface can generate a first action data structure, and a content selector component can select a content item. An interface management component can identify first and second candidate interfaces, and respective resource utilization values. The interface management component can select, based on the resource utilization values, the first candidate interface to present the content item. The interface management component can provide the first action data structure to the client computing device for rendering as audio output, and can transmit the content item converted for a first modality to deliver the content item for rendering from the selected interface.
    Type: Grant
    Filed: July 18, 2018
    Date of Patent: January 14, 2020
    Assignee: Google LLC
    Inventors: Gaurav Bhaya, Robert Stets
  • Patent number: 10529322
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for tagging during speech recognition. A word lattice that indicates probabilities for sequences of words in an utterance is obtained. A conditional probability transducer that indicates a frequency that sequences of both the words and semantic tags for the words appear is obtained. The word lattice and the conditional probability transducer are composed to construct a word lattice that indicates probabilities for sequences of both the words in the utterance and the semantic tags for the words. The word lattice that indicates probabilities for sequences of both the words in the utterance and the semantic tags for the words is used to generate a transcription that includes the words in the utterance and the semantic tags for the words.
    Type: Grant
    Filed: August 21, 2017
    Date of Patent: January 7, 2020
    Assignee: Google LLC
    Inventors: Petar Aleksic, Michael D. Riley, Pedro J. Moreno Mengibar, Leonid Velikovich
  • Patent number: 10518607
    Abstract: A sound of a second vehicle engine is detected. Upon predicting a pollution event by comparing the detected sound to a stored sound model, a countermeasure is actuated in a first vehicle.
    Type: Grant
    Filed: August 28, 2017
    Date of Patent: December 31, 2019
    Assignee: FORD GLOBAL TECHNOLOGIES, LLC
    Inventors: Howard E. Churchwell, II, Mahmoud Yousef Ghannam
  • Patent number: 10496758
    Abstract: According to one embodiment, According to one embodiment, a machine translation apparatus includes a circuitry and a memory. The circuitry is configured to input a sentence of a first language, to segment the sentence to obtain a plurality of phrases, to search a translation model for translation options of a second language of each of the plurality of phrases, and to select top N translation options with high probabilities for decoding. N is an integer equal to or larger than 1. Furthermore, the circuitry is configured to combine the top N translation options of the plurality of phases to obtain a plurality of translation hypotheses, to search user history phrase pairs for the translation hypotheses, and to increase a score of a translation hypothesis existing in the user history phrase pairs. The memory is configured to store the score of the translation hypothesis.
    Type: Grant
    Filed: August 31, 2017
    Date of Patent: December 3, 2019
    Assignee: Kabushiki Kaisha Toshiba
    Inventors: Zhengshan Xue, Dakun Zhang, Jichong Guo, Jie Hao
  • Patent number: 10496693
    Abstract: Systems and methods provide for classification and ranking of features for a hierarchical dataset. A hierarchical schema of features from the dataset is accessed. A hierarchical rank is assigned to each feature based on its schema level in the hierarchical schema. Additionally, a semantic rank is assigned to each feature using a semantic model having ranked semantic contexts. The semantic rank of a feature is assigned by identifying a semantic context of the feature and assigning the rank of the semantic context as the semantic rank of the feature. A rank is computed for each feature as a function of its hierarchical rank and semantic rank.
    Type: Grant
    Filed: May 31, 2016
    Date of Patent: December 3, 2019
    Assignee: ADOBE INC.
    Inventors: Shiladitya Bose, Wei Zhang, Arvind Heda