Probability Patents (Class 704/240)
  • Patent number: 10714122
    Abstract: Speech or non-speech detection techniques are discussed and include updating a speech pattern model using probability scores from an acoustic model to generate a score for each state of the speech pattern model, such that the speech pattern model includes a first non-speech state having multiple self loops each associated with a non-speech probability score of the probability scores, a plurality of speech states following the first non-speech state, and a second non-speech state following the speech states, and detecting speech based on a comparison of a score of the first non-speech state and a score of the last speech state of the multiple speech states.
    Type: Grant
    Filed: June 6, 2018
    Date of Patent: July 14, 2020
    Assignee: Intel Corporation
    Inventors: Maciej Muchlinski, Tobias Bocklet
  • Patent number: 10714080
    Abstract: A weighted finite-state transducer (WFST) decoding system is provided. The WFST decoding system includes a memory that stores WFST data and a WFST decoder including a data fetch logic. The WFST data has a structure including states, and arcs connecting the states with directivity. The WFST data is compressed in the memory. The WFST data includes body data, and header data including state information for each states that is aligned discontinuously. The body data includes arc information of the arcs that is aligned continuously. The state information includes an arc index of the arcs, a number of the arcs, and compression information of the arcs, and the data fetch logic de-compresses the WFST data using the compression information, and retrieves the WFST data from the memory.
    Type: Grant
    Filed: September 8, 2017
    Date of Patent: July 14, 2020
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Jae Sung Yoon, Jun Seok Park
  • Patent number: 10713519
    Abstract: The present invention is directed towards providing automated workflows for the identification of a reading order from text segments extracted from a document. Ordering the text segments is based on trained natural language models. In some embodiments, the workflows are enabled to perform a method for identifying a sequence associated with a portable document. The methods includes iteratively generating a probabilistic language model, receiving the portable document, and selectively extracting features (such as but not limited to text segments) from the document. The method may generate pairs of features (or feature pair from the extracted features). The method may further generate a score for each of the pairs based on the probabilistic language model and determine an order to features based on the scores. The method may provide the extracted features in the determined order.
    Type: Grant
    Filed: June 22, 2017
    Date of Patent: July 14, 2020
    Assignee: ADOBE INC.
    Inventors: Trung Huu Bui, Hung Hai Bui, Shawn Alan Gaither, Walter Wei-Tuh Chang, Michael Frank Kraley, Pranjal Daga
  • Patent number: 10713289
    Abstract: Systems, methods, and devices for performing interactive question answering using data source credibility and conversation entropy are disclosed. A speech-controlled device captures audio including a spoken question, and sends audio data corresponding thereto to a server(s). The server(s) performs speech processing on the audio data, and determines various stored data that can be used to determine an answer to the question. The server(s) determines which stored data to use based on the credibility of the source from which the stored data was received. The server(s) may also determine a number of user interactions needed to obtain data in order to fully answer the question and may select a question for a dialog soliciting further data based on the number of user interactions.
    Type: Grant
    Filed: March 31, 2017
    Date of Patent: July 14, 2020
    Assignee: Amazon Technologies, Inc.
    Inventors: Nina Mishra, Yonatan Naamad
  • Patent number: 10706852
    Abstract: The described technology provides arbitration between speech recognition results generated by different automatic speech recognition (ASR) engines, such as ASR engines trained according to different language or acoustic models. The system includes an arbitrator that selects between a first speech recognition result representing an acoustic utterance as transcribed by a first ASR engine and a second speech recognition result representing the acoustic utterance as transcribed by a second ASR engine. This selection is based on a set of confidence features that is initially used by the first ASR engine or the second ASR engine to generate the first and second speech recognition results.
    Type: Grant
    Filed: November 13, 2015
    Date of Patent: July 7, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Kshitiz Kumar, Hosam Khalil, Yifan Gong, Ziad Al-Bawab, Chaojun Liu
  • Patent number: 10699712
    Abstract: An information processing method and an electronic device are provided. The method includes an electronic device obtaining an input information through a second collection manner when the electronic device is in a speech collection state for obtaining speech information through a first collection manner, and determining a logic boundary position in relation to a first speech information in accordance with the input information, the first speech information is obtained by the electronic device through the first collection manner which is different from the second collection manner. An electronic device corresponding thereto is also disclosed.
    Type: Grant
    Filed: March 4, 2015
    Date of Patent: June 30, 2020
    Assignee: LENOVO (BEIJING) CO., LTD.
    Inventors: Haisheng Dai, Zhepeng Wang
  • Patent number: 10699696
    Abstract: The present disclosure provides a method and apparatus for correcting a speech recognition error based on artificial intelligence, and a storage medium, wherein the method comprises: obtaining a second speech recognition result of a second speech query input by the user; performing error-correcting intention recognition according to the second speech recognition result; extracting error-correcting information from the second speech recognition result when it is determined that the user has an error-correcting intention; screening error-correcting resources according to the error-correcting information, and using a selected best-matched error-correcting resource to perform error correction for the first speech recognition result, the first speech recognition result being a speech recognition result of a first speech query which is input before the second speech query.
    Type: Grant
    Filed: May 22, 2018
    Date of Patent: June 30, 2020
    Assignee: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY CO., LTD.
    Inventors: Hehan Li, Wensong He
  • Patent number: 10649850
    Abstract: Techniques and systems for storing and retrieving data storage devices of a data storage system are disclosed. In some embodiments, inventory holders are used to store data storage devices used by a data storage system. When data is to be transacted with the data storage devices, mobile drive units locate appropriate inventory holders and transport them to a device reading station, where an appropriate device retrieval unit transacts the data. In some embodiments, each inventory holder includes a heterogenous mix of data storage device types, the layout of which may be calculated according to the specific mix allocated to a given inventory holder. After the data has been transacted, the data storage devices are returned to the appropriate inventory holders, and the inventory holders are placed by the mobile drive units in locations where they may be accessed in response to further data transactions.
    Type: Grant
    Filed: June 29, 2015
    Date of Patent: May 12, 2020
    Assignee: Amazon Technologies, Inc.
    Inventors: James Raymond Allard, Paul David Franklin, Samuel Rubin Barrett, Jeremiah Brazeau, Jeffrey Allen Dzado, James Caleb Kirschner, David Levy, Brent James Lutz, Andrew Brendan Tinka, Colin Laird Lazier
  • Patent number: 10652592
    Abstract: Methods and systems are disclosed for enriching a viewing experience of a user watching video content on a screen of a client terminal by increasing the relevance of additional media content proposed or provided to the user. Disambiguation of named entities detected in a video content item being played is performed by identifying and accessing an information source directly associated with the video content item, and/or by analyzing visual content of a segment of the video content item. Selecting, proposing and/or providing an additional media content item is based on the information source and/or on the analyzing.
    Type: Grant
    Filed: March 25, 2018
    Date of Patent: May 12, 2020
    Assignee: Comigo Ltd.
    Inventors: Guy Geva, Menahem Lasser
  • Patent number: 10643621
    Abstract: An electronic device is provided. The electronic device includes a processor configured to perform automatic speech recognition (ASR) on a speech input by using a speech recognition model that is stored in a memory and a communication module configured to provide the speech input to a server and receive a speech instruction, which corresponds to the speech input, from the server. The electronic device may perform different operations according to a confidence score of a result of the ASR. Besides, it may be permissible to prepare other various embodiments speculated through the specification.
    Type: Grant
    Filed: September 11, 2018
    Date of Patent: May 5, 2020
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Seok Yeong Jung, Kyung Tae Kim
  • Patent number: 10621282
    Abstract: A computer-implemented method for providing agent assisted transcriptions of user utterances. A user utterance is received in response to a prompt provided to the user at a remote client device. An automatic transcription is generated from the utterance using a language model based upon an application or context, and presented to a human agent. The agent reviews the transcription and may replace at least a portion of the transcription with a corrected transcription. As the agent inputs the corrected transcription, accelerants are presented to the user comprising suggested texted to be inputted. The accelerants may be determined based upon an agent input, an application or context of the transcription, the portion of the transcription being replaced, or any combination thereof. In some cases, the user provides textual input, to which the agent transcribes an intent associated with the input with the aid of one or more accelerants.
    Type: Grant
    Filed: April 26, 2018
    Date of Patent: April 14, 2020
    Assignee: Interactions LLC
    Inventors: Ethan Selfridge, Michael Johnston, Robert Lifgren, James Dreher, John Leonard
  • Patent number: 10622009
    Abstract: A system configured to improve double-talk detection. The system detects when double-talk is present in a voice conversation using two or more speaker models. The system extracts feature data from microphone audio data and compares the feature data to each speaker model. For example, the system may generate a first distance score indicating a likelihood that the feature data corresponds to a far-end speaker model and a second distance score indicating a likelihood that the feature data corresponds to a universal speaker model. The system may determine current system conditions based on the distance scores and may change settings to improve speech quality during the voice conversation. For example, during far-end single-talk the system may aggressively reduce an echo signal, whereas during near-end single-talk and double-talk the system may apply minimal echo cancellation to improve a quality of the local speech.
    Type: Grant
    Filed: September 10, 2018
    Date of Patent: April 14, 2020
    Assignee: Amazon Technologies, Inc.
    Inventors: Xianxian Zhang, Philip Ryan Hilmes, Trausti Thor Kristjansson
  • Patent number: 10614121
    Abstract: Content from multiple different stations can be divided into segments based on time. Matched segments associated with each station can be identified by comparing content included in a first segment associated with a first station, to content included in a second segment associated with a second station. Syndicated content can be identified and tagged based, at least in part, on a relationship between sequences of matched segments on different stations. Various embodiments also include identifying main sequences associated with each station under consideration, removing some of the main sequences, and consolidating remaining main sequences based on various threshold criteria.
    Type: Grant
    Filed: May 26, 2015
    Date of Patent: April 7, 2020
    Assignee: IHEARTMEDIA MANAGEMENT SERVICES, INC.
    Inventors: Periklis Beltas, Philippe Generali, David C. Jellison, Jr.
  • Patent number: 10593329
    Abstract: A system of multi-modal transmission of packetized data in a voice activated data packet based computer network environment is provided. A natural language processor component can parse an input audio signal to identify a request and a trigger keyword. Based on the input audio signal, a direct action application programming interface can generate a first action data structure, and a content selector component can select a content item. An interface management component can identify first and second candidate interfaces, and respective resource utilization values. The interface management component can select, based on the resource utilization values, the first candidate interface to present the content item. The interface management component can provide the first action data structure to the client computing device for rendering as audio output, and can transmit the content item converted for a first modality to deliver the content item for rendering from the selected interface.
    Type: Grant
    Filed: July 18, 2018
    Date of Patent: March 17, 2020
    Assignee: Google LLC
    Inventors: Gaurav Bhaya, Robert Stets, Umesh Patil
  • Patent number: 10573300
    Abstract: The invention provides a method of automatic speech recognition. The method includes receiving a speech signal, dividing the speech signal into time windows, for each time window determining acoustic parameters of the speech signal within that window, and identifying phonological features from the acoustic parameters, such that a sequence of phonological features are generated for the speech signal, separating the sequence of phonological features into a sequence of zones, and comparing the sequences of zones to a lexical entry comprising a sequence of phonological segments to a stored lexicon to identify one or more words in the speech signal.
    Type: Grant
    Filed: August 22, 2018
    Date of Patent: February 25, 2020
    Assignee: Oxford University Innovation Limited
    Inventors: Aditi Lahiri, Henning Reetz, Philip Roberts
  • Patent number: 10535348
    Abstract: A system of multi-modal transmission of packetized data in a voice activated data packet based computer network environment is provided. A natural language processor component can parse an input audio signal to identify a request and a trigger keyword. Based on the input audio signal, a direct action application programming interface can generate a first action data structure, and a content selector component can select a content item. An interface management component can identify first and second candidate interfaces, and respective resource utilization values. The interface management component can select, based on the resource utilization values, the first candidate interface to present the content item. The interface management component can provide the first action data structure to the client computing device for rendering as audio output, and can transmit the content item converted for a first modality to deliver the content item for rendering from the selected interface.
    Type: Grant
    Filed: July 18, 2018
    Date of Patent: January 14, 2020
    Assignee: Google LLC
    Inventors: Gaurav Bhaya, Robert Stets
  • Patent number: 10535342
    Abstract: Techniques and systems are disclosed for context-dependent speech recognition. The techniques and systems described enable accurate recognition of speech by accessing sub-libraries associated with the context of the speech to be recognized. These techniques translate audible input into audio data at a smart device and determine context for the speech, such as location-based, temporal-based, recipient-based, and application based context. The smart device then accesses a context-dependent library to compare the audio data with phrase-associated translation data in one or more sub-libraries of the context-dependent library to determine a match. In this way, the techniques allow access to a large quantity of phrases while reducing incorrect matching of the audio data to translation data caused by organizing the phrases into context-dependent sub-libraries.
    Type: Grant
    Filed: April 10, 2017
    Date of Patent: January 14, 2020
    Assignee: Microsoft Technology Licensing, LLC
    Inventor: Christian Liensberger
  • Patent number: 10529322
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for tagging during speech recognition. A word lattice that indicates probabilities for sequences of words in an utterance is obtained. A conditional probability transducer that indicates a frequency that sequences of both the words and semantic tags for the words appear is obtained. The word lattice and the conditional probability transducer are composed to construct a word lattice that indicates probabilities for sequences of both the words in the utterance and the semantic tags for the words. The word lattice that indicates probabilities for sequences of both the words in the utterance and the semantic tags for the words is used to generate a transcription that includes the words in the utterance and the semantic tags for the words.
    Type: Grant
    Filed: August 21, 2017
    Date of Patent: January 7, 2020
    Assignee: Google LLC
    Inventors: Petar Aleksic, Michael D. Riley, Pedro J. Moreno Mengibar, Leonid Velikovich
  • Patent number: 10518607
    Abstract: A sound of a second vehicle engine is detected. Upon predicting a pollution event by comparing the detected sound to a stored sound model, a countermeasure is actuated in a first vehicle.
    Type: Grant
    Filed: August 28, 2017
    Date of Patent: December 31, 2019
    Assignee: FORD GLOBAL TECHNOLOGIES, LLC
    Inventors: Howard E. Churchwell, II, Mahmoud Yousef Ghannam
  • Patent number: 10496758
    Abstract: According to one embodiment, According to one embodiment, a machine translation apparatus includes a circuitry and a memory. The circuitry is configured to input a sentence of a first language, to segment the sentence to obtain a plurality of phrases, to search a translation model for translation options of a second language of each of the plurality of phrases, and to select top N translation options with high probabilities for decoding. N is an integer equal to or larger than 1. Furthermore, the circuitry is configured to combine the top N translation options of the plurality of phases to obtain a plurality of translation hypotheses, to search user history phrase pairs for the translation hypotheses, and to increase a score of a translation hypothesis existing in the user history phrase pairs. The memory is configured to store the score of the translation hypothesis.
    Type: Grant
    Filed: August 31, 2017
    Date of Patent: December 3, 2019
    Assignee: Kabushiki Kaisha Toshiba
    Inventors: Zhengshan Xue, Dakun Zhang, Jichong Guo, Jie Hao
  • Patent number: 10496693
    Abstract: Systems and methods provide for classification and ranking of features for a hierarchical dataset. A hierarchical schema of features from the dataset is accessed. A hierarchical rank is assigned to each feature based on its schema level in the hierarchical schema. Additionally, a semantic rank is assigned to each feature using a semantic model having ranked semantic contexts. The semantic rank of a feature is assigned by identifying a semantic context of the feature and assigning the rank of the semantic context as the semantic rank of the feature. A rank is computed for each feature as a function of its hierarchical rank and semantic rank.
    Type: Grant
    Filed: May 31, 2016
    Date of Patent: December 3, 2019
    Assignee: ADOBE INC.
    Inventors: Shiladitya Bose, Wei Zhang, Arvind Heda
  • Patent number: 10475442
    Abstract: A method and a device for recognition, and a method and a device for constructing a recognition model are disclosed. A device for constructing a recognition model includes a training data inputter configured to receive additional training data, a model learner configured to train a first recognition model constructed based on basic training data to learn the additional training data, and a model constructor configured to construct a final recognition model by integrating the first recognition model with a second recognition model generated by the training of the first recognition model.
    Type: Grant
    Filed: October 24, 2016
    Date of Patent: November 12, 2019
    Assignee: Samsung Electronics Co., Ltd.
    Inventor: Ho Shik Lee
  • Patent number: 10453460
    Abstract: Systems and methods for determining that artificial commands, in excess of a threshold value, are detected by multiple voice activated electronic devices is described herein. In some embodiments, numerous voice activated electronic devices may send audio data representing a phrase to a backend system at a substantially same time. Text data representing the phrase, and counts for instances of that text data, may be generated. If the number of counts exceeds a predefined threshold, the backend system may cause any remaining response generation functionality that particular command that is in excess of the predefined threshold to be stopped, and those devices returned to a sleep state. In some embodiments, a sound profile unique to the phrase that caused the excess of the predefined threshold may be generated such that future instances of the same phrase may be recognized prior to text data being generated, conserving the backend system's resources.
    Type: Grant
    Filed: March 30, 2016
    Date of Patent: October 22, 2019
    Assignee: Amazon Technologies, Inc.
    Inventors: Colin Wills Wightman, Naresh Narayanan, Daniel Robert Rashid
  • Patent number: 10417329
    Abstract: A dialog act estimation method includes acquiring learning data including a first sentence to be estimated in the form of text data of a first uttered sentence uttered at a first time point, a second sentence which is text data of a second uttered sentence uttered, at a time point before the first time point, successively after the first uttered sentence, act information indicating an act associated to the first sentence, property information indicating a property information associated to the first sentence, and dialog act information indicating a dialog act in the form of a combination of an act and a property associated to the first sentence, making a particular model learn three or more tasks at the same time using the learning data, and storing a result of the learning as learning result information in a memory.
    Type: Grant
    Filed: August 1, 2017
    Date of Patent: September 17, 2019
    Assignee: PANASONIC INTELLECTUAL PROPERTY MANAGEMENT CO., LTD.
    Inventors: Takashi Ushio, Hongjie Shi, Mitsuru Endo, Katsuyoshi Yamagami
  • Patent number: 10417328
    Abstract: Methods and processes evaluate a quality score of a text. The text includes a plurality of words. The methods compute first probability characteristics of groups of words in a reference text which is known to be a high-quality text. The methods also compute second probability characteristics of groups of words in a text to be scored. The methods also compute the quality score based on a difference between the first probability characteristics and the second probability characteristics.
    Type: Grant
    Filed: January 5, 2018
    Date of Patent: September 17, 2019
    Assignee: Searchmetrics GmbH
    Inventors: Ahmet Anil Pala, Alexander Kagoshima, Marcus Tober
  • Patent number: 10409909
    Abstract: Techniques are disclosed for building a dictionary of words from combinations of symbols generated based on input data. A neuro-linguistic behavior recognition system includes a neuro-linguistic module that generates a linguistic model that describes data input from a source (e.g., video data, SCADA data, etc.). To generate words for the linguistic model, a lexical analyzer component in the neuro-linguistic module receives a stream of symbols, each symbol generated based on an ordered stream of normalized vectors generated from input data. The lexical analyzer component determines words from combinations of the symbols based on a hierarchical learning model having one or more levels. Each level indicates a length of the words to be identified at that level. Statistics are evaluated for the words identified at each level. The lexical analyzer component identifies one or more of the words having statistical significance.
    Type: Grant
    Filed: December 12, 2014
    Date of Patent: September 10, 2019
    Assignee: Omni AI, Inc.
    Inventors: Gang Xu, Ming-Jung Seow, Tao Yang, Wesley Kenneth Cobb
  • Patent number: 10409910
    Abstract: Techniques are disclosed for generating a syntax for a neuro-linguistic model of input data obtained from one or more sources. A stream of words of a dictionary built from a sequence of symbols are received. The symbols are generated from an ordered stream of normalized vectors generated from input data. Statistics for combinations of words co-occurring in the stream are evaluated. The statistics includes a frequency upon which the combinations of words co-occur. A model of combinations of words based on the evaluated statistics is updated. The model identifies statistically relevant words. A connected graph is generated. Each node in the connected graph represents one of the words in the stream. Edges connecting the nodes represent a probabilistic relationship between words in the stream. Phrases are identified based on the connected graph.
    Type: Grant
    Filed: December 12, 2014
    Date of Patent: September 10, 2019
    Assignee: Omni AI, Inc.
    Inventors: Ming-Jung Seow, Gang Xu, Tao Yang, Wesley Kenneth Cobb
  • Patent number: 10375224
    Abstract: A mobile device providing integrated management of message information and service provision through artificial intelligence is disclosed. The mobile device includes an integrated message management unit comprising a message monitoring unit configured to monitor voice call information and text message information in association with the voice call management part and the text message management part, a message information managing unit configured to generate integrated message information, which is to be provided to a user, based on the voice call information and the text message information, an interface managing unit configured to generate an integrated message management user interface displaying the integrated message information, and an artificial intelligence agent analyzing the voice call information and the text message information and providing a service associated with at least one additional function in association with the additional function process part based on the analyzed result.
    Type: Grant
    Filed: March 28, 2018
    Date of Patent: August 6, 2019
    Assignee: NHN Entertainment Corporation
    Inventor: Dong Wook Kim
  • Patent number: 10372737
    Abstract: According to one embodiment, a method, computer system, and computer program product for retraining a classifier-based automatic dialog system with recorded user interactions is provided. The present invention may include receiving recorded interactions, where the interactions are between a user and an automatic dialog system; determining, based on the recorded interactions, whether to pair a given input with one or more classes; pairing inputs with one or more classes; assessing the reliability of the paired inputs and classes; determining whether the reliable paired inputs and classes can be consistently mapped; and merging all consistently mapped reliable pairs with an initial training set.
    Type: Grant
    Filed: November 16, 2017
    Date of Patent: August 6, 2019
    Assignee: International Business Machines Corporation
    Inventors: Allen Ginsberg, Edward G. Katz, Alexander C. Tonetti
  • Patent number: 10366159
    Abstract: A system for identifying address components includes an interface and a processor. The interface is to receive an address for parsing. The processor is to determine a matching model of a set of models based at least in part on a matching probability for each model for a tokenized address, which is based on the address for parsing, and associate each component of the tokenized address with an identifier based at least in part on the matching model, wherein each component of the set of components is associated with an identifier, and wherein probabilities of each component of the set of components are determined using training addresses.
    Type: Grant
    Filed: October 14, 2016
    Date of Patent: July 30, 2019
    Assignee: Workday, Inc.
    Inventors: Parag Avinash Namjoshi, Shuangshuang Jiang, Mohammad Sabah
  • Patent number: 10360898
    Abstract: A system and method are presented for predicting speech recognition performance using accuracy scores in speech recognition systems within the speech analytics field. A keyword set is selected. Figure of Merit (FOM) is computed for the keyword set. Relevant features that describe the word individually and in relation to other words in the language are computed. A mapping from these features to FOM is learned. This mapping can be generalized via a suitable machine learning algorithm and be used to predict FOM for a new keyword. In at least embodiment, the predicted FOM may be used to adjust internals of speech recognition engine to achieve a consistent behavior for all inputs for various settings of confidence values.
    Type: Grant
    Filed: June 5, 2018
    Date of Patent: July 23, 2019
    Inventors: Aravind Ganapathiraju, Yingyi Tan, Felix Immanuel Wyss, Scott Allen Randal
  • Patent number: 10360904
    Abstract: Methods and apparatus for performing speech recognition using a garbage model. The method comprises receiving audio comprising speech and processing at least some of the speech using a garbage model to produce a garbage speech recognition result. The garbage model includes a plurality of sub-words, each of which corresponds to a possible combination of phonemes in a particular language.
    Type: Grant
    Filed: May 9, 2014
    Date of Patent: July 23, 2019
    Assignee: Nuance Communications, Inc.
    Inventors: Cosmin Popovici, Kenneth W. D. Smith, Petrus C. Cools
  • Patent number: 10354647
    Abstract: Implementations of the present disclosure include actions of providing first text for display on a computing device of a user, the first text being provided from a first speech recognition engine based on first speech received from the computing device, and being displayed as a search query, receiving a speech correction indication from the computing device, the speech correction indication indicating a portion of the first text that is to be corrected, receiving second speech from the computing device, receiving second text from a second speech recognition engine based on the second speech, the second speech recognition engine being different from the first speech recognition engine, replacing the portion of the first text with the second text to provide a combined text, and providing the combined text for display on the computing device as a revised search query.
    Type: Grant
    Filed: April 28, 2016
    Date of Patent: July 16, 2019
    Assignee: Google LLC
    Inventors: Dhruv Bakshi, Zaheed Sabur, Tilke Mary Judd, Nicholas G. Fey
  • Patent number: 10347245
    Abstract: Either or both of voice speaker identification or utterance classification such as by age, gender, accent, mood, and prosody characterize speech utterances in a system that performs automatic speech recognition (ASR) and natural language processing (NLP). The characterization conditions NLP, either through application to interpretation hypotheses or to specific grammar rules. The characterization also conditions language models of ASR. Conditioning may comprise enablement and may comprise reweighting of hypotheses.
    Type: Grant
    Filed: January 20, 2017
    Date of Patent: July 9, 2019
    Assignee: SOUNDHOUND, INC.
    Inventor: Karl Stahl
  • Patent number: 10332016
    Abstract: The invention concerns a method to compare two data obtained from a sensor or interface, carried out by processing means of a processing unit, the method comprising the computing of a similarity function between two feature vectors of the data to be compared, characterized in that each feature vector of a datum is modelled as the summation of Gaussian variables, said variables comprising: a mean of a class to which the vector belongs, an intrinsic deviation, and an observation noise of the vector, each feature vector being associated with a quality vector comprising information on the observation noise of the feature vector, and in that the similarity function is computed from the feature vectors and associated quality vectors.
    Type: Grant
    Filed: November 3, 2015
    Date of Patent: June 25, 2019
    Assignee: IDEMIA IDENTITY & SECURITY
    Inventors: Julien Bohne, Stephane Gentric
  • Patent number: 10318009
    Abstract: For the purpose of enhancing accessibility of a user with respect to various applications, the disclosed technique provides a method for controlling a user-interface that provides an instruction to an application through a user-operation which is performed on a display provided on a device. The method includes a process performed by the device. The process includes: acquiring information displayed by the display; extracting at least one feature existing on the acquired information; receiving an action of a user; searching a database to identify a predetermined operation, which corresponds to the received action and the extracted at least one feature; and providing an instruction to the application through applying the identified predetermined operation, not the received action, to the user-interface.
    Type: Grant
    Filed: April 13, 2017
    Date of Patent: June 11, 2019
    Assignee: HI CORPORATION
    Inventors: Tomonobu Aoyama, Tatsuo Sasaki, Seiichi Kataoka
  • Patent number: 10319209
    Abstract: A system and method of motion analysis, fall detection, and fall prediction using machine learning and classifiers. A wearable motion sensor for collecting and transmitting motion data for use in a fall prediction model using features and parameters to classify the motion data and notify when a fall is emergent. Using machine learning, the fall prediction model can be created, implemented, evaluated, and it can evolve over time with additional data. The system and method can use individual data or pool data from various individuals for use in fall prediction.
    Type: Grant
    Filed: June 5, 2017
    Date of Patent: June 11, 2019
    Inventor: John Carlton-Foss
  • Patent number: 10305828
    Abstract: A computing device is described that includes at least one processor and a memory including instructions that when executed cause the at least one processor to output, for display, a graphical keyboard comprising a plurality of keys, and determine, based on an indication of a selection of one or more keys from the plurality of keys, text of an electronic communication. The instructions, when executed, further cause the at least one processor to identify, based at least in part on the text, a searchable entity or trigger phrase, generate, based on the searchable entity or trigger phrase, a search query, and output, for display, within the graphical keyboard, a graphical indication to indicate that the computing device generated the search query.
    Type: Grant
    Filed: April 20, 2016
    Date of Patent: May 28, 2019
    Assignee: Google LLC
    Inventors: Jing Cao, Alexa Greenberg, Abhanshu Sharma, Yanchao Su, Nicholas Kong, Muhammad Mohsin, Jacek Jurewicz, Wei Huang, Matthew Sharifi, Benjamin Sidhom
  • Patent number: 10296633
    Abstract: A system includes a storage system configured to store data objects as a plurality of shards according to a redundancy encoding technique at a plurality of availability zones. The system further includes a redundancy reduction manager configured to perform a shard spreading process and a shard pruning process. The shard spreading process involves identifying an underutilized availability zone for a particular data object and moving at least one shard of the particular data object from another availability zone to the underutilized availability zone. The shard pruning process involves identifying a pruning candidate availability zone and deleting a shard of a particular data object at the pruning candidate availability zone in response to determining that deleting the shard would not violate a durability model for the particular data object.
    Type: Grant
    Filed: March 23, 2016
    Date of Patent: May 21, 2019
    Assignee: Amazon Technologies, Inc.
    Inventor: Jonathan Robert Collins
  • Patent number: 10283168
    Abstract: Provided are an audio file re-recording method and device, and a storage medium. The method includes: determining first time, the first time being start time of a recorded clip to be re-recorded in an audio file; playing a first recorded clip that has been recorded, the first recorded clip using the first time as end time in the audio file; upon arrival of the first time, collecting first voice data of a user to obtain a second recorded clip; and processing the first recorded clip and the second recorded clip to obtain a re-recorded audio file.
    Type: Grant
    Filed: May 1, 2018
    Date of Patent: May 7, 2019
    Assignee: GUANGZHOU KUGOU COMPUTER TECHNOLOGY CO., LTD.
    Inventor: Suiyu Feng
  • Patent number: 10242668
    Abstract: An apparatus includes a language model group identifier configured to identify a language model group based on determined characteristic data of a user, and a language model generator configured to generate a user-based language model by interpolating a general language model for speech recognition based on the identified language model group.
    Type: Grant
    Filed: August 3, 2016
    Date of Patent: March 26, 2019
    Assignee: Samsung Electronics Co., Ltd.
    Inventor: Min Young Mun
  • Patent number: 10235996
    Abstract: A system and method for providing a voice assistant including receiving, at a first device, a first audio input from a user requesting a first action; performing automatic speech recognition on the first audio input; obtaining a context of user; performing natural language understanding based on the speech recognition of the first audio input; and taking the first action based on the context of the user and the natural language understanding.
    Type: Grant
    Filed: September 30, 2015
    Date of Patent: March 19, 2019
    Assignee: Xbrain, Inc.
    Inventors: Gregory Renard, Mathias Herbaux
  • Patent number: 10199040
    Abstract: A method of automatic speech recognition, the method comprising the steps of receiving a speech signal, dividing the speech signal into time windows, for each time window, determining acoustic parameters of the speech signal within that window, and identifying speech features from the acoustic parameters, such that a sequence of speech features are generated for the speech signal, separating the sequence of speech features into a sequence of phonological segments, and comparing the sequential phonological segments to a stored lexicon to identify one or more words in the speech signal.
    Type: Grant
    Filed: December 17, 2014
    Date of Patent: February 5, 2019
    Assignee: OXFORD UNIVERSITY INNOVATION LIMITED
    Inventors: Aditi Lahiri, Henning Reetz, Philip Roberts
  • Patent number: 10152507
    Abstract: Methods and systems are provided for finding a target document in spoken language processing. One of the methods includes calculating a score of each document in a document set, in response to a receipt of first n words of output of an automatic speech recognition (ASR) system, n being equal or greater than zero. The method further includes reading a prior distribution of each document in the document set from a memory device, and updating, for each document in the document set, the score, using the prior distribution, and a weight for interpolation, the weight for interpolation being set based on a confidence score of output of the ASR system. The method additionally includes finding a target document among the document set, based on the updated score of each document.
    Type: Grant
    Filed: March 22, 2016
    Date of Patent: December 11, 2018
    Assignee: International Business Machines Corporation
    Inventors: Gakuto Kurata, Masayuki A. Suzuki, Ryuki Tachibana
  • Patent number: 10140581
    Abstract: Features are disclosed for generating models, such as conditional random field (“CRF”) models, that consume less storage space and/or transmission bandwidth than conventional models. In some embodiments, the generated CRF models are composed of fewer or alternate components in comparison with conventional CRF models. For example, a system generating such CRF models may forgo the use of large dictionaries or other cross-reference lists that map information extracted from input (e.g., “features”) to model parameters; reduce in weight (or exclude altogether) certain model parameters that may not have a significant effect on model accuracy; and/or reduce the numerical precision of model parameters.
    Type: Grant
    Filed: December 22, 2014
    Date of Patent: November 27, 2018
    Assignee: Amazon Technologies, Inc.
    Inventors: Imre Attila Kiss, Wei Chen, Anjishnu Kumar
  • Patent number: 10127224
    Abstract: Technologies for extensible, context-aware natural language interactions include a computing device having a number of context source modules. Context source modules may be developed or installed after deployment of the computing device to a user. Each context source module includes a context capture module, a language model, one or more database query mappings, and may include one or more user interface element mappings. The context capture module interprets, generates, and stores context data. A virtual personal assistant (VPA) of the computing device indexes the language models and generates a semantic representation of a user request that associates each word of the request to a language model. The VPA translates the user request into a database query, and may generate a user interface element for the request. The VPA may execute locally on the computing device or remotely on a cloud server. Other embodiments are described and claimed.
    Type: Grant
    Filed: August 30, 2013
    Date of Patent: November 13, 2018
    Assignee: Intel Corporation
    Inventor: William C. Deleeuw
  • Patent number: 10129135
    Abstract: A flow of packets is communicated through a data center. The data center includes multiple racks, where each rack includes multiple network devices. A group of packets of the flow is received onto an integrated circuit located in a first network device. The integrated circuit includes a neural network. The neural network analyzes the group of packets and in response outputs a neural network output value. The neural network output value is used to determine how the packets of the flow are to be output from a second network device. In one example, each packet of the flow output by the first network device is output along with a tag. The tag is indicative of the neural network output value. The second device uses the tag to determine which output port located on the second device is to be used to output each of the packets.
    Type: Grant
    Filed: September 1, 2015
    Date of Patent: November 13, 2018
    Assignee: Netronome Systems, Inc.
    Inventor: Nicolaas J. Viljoen
  • Patent number: 10114668
    Abstract: Techniques are described for managing execution of programs, including using excess program execution capacity of one or more computing systems. For example, a private pool of excess computing capacity may be maintained for a user based on unused dedicated program execution capacity allocated for that user, with the private pool of excess capacity being available for priority use by that user. Such private excess capacity pools may further in some embodiments be provided in addition to a general, non-private excess computing capacity pool that is available for use by multiple users, optionally including users who are associated with the private excess capacity pools. In some such situations, excess computing capacity may be made available to execute programs on a temporary basis, such that the programs executing using the excess capacity may be terminated at any time if other preferred use for the excess capacity arises.
    Type: Grant
    Filed: December 29, 2014
    Date of Patent: October 30, 2018
    Assignee: Amazon Technologies, Inc.
    Inventors: Eric Jason Brandwine, James Alfred Gordon Greenfield
  • Patent number: 10115394
    Abstract: An object is to provide a technique which can provide a highly valid recognition result while preventing unnecessary processing. A voice recognition device includes first to third voice recognition units, and a control unit. When it is decided based on recognition results obtained by the first and second voice recognition units to cause the third voice recognition unit to recognize an input voice, the control unit causes the third voice recognition unit to recognize the input voice by using a dictionary including a candidate character string obtained by at least one of the first and second voice recognition units.
    Type: Grant
    Filed: July 8, 2014
    Date of Patent: October 30, 2018
    Assignee: MITSUBISHI ELECTRIC CORPORATION
    Inventors: Naoya Sugitani, Yohei Okato, Michihiro Yamazaki
  • Patent number: 10078673
    Abstract: A computing device is described that includes at least one processor and a memory including instructions that when executed cause the at least one processor to output for display a graphical keyboard comprising a plurality of keys, determine, based on an indication of a selection of one or more keys from the plurality of keys, inputted, determine, based on the inputted text, an information category associated with the inputted text, determine, based on the information category, a graphical symbol associated with the information category, and output, for display, the graphical symbol in a suggestion region of the graphical keyboard.
    Type: Grant
    Filed: April 20, 2016
    Date of Patent: September 18, 2018
    Assignee: Google LLC
    Inventors: Jens Nagel, Alexa Greenberg, Christian Paul Charsagua