Probability Patents (Class 704/240)
-
Patent number: 10957308Abstract: Provided is a method and device to personalize a speech recognition model, the device that personalizes a speech recognition model by identifying a language group corresponding to a user, and generating a personalized speech recognition model by applying a group scale matrix corresponding to the identified language group to at least a layer of a speech recognition model.Type: GrantFiled: August 31, 2018Date of Patent: March 23, 2021Assignee: Samsung Electronics Co., Ltd.Inventors: Ki Soo Kwon, Inchul Song, YoungSang Choi
-
Patent number: 10943583Abstract: A system to perform automatic speech recognition (ASR) using a dynamic language model. Portions of the language model can include a group of probabilities rather than a single probability. At runtime individual probabilities of the group are weighted and combined to create an adjusted probability for the portion of the language model. The adjusted probability can be used for ASR processing. The weights can be determined based on a characteristic of the utterance, for example an associated speechlet/application, the specific user speaking, or other characteristic. By applying the weights at runtime the system can use a single language model to dynamically adjust to different utterance conditions.Type: GrantFiled: March 23, 2018Date of Patent: March 9, 2021Assignee: Amazon Technologies, Inc.Inventors: Ankur Gandhe, Ariya Rastrow, Shaswat Pratap Shah
-
Patent number: 10943143Abstract: Techniques are disclosed relating to scoring partial matches between words. In certain embodiments, a method may include receiving a request to determine a similarity between an input text data and a stored text data. The method also includes determining, based on comparing one or more words included in the input text data with one or more words included in the stored text data, a set of word pairs and a set of unpaired words. Further, in response to determining that the set of unpaired words passes elimination criteria, the method includes calculating a base similarity score between the input text data and the stored text data based on the set of word pairs. The method also includes determining a scoring penalty based on the set of unpaired words and generating a final similarity score between the input text data and the stored text data by modifying the base similarity score based on the scoring penalty.Type: GrantFiled: December 28, 2018Date of Patent: March 9, 2021Assignee: PAYPAL, INC.Inventors: Rushik Upadhyay, Dhamodharan Lakshmipathy, Nandhini Ramesh, Aditya Kaulagi
-
Patent number: 10937415Abstract: There is provided an information processing device to further improve the operability of user interfaces that use a voice as an input, the information processing device including: an acquisition unit configured to acquire context information in a period for collection of a voice; and a control unit configured to cause a predetermined output unit to present a candidate for character information obtained by converting the voice in a mode in accordance with the context information.Type: GrantFiled: March 15, 2017Date of Patent: March 2, 2021Assignee: SONY CORPORATIONInventors: Ayumi Kato, Shinichi Kawano, Yuhei Taki, Yusuke Nakagawa
-
Patent number: 10922990Abstract: A display apparatus and a method for questions and answers includes a display unit includes an input unit configured to receive user's speech voice; a communication unit configured to perform data communication with an answer server; and a processor configured to create and display one or more question sentences using the speech voice in response to the speech voice being a word speech, create a question language corresponding to the question sentence selected from among the displayed one or more question sentences, transmit the created question language to the answer server via the communication unit, and, in response to one or more answer results related to the question language being received from the answer server, display the received one or more answer results. Accordingly, the display apparatus may provide an answer result appropriate to a user's question intention although a non-sentence speech is input.Type: GrantFiled: May 23, 2019Date of Patent: February 16, 2021Assignee: SAMSUNG ELECTRONICS CO., LTD.Inventor: Eun-sang Bak
-
Patent number: 10896293Abstract: Provided is an information processing apparatus including a processing unit configured to determine, on a basis of a word of a predetermined unit selected in a text string indicated by text string information, another word connected to the selected word and included in the text string and to set a delimitation in the text string with regard to the selected word.Type: GrantFiled: April 19, 2017Date of Patent: January 19, 2021Assignee: SONY CORPORATIONInventors: Yuhei Taki, Shinichi Kawano
-
Patent number: 10896681Abstract: This document describes, among other things, a computer-implemented method for transcribing an utterance. The method can include receiving, at a computing system, speech data that characterizes an utterance of a user. A first set of candidate transcriptions of the utterance can be generated using a static class-based language model that includes a plurality of classes that are each populated with class-based terms selected independently of the utterance or the user. The computing system can then determine whether the first set of candidate transcriptions includes class-based terms. Based on whether the first set of candidate transcriptions includes class-based terms, the computing system can determine whether to generate a dynamic class-based language model that includes at least one class that is populated with class-based terms selected based on a context associated with at least one of the utterance and the user.Type: GrantFiled: December 29, 2015Date of Patent: January 19, 2021Assignee: Google LLCInventors: Petar Aleksic, Pedro J. Moreno Mengibar
-
Patent number: 10885909Abstract: A speech recognition method to be performed by a computer, the method including: detecting a first keyword uttered by a user from an audio signal representing voice of the user; detecting a term indicating a request of the user from sections that follow the first keyword in the audio signal; and determining a type of speech recognition processing applied to the following sections in accordance with the detected term indicating the request of the user.Type: GrantFiled: February 6, 2018Date of Patent: January 5, 2021Assignee: FUJITSU LIMITEDInventors: Chikako Matsumoto, Naoshi Matsuo
-
Patent number: 10872613Abstract: A method includes generating a synthesized non-reference high-band channel based on a non-reference high-band excitation corresponding to a non-reference target channel. The method further includes estimating one or more spectral mapping parameters based on the synthesized non-reference high-band channel and a high-band portion of the non-reference target channel. The method also includes applying the one or more spectral mapping parameters to the synthesized non-reference high-band channel to generate a spectrally shaped synthesized non-reference high-band channel. The method further includes generating an encoded bitstream based on the one or more spectral mapping parameters and the spectrally shaped synthesized non-reference high-band channel.Type: GrantFiled: November 4, 2019Date of Patent: December 22, 2020Assignee: QUALCOMM IncorporatedInventors: Venkata Subrahmanyam Chandra Sekhar Chebiyyam, Venkatraman Atti
-
Patent number: 10867598Abstract: A semantic analysis method, semantic analysis and non-transitory computer-readable medium are provided in this disclosure.Type: GrantFiled: December 10, 2018Date of Patent: December 15, 2020Assignee: INSTITUTE FOR INFORMATION INDUSTRYInventors: Yu-Shian Chiu, Wei-Jen Yang
-
Patent number: 10847137Abstract: An approach to speech recognition, and in particular trigger word detection, implements fixed feature extraction form waveform samples with a neural network (NN). For example, rather than computing Log Frequency Band Energies (LFBEs), a convolutional neural network is used. In some implementations, this NN waveform processing is combined with a trained secondary classification that makes use of phonetic segmentation of a possible trigger word occurrence.Type: GrantFiled: December 12, 2017Date of Patent: November 24, 2020Assignee: Amazon Technologies, Inc.Inventors: Arindam Mandal, Nikko Strom, Kenichi Kumatani, Sankaran Panchapagesan
-
Patent number: 10847147Abstract: Automatic speech recognition systems can benefit from cues in user voice such as hyperarticulation. Traditional approaches typically attempt to define and detect an absolute state of hyperarticulation, which is very difficult, especially on short voice queries. This disclosure provides for an approach for hyperarticulation detection using pair-wise comparisons and on a real-world speech recognition system. The disclosed approach uses delta features extracted from a pair of repetitive user utterances. The improvements provided by the disclosed systems and methods include improvements in word error rate by using hyperarticulation information as a feature in a second pass N-best hypotheses rescoring setup.Type: GrantFiled: May 24, 2019Date of Patent: November 24, 2020Assignee: Microsoft Technology Licensing, LLCInventors: Ranjitha Gurunath Kulkarni, Ahmed Moustafa El Kholy, Ziad Al Bawab, Noha Alon, Imed Zitouni
-
Patent number: 10841411Abstract: Systems, methods, and devices for establishing communications sessions with contacts are disclosed. In some embodiments, a first request may be received from a first device. The first request may be to communicate with a contact name. A user account associated with the first device may then be identified, and a contact list associated with the user account may be accessed to determine contacts associated with the contact name. Based on the contact list, a first contact and a second contact associated with the contact name may be identified. It may be determined, from memory, that the first contact is a first preferred contact. However, based on an intervening event, the second contact, rather than the preferred contact, may be selected for communicating with the contact.Type: GrantFiled: November 9, 2017Date of Patent: November 17, 2020Assignee: Amazon Technologies, Inc.Inventor: Aparna Nandyal
-
Patent number: 10839796Abstract: Multi-turn conversation systems that are personalized to a user based on insights derived from big data are described. A method includes: receiving, by a computer device, input from a user; obtaining, by the computer device, insights about the user; generating, by the computer device, a response based on the insights and the input; and outputting, by the computer device, the response.Type: GrantFiled: December 15, 2017Date of Patent: November 17, 2020Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Faried Abrahams, Lalit Agarwalla, Gandhi Sivakumar
-
Patent number: 10832664Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for language models using domain-specific model components. In some implementations, context data for an utterance is obtained. A domain-specific model component is selected from among multiple domain-specific model components of a language model based on the non-linguistic context of the utterance. A score for a candidate transcription for the utterance is generated using the selected domain-specific model component and a baseline model component of the language model that is domain-independent. A transcription for the utterance is determined using the score the transcription is provided as output of an automated speech recognition system.Type: GrantFiled: August 21, 2017Date of Patent: November 10, 2020Assignee: Google LLCInventors: Fadi Biadsy, Diamantino Antionio Caseiro
-
Patent number: 10832658Abstract: A method, program product and computer system to predict utterances in a dialog system includes receiving a set of utterances associated with a dialog between a client device and a dialog system, mapping the utterances to vector representations of the utterances, and identifying at least one cluster to which the utterances belong from among a plurality of possible clusters. A next cluster is predicted based upon a conditional probability of the next cluster following a set of a predetermined number of previous clusters using a language model. A next utterance is predicted from among a plurality of possible utterances within the predicted next cluster.Type: GrantFiled: March 8, 2018Date of Patent: November 10, 2020Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Chulaka Gunasekara, David Nahamoo, Lazaros Polymenakos, Kshitij Fadnis, David Echeverria Ciaurri, Jatin Ganhotra
-
Patent number: 10810472Abstract: Techniques are provided for performing sentiment analysis on words in a first data set. An example embodiment includes generating a word embedding model including a first plurality of features. A value indicating sentiment for the words in the first data set can be determined using a convolutional neural network (CNN). A second plurality of features are generated based on bigrams identified in the data set. The bigrams can be generated using a co-occurrence graph. The model is updated to include the second plurality of features, and sentiment analysis can be performed on a second data set using the updated model.Type: GrantFiled: May 10, 2018Date of Patent: October 20, 2020Assignee: ORACLE INTERNATIONAL CORPORATIONInventors: Michael Malak, Mark L. Kreider
-
Patent number: 10789946Abstract: Systems and methods are provided for speech recognition. An example method may be implementable by a server. The method may comprise adding a key phrase into a dictionary comprising a plurality of dictionary phrases, and for each one or more of the dictionary phrases, obtaining a first probability that the dictionary phrase is after the key phrase in a phrase sequence. The key phrase and the dictionary phrase may each comprise one or more words. The first probability may be independent of the key phrase.Type: GrantFiled: December 27, 2019Date of Patent: September 29, 2020Assignee: BEIJING DIDI INFINITY TECHNOLOGY AND DEVELOPMENT CO., LTD.Inventor: Chen Huang
-
Patent number: 10713519Abstract: The present invention is directed towards providing automated workflows for the identification of a reading order from text segments extracted from a document. Ordering the text segments is based on trained natural language models. In some embodiments, the workflows are enabled to perform a method for identifying a sequence associated with a portable document. The methods includes iteratively generating a probabilistic language model, receiving the portable document, and selectively extracting features (such as but not limited to text segments) from the document. The method may generate pairs of features (or feature pair from the extracted features). The method may further generate a score for each of the pairs based on the probabilistic language model and determine an order to features based on the scores. The method may provide the extracted features in the determined order.Type: GrantFiled: June 22, 2017Date of Patent: July 14, 2020Assignee: ADOBE INC.Inventors: Trung Huu Bui, Hung Hai Bui, Shawn Alan Gaither, Walter Wei-Tuh Chang, Michael Frank Kraley, Pranjal Daga
-
Patent number: 10714080Abstract: A weighted finite-state transducer (WFST) decoding system is provided. The WFST decoding system includes a memory that stores WFST data and a WFST decoder including a data fetch logic. The WFST data has a structure including states, and arcs connecting the states with directivity. The WFST data is compressed in the memory. The WFST data includes body data, and header data including state information for each states that is aligned discontinuously. The body data includes arc information of the arcs that is aligned continuously. The state information includes an arc index of the arcs, a number of the arcs, and compression information of the arcs, and the data fetch logic de-compresses the WFST data using the compression information, and retrieves the WFST data from the memory.Type: GrantFiled: September 8, 2017Date of Patent: July 14, 2020Assignee: SAMSUNG ELECTRONICS CO., LTD.Inventors: Jae Sung Yoon, Jun Seok Park
-
Patent number: 10713289Abstract: Systems, methods, and devices for performing interactive question answering using data source credibility and conversation entropy are disclosed. A speech-controlled device captures audio including a spoken question, and sends audio data corresponding thereto to a server(s). The server(s) performs speech processing on the audio data, and determines various stored data that can be used to determine an answer to the question. The server(s) determines which stored data to use based on the credibility of the source from which the stored data was received. The server(s) may also determine a number of user interactions needed to obtain data in order to fully answer the question and may select a question for a dialog soliciting further data based on the number of user interactions.Type: GrantFiled: March 31, 2017Date of Patent: July 14, 2020Assignee: Amazon Technologies, Inc.Inventors: Nina Mishra, Yonatan Naamad
-
Patent number: 10714122Abstract: Speech or non-speech detection techniques are discussed and include updating a speech pattern model using probability scores from an acoustic model to generate a score for each state of the speech pattern model, such that the speech pattern model includes a first non-speech state having multiple self loops each associated with a non-speech probability score of the probability scores, a plurality of speech states following the first non-speech state, and a second non-speech state following the speech states, and detecting speech based on a comparison of a score of the first non-speech state and a score of the last speech state of the multiple speech states.Type: GrantFiled: June 6, 2018Date of Patent: July 14, 2020Assignee: Intel CorporationInventors: Maciej Muchlinski, Tobias Bocklet
-
Patent number: 10706852Abstract: The described technology provides arbitration between speech recognition results generated by different automatic speech recognition (ASR) engines, such as ASR engines trained according to different language or acoustic models. The system includes an arbitrator that selects between a first speech recognition result representing an acoustic utterance as transcribed by a first ASR engine and a second speech recognition result representing the acoustic utterance as transcribed by a second ASR engine. This selection is based on a set of confidence features that is initially used by the first ASR engine or the second ASR engine to generate the first and second speech recognition results.Type: GrantFiled: November 13, 2015Date of Patent: July 7, 2020Assignee: Microsoft Technology Licensing, LLCInventors: Kshitiz Kumar, Hosam Khalil, Yifan Gong, Ziad Al-Bawab, Chaojun Liu
-
Patent number: 10699712Abstract: An information processing method and an electronic device are provided. The method includes an electronic device obtaining an input information through a second collection manner when the electronic device is in a speech collection state for obtaining speech information through a first collection manner, and determining a logic boundary position in relation to a first speech information in accordance with the input information, the first speech information is obtained by the electronic device through the first collection manner which is different from the second collection manner. An electronic device corresponding thereto is also disclosed.Type: GrantFiled: March 4, 2015Date of Patent: June 30, 2020Assignee: LENOVO (BEIJING) CO., LTD.Inventors: Haisheng Dai, Zhepeng Wang
-
Patent number: 10699696Abstract: The present disclosure provides a method and apparatus for correcting a speech recognition error based on artificial intelligence, and a storage medium, wherein the method comprises: obtaining a second speech recognition result of a second speech query input by the user; performing error-correcting intention recognition according to the second speech recognition result; extracting error-correcting information from the second speech recognition result when it is determined that the user has an error-correcting intention; screening error-correcting resources according to the error-correcting information, and using a selected best-matched error-correcting resource to perform error correction for the first speech recognition result, the first speech recognition result being a speech recognition result of a first speech query which is input before the second speech query.Type: GrantFiled: May 22, 2018Date of Patent: June 30, 2020Assignee: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY CO., LTD.Inventors: Hehan Li, Wensong He
-
Patent number: 10649850Abstract: Techniques and systems for storing and retrieving data storage devices of a data storage system are disclosed. In some embodiments, inventory holders are used to store data storage devices used by a data storage system. When data is to be transacted with the data storage devices, mobile drive units locate appropriate inventory holders and transport them to a device reading station, where an appropriate device retrieval unit transacts the data. In some embodiments, each inventory holder includes a heterogenous mix of data storage device types, the layout of which may be calculated according to the specific mix allocated to a given inventory holder. After the data has been transacted, the data storage devices are returned to the appropriate inventory holders, and the inventory holders are placed by the mobile drive units in locations where they may be accessed in response to further data transactions.Type: GrantFiled: June 29, 2015Date of Patent: May 12, 2020Assignee: Amazon Technologies, Inc.Inventors: James Raymond Allard, Paul David Franklin, Samuel Rubin Barrett, Jeremiah Brazeau, Jeffrey Allen Dzado, James Caleb Kirschner, David Levy, Brent James Lutz, Andrew Brendan Tinka, Colin Laird Lazier
-
Patent number: 10652592Abstract: Methods and systems are disclosed for enriching a viewing experience of a user watching video content on a screen of a client terminal by increasing the relevance of additional media content proposed or provided to the user. Disambiguation of named entities detected in a video content item being played is performed by identifying and accessing an information source directly associated with the video content item, and/or by analyzing visual content of a segment of the video content item. Selecting, proposing and/or providing an additional media content item is based on the information source and/or on the analyzing.Type: GrantFiled: March 25, 2018Date of Patent: May 12, 2020Assignee: Comigo Ltd.Inventors: Guy Geva, Menahem Lasser
-
Patent number: 10643621Abstract: An electronic device is provided. The electronic device includes a processor configured to perform automatic speech recognition (ASR) on a speech input by using a speech recognition model that is stored in a memory and a communication module configured to provide the speech input to a server and receive a speech instruction, which corresponds to the speech input, from the server. The electronic device may perform different operations according to a confidence score of a result of the ASR. Besides, it may be permissible to prepare other various embodiments speculated through the specification.Type: GrantFiled: September 11, 2018Date of Patent: May 5, 2020Assignee: Samsung Electronics Co., Ltd.Inventors: Seok Yeong Jung, Kyung Tae Kim
-
Patent number: 10621282Abstract: A computer-implemented method for providing agent assisted transcriptions of user utterances. A user utterance is received in response to a prompt provided to the user at a remote client device. An automatic transcription is generated from the utterance using a language model based upon an application or context, and presented to a human agent. The agent reviews the transcription and may replace at least a portion of the transcription with a corrected transcription. As the agent inputs the corrected transcription, accelerants are presented to the user comprising suggested texted to be inputted. The accelerants may be determined based upon an agent input, an application or context of the transcription, the portion of the transcription being replaced, or any combination thereof. In some cases, the user provides textual input, to which the agent transcribes an intent associated with the input with the aid of one or more accelerants.Type: GrantFiled: April 26, 2018Date of Patent: April 14, 2020Assignee: Interactions LLCInventors: Ethan Selfridge, Michael Johnston, Robert Lifgren, James Dreher, John Leonard
-
Patent number: 10622009Abstract: A system configured to improve double-talk detection. The system detects when double-talk is present in a voice conversation using two or more speaker models. The system extracts feature data from microphone audio data and compares the feature data to each speaker model. For example, the system may generate a first distance score indicating a likelihood that the feature data corresponds to a far-end speaker model and a second distance score indicating a likelihood that the feature data corresponds to a universal speaker model. The system may determine current system conditions based on the distance scores and may change settings to improve speech quality during the voice conversation. For example, during far-end single-talk the system may aggressively reduce an echo signal, whereas during near-end single-talk and double-talk the system may apply minimal echo cancellation to improve a quality of the local speech.Type: GrantFiled: September 10, 2018Date of Patent: April 14, 2020Assignee: Amazon Technologies, Inc.Inventors: Xianxian Zhang, Philip Ryan Hilmes, Trausti Thor Kristjansson
-
Patent number: 10614121Abstract: Content from multiple different stations can be divided into segments based on time. Matched segments associated with each station can be identified by comparing content included in a first segment associated with a first station, to content included in a second segment associated with a second station. Syndicated content can be identified and tagged based, at least in part, on a relationship between sequences of matched segments on different stations. Various embodiments also include identifying main sequences associated with each station under consideration, removing some of the main sequences, and consolidating remaining main sequences based on various threshold criteria.Type: GrantFiled: May 26, 2015Date of Patent: April 7, 2020Assignee: IHEARTMEDIA MANAGEMENT SERVICES, INC.Inventors: Periklis Beltas, Philippe Generali, David C. Jellison, Jr.
-
Patent number: 10593329Abstract: A system of multi-modal transmission of packetized data in a voice activated data packet based computer network environment is provided. A natural language processor component can parse an input audio signal to identify a request and a trigger keyword. Based on the input audio signal, a direct action application programming interface can generate a first action data structure, and a content selector component can select a content item. An interface management component can identify first and second candidate interfaces, and respective resource utilization values. The interface management component can select, based on the resource utilization values, the first candidate interface to present the content item. The interface management component can provide the first action data structure to the client computing device for rendering as audio output, and can transmit the content item converted for a first modality to deliver the content item for rendering from the selected interface.Type: GrantFiled: July 18, 2018Date of Patent: March 17, 2020Assignee: Google LLCInventors: Gaurav Bhaya, Robert Stets, Umesh Patil
-
Patent number: 10573300Abstract: The invention provides a method of automatic speech recognition. The method includes receiving a speech signal, dividing the speech signal into time windows, for each time window determining acoustic parameters of the speech signal within that window, and identifying phonological features from the acoustic parameters, such that a sequence of phonological features are generated for the speech signal, separating the sequence of phonological features into a sequence of zones, and comparing the sequences of zones to a lexical entry comprising a sequence of phonological segments to a stored lexicon to identify one or more words in the speech signal.Type: GrantFiled: August 22, 2018Date of Patent: February 25, 2020Assignee: Oxford University Innovation LimitedInventors: Aditi Lahiri, Henning Reetz, Philip Roberts
-
Patent number: 10535342Abstract: Techniques and systems are disclosed for context-dependent speech recognition. The techniques and systems described enable accurate recognition of speech by accessing sub-libraries associated with the context of the speech to be recognized. These techniques translate audible input into audio data at a smart device and determine context for the speech, such as location-based, temporal-based, recipient-based, and application based context. The smart device then accesses a context-dependent library to compare the audio data with phrase-associated translation data in one or more sub-libraries of the context-dependent library to determine a match. In this way, the techniques allow access to a large quantity of phrases while reducing incorrect matching of the audio data to translation data caused by organizing the phrases into context-dependent sub-libraries.Type: GrantFiled: April 10, 2017Date of Patent: January 14, 2020Assignee: Microsoft Technology Licensing, LLCInventor: Christian Liensberger
-
Patent number: 10535348Abstract: A system of multi-modal transmission of packetized data in a voice activated data packet based computer network environment is provided. A natural language processor component can parse an input audio signal to identify a request and a trigger keyword. Based on the input audio signal, a direct action application programming interface can generate a first action data structure, and a content selector component can select a content item. An interface management component can identify first and second candidate interfaces, and respective resource utilization values. The interface management component can select, based on the resource utilization values, the first candidate interface to present the content item. The interface management component can provide the first action data structure to the client computing device for rendering as audio output, and can transmit the content item converted for a first modality to deliver the content item for rendering from the selected interface.Type: GrantFiled: July 18, 2018Date of Patent: January 14, 2020Assignee: Google LLCInventors: Gaurav Bhaya, Robert Stets
-
Patent number: 10529322Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for tagging during speech recognition. A word lattice that indicates probabilities for sequences of words in an utterance is obtained. A conditional probability transducer that indicates a frequency that sequences of both the words and semantic tags for the words appear is obtained. The word lattice and the conditional probability transducer are composed to construct a word lattice that indicates probabilities for sequences of both the words in the utterance and the semantic tags for the words. The word lattice that indicates probabilities for sequences of both the words in the utterance and the semantic tags for the words is used to generate a transcription that includes the words in the utterance and the semantic tags for the words.Type: GrantFiled: August 21, 2017Date of Patent: January 7, 2020Assignee: Google LLCInventors: Petar Aleksic, Michael D. Riley, Pedro J. Moreno Mengibar, Leonid Velikovich
-
Patent number: 10518607Abstract: A sound of a second vehicle engine is detected. Upon predicting a pollution event by comparing the detected sound to a stored sound model, a countermeasure is actuated in a first vehicle.Type: GrantFiled: August 28, 2017Date of Patent: December 31, 2019Assignee: FORD GLOBAL TECHNOLOGIES, LLCInventors: Howard E. Churchwell, II, Mahmoud Yousef Ghannam
-
Patent number: 10496758Abstract: According to one embodiment, According to one embodiment, a machine translation apparatus includes a circuitry and a memory. The circuitry is configured to input a sentence of a first language, to segment the sentence to obtain a plurality of phrases, to search a translation model for translation options of a second language of each of the plurality of phrases, and to select top N translation options with high probabilities for decoding. N is an integer equal to or larger than 1. Furthermore, the circuitry is configured to combine the top N translation options of the plurality of phases to obtain a plurality of translation hypotheses, to search user history phrase pairs for the translation hypotheses, and to increase a score of a translation hypothesis existing in the user history phrase pairs. The memory is configured to store the score of the translation hypothesis.Type: GrantFiled: August 31, 2017Date of Patent: December 3, 2019Assignee: Kabushiki Kaisha ToshibaInventors: Zhengshan Xue, Dakun Zhang, Jichong Guo, Jie Hao
-
Patent number: 10496693Abstract: Systems and methods provide for classification and ranking of features for a hierarchical dataset. A hierarchical schema of features from the dataset is accessed. A hierarchical rank is assigned to each feature based on its schema level in the hierarchical schema. Additionally, a semantic rank is assigned to each feature using a semantic model having ranked semantic contexts. The semantic rank of a feature is assigned by identifying a semantic context of the feature and assigning the rank of the semantic context as the semantic rank of the feature. A rank is computed for each feature as a function of its hierarchical rank and semantic rank.Type: GrantFiled: May 31, 2016Date of Patent: December 3, 2019Assignee: ADOBE INC.Inventors: Shiladitya Bose, Wei Zhang, Arvind Heda
-
Patent number: 10475442Abstract: A method and a device for recognition, and a method and a device for constructing a recognition model are disclosed. A device for constructing a recognition model includes a training data inputter configured to receive additional training data, a model learner configured to train a first recognition model constructed based on basic training data to learn the additional training data, and a model constructor configured to construct a final recognition model by integrating the first recognition model with a second recognition model generated by the training of the first recognition model.Type: GrantFiled: October 24, 2016Date of Patent: November 12, 2019Assignee: Samsung Electronics Co., Ltd.Inventor: Ho Shik Lee
-
Patent number: 10453460Abstract: Systems and methods for determining that artificial commands, in excess of a threshold value, are detected by multiple voice activated electronic devices is described herein. In some embodiments, numerous voice activated electronic devices may send audio data representing a phrase to a backend system at a substantially same time. Text data representing the phrase, and counts for instances of that text data, may be generated. If the number of counts exceeds a predefined threshold, the backend system may cause any remaining response generation functionality that particular command that is in excess of the predefined threshold to be stopped, and those devices returned to a sleep state. In some embodiments, a sound profile unique to the phrase that caused the excess of the predefined threshold may be generated such that future instances of the same phrase may be recognized prior to text data being generated, conserving the backend system's resources.Type: GrantFiled: March 30, 2016Date of Patent: October 22, 2019Assignee: Amazon Technologies, Inc.Inventors: Colin Wills Wightman, Naresh Narayanan, Daniel Robert Rashid
-
Patent number: 10417329Abstract: A dialog act estimation method includes acquiring learning data including a first sentence to be estimated in the form of text data of a first uttered sentence uttered at a first time point, a second sentence which is text data of a second uttered sentence uttered, at a time point before the first time point, successively after the first uttered sentence, act information indicating an act associated to the first sentence, property information indicating a property information associated to the first sentence, and dialog act information indicating a dialog act in the form of a combination of an act and a property associated to the first sentence, making a particular model learn three or more tasks at the same time using the learning data, and storing a result of the learning as learning result information in a memory.Type: GrantFiled: August 1, 2017Date of Patent: September 17, 2019Assignee: PANASONIC INTELLECTUAL PROPERTY MANAGEMENT CO., LTD.Inventors: Takashi Ushio, Hongjie Shi, Mitsuru Endo, Katsuyoshi Yamagami
-
Patent number: 10417328Abstract: Methods and processes evaluate a quality score of a text. The text includes a plurality of words. The methods compute first probability characteristics of groups of words in a reference text which is known to be a high-quality text. The methods also compute second probability characteristics of groups of words in a text to be scored. The methods also compute the quality score based on a difference between the first probability characteristics and the second probability characteristics.Type: GrantFiled: January 5, 2018Date of Patent: September 17, 2019Assignee: Searchmetrics GmbHInventors: Ahmet Anil Pala, Alexander Kagoshima, Marcus Tober
-
Patent number: 10409909Abstract: Techniques are disclosed for building a dictionary of words from combinations of symbols generated based on input data. A neuro-linguistic behavior recognition system includes a neuro-linguistic module that generates a linguistic model that describes data input from a source (e.g., video data, SCADA data, etc.). To generate words for the linguistic model, a lexical analyzer component in the neuro-linguistic module receives a stream of symbols, each symbol generated based on an ordered stream of normalized vectors generated from input data. The lexical analyzer component determines words from combinations of the symbols based on a hierarchical learning model having one or more levels. Each level indicates a length of the words to be identified at that level. Statistics are evaluated for the words identified at each level. The lexical analyzer component identifies one or more of the words having statistical significance.Type: GrantFiled: December 12, 2014Date of Patent: September 10, 2019Assignee: Omni AI, Inc.Inventors: Gang Xu, Ming-Jung Seow, Tao Yang, Wesley Kenneth Cobb
-
Patent number: 10409910Abstract: Techniques are disclosed for generating a syntax for a neuro-linguistic model of input data obtained from one or more sources. A stream of words of a dictionary built from a sequence of symbols are received. The symbols are generated from an ordered stream of normalized vectors generated from input data. Statistics for combinations of words co-occurring in the stream are evaluated. The statistics includes a frequency upon which the combinations of words co-occur. A model of combinations of words based on the evaluated statistics is updated. The model identifies statistically relevant words. A connected graph is generated. Each node in the connected graph represents one of the words in the stream. Edges connecting the nodes represent a probabilistic relationship between words in the stream. Phrases are identified based on the connected graph.Type: GrantFiled: December 12, 2014Date of Patent: September 10, 2019Assignee: Omni AI, Inc.Inventors: Ming-Jung Seow, Gang Xu, Tao Yang, Wesley Kenneth Cobb
-
Patent number: 10372737Abstract: According to one embodiment, a method, computer system, and computer program product for retraining a classifier-based automatic dialog system with recorded user interactions is provided. The present invention may include receiving recorded interactions, where the interactions are between a user and an automatic dialog system; determining, based on the recorded interactions, whether to pair a given input with one or more classes; pairing inputs with one or more classes; assessing the reliability of the paired inputs and classes; determining whether the reliable paired inputs and classes can be consistently mapped; and merging all consistently mapped reliable pairs with an initial training set.Type: GrantFiled: November 16, 2017Date of Patent: August 6, 2019Assignee: International Business Machines CorporationInventors: Allen Ginsberg, Edward G. Katz, Alexander C. Tonetti
-
Patent number: 10375224Abstract: A mobile device providing integrated management of message information and service provision through artificial intelligence is disclosed. The mobile device includes an integrated message management unit comprising a message monitoring unit configured to monitor voice call information and text message information in association with the voice call management part and the text message management part, a message information managing unit configured to generate integrated message information, which is to be provided to a user, based on the voice call information and the text message information, an interface managing unit configured to generate an integrated message management user interface displaying the integrated message information, and an artificial intelligence agent analyzing the voice call information and the text message information and providing a service associated with at least one additional function in association with the additional function process part based on the analyzed result.Type: GrantFiled: March 28, 2018Date of Patent: August 6, 2019Assignee: NHN Entertainment CorporationInventor: Dong Wook Kim
-
Patent number: 10366159Abstract: A system for identifying address components includes an interface and a processor. The interface is to receive an address for parsing. The processor is to determine a matching model of a set of models based at least in part on a matching probability for each model for a tokenized address, which is based on the address for parsing, and associate each component of the tokenized address with an identifier based at least in part on the matching model, wherein each component of the set of components is associated with an identifier, and wherein probabilities of each component of the set of components are determined using training addresses.Type: GrantFiled: October 14, 2016Date of Patent: July 30, 2019Assignee: Workday, Inc.Inventors: Parag Avinash Namjoshi, Shuangshuang Jiang, Mohammad Sabah
-
Patent number: 10360898Abstract: A system and method are presented for predicting speech recognition performance using accuracy scores in speech recognition systems within the speech analytics field. A keyword set is selected. Figure of Merit (FOM) is computed for the keyword set. Relevant features that describe the word individually and in relation to other words in the language are computed. A mapping from these features to FOM is learned. This mapping can be generalized via a suitable machine learning algorithm and be used to predict FOM for a new keyword. In at least embodiment, the predicted FOM may be used to adjust internals of speech recognition engine to achieve a consistent behavior for all inputs for various settings of confidence values.Type: GrantFiled: June 5, 2018Date of Patent: July 23, 2019Inventors: Aravind Ganapathiraju, Yingyi Tan, Felix Immanuel Wyss, Scott Allen Randal
-
Patent number: 10360904Abstract: Methods and apparatus for performing speech recognition using a garbage model. The method comprises receiving audio comprising speech and processing at least some of the speech using a garbage model to produce a garbage speech recognition result. The garbage model includes a plurality of sub-words, each of which corresponds to a possible combination of phonemes in a particular language.Type: GrantFiled: May 9, 2014Date of Patent: July 23, 2019Assignee: Nuance Communications, Inc.Inventors: Cosmin Popovici, Kenneth W. D. Smith, Petrus C. Cools