Specialized Models Patents (Class 704/255)
  • Patent number: 11188594
    Abstract: Techniques herein improve computational efficiency for wildcard searches by using numeric string hashes. In an embodiment, a plurality of query K-gram tokens for a term in a query are generated. Using a first index, an intersection of hash tokens is determined, wherein said first index indexes each query K-gram token of said K-gram tokens to a respective subset of hash tokens of a plurality of hash tokens, each of hash token of said plurality of hash tokens corresponding to a term found in one or more documents of a corpus of documents. The intersection of hash tokens comprises only hash tokens indexed to all of said plurality of query K-gram tokens by said first index. Using a second index, documents of said corpus of documents that contain said term are determined, said second index indexing said hash tokens to a plurality of terms in said corpus of documents and for each term of said plurality of terms, a respective subset of documents of corpus of documents that contain said each term.
    Type: Grant
    Filed: February 7, 2018
    Date of Patent: November 30, 2021
    Assignee: ORACLE INTERNATIONAL CORPORATION
    Inventors: Rahul Manohar Kadwe, Saurabh Naresh Netravalkar
  • Patent number: 11182706
    Abstract: A method, system and computer program product for improving collaboration among participants in a collaboration system. In one embodiment of the present invention, a system, referred to herein as the “integration system,” connected to a collaboration system monitors for comments or updates pertaining to a work item involving a customer problem to be resolved by different participants of the collaboration system. These comments or updates for completing the work item are analyzed. After analyzing the comments or updates, strategies are derived for completing the work item based on the analysis of the comments or updates as well as based on stored data of previously resolved work items. The derived strategies are then presented to the appropriate participant(s) to resolve the work item, such as based on the roles of the participant(s) that would most effectively and efficiently perform the strategy.
    Type: Grant
    Filed: November 13, 2017
    Date of Patent: November 23, 2021
    Assignee: International Business Machines Corporation
    Inventors: Abhishek Shetty Balakrishna, Sivaranjani Kathirvel, Shunmugaraja Periadurai, Sriharidatta Sriharidatta
  • Patent number: 11176924
    Abstract: A computer-implemented method includes: comparing features extracted from a first document that include a sound to features extracted from acoustic files related to the sound; designating the sound in a document of the plurality of documents as a true; designating the sound in the first document as a false negative; generating a first sound vector for the sound in the first document in response to the sound in the first document being designated a false negative; generating a sound vector for each of the documents designated as a true positive; creating a centroid vector for the sound vectors of the documents designated as a true positive; and redesignating the sound in the first document from a false negative to a true positive in response to the first sound vector and the centroid vector being a Banach space.
    Type: Grant
    Filed: January 9, 2020
    Date of Patent: November 16, 2021
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Craig M. Trim, Aaron K. Baughman, Micah Forster, Shikhar Kwatra
  • Patent number: 11176943
    Abstract: According to an embodiment, a voice recognition device includes one or more processors. The one or more processors are configured to: recognize a voice signal representing a voice uttered by an object speaker, to generate text and meta information representing information that is not included in the text and included in the voice signal; generate an object presentation vector including a plurality of parameters representing a feature of a presentation uttered by the object speaker; calculate a similarity between the object presentation vector and a reference presentation vector including a plurality of parameters representing a feature of a presentation uttered by a reference speaker; and output the text. The one or more processors are further configured to determine whether to output the meta information based on the similarity, and upon determining to output the meta information, add the meta information to the text and output the meta information.
    Type: Grant
    Filed: February 14, 2018
    Date of Patent: November 16, 2021
    Assignee: KABUSHIKI KAISHA TOSHIBA
    Inventors: Kosei Fume, Masahiro Yamamoto
  • Patent number: 11170177
    Abstract: A method is described comprising receiving a conversational transcript of a conversational interaction among a plurality of participants, wherein each participant contributes a sequence of contributions to the conversational interaction. The method includes projecting contributions of the plurality of participants into a semantic space using a natural language vectorization, wherein the semantic space describes semantic relationships among words of the conversational interaction. The method includes computing interaction process measures using information of the conversational transcript, the conversational interaction, and the natural language vectorization.
    Type: Grant
    Filed: July 30, 2018
    Date of Patent: November 9, 2021
    Inventors: Nia Marcia Maria Dowell, Tristan Nixon
  • Patent number: 11138253
    Abstract: Example methods, apparatus, systems and articles of manufacture are disclosed to determine tags for unknown media using multiple media features. Disclosed examples extract features from audio and image portions of the unknown media. Disclosed examples weight the features based at least partially on respective recognition technologies used to extract the features to determine corresponding weighted features, wherein disclosed examples assign a first weight to a first feature extracted by an image-based recognition technology, and assign a second weight, different from the first weight, to a second feature extracted by an audio-based recognition technology. Disclosed examples search a database of pre-tagged media with a combination of the weighted features to generate a list of suggested tags for the unknown media. Disclosed examples assign one or more tags from the list of suggested tags to the unknown media.
    Type: Grant
    Filed: June 28, 2019
    Date of Patent: October 5, 2021
    Assignee: The Nielsen Company (US), LLC
    Inventor: Morris Lee
  • Patent number: 11132996
    Abstract: Embodiments of the present disclosure relate to a method and apparatus for outputting information. The method includes: outputting a to-be-read audio in response to receiving a reading instruction from a user; acquiring an actually read audio obtained by reading the to-be-read audio by the user; performing speech recognition on the actually read audio to obtain a recognition result; calculating a similarity between the actually read audio and the to-be-read audio based on a character string corresponding to the recognition result and a character string corresponding to the to-be-read audio; determining, from a predetermined set of similarity intervals, a similarity interval to which the calculated similarity belongs; and outputting a reading evaluation corresponding to the determined similarity interval. The embodiment may help a reader to improve the learning efficiency and learning interest, thereby improving the rate of a user using a device.
    Type: Grant
    Filed: October 4, 2019
    Date of Patent: September 28, 2021
    Assignee: BAIDU ONLINE NETWORK TECHNOLOGY (BEIJING) CO., LTD.
    Inventor: Yongshuai Lu
  • Patent number: 11120063
    Abstract: There is provided an information processing apparatus including: a processing unit configured to perform a summarization process of summarizing content of speech indicated by voice information based on speech of a user on a basis of acquired information indicating a weight related to a summary.
    Type: Grant
    Filed: October 14, 2016
    Date of Patent: September 14, 2021
    Assignee: SONY CORPORATION
    Inventors: Shinichi Kawano, Keisuke Touyama, Nobuki Furue, Keisuke Saito, Daisuke Sato, Mitani Ryosuke, Miwa Ichikawa
  • Patent number: 11114092
    Abstract: A computer-implemented method and supporting system transcribes spoken words being monitored from a telephonic interaction among two or more individuals. Telephonic interactions among the individuals are monitored, and at least two of the individuals are each assigned to a separate channel. While still being monitored, each of the channels is assigned a context-based speech recognition models, and in substantially real-time, the monitored telephonic interaction is transcribed from speech to text based on the different assigned models.
    Type: Grant
    Filed: April 10, 2020
    Date of Patent: September 7, 2021
    Assignee: Groupe Allo Media SAS
    Inventor: Romain Sambarino
  • Patent number: 11113286
    Abstract: A query directed at a source table organized into a set of batch units is received. The query includes a pattern matching predicate that specifies a search pattern. A set of N-grams are generated based on the search pattern. A pruning index associated with the source table is accessed. The pruning index comprises a set of filters that index distinct N-grams in each column of the source table. The pruning index is used to identify a subset of batch units to scan for matching data based on the set of N-grams generated for the search pattern. The query is processed by scanning the subset of batch units.
    Type: Grant
    Filed: March 31, 2021
    Date of Patent: September 7, 2021
    Assignee: Snowflake Inc.
    Inventors: Thierry Cruanes, Benoit Dageville, Ismail Oukid, Stefan Richter
  • Patent number: 11100916
    Abstract: A speech recognition method and apparatus are disclosed. The speech recognition method includes determining a first score of candidate texts based on an input speech, determining a weight for an output of a language model based on the input speech, applying the weight to a second score of the candidate texts output from the language model to obtain a weighted second score, selecting a target candidate text from among the candidate texts based on the first score and the weighted second score corresponding to the target candidate text, and determining the target candidate text to correspond to a portion of the input speech.
    Type: Grant
    Filed: April 30, 2019
    Date of Patent: August 24, 2021
    Assignee: Samsung Electronics Co., Ltd.
    Inventor: Jihyun Lee
  • Patent number: 11048356
    Abstract: A game controller includes a touchpad that a user, viewing a virtual keyboard on a screen, can soft-touch to move a cursor on the screen and then hard-touch to move the cursor and also send location data to a processor for inputting a letter from the virtual keyboard. A microphone on the touchpad can be used to receive voice signals for training a machine learning module to predict a next letter or next word, or to insert special characters/punctuations/graphics such as “smileys” during the swipe, or to indicate a tone of a Chinese character while typing with Chinese Pinyin.
    Type: Grant
    Filed: July 31, 2019
    Date of Patent: June 29, 2021
    Assignee: Sony Interactive Entertainment LLC
    Inventors: Cen Zhao, Chung-Hsien Yu, Samuel Ian Matthews
  • Patent number: 11024298
    Abstract: Methods and apparatus for performing speech recognition using a garbage model. The method comprises receiving audio comprising speech and processing at least some of the speech using a garbage model to produce a garbage speech recognition result. The garbage model includes a plurality of sub-words, each of which corresponds to a possible combination of phonemes in a particular language.
    Type: Grant
    Filed: July 18, 2019
    Date of Patent: June 1, 2021
    Assignee: Nuance Communications, Inc.
    Inventors: Cosmin Popovici, Kenneth W. D. Smith, Petrus C. Cools
  • Patent number: 10999624
    Abstract: A multimedia device of a vehicle may include: a broadcast receiver configured to receive broadcast information of a plurality of channels from a radio signal; a controller configured to divide the received broadcast information into sound information and subtitle information, and to simultaneously output the sound and subtitle information; a sound outputter configured to output the divided sound information; and a display configured to display the divided subtitle information. The controller is further configured to determine electronic program guide (EPG) information of the radio signal, and to control the display so as to display a text-changeable channel according to the EPG information.
    Type: Grant
    Filed: October 29, 2018
    Date of Patent: May 4, 2021
    Assignees: Hyundai Motor Company, Kia Motors Corporation
    Inventors: Hyunwoo Lee, Daebong An
  • Patent number: 10983853
    Abstract: Provided are methods and systems for automatically generating input grammars for grammar-based fuzzing by utilizing machine-learning techniques and sample inputs. Neural-network-based statistical learning techniques are used for the automatic generation of input grammars. Recurrent neural networks are used for learning a statistical input model that is also generative in that the model is used to generate new inputs based on the probability distribution of the learnt model.
    Type: Grant
    Filed: June 30, 2017
    Date of Patent: April 20, 2021
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Patrice Godefroid, Rishabh Singh, Hila Peleg
  • Patent number: 10964315
    Abstract: An approach to wakeword detection uses an explicit representation of non-wakeword speech in the form of subword (e.g., phonetic monophone) units that do not necessarily occur in the wakeword and that broadly represent general speech. These subword units are arranged in a “background” model, which at runtime essentially competes with the wakeword model such that a wakeword is less likely to be declare as occurring when the input matches that background model well. An HMM may be used with the model to locate possible occurrences of the wakeword. Features are determined from portions of the input corresponding to subword units of the wakeword detected using the HMM. A secondary classifier is then used to process the features to yield a decision of whether the wakeword occurred.
    Type: Grant
    Filed: June 30, 2017
    Date of Patent: March 30, 2021
    Assignee: Amazon Technologies, Inc.
    Inventors: Minhua Wu, Sankaran Panchapagesan, Ming Sun, Shiv Naga Prasad Vitaladevuni, Bjorn Hoffmeister, Ryan Paul Thomas, Arindam Mandal
  • Patent number: 10936828
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for neural translation systems with rare word processing. One of the methods is a method training a neural network translation system to track the source in source sentences of unknown words in target sentences, in a source language and a target language, respectively and includes deriving alignment data from a parallel corpus, the alignment data identifying, in each pair of source and target language sentences in the parallel corpus, aligned source and target words; annotating the sentences in the parallel corpus according to the alignment data and a rare word model to generate a training dataset of paired source and target language sentences; and training a neural network translation model on the training dataset.
    Type: Grant
    Filed: November 16, 2018
    Date of Patent: March 2, 2021
    Assignee: Google LLC
    Inventors: Quoc V. Le, Minh-Thang Luong, Ilya Sutskever, Oriol Vinyals, Wojciech Zaremba
  • Patent number: 10915118
    Abstract: A method for controlling a drone includes receiving a request for information about a spatial location, generating data requests, configuring a flight plan and controlling one or more drones to fly over the spatial location to obtain data types based on the data requests, storing heterogeneous data captured by the one or more drones and creating spatio-temporal indices for identifying spatial or temporal coverage gaps in the data necessary to answer the request, controlling the one or more drones to fly over the spatial location to obtain a plurality of data types from the identified spatial or temporal coverage gaps and extracting and analyzing data to answer the request.
    Type: Grant
    Filed: December 23, 2019
    Date of Patent: February 9, 2021
    Assignee: International Business Machines Corporation
    Inventors: Gregory F. Boland, James R. Kozloski, Yu Ma, Justin G. Manweiler, Kevin E. Siemonsen, Umut Topkara, Katherine Vogt, Justin D. Weisz
  • Patent number: 10909171
    Abstract: Systems and processes for operating an intelligent automated assistant are provided. In accordance with one example, a method includes, at an electronic device with one or more processors and memory, receiving a first natural-language speech input indicative of a request for media, where the first natural-language speech input comprises a first search parameter; providing, by a digital assistant, a first media item identified based on the first search parameter. The method further includes, while providing the first media item, receiving a second natural-language speech input and determining whether the second input corresponds to a user intent of refining the request for media. The method further includes, in accordance with a determination that the second speech input corresponds to a user intent of refining the request for media: identifying, based on the first parameter and the second speech input, a second media item and providing the second media item.
    Type: Grant
    Filed: March 1, 2019
    Date of Patent: February 2, 2021
    Assignee: Apple Inc.
    Inventors: David Chance Graham, Cyrus Daniel Irani, Aimee Piercy, Thomas Alsina
  • Patent number: 10902848
    Abstract: A method for providing a telematics service by using voice recognition is provided. The method includes steps of: (a) a telematics server, if a text command signal converted from a voice command signal for a control of a vehicle from a voice input device and a user's voice sample are acquired from a voice recognition service-providing device, analyzing the user's voice sample; and (b) the telematics server, if the user's voice sample satisfies at least one preset condition, allowing the vehicle to be controlled according to the voice command signal by supporting a telematics device placed in the vehicle to run a telematics API corresponding to the text command signal.
    Type: Grant
    Filed: July 20, 2017
    Date of Patent: January 26, 2021
    Assignee: HYUNDAI AUTOEVER AMERICA, LLC.
    Inventors: Changki Sohn, Bongseok Kim
  • Patent number: 10878339
    Abstract: Systems and methods of leveraging machine learning to predict user generated content are provided. For instance, first entity data associated with an entity can be received. The first entity data can include user specified data associated with an attribute of the entity. The first entity data can be input into a machine-learned content prediction model. Inferred entity data can be received as output of the machine-learned content prediction model. The inferred entity data can include inferred data descriptive of the attribute of the entity.
    Type: Grant
    Filed: January 27, 2017
    Date of Patent: December 29, 2020
    Assignee: Google LLC
    Inventors: Arun Mathew, Kaleigh Smith, Per Anderson, Ian Langmore
  • Patent number: 10853747
    Abstract: An example method includes receiving, by a computational assistant executing at one or more processors, a representation of an utterance spoken at a computing device; identifying, based on the utterance, a task to be performed; determining a capability level of a first party (1P) agent to perform the task; determining capability levels of respective third party (3P) agents of a plurality of 3P agents to perform the task; responsive to determining that the capability level of the 1P agent does not satisfy a threshold capability level, that a capability level of a particular 3P agent of the plurality of 3P agents is a greatest of the determined capability levels, and that the capability level of the particular 3P agent satisfies the threshold capability level, selecting the particular 3P agent to perform the task; and performing one or more actions determined by the selected agent to perform the task.
    Type: Grant
    Filed: November 16, 2017
    Date of Patent: December 1, 2020
    Assignee: GOOGLE LLC
    Inventors: Bo Wang, Lei Zhong, Barnaby John James, Saisuresh Krishnakumaran, Robert Stets, Bogdan Caprita, Valerie Nygaard
  • Patent number: 10824819
    Abstract: Implementations of the present specification disclose methods, apparatuses, and devices for generating word vectors. The method includes: obtaining individual words by segmenting a corpus; establishing a feature vector of each word based on n-ary characters; training a recurrent neural network based on the feature vectors of the obtained words and feature vectors of context words associated with the obtained words in the corpus; and generating a word vector for each obtained word based on the feature vector of the obtained word and the trained recurrent neural network.
    Type: Grant
    Filed: May 20, 2020
    Date of Patent: November 3, 2020
    Assignee: ALIBABA GROUP HOLDING LIMITED
    Inventors: Shaosheng Cao, Jun Zhou
  • Patent number: 10810485
    Abstract: A convolutional neural network for classifying time series data uses a dynamic context selection. In one example a method includes receiving a plurality of inputs of different sizes at a convolutional neural network, applying convolution and pooling to each of the inputs to provide a plurality of outputs of different sizes, changing the size of each of the outputs to a selected uniform size, reshaping each of the outputs to a vector, and fully connecting the vectors.
    Type: Grant
    Filed: September 13, 2016
    Date of Patent: October 20, 2020
    Assignee: Intel Corporation
    Inventors: Freddie H. Huang, Omar U. Florez, Jonathan J. Huang, Yen-Kuang Chen
  • Patent number: 10769191
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for classifying data objects. One of the methods includes obtaining data that associates each term in a vocabulary of terms with a respective high-dimensional representation of the term; obtaining classification data for a data object, wherein the classification data includes a respective score for each of a plurality of categories, and wherein each of the categories is associated with a respective category label; computing an aggregate high-dimensional representation for the data object from high-dimensional representations for the category labels associated with the categories and the respective scores; identifying a first term in the vocabulary of terms having a high-dimensional representation that is closest to the aggregate high-dimensional representation; and selecting the first term as a category label for the data object.
    Type: Grant
    Filed: December 19, 2014
    Date of Patent: September 8, 2020
    Assignee: Google LLC
    Inventors: Gregory Sean Corrado, Tomas Mikolov, Samy Bengio, Yoram Singer, Jonathon Shlens, Andrea L. Frome, Jeffrey Adgate Dean, Mohammad Norouzi
  • Patent number: 10769374
    Abstract: Disclosed are an answer selection method for a question answering system. With the method and system, a sentence distributed representation of a question and an answer is generated by adopting a hierarchical length-adaptive neural network structure with a purpose of extracting high-quality sentence features by using different neural feature extractors according to a length of an input sentence. First, one word distributed representation is generated for each word in the input sentence by connecting one fixed word embedding and one finely-adjusted word embedding. Next, a BiLSTM-based feature extractor is used for a short sentence and a Transformer-based feature extractor is used for a long sentence. Finally, a sentence vector representation for measuring a correlation between a question and a candidate answer is generated by using an attention pooling layer considering an interaction between a question sentence and an answer sentence.
    Type: Grant
    Filed: March 26, 2020
    Date of Patent: September 8, 2020
    Inventor: Honghui Chen
  • Patent number: 10769382
    Abstract: A method, system and computer program product for disambiguating meaning of terms. A natural language processing system obtains a first corpora of words used in a first sense and a second corpora of the same words that are used in a second sense. Each of these corpora of words may be associated with different linguistic domains. The natural language processing system generates a first and a second set of patterns using both the first and second corpora of words, respectively. A question passage is then received by the natural language processing system. The natural language processing system examines a word of interest in the question passage. The user is then notified that the word of interest is being used in the first sense or the second sense in response to identifying the word of interest in only the first set of patterns or the second set of patterns, respectively.
    Type: Grant
    Filed: January 28, 2019
    Date of Patent: September 8, 2020
    Assignee: International Business Machines Corporation
    Inventors: Alfredo Alba, Clemens Drews, Daniel F. Gruhl, Linda H. Kato, Christian B. Kau, Neal R. Lewis, Pablo N. Mendes, Meenakshi Nagarajan, Cartic Ramakrishnan
  • Patent number: 10755028
    Abstract: An analysis device includes a memory and a processor configured to acquire a document for analyzing, generate a coded document including a plurality of coded words by coding the document for each word unit, determine a vector value according to a condition of appearance of a first coded phrase in the coded document, the first coded phrase corresponding to a combination of two or more coded words included in the plurality of coded words, and assign the vector value to the first coded phrase.
    Type: Grant
    Filed: February 22, 2018
    Date of Patent: August 25, 2020
    Assignee: FUJITSU LIMITED
    Inventors: Masahiro Kataoka, Satoshi Onoue, Hiroyuki Yoshida
  • Patent number: 10706215
    Abstract: A formula system can transform a formula natural language representation (“NLR”) into a representation which shows the formula in traditional mathematical notation. This transformation can include creating a state machine with transition mappings between states that match to initial parts of the NLR. These transition mappings can include global transition mappings that are first attempted to be matched to the beginning of the NLR and then state specific transition mappings can be matched to the NLR. The formula system can consume the NLR, transitioning from state to state as indicated by the transition mappings and removing the matched initial part from the NLR, until the NLR has been fully consumed. In some cases, the formula system can recursively or iteratively create additional state machines to consume portions of the NLR. Some states provide a result (e.g. portion of a formula representation) which are combined to create the final formula representation.
    Type: Grant
    Filed: April 4, 2018
    Date of Patent: July 7, 2020
    Assignee: Parsegon
    Inventors: Mathew Pregasen, Rikhav Shah
  • Patent number: 10692016
    Abstract: Methods and apparatuses are described for analyzing unstructured computer text for identification and classification of complaint-specific interactions. A computer data stores unstructured text. A server computing device splits the unstructured text into phrases of words. The server generates a set of tokens from each phrase and removes tokens that are stopwords. The server generates a normalized sentiment score for each set of tokens. The server uses a rules-based classification engine to generate a rules-based complaint score for each set of tokens. The server uses an artificial intelligence machine learning model to generate a model-based complaint score for each set of tokens. The server determines determine whether each set of tokens corresponds to a complaint-specific interaction based upon the rules-based complaint score and the model-based complaint score.
    Type: Grant
    Filed: February 7, 2017
    Date of Patent: June 23, 2020
    Assignee: FMR LLC
    Inventors: Aravind Chandramouli, Nitin Hardeniya, Sunil Kumar
  • Patent number: 10685647
    Abstract: A speech recognition method and a speech recognition device are disclosed. The speech recognition method includes: obtaining features of a speech signal to be recognized; performing a path search in a search space generated by establishing a map according to the features to output a decoding result; judging whether a rejection is needed according to the decoding result; and when the rejection is needed, determining that a speech recognition result is the rejection, and when the rejection is not needed, obtaining the speech recognition result according to the decoding result. The method has a good recognition rejection effect.
    Type: Grant
    Filed: June 24, 2016
    Date of Patent: June 16, 2020
    Assignee: BAIDU ONLINE NETWORK TECHNOLOGY (BEIJING) CO., LTD.
    Inventors: Bin Yuan, Shouye Peng, Yangfei Xu
  • Patent number: 10657328
    Abstract: The present disclosure generally relates to systems and processes for morpheme-based word prediction. An example method includes receiving a current word; determining a context of the current word based on the current word and a context of a previous word; determining, using a morpheme-based language model, a likelihood of a prefix based on the context of the current word; determining, using the morpheme-based language model, a likelihood of a stem based on the context of the current word; determining, using the morpheme-based language model, a likelihood of a suffix based on the context of the current word; determining a next word based on the likelihood of the prefix, the likelihood of the stem, and the likelihood of the suffix; and providing an output including the next word.
    Type: Grant
    Filed: December 21, 2017
    Date of Patent: May 19, 2020
    Assignee: Apple Inc.
    Inventors: Jerome R. Bellegarda, Jannes G. Dolfing
  • Patent number: 10657957
    Abstract: A computer-implemented method and supporting system transcribes spoken words being monitored from a telephonic interaction among two or more individuals. Telephonic interactions among the individuals are monitored, and at least two of the individuals are each assigned to a separate channel. While still being monitored, each of the channels is assigned a context-based speech recognition models, and in substantially real-time, the monitored telephonic interaction is transcribed from speech to text based on the different assigned models.
    Type: Grant
    Filed: November 22, 2019
    Date of Patent: May 19, 2020
    Assignee: Groupe Allo Media SAS
    Inventor: Romain Sambarino
  • Patent number: 10636428
    Abstract: Systems, methods, and devices for determining a target device for a voice command are provided. A voice command is detected at a plurality of devices. A weight is determined for the detected voice command at each device of the plurality of devices. The determined weight is exchanged among the plurality of devices. A highest weight among the exchanged weights is determined. The device associated with the highest weight is determined as the target device for the voice command.
    Type: Grant
    Filed: June 29, 2017
    Date of Patent: April 28, 2020
    Assignee: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventor: Aniket Krishnat Naravanekar
  • Patent number: 10629197
    Abstract: A voice processing system includes a voice acquisition unit that acquires an utterance of a user, a voice recognition unit that recognizes contents of the utterance acquired by the voice acquisition unit, an execution unit that executes processing based on a result of the voice recognition unit, and a determination unit that determines a candidate of a word to substitute a certain word. When the voice recognition unit recognizes a request including a first word, the execution unit executes the request including the first word and provides the user with a processing result. The determination unit determines a second word that is a substitute candidate of the first word. The execution unit also executes a request including the second word instead of the first word and stores a processing result in a storage unit.
    Type: Grant
    Filed: January 19, 2017
    Date of Patent: April 21, 2020
    Assignee: TOYOTA JIDOSHA KABUSHIKI KAISHA
    Inventors: Sei Kato, Junichi Ito, Takuma Minemura, Junya Masui, Toshiyuki Namba
  • Patent number: 10630619
    Abstract: Disclosed is a method of extracting and using a semantic entity from a text message by an electronic device. The method includes: recognizing a text-based input; extracting a semantic entity from the text-based input; and providing the extracted semantic entity through an application in response to the application having a semantic setting that corresponds to the extracted semantic entity.
    Type: Grant
    Filed: February 18, 2015
    Date of Patent: April 21, 2020
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Muwoong Lee, Yoonkyong Lee, Jaewon Lee, Seokjin Chang, Sunghoon Cho, Sangmi Kim, Donghyun Roh
  • Patent number: 10614811
    Abstract: A system, method, apparatus and computer readable medium for hierarchical speech recognition resolution. The method of hierarchical speech recognition resolution on a platform includes receiving a speech stream from a microphone. The speech stream is resolved using a lowest possible level automatic speech recognition (ASR) engine of multi-level ASR engines. The selection of the lowest possible level ASR engine is based on policies defined for the platform. If resolution of the speech stream is rated less than a predetermined confidence level, the resolution of the speech stream is pushed to a next higher-level ASR engine of the multi-level ASR engines until the resolution of the speech stream meets the predetermined confidence level without violating one or more policies.
    Type: Grant
    Filed: December 29, 2017
    Date of Patent: April 7, 2020
    Assignee: Intel Corporation
    Inventors: Douglas Gabel, Jonathan Huang, Sylvia J. Downing, Narayan Biswal, Binuraj Ravindran, Willem Beltman, Vered Bar Bracha, Ze'Ev Rivlin
  • Patent number: 10600407
    Abstract: A generation device includes a receiving unit and a generating unit. The receiving unit receives a model representing correspondence between one or more phonetic symbols and one or more words. The generating unit generates a first finite state transducer based on the model, the first finite state transducer at least including, as outgoing transitions from a first state representing transition destination of a first transition which has a first phonetic symbol of a predetermined type as input symbol, a second transition that has a second phonetic symbol, which is different than a particular symbol representing part or whole of input symbol of the first transition, as input symbol, and a third transition that has a third phonetic symbol, which represents the particular symbol or silence, as input symbol.
    Type: Grant
    Filed: February 9, 2017
    Date of Patent: March 24, 2020
    Assignee: Kabushiki Kaisha Toshiba
    Inventor: Manabu Nagao
  • Patent number: 10599776
    Abstract: A mechanism is provided for improving predicate parses (or logical representations of a passage) using semantic knowledge. In response to encountering an ambiguous decision point during a syntactic analysis of a portion of natural language content, a candidate meaning of the ambiguous decision point is generated. Characteristics of the ambiguous decision point are evaluated based on a semantic knowledge base to determine a semantic meaning associated with the ambiguous decision point. A determination is made as to whether the semantic meaning supports or refutes the candidate meaning. In response to determining that the semantic meaning refutes the candidate meaning, the candidate meaning of the ambiguous decision point is overridden based on the semantic meaning to include the semantic meaning as a final meaning for the ambiguous decision point. The portion of natural language content is then processed based on the final meaning for the ambiguous decision point.
    Type: Grant
    Filed: October 12, 2018
    Date of Patent: March 24, 2020
    Assignee: International Business Machines Corporation
    Inventors: Brendan C. Bull, David Contreras, Paul L. Felt
  • Patent number: 10565994
    Abstract: A method, computer-readable medium, and system including a speech-to-text module to receive an input of speech including one or more words generated by a human and to output data including text, sentiment information, and other parameters corresponding to the speech input; a processing module like Artificial Intelligence to generate a reply to the speech input, the reply including a textual component, sentimental information associated with the textual component, and contextual information associated with the textual component; and a text-to-speech module to receive the textual component, sentimental information, and contextual information and to generate, based on the received textual component and its associated sentimental information and contextual information, a speech output including one or more spoken words, the spoken words to be presented with at least one of a pace, a tone, a volume, and an emphasis representative of the sentimental information and contextual information associated with the textual
    Type: Grant
    Filed: November 30, 2017
    Date of Patent: February 18, 2020
    Assignee: General Electric Company
    Inventors: Ching-Ling Huang, Raju Venkataramana, Yoshifumi Nishida
  • Patent number: 10552534
    Abstract: Aspects of the present invention disclose a method, computer program product, and system for parsing a domain-specific language (DSL) statement. The method includes one or more processors accessing a DSL statement that includes contracted phrases. The method further includes one or more processors identifying one or more contracted phrases in the DSL statement utilizing an annotated domain vocabulary for a DSL associated with the DSL statement and grammar rules for the DSL. The method further includes one or more processors determining expanded phrases corresponding to the identified one or more contracted phrases based on the annotated domain vocabulary and the grammar rules. The method further includes one or more processors creating an expanded abstract syntax tree (AST) that is representative of the DSL statement with the determined expanded phrases replacing the identified one or more contracted phrases.
    Type: Grant
    Filed: January 19, 2017
    Date of Patent: February 4, 2020
    Assignee: International Business Machines Corporation
    Inventors: Jean-Michel G. B. Bernelas, Ulrich M. Junker, Stephane Mery
  • Patent number: 10545512
    Abstract: A method for controlling a drone includes receiving a request for information about a spatial location, generating data requests, configuring a flight plan and controlling one or more drones to fly over the spatial location to obtain data types based on the data requests, and extracting and analyzing data to answer the request. The method can include extracting data points from the data types, obtaining labels from a user for one or more of the data points, predicting labels for unlabeled data points from a learning algorithm using the labels obtained from the user, determining the predicted labels are true labels for the unlabeled data points and combining the extracted data, the user labeled data points and the true labeled data points to answer the request for information. The learning algorithm may be active learning using a support vector machine.
    Type: Grant
    Filed: April 17, 2019
    Date of Patent: January 28, 2020
    Assignee: International Business Machines Corporation
    Inventors: Gregory F. Boland, James R. Kozloski, Yu Ma, Justin G. Manweiler, Kevin E. Siemonsen, Umut Topkara, Katherine Vogt, Justin D. Weisz
  • Patent number: 10546595
    Abstract: Disclosed herein are systems, methods, and computer-readable storage media for improving speech recognition accuracy using textual context. The method includes retrieving a recorded utterance, capturing text from a device display associated with the spoken dialog and viewed by one party to the recorded utterance, and identifying words in the captured text that are relevant to the recorded utterance. The method further includes adding the identified words to a dynamic language model, and recognizing the recorded utterance using the dynamic language model. The recorded utterance can be a spoken dialog. A time stamp can be assigned to each identified word. The method can include adding identified words to and/or removing identified words from the dynamic language model based on their respective time stamps. A screen scraper can capture text from the device display associated with the recorded utterance. The device display can contain customer service data.
    Type: Grant
    Filed: March 5, 2018
    Date of Patent: January 28, 2020
    Assignee: NUANCE COMMUNICATIONS, INC.
    Inventors: Dan Melamed, Srinivas Bangalore, Michael Johnston
  • Patent number: 10540966
    Abstract: A method includes: loading, by a processor, a grammar specification defining at least one parameterizable grammar including a plurality of rules; setting, by the processor, an initial state of a grammar processor as a current state, the current state including parameters supplied to the rules; selecting, by the processor, a rule of the plurality of rules matching the parameters of the current state of the grammar processor; applying, by the processor, the selected rule to the audio and updating the current state; determining, by the processor, whether termination conditions have been met; in response to determining the termination conditions are not met, selecting, by the processor, from the plurality of rules in accordance with parameters of the updated state; and in response to determining the termination conditions are met, outputting, by the processor, a recognizer result of the current state.
    Type: Grant
    Filed: November 2, 2017
    Date of Patent: January 21, 2020
    Inventor: Felix Immanuel Wyss
  • Patent number: 10475438
    Abstract: A text-to-speech (TTS) system that is capable of considering characteristics of various portions of text data in order to create continuity between segments of synthesized speech. The system can analyze text portions of a work and create feature vectors including data corresponding to characteristics of the individual portions and/or the overall work. A TTS processing component can then consider feature vector(s) from other portions when performing TTS processing on text of a first portion, thus giving the TTS component some intelligence regarding other portions of the work, which can then result in more continuity between synthesized speech segments.
    Type: Grant
    Filed: March 2, 2017
    Date of Patent: November 12, 2019
    Assignee: Amazon Technologies, Inc.
    Inventors: Roberto Barra Chicote, Javier Latorre, Adam Franciszek Nadolski, Viacheslav Klimkov, Thomas Edward Merritt
  • Patent number: 10437933
    Abstract: A machine translation system capable of clustering training data and performing dynamic domain adaptation is disclosed. An unsupervised domain clustering process is utilized to identify domains in general training data that can include in-domain training data and out-of-domain training data. Segments in the general training data are then assigned to the domains in order to create domain-specific training data. The domain-specific training data is then utilized to create domain-specific language models, domain-specific translation models, and domain-specific model weights for the domains. An input segment to be translated can be assigned to a domain at translation time. The domain-specific model weights for the assigned domain can be utilized to translate the input segment.
    Type: Grant
    Filed: August 16, 2016
    Date of Patent: October 8, 2019
    Assignee: Amazon Technologies, Inc.
    Inventors: Ann Clifton, Michael Denkowski, Alon Lavie
  • Patent number: 10403276
    Abstract: Techniques for controlling a voice activated feature of a voice activated device are described. Data from one or more sensors and data indicative of a status of a user are received. Based on the analyzing the data, a proximity of the user relative to the location of the voice activated device is determined. One or more voice activated features of the voice activated device are enabled based at least in part on the determined proximity, one or more rules, and one or more user preferences.
    Type: Grant
    Filed: March 17, 2017
    Date of Patent: September 3, 2019
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Suma SaganeGowda, Louis Amadio, Artem Zhurid
  • Patent number: 10395109
    Abstract: According to an embodiment, a recognition apparatus includes one or more processors. The one or more processors are configured to calculate, based on the input signal, a score vector sequence in which a plurality of score vectors each including respective scores of symbols are arranged; and cause, among: a first score vector in which a representative symbol corresponding to a best score is a recognition-target symbol; a second score vector in which a representative symbol is a non-target symbol, and a score of the representative symbol is worse than a first threshold; and a third score vector in which a representative symbol is a non-target symbol, and a score of the representative symbol is equal to the first threshold or better than the first threshold, a third score vector satisfying a predefined first condition, to pass through to filter the score vector sequence.
    Type: Grant
    Filed: August 17, 2017
    Date of Patent: August 27, 2019
    Assignee: Kabushiki Kaisha Toshiba
    Inventor: Manabu Nagao
  • Patent number: 10387717
    Abstract: Embodiments of the present invention disclose a transmission method, where the method includes: starting a session between a terminal and a peer terminal; detecting whether information transmitted in the session satisfies a trigger condition; and if the information satisfies the trigger condition, acquiring an image, and sending the image to the peer terminal, where the information transmitted in the session includes text information or voice information. According to the present invention, information transmission efficiency can be improved.
    Type: Grant
    Filed: July 2, 2014
    Date of Patent: August 20, 2019
    Assignee: Huawei Technologies Co., Ltd.
    Inventors: Xiaojuan Li, Zhu Liu, Wenmei Gao
  • Patent number: 10366163
    Abstract: Systems and methods for determining knowledge-guided information for a recurrent neural networks (RNN) to guide the RNN in semantic tagging of an input phrase are presented. A knowledge encoding module of a Knowledge-Guided Structural Attention Process (K-SAP) receives an input phrase and, in conjunction with additional sub-components or cooperative components generates a knowledge-guided vector that is provided with the input phrase to the RNN for linguistic semantic tagging. Generating the knowledge-guided vector comprises at least parsing the input phrase and generating a corresponding hierarchical linguistic structure comprising one or more discrete sub-structures. The sub-structures may be encoded into vectors along with attention weighting identifying those sub-structures that have greater importance in determining the semantic meaning of the input phrase.
    Type: Grant
    Filed: September 7, 2016
    Date of Patent: July 30, 2019
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Yun-Nung Chen, Dilek Z. Hakkani-Tur, Gokhan Tur, Asli Celikyilmaz, Jianfeng Gao, Li Deng