Natural Language Patents (Class 704/257)
  • Patent number: 11922935
    Abstract: A voice interaction method, where a service type set on which a user has a voice interaction intention is predicted based on a target event that can trigger voice interaction, and when a service type of a first service expressed by a voice instruction is a target service type in the service type set, the first service is executed.
    Type: Grant
    Filed: February 19, 2021
    Date of Patent: March 5, 2024
    Assignee: HUAWEI TECHNOLOGIES CO., LTD.
    Inventors: Zijuan Shi, Weiran Nie
  • Patent number: 11900072
    Abstract: Offered is a system that presents on a display screen a translation of a sentence together with an untranslated version of the sentence, and that can cause both of the displayed sentences to break apart into component parts in response to a simple user action, e.g., double-tapping on one of them. When the user selects (e.g., taps on) any portion of either version of the sentence, the system can identify a corresponding portion of the other version (in the other language). In some implementations, a user device can include both a microphone and a display screen, and an automatic speech recognition (ASR) engine can be used to transcribe the user's speech in one language (e.g., English) into text. The system can translate the resulting text into another language (e.g., Spanish) and display the translated text on the display screen along with the untranslated text. When a user selects a portion of a sentence, the system can also present information about the selected portion (e.g.
    Type: Grant
    Filed: July 18, 2017
    Date of Patent: February 13, 2024
    Assignee: Amazon Technologies, Inc.
    Inventors: Lee Michael Bossio, Kartik Suchindra Babu, John Thomas Beck
  • Patent number: 11900931
    Abstract: An information processing apparatus that includes a deciding unit that decides character representation related to speech-to-text conversion of a verbal action performed by a user is provided. The deciding unit decides the character representation for each detected verbal action section based on a detected non-verbal action of the user. Furthermore, an information processing method that causes a processor to execute a process for deciding character representation related to speech-to-text conversion of a verbal action performed by the user is provided. The process further includes a process for deciding the character representation for each detected verbal action section based on a detected non-verbal action performed by the user.
    Type: Grant
    Filed: October 23, 2019
    Date of Patent: February 13, 2024
    Assignee: SONY GROUP CORPORATION
    Inventors: Shinichi Kawano, Ryouhei Yasuda
  • Patent number: 11876758
    Abstract: Systems and methods for configuring a dialogue guidance graph that governs a set of operations of an automated dialogue system and that includes encoding a recall operation to a target graphical node of the graph, wherein the recall operation causes: an accessing of a temporary data storage storing a log of data of an active dialogue session between a user and the automated dialogue system, an assessment of a terminality attribute associated with the target graphical node, a determination of whether the terminality attribute is disabled or enabled, wherein if the terminality attribute is disabled, the recall operation, causes the automated dialogue system to revert the active dialogue session to the target graphical node to perform one or more dialogue-based operations between the user and the automated dialogue system that converts the terminality attribute of the target graphical node from the disabled state to the enabled state.
    Type: Grant
    Filed: September 18, 2023
    Date of Patent: January 16, 2024
    Assignee: Clinc, Inc.
    Inventors: Matthew Mueller, Connor Witt, Jamal El-Mokadem
  • Patent number: 11869491
    Abstract: A speech recognition unit converts an input utterance sequence into a confusion network sequence constituted by a k-best of candidate words of speech recognition results; a lattice generating unit generates a lattice sequence having the candidate words as internal nodes and a combination of k words among the candidate words for an identical speech as an external node, in which edges are extended between internal nodes other than internal nodes included in an identical external node, from the confusion network sequence; an integer programming problem generating unit generates an integer programming problem for selecting a path that maximizes an objective function including at least a coverage score of an important word, of paths following the internal nodes with the edges extended, in the lattice sequence; and the summary generating unit generates a high-quality summary having less speech recognition errors and low redundancy using candidate words indicated by the internal nodes included in the path selected b
    Type: Grant
    Filed: January 16, 2020
    Date of Patent: January 9, 2024
    Assignee: NIPPON TELEGRAPH AND TELEPHONE CORPORATION
    Inventors: Tsutomu Hirao, Atsunori Ogawa, Tomohiro Nakatani, Masaaki Nagata
  • Patent number: 11842727
    Abstract: Multi-modal natural language processing systems are provided. Some systems are context-aware systems that use multi-modal data to improve the accuracy of natural language understanding as it is applied to spoken language input. Machine learning architectures are provided that jointly model spoken language input (“utterances”) and information displayed on a visual display (“on-screen information”). Such machine learning architectures can improve upon, and solve problems inherent in, existing spoken language understanding systems that operate in multi-modal contexts.
    Type: Grant
    Filed: April 18, 2022
    Date of Patent: December 12, 2023
    Assignee: Amazon Technologies, Inc.
    Inventors: Angeliki Metallinou, Rahul Goel, Vishal Ishwar
  • Patent number: 11830497
    Abstract: A multi-tier domain is provided for processing user voice queries and making routing decisions for generating responses, including for user voice queries that include multi-domain trigger words or phrases. When an utterance is recognized as different intents in different domains, a routing system for a domain may consider contextual signals, including those associated with other domains, to determine whether the domain is the proper one to handle the request. This determination can be performed with a statistical model specifically trained to make such determinations using the available contextual data.
    Type: Grant
    Filed: June 24, 2021
    Date of Patent: November 28, 2023
    Assignee: Amazon Technologies, Inc.
    Inventors: Ponnu Jacob, Jingqian Zhao, Prathap Ramachandra, Krupal Maddipati, Jinning Wu, Charlotte Alizerine Dzialo, Daksh Gautam, Wenbo Yan, Liu Yang, Uday Kumar Kollu
  • Patent number: 11830489
    Abstract: A system for determining intent in a voice signal receives a first voice signal that indicates to perform a task. The system sends a first response that comprises a hyperlink associated with a particular webpage used to perform the task. The system receives a second voice signal that indicates whether to access the hyperlink. The system determines intent of the second voice signal by comparing keywords of the second voice signal with keywords of the first response. The system activates the hyperlink in response to determining that the keywords of the second voice signal correspond to the keywords of the first response.
    Type: Grant
    Filed: June 30, 2021
    Date of Patent: November 28, 2023
    Assignee: Bank of America Corporation
    Inventor: Emad Noorizadeh
  • Patent number: 11816246
    Abstract: Modeling techniques to classify data sets containing personal identifiable information (PII) comprising identifiers are provided. In one technique, multiple data sets are identified, each data set containing identifiers that were generated by a computer system and that qualify as PII of a known identifier (ID) type. For each of the multiple data sets, a model is generated based on that data set and added to a set of models. A target data set that contains identifiers that were generated by the computer system and that qualify as PII of an unknown ID type is identified. A target model is generated based on the target data set. For at least one model in the set of models, a similarity operation of that model and the target model is performed. Based on the similarity operation, it is determined whether to associate the ID type of that model with the target data set.
    Type: Grant
    Filed: July 26, 2021
    Date of Patent: November 14, 2023
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Chang-Shing Perng, Vishal Khushal Gosar, Harish Ankam, Haoji Liu
  • Patent number: 11803884
    Abstract: A system is provided for generating compliance manuals from modularized data and taxonomy-based classifications of regulatory obligations. The system comprises a plurality of databases storing regulatory compliance data and a plurality of processors that process the regulatory compliance data to generate business requirements for complying with regulatory obligations and corresponding compliance information related to the business requirements. A taxonomy engine receives business operating parameters related to a first business and identifies a subset of the business requirements and compliance information related to the business operating parameters of the first business. The taxonomy engine further aggregates the business requirements and the compliance information related to the business operation parameters and generates a compliance manual containing the business requirements and the compliance information for use by the first business.
    Type: Grant
    Filed: April 12, 2016
    Date of Patent: October 31, 2023
    Assignee: Ascent Technologies Inc.
    Inventor: Brian T. Clark
  • Patent number: 11804218
    Abstract: This document generally describes systems and methods for dynamically adapting speech recognition for individual voice queries of a user using class-based language models. The method may include receiving a voice query from a user that includes audio data corresponding to an utterance of the user, and context data associated with the user. One or more class models are then generated that collectively identify a first set of terms determined based on the context data, and a respective class to which the respective term is assigned for each respective term in the first set of terms. A language model that includes a residual unigram may then be accessed and processed for each respective class to insert a respective class symbol at each instance of the residual unigram that occurs within the language model. A transcription of the utterance of the user is then generated using the modified language model.
    Type: Grant
    Filed: February 10, 2021
    Date of Patent: October 31, 2023
    Assignee: Google LLC
    Inventors: Justin Max Scheiner, Petar Aleksic
  • Patent number: 11790899
    Abstract: Determining a dialog state of an electronic dialog that includes an automated assistant and at least one user, and performing action(s) based on the determined dialog state. The dialog state can be represented as one or more slots and, for each of the slots, one or more candidate values for the slot and a corresponding score (e.g., a probability) for each of the candidate values. Candidate values for a slot can be determined based on language processing of user utterance(s) and/or system utterance(s) during the dialog. In generating scores for candidate value(s) of a given slot at a given turn of an electronic dialog, various features are determined based on processing of the user utterance and the system utterance using a memory network. The various generated features can be processed using a scoring model to generate scores for candidate value(s) of the given slot at the given turn.
    Type: Grant
    Filed: November 19, 2020
    Date of Patent: October 17, 2023
    Assignee: GOOGLE LLC
    Inventors: Abhinav Rastogi, Larry Paul Heck, Dilek Hakkani-Tur
  • Patent number: 11741955
    Abstract: A method to select a response in a multi-turn conversation between a user and a conversational bot. The conversation is composed of a set of events, wherein an event is a linear sequence of observations that are user speech or physical actions. Queries are processed against a set of conversations that are organized as a set of inter-related data tables, with events and observations stored in distinct tables. As the multi-turn conversation proceeds, a data model comprising an observation history, together with a hierarchy of events determined to represent the conversation up to at least one turn, is persisted. When a new input (speech or physical action) is received, it is classified using a statistical model to generate a result. The result is then mapped to an observation in the data model. Using the mapped observation, a look-up is performed into the data tables to retrieve a possible response.
    Type: Grant
    Filed: February 22, 2021
    Date of Patent: August 29, 2023
    Assignee: Drift.com, Inc.
    Inventors: Jeffrey D. Orkin, Christopher M. Ward
  • Patent number: 11735165
    Abstract: An automatic agent may be improved through feedback. A user input may be received through a user interface. A plurality of current utterance variables may be obtained by tokenizing the user input. The automatic agent may execute a machine learning policy to generate a reply to the user input based on the plurality of current utterance variables. A different reply may be obtained in response to an indication that the reply will lead to a breakdown, wherein the breakdown comprises an unhuman response from the automatic agent according to the machine learning policy. The machine learning policy may be adjusted based on the plurality of current utterance variables and the different reply.
    Type: Grant
    Filed: December 11, 2020
    Date of Patent: August 22, 2023
    Assignee: Beijing DiDi Infinity Technology and Development Co., Ltd.
    Inventor: Scot Fang
  • Patent number: 11727460
    Abstract: Systems and methods for obtaining product information via a conversational user interface. The communication channel receives communication from a user, the intent and entities of which are deduced by the NLP. These are communicated by the fulfillment API to the knowledge engine which retrieves information that fulfills the intent. The information is communicated to the fulfillment API, which converts the intent into a response, which in turn is forwarded by the NLP to the communication channel, and back to the user.
    Type: Grant
    Filed: July 31, 2020
    Date of Patent: August 15, 2023
    Assignee: Kinaxis Inc.
    Inventors: Marcio Oliveira Almeida, Zhen Lin, Casey Bigelow, Liam Meade, Akshatha Mummigatti
  • Patent number: 11727921
    Abstract: A method, a system, and a computer program product for executing intent classification based on user feedback in a digital assistant environment. Using a natural language processor, an audio input received from user is processed. At least one implicit feedback parameter is extracted from the processed audio input. The feedback parameter classifies an intent derived from the audio input received from the user. The extracted feedback parameter is stored in a replay memory. The replay memory stores labeled data associated with the audio input received from the user. Based on the processed audio input and the labeled data, an initial response to the received audio input is determined. Modeling of the extracted implicit feedback parameter and the stored labeled data is executed. An updated response to the received audio input is generated.
    Type: Grant
    Filed: March 29, 2021
    Date of Patent: August 15, 2023
    Assignee: SAP SE
    Inventors: Sebastian Schuetz, Christian Pretzsch, Gil Katz
  • Patent number: 11714857
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for determining, for each of one or more categorical features, a respective vocabulary of categorical feature values of the categorical feature that should be active during processing of inputs by a machine learning model. In one aspect, a method comprises: generating a batch of output sequences, each output sequence in the batch specifying, for each of the categorical features, a respective vocabulary of categorical feature values of the categorical feature that should be active; for each output sequence in the batch, determining a performance metric of the machine learning model on a machine learning task after the machine learning model has been trained to perform the machine learning task with only the respective vocabulary of categorical feature values of each categorical feature specified by the output sequence being active.
    Type: Grant
    Filed: December 7, 2022
    Date of Patent: August 1, 2023
    Assignee: Google LLC
    Inventors: Cong Li, Jay Adams, Manas Joglekar, Pranav Khaitan, Quoc V. Le, Mei Chen
  • Patent number: 11710479
    Abstract: Techniques for implementing a chatbot that utilizes context embeddings are described. An exemplary method includes determining a next turn by: applying a language model to the utterance to determine a probability of a sequence of words, generating a context embedding for the utterance based at least on one or more of: a dialog act as defined by a chatbot definition of the chatbot, a topic vector identifying a domain of the chatbot, a previous chatbot response, and one or more slot options; performing neural language model rescoring using the determined probability of a sequence of words as a word embedding and the generated context embedding to predict an hypothesis; determining at least a name of a slot and type to be fulfilled based at least in part on the hypothesis and the chatbot definition; and determining a next turn based at least in part on the chatbot definition, any previous state, and the name of the slot and type to be fulfilled.
    Type: Grant
    Filed: March 31, 2021
    Date of Patent: July 25, 2023
    Assignee: Amazon Technologies, Inc.
    Inventors: Ashish Vishwanath Shenoy, Sravan Babu Bodapati, Katrin Kirchhoff
  • Patent number: 11710194
    Abstract: Systems, media, and methods for automated response to social queries comprising: monitoring queries from users, each query submitted to a vendor via an interactive chat feature of an external electronic communication platform, monitoring human responses to the queries, monitoring subsequent communications conducted via the electronic communication platform until each query is resolved; applying a first machine learning algorithm to the monitored communications to identify a query susceptible to response automation; applying a second machine learning algorithm to the query susceptible to response automation to identify one or more responses likely to resolve the query; and either i) notifying a human to respond to the query susceptible to response automation with the one or more responses likely to resolve the query, or ii) instantiating an autonomous software agent configured to respond to the query susceptible to response automation with the one or more responses likely to resolve the query.
    Type: Grant
    Filed: August 30, 2017
    Date of Patent: July 25, 2023
    Assignee: LIVEPERSON, INC.
    Inventors: Andrew Thomas Busey, Anthony Dan Chen, Isao Uchida Jonas, Douglas James Daniels, Jr., Benjamin Edward Lamm
  • Patent number: 11705117
    Abstract: Acoustic features are batched into two different batches. The second batch of the two batches is made in response to a detection of a word hypothesis output by a speech recognition network that received the first batch. The number of acoustic feature frames of the second batch is equal to a second batch size greater than the first batch size. The second batch is also to the speech recognition network for processing.
    Type: Grant
    Filed: October 13, 2021
    Date of Patent: July 18, 2023
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Hosam A. Khalil, Emilian Y. Stoimenov, Yifan Gong, Chaojun Liu, Christopher H. Basoglu, Amit K. Agarwal, Naveen Parihar, Sayan Pathak
  • Patent number: 11700311
    Abstract: An electronic apparatus and an operation method thereof for determining a session duration time according to a characteristic of a response message in a chatbot are provided. The electronic apparatus is configured to determine a default session time based on a difficulty level of the response message, determine an additional session time based on conversation history information through the chatbot prior to outputting of the response message, determine a session duration time, which is a sum of the default session time and the additional session time, and provide the session duration time.
    Type: Grant
    Filed: January 26, 2022
    Date of Patent: July 11, 2023
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Hoyoung Kim, Dongil Yang, Jaehyung An, Jiyeon Lee, Cheolseung Jung
  • Patent number: 11615790
    Abstract: Technologies are disclosed for disambiguating contact information using relationship data using a virtual assistant. A user interacts with a virtual assistant to obtain and utilize contact information. For example, a virtual assistant may allow users to perform an action that utilizes contact information (e.g., make a call to a contact). The virtual assistant utilizes a contact service to identify candidate contacts that are related to the requesting user. The contact service identifies candidate contacts based on relationship data between the requesting user and the stored contacts. For example, the relationship data may indicate that the requesting user is on the same project or team as another contact, that the requesting user has the same role as another contact, that the requesting user is a manager of another contact, as well as other attributes. In some examples, the contact service limits the number of candidate contacts provided to requesting user.
    Type: Grant
    Filed: September 30, 2019
    Date of Patent: March 28, 2023
    Assignee: Amazon Technologies, Inc.
    Inventors: Inchara Shivalingaiah, Milo Oostergo, Gary Zhong, Aakarsh Nair, Sushant Bhatia
  • Patent number: 11580960
    Abstract: Exemplary embodiments relate to a system for recovering a conversation between a user and the system when the system is unable to properly respond to a user's input. The system may process the user input and determine an error condition exists. The system may query one or more storage systems to identify candidate text data based on their semantic similarity to the user input. The storage systems may store data related to past frequently entered inputs and/or user-generated inputs. Alternative text data is selected from the candidate text data, and presented to the user for confirmation.
    Type: Grant
    Filed: December 2, 2020
    Date of Patent: February 14, 2023
    Assignee: Amazon Technologies, Inc.
    Inventors: Ravi Chandra Reddy Yasa, Sai Rahul Reddy Pulikunta, Eliav Kahan, Gregory Newell
  • Patent number: 11580094
    Abstract: An audio stream is detected during a communication session with a user. Natural language processing on the audio stream is performed to update a set of attributes by supplementing the set of attributes based on attributes derived from the audio stream. A set of filter values is updated based on the updated set of attributes. The updated set of filter values is used to query a set of databases to obtain datasets. A probabilistic program is executed during the communication session by determining a set of probability parameters characterizing a probability of an anomaly occurring based on the datasets and the set of attributes. A determination is made if whether the probability satisfies a threshold. In response to a determination that the probability satisfies the threshold, a record is updated to identify the communication session to indicate that the threshold is satisfied.
    Type: Grant
    Filed: May 27, 2021
    Date of Patent: February 14, 2023
    Assignee: Capital One Services, LLC
    Inventors: David Beilis, Alexey Shpurov
  • Patent number: 11580977
    Abstract: A conversation engine performs conversations with users using chatbots customized for performing a set of tasks that can be performed using an online system. The conversation engine loads a chatbot configuration that specifies the behavior of a chatbot including the tasks that can be performed by the chatbot, the types of entities relevant to each task, and so on. The conversation may be voice based and use natural language. The conversation engine may load different chatbot configurations to implement different chatbots. The conversation engine receives a conversation engine configuration that specifies the behavior of the conversation engine across chatbots. The system may be a multi-tenant system that allows customization of the chatbots for each tenant.
    Type: Grant
    Filed: September 29, 2020
    Date of Patent: February 14, 2023
    Assignee: Salesforce, Inc.
    Inventors: Xinyi Yang, Tian Xie, Caiming Xiong, Wenhao Liu, Huan Wang, Kazuma Hashimoto, Yingbo Zhou, Xugang Ye, Jin Qu, Feihong Wu
  • Patent number: 11574131
    Abstract: The present disclosure is directed to systems and methods that include and/or leverage one or more machine-learned language models that generate intermediate textual analysis (e.g., including usage of structural tools such as APIs) in service of contextual text generation. For example, a computing system can obtain a contextual text string that includes one or more contextual text tokens. The computing system can process the contextual text string with the machine-learned language model to generate one or more intermediate text strings that include one or more intermediate text tokens. The computing system can process the one or more intermediate text strings with the machine-learned language model to generate an output text string comprising one or more output text tokens. The one or more intermediate text strings can include textual analysis of the contextual text string that supports the output text string.
    Type: Grant
    Filed: May 20, 2022
    Date of Patent: February 7, 2023
    Assignee: GOOGLE LLC
    Inventors: Noam Shazeer, Daniel De Freitas Adiwardana
  • Patent number: 11562737
    Abstract: Speech recognition may be improved by generating and using a topic specific language model. A topic specific language model may be created by performing an initial pass on an audio signal using a generic or basis language model. A speech recognition device may then determine topics relating to the audio signal based on the words identified in the initial pass and retrieve a corpus of text relating to those topics. Using the retrieved corpus of text, the speech recognition device may create a topic specific language model. In one example, the speech recognition device may adapt or otherwise modify the generic language model based on the retrieved corpus of text.
    Type: Grant
    Filed: December 27, 2019
    Date of Patent: January 24, 2023
    Assignee: TIVO CORPORATION
    Inventors: David F. Houghton, Seth Michael Murray, Sibley Verbeck Simon
  • Patent number: 11545139
    Abstract: Systems and methods of script identification in audio data obtained from audio data. The audio data is segmented into a plurality of utterances. A script model representative of a script text is obtained. The plurality of utterances are decoded with the script model. A determination is made if the script text occurred in the audio data.
    Type: Grant
    Filed: February 3, 2020
    Date of Patent: January 3, 2023
    Assignee: VERINT SYSTEMS INC.
    Inventors: Jeffrey Michael Iannone, Ron Wein, Omer Ziv
  • Patent number: 11537664
    Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for determining, for each of one or more categorical features, a respective vocabulary of categorical feature values of the categorical feature that should be active during processing of inputs by a machine learning model. In one aspect, a method comprises: generating a batch of output sequences, each output sequence in the batch specifying, for each of the categorical features, a respective vocabulary of categorical feature values of the categorical feature that should be active; for each output sequence in the batch, determining a performance metric of the machine learning model on a machine learning task after the machine learning model has been trained to perform the machine learning task with only the respective vocabulary of categorical feature values of each categorical feature specified by the output sequence being active.
    Type: Grant
    Filed: May 20, 2020
    Date of Patent: December 27, 2022
    Assignee: Google LLC
    Inventors: Cong Li, Jay Adams, Manas Joglekar, Pranav Khaitan, Quoc V. Le, Mei Chen
  • Patent number: 11527236
    Abstract: Systems and methods of script identification in audio data obtained from audio data. The audio data is segmented into a plurality of utterances. A script model representative of a script text is obtained. The plurality of utterances are decoded with the script model. A determination is made if the script text occurred in the audio data.
    Type: Grant
    Filed: February 3, 2020
    Date of Patent: December 13, 2022
    Assignee: Verint Systems Ltd.
    Inventors: Jeffrey Michael Iannone, Ron Wein, Omer Ziv
  • Patent number: 11514893
    Abstract: Techniques performed by a data processing system for processing voice content received from a user herein include receiving a first audio input from a user comprising spoken content, analyzing the first audio input using one or more natural language processing models to produce a first textual output comprising a textual representation of the first audio input, analyzing the first textual output using one or more machine learning models to determine first context information of the first textual output, and processing the first textual output in the application based on the first context information.
    Type: Grant
    Filed: March 13, 2020
    Date of Patent: November 29, 2022
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Erez Kikin-Gil, Emily Tran, Benjamin David Smith, Alan Liu, Erik Thomas Oveson
  • Patent number: 11488602
    Abstract: A collaborative content management system allows multiple users to access and modify collaborative documents. When audio data is recorded by or uploaded to the system, the audio data may be transcribed or summarized to improve accessibility and user efficiency. Text transcriptions are associated with portions of the audio data representative of the text, and users can search the text transcription and access the portions of the audio data corresponding to search queries for playback. An outline can be automatically generated based on a text transcription of audio data and embedded as a modifiable object within a collaborative document. The system associates hot words with actions to modify the collaborative document upon identifying the hot words in the audio data. Collaborative content management systems can also generate custom lexicons for users based on documents associated with the user for use in transcribing audio data, ensuring that text transcription is more accurate.
    Type: Grant
    Filed: February 20, 2018
    Date of Patent: November 1, 2022
    Assignee: Dropbox, Inc.
    Inventors: Timo Mertens, Bradley Neuberg
  • Patent number: 11468889
    Abstract: A speech recognition platform configured to receive an audio signal that includes speech from a user and perform automatic speech recognition (ASR) on the audio signal to identify ASR results. The platform may identify: (i) a domain of a voice command within the speech based on the ASR results and based on context information associated with the speech or the user, and (ii) an intent of the voice command. In response to identifying the intent, the platform may perform a corresponding action, such as streaming audio to the device, setting a reminder for the user, purchasing an item on behalf of the user, making a reservation for the user or launching an application for the user. The speech recognition platform, in combination with the device, may therefore facilitate efficient interactions between the user and a voice-controlled device.
    Type: Grant
    Filed: March 2, 2020
    Date of Patent: October 11, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Gregory Michael Hart, Peter Paul Henri Carbon, John Daniel Thimsen, Vikram Kumar Gundeti, Scott Ian Blanksteen, Allan Timothy Lindsay, Frederic Johan Georges Deramat
  • Patent number: 11468286
    Abstract: A computerized prediction guided learning method for classification of sequential data performs a prediction learning and a prediction guided learning by a computer program of a computerized machine learning tool. The prediction learning uses an input data sequence to generate an initial classifier. The prediction guided learning may be a semantic learning, an update learning, or an update and semantic learning. The prediction guided semantic learning uses the input data sequence, the initial classifier and semantic label data to generate an output classifier and a semantic classification. The prediction guided update learning uses the input data sequence, the initial classifier and label data to generate an output classifier and a data classification. The prediction guided update and semantic learning uses the input data sequence, the initial classifier and semantic and label data to generate an output classifier, a semantic classification and a data classification.
    Type: Grant
    Filed: May 30, 2017
    Date of Patent: October 11, 2022
    Assignee: Leica Microsystems CMS GmbH
    Inventors: Shih-Jong James Lee, Hideki Sasaki
  • Patent number: 11468694
    Abstract: Disclosed herein are embodiments of an optical character recognition pre-processing software system, which is integrated into a language translation system to provide automated cleaning and correction of noisy and degraded document images to enable seamless and efficient optical character recognition processing and machine translation of information within the document images.
    Type: Grant
    Filed: November 9, 2020
    Date of Patent: October 11, 2022
    Assignee: ARCHITECTURE TECHNOLOGY CORPORATION
    Inventors: Jafar Al-Gharaibeh, William Estey, Paul Nicotera
  • Patent number: 11461681
    Abstract: Methods and systems for multi-modality soft-agents for an enterprise virtual assistant tool are disclosed. An exemplary method comprises capturing, with a computing device, one or more user requests based on at least one multi-modality interaction, populating, with a computing device, soft-queries to access associated data sources and applications, and mining information retrieved by executing at least one populated soft-query. A soft-query is created from user requests. A multi-modality user interface engine annotates the focus of user requests received via text, speech, touch, image, video, or object scanning. A query engine populates queries by identifying the sequence of multi-modal interactions, executes queries and provides results by mining the query results. The multi-modality interactions identify specific inputs for query building and specific parameters associated with the query. A query is populated and used to generate micro-queries associated with the applications involved.
    Type: Grant
    Filed: October 14, 2020
    Date of Patent: October 4, 2022
    Assignee: Openstream Inc.
    Inventor: Rajasekhar Tumuluri
  • Patent number: 11442975
    Abstract: The present invention relates to a method for generating abstractive summary. The method comprises receiving a query for generating an abstractive summary from a document and splitting the query into one or more lexical units. Further, a semantic graph and a graph index is generated based on a role assigned to the one or more lexical units. Furthermore, a measure of information is determined for the retrieved one or more sentences. The one or more sentences having a semantic graph analogous to the generated semantic graph of the query are retrieved from the document. Finally, at least one of re-ordering and re-phrasing is performed on at least one of the retrieved one or more sentences based on the computed measure of information and the one or more lexical units in the retrieved one or more sentences to generate the abstractive summary.
    Type: Grant
    Filed: July 18, 2019
    Date of Patent: September 13, 2022
    Assignee: Wipro Limited
    Inventors: Gopichand Agnihotram, Meenakshi Sundaram Murugeshan
  • Patent number: 11416688
    Abstract: Embodiments described in this disclosure illustrate the use of self-/semi supervised approaches for label-efficient DST in task-oriented dialogue systems. Conversational behavior is modeled by next response generation and turn utterance generation tasks. Prediction consistency is strengthened by augmenting data with stochastic word dropout and label guessing. Experimental results show that by exploiting self-supervision the joint goal accuracy can be boosted with limited labeled data.
    Type: Grant
    Filed: May 8, 2020
    Date of Patent: August 16, 2022
    Assignee: salesforce.com, inc.
    Inventors: Chien-Sheng Wu, Chu Hong Hoi, Caiming Xiong
  • Patent number: 11410650
    Abstract: Methods and systems for processing of voice input to identify intents and mapped standard terminologies are provided. Using natural language processing, an intent of a voice input is identified. The intent is utilized to identify a standard terminology that maps to the intent. The standard terminology is utilized to identify information relevant to the standard terminology in a patient's electronic health record.
    Type: Grant
    Filed: December 19, 2019
    Date of Patent: August 9, 2022
    Assignee: CERNER INNOVATION, INC.
    Inventors: Emin Agassi, Jodi Kodish-Wachs
  • Patent number: 11397859
    Abstract: In an approach to training a corpus link model and generating collocated terms for intra-channel and inter-channel activity, one or more computer processors train a corpus link model based on an analysis of a linguistic corpus and an analysis of one or more author metrics. The one or more computer processors generate one or more collocated terms based on one or more calculations by the trained corpus link model. The one or more computer processors generate a co-occurrence rating for each of the one or more generated collocated terms. The one or more computer processors display the one or more generated collocated terms according to the generated co-occurrence rating of each collocated term.
    Type: Grant
    Filed: September 11, 2019
    Date of Patent: July 26, 2022
    Assignee: International Business Machines Corporation
    Inventors: Jeremy R. Fox, Kelley Anders, Jonathan D. Dunne, Liam S. Harpur
  • Patent number: 11386269
    Abstract: A computer-implemented method for automatically analyzing a natural language input for information extraction comprises (i) a step of receiving the natural language input; (ii) a step of providing a grammar model comprising: a local grammar model, a set of external functions, and a finite set of read/write shared memory registers used by a parsing engine and the external functions; (iii) a step of applying the grammar model to the natural language input using the parsing engine, and (iv) a step of extracting information from the natural language input using at least one new output of the grammar model, the new output of the grammar model being built based: on at least one return value of the external functions from evaluating the one or more external functions in step (iii), and one or more input labels and/or output labels.
    Type: Grant
    Filed: May 11, 2018
    Date of Patent: July 12, 2022
    Assignees: Université Paris-Est Marne-la-Vallée, ESIEE Paris, Chambre de commerce et d'industrie de région Paris Ile de France, Centre National de la Recherche Scientifique, École des ponts ParisTech
    Inventors: Cristian Martinez, Claude Martineau, Antoine Schoen, Tita Kyriacopoulou
  • Patent number: 11366574
    Abstract: Embodiments of the disclosure disclose a human-machine conversation method, a client, an electronic device, and a storage medium. The method comprises: providing a human-machine conversation interface; displaying, upon receiving first input, input information of a user; displaying, in response to the input information, a response result; displaying, upon receiving second input, a function identifier based on the second input, a service function corresponding to the function identifier being configured to respond to the input information; and determining, upon receiving third input, a target function identifier based on the third input, the target function identifier indicating a service function for responding to the input information.
    Type: Grant
    Filed: April 26, 2019
    Date of Patent: June 21, 2022
    Assignee: ALIBABA GROUP HOLDING LIMITED
    Inventor: Ming Hon Chow
  • Patent number: 11356788
    Abstract: Various implementations include systems for rendering 3D audio signals for a vehicle operator. In particular implementations, a method of spatializing audio for an audio device worn on a head of an operator includes: receiving audio data and location data, the location data indicating where the audio data should be rendered relative to the vehicle; tracking a head position of the operator of the vehicle; calculating a compensated location of where the audio data should be rendered based on the location data and the head position of the operator of the vehicle; and spatially rendering, using the audio device, the audio data at the compensated location.
    Type: Grant
    Filed: July 16, 2020
    Date of Patent: June 7, 2022
    Assignee: Bose Corporation
    Inventors: Paul G. Yamkovoy, Matthew Neil Ruwe, Dylan Armand Baranik
  • Patent number: 11355101
    Abstract: Disclosed is an artificial intelligence (AI) apparatus for training an acoustic model, and more particularly, an AI apparatus for training an acoustic model including a shared network and a branch network connected to the shared network using speech data and phonemes corresponding to the speech data.
    Type: Grant
    Filed: March 10, 2020
    Date of Patent: June 7, 2022
    Assignee: LG ELECTRONICS INC.
    Inventor: Jeehye Lee
  • Patent number: 11348573
    Abstract: Systems and processes for operating an intelligent automated assistant are provided. An example process for determining user intent includes receiving a natural language input and detecting an event. The process further includes, determining, at a first time, based on the natural language input, a first value for a first node of a parsing structure; and determining, at a second time, based on the detected data event, a second value for a second node of the parsing structure. The process further includes in accordance with a determination that the first time and the second time are within the predetermined time: determining, using the parsing structure, the first value, and the second value, a user intent associated with the natural language input; initiating a task based on the determined intent; and providing an output indicative of the task.
    Type: Grant
    Filed: June 14, 2019
    Date of Patent: May 31, 2022
    Assignee: Apple Inc.
    Inventors: Pierre P. Greborio, Didier Rene Guzzoni, Philippe P. Piernot
  • Patent number: 11341972
    Abstract: In one aspect, a method comprises accessing audio data generated by a computing device based on audio input from a user, the audio data encoding one or more user utterances. The method further comprises generating a first transcription of the utterances by performing speech recognition on the audio data using a first speech recognizer that employs a language model based on user-specific data. The method further comprises generating a second transcription of the utterances by performing speech recognition on the audio data using a second speech recognizer that employs a language model independent of user-specific data. The method further comprises determining that the second transcription of the utterances includes a term from a predefined set of one or more terms. The method further comprises, based on determining that the second transcription of the utterance includes the term, providing an output of the first transcription of the utterance.
    Type: Grant
    Filed: October 22, 2020
    Date of Patent: May 24, 2022
    Assignee: Google LLC
    Inventors: Alexander H. Gruenstein, Petar Aleksic
  • Patent number: 11335346
    Abstract: Techniques for processing a user input are described. Text data representing a user input is processed with respect to at least one finite state transducer (FST) to generate at least one FST hypothesis. Context information may be required to traverse one or more paths of the at least one FST. The text data is also processed using at least one statistical model (e.g., perform intent classification, named entity recognition, and/or domain classification processing) to generate at least one statistical model hypothesis. The at least one FST hypothesis and the at least one statistical model hypothesis are input to a reranker that determines a most likely interpretation of the user input.
    Type: Grant
    Filed: December 10, 2018
    Date of Patent: May 17, 2022
    Assignee: Amazon Technologies, Inc.
    Inventors: Chengwei Su, Spyridon Matsoukas, Sankaranarayanan Ananthakrishnan, Shirin Saleem, Chungnam Chan, Yugang Li, Mallory McManamon, Rahul Gupta, Luca Soldaini
  • Patent number: 11331800
    Abstract: Apparatus and methods for training and operating of robotic devices. Robotic controller may comprise a predictor apparatus configured to generate motor control output. The predictor may be operable in accordance with a learning process based on a teaching signal comprising the control output. An adaptive controller block may provide control output that may be combined with the predicted control output. The predictor learning process may be configured to learn the combined control signal. Predictor training may comprise a plurality of trials. During initial trial, the control output may be capable of causing a robot to perform a task. During intermediate trials, individual contributions from the controller block and the predictor may be inadequate for the task. Upon learning, the control knowledge may be transferred to the predictor so as to enable task execution in absence of subsequent inputs from the controller. Control output and/or predictor output may comprise multi-channel signals.
    Type: Grant
    Filed: June 22, 2020
    Date of Patent: May 17, 2022
    Assignee: Brain Corporation
    Inventors: Eugene Izhikevich, Oleg Sinyavskiy, Jean-Baptiste Passot
  • Patent number: 11322136
    Abstract: A method includes performing, using at least one processor, feature extraction of input audio data to identify extracted features associated with the input audio data. The method also includes detecting, using the at least one processor, a language associated with each of multiple portions of the input audio data by processing the extracted features using a plurality of language models, where each language model is associated with a different language. In addition, the method includes directing, using the at least one processor, each portion of the input audio data to one of a plurality of automatic speech recognition (ASR) models based on the language associated with the portion of the input audio data.
    Type: Grant
    Filed: December 31, 2019
    Date of Patent: May 3, 2022
    Assignee: Samsung Electronics Co., Ltd.
    Inventors: Vijendra R. Apsingekar, Pu Song, Mohammad M. Moazzami, Asif Ali
  • Patent number: 11321105
    Abstract: Example embodiments described herein relate to an interactive interface system to maintain a user support profile, wherein the user support profile comprises a plurality of media content that includes user support content, receive a message request from a client device wherein the message request includes an identification of the user support profile and corresponding message content that includes a natural language request, convert the natural language request to a query term, perform a query upon the plurality of media content associated with the user support profile based on the query term, identify relevant media content based on the query, and cause display of the relevant media content within a chat interface at the client device. The interfaces generated and displayed by the interactive interface system therefore enable a user to access user support without having to navigate to a separate interface.
    Type: Grant
    Filed: September 22, 2020
    Date of Patent: May 3, 2022
    Assignee: Snap Inc.
    Inventor: Jeremy Voss