Patents by Inventor Yun-Nung Chen
Yun-Nung Chen has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20230401445Abstract: A processing unit can train a model as a joint multi-domain recurrent neural network (JRNN), such as a bi-directional recurrent neural network (bRNN) and/or a recurrent neural network with long-short term memory (RNN-LSTM) for spoken language understanding (SLU). The processing unit can use the trained model to, e.g., jointly model slot filling, intent determination, and domain classification. The joint multi-domain model described herein can estimate a complete semantic frame per query, and the joint multi-domain model enables multi-task deep learning leveraging the data from multiple domains. The joint multi-domain recurrent neural (JRNN) can leverage semantic intents (such as, finding or identifying, e.g., a domain specific goal) and slots (such as, dates, times, locations, subjects, etc.) across multiple domains.Type: ApplicationFiled: August 29, 2023Publication date: December 14, 2023Inventors: Dilek Z. Hakkani-Tur, Asli Celikyilmaz, Yun-Nung Chen, Li Deng, Jianfeng Gao, Gokhan Tur, Ye Yi Wang
-
Patent number: 11783173Abstract: A processing unit can train a model as a joint multi-domain recurrent neural network (JRNN), such as a bi-directional recurrent neural network (bRNN) and/or a recurrent neural network with long-short term memory (RNN-LSTM) for spoken language understanding (SLU). The processing unit can use the trained model to, e.g., jointly model slot filling, intent determination, and domain classification. The joint multi-domain model described herein can estimate a complete semantic frame per query, and the joint multi-domain model enables multi-task deep learning leveraging the data from multiple domains. The joint multi-domain recurrent neural network (JRNN) can leverage semantic intents (such as, finding or identifying, e.g., a domain specific goal) and slots (such as, dates, times, locations, subjects, etc.) across multiple domains.Type: GrantFiled: August 4, 2016Date of Patent: October 10, 2023Assignee: Microsoft Technology Licensing, LLCInventors: Dilek Z Hakkani-Tur, Asli Celikyilmaz, Yun-Nung Chen, Li Deng, Jianfeng Gao, Gokhan Tur, Ye-Yi Wang
-
Patent number: 11449744Abstract: A processing unit can extract salient semantics to model knowledge carryover, from one turn to the next, in multi-turn conversations. Architecture described herein can use the end-to-end memory networks to encode inputs, e.g., utterances, with intents and slots, which can be stored as embeddings in memory, and in decoding the architecture can exploit latent contextual information from memory, e.g., demographic context, visual context, semantic context, etc. e.g., via an attention model, to leverage previously stored semantics for semantic parsing, e.g., for joint intent prediction and slot tagging. In examples, architecture is configured to build an end-to-end memory network model for contextual, e.g., multi-turn, language understanding, to apply the end-to-end memory network model to multiple turns of conversational input; and to fill slots for output of contextual, e.g., multi-turn, language understanding of the conversational input.Type: GrantFiled: August 4, 2016Date of Patent: September 20, 2022Assignee: Microsoft Technology Licensing, LLCInventors: Yun-Nung Chen, Dilek Z. Hakkani-Tur, Gokhan Tur, Li Deng, Jianfeng Gao
-
Patent number: 10546066Abstract: Described herein are systems, methods, and techniques by which a processing unit can build an end-to-end dialogue agent model for end-to-end learning of dialogue agents for information access and apply the end-to-end dialogue agent model with soft attention over knowledge base entries to make the dialogue system differentiable. In various examples the processing unit can apply the end-to-end dialogue agent model to a source of input, fill slots for output from the knowledge base entries, induce a posterior distribution over the entities in a knowledge base or induce a posterior distribution of a target of the requesting user over entities from a knowledge base, develop an end-to-end differentiable model of a dialogue agent, use supervised and/or imitation learning to initialize network parameters, calculate a modified version of an episodic algorithm. e.g., the REINFORCE algorithm, for training an end-to-end differentiable model based on user feedback.Type: GrantFiled: January 13, 2017Date of Patent: January 28, 2020Assignee: Microsoft Technology Licensing, LLCInventors: Lihong Li, Bhuwan Dhingra, Jianfeng Gao, Xiujun Li, Yun-Nung Chen, Li Deng, Faisal Ahmed
-
Patent number: 10366163Abstract: Systems and methods for determining knowledge-guided information for a recurrent neural networks (RNN) to guide the RNN in semantic tagging of an input phrase are presented. A knowledge encoding module of a Knowledge-Guided Structural Attention Process (K-SAP) receives an input phrase and, in conjunction with additional sub-components or cooperative components generates a knowledge-guided vector that is provided with the input phrase to the RNN for linguistic semantic tagging. Generating the knowledge-guided vector comprises at least parsing the input phrase and generating a corresponding hierarchical linguistic structure comprising one or more discrete sub-structures. The sub-structures may be encoded into vectors along with attention weighting identifying those sub-structures that have greater importance in determining the semantic meaning of the input phrase.Type: GrantFiled: September 7, 2016Date of Patent: July 30, 2019Assignee: Microsoft Technology Licensing, LLCInventors: Yun-Nung Chen, Dilek Z. Hakkani-Tur, Gokhan Tur, Asli Celikyilmaz, Jianfeng Gao, Li Deng
-
Patent number: 10268679Abstract: A processing unit can operate an end-to-end recurrent neural network (RNN) with limited contextual dialog memory that can be jointly trained by supervised signals-user slot tagging, intent prediction and/or system action prediction. The end-to-end RNN, or joint model has shown advantages over separate models for natural language understanding (NLU) and dialog management and can capture expressive feature representations beyond conventional aggregation of slot tags and intents, to mitigate effects of noisy output from NLU. The joint model can apply a supervised signal from system actions to refine the NLU model. By back-propagating errors associated with system action prediction to the NLU model, the joint model can use machine learning to predict user intent by a binary classification obtained by both forward and backward output, and perform slot tagging, and make system action predictions based on user input, e.g., utterances across a number of domains.Type: GrantFiled: December 2, 2016Date of Patent: April 23, 2019Assignee: MICROSOFT TECHNOLOGY LICENSING, LLCInventors: Xiujun Li, Paul Anthony Crook, Li Deng, Jianfeng Gao, Yun-Nung Chen, Xuesong Yang
-
Patent number: 10262654Abstract: A computer-implemented technique is described herein for detecting actionable items in speech. In one manner of operation, the technique can include receiving utterance information that expresses at least one utterance made by one participant of a conversation to at least one other participant of the conversation. The technique can also include converting the utterance information into recognized speech information and using a machine-trained model to recognize at least one actionable item associated with the recognized speech information. The technique can also include performing at least one computer-implemented action associated the actionable item(s). The machine-trained model may correspond to a deep-structured convolutional neural network. The technique can produce the machine-trained model using a source environment corpus that is not optimally suited for a target environment in which the model is intended to be applied.Type: GrantFiled: September 24, 2015Date of Patent: April 16, 2019Assignee: Microsoft Technology Licensing, LLCInventors: Dilek Zeynep Hakkani-Tur, Xiaodong He, Yun-Nung Chen
-
Publication number: 20180157638Abstract: A processing unit can operate an end-to-end recurrent neural network (RNN) with limited contextual dialogue memory that can be jointly trained by supervised signals—user slot tagging, intent prediction and/or system action prediction. The end-to-end RNN, or joint model has shown advantages over separate models for natural language understanding (NLU) and dialogue management and can capture expressive feature representations beyond conventional aggregation of slot tags and intents, to mitigate effects of noisy output from NLU. The joint model can apply a supervised signal from system actions to refine the NLU model. By back-propagating errors associated with system action prediction to the NLU model, the joint model can use machine learning to predict user intent, and perform slot tagging, and make system action predictions based on user input, e.g., utterances across a number of domains.Type: ApplicationFiled: December 2, 2016Publication date: June 7, 2018Inventors: Xiujun Li, Paul Anthony Crook, Li Deng, Jianfeng Gao, Yun-Nung Chen, Xuesong Yang
-
Publication number: 20180067923Abstract: Systems and methods for determining knowledge-guided information for a recurrent neural networks (RNN) to guide the RNN in semantic tagging of an input phrase are presented. A knowledge encoding module of a Knowledge-Guided Structural Attention Process (K-SAP) receives an input phrase and, in conjunction with additional sub-components or cooperative components generates a knowledge-guided vector that is provided with the input phrase to the RNN for linguistic semantic tagging. Generating the knowledge-guided vector comprises at least parsing the input phrase and generating a corresponding hierarchical linguistic structure comprising one or more discrete sub-structures. The sub-structures may be encoded into vectors along with attention weighting identifying those sub-structures that have greater importance in determining the semantic meaning of the input phrase.Type: ApplicationFiled: September 7, 2016Publication date: March 8, 2018Inventors: Yun-Nung Chen, Dilek Z. Hakkani-Tur, Gokhan Tur, Asli Celikyilmaz, Jianfeng Gao, Li Deng
-
Publication number: 20180060301Abstract: Described herein are systems, methods, and techniques by which a processing unit can build an end-to-end dialogue agent model for end-to-end learning of dialogue agents for information access and apply the end-to-end dialogue agent model with soft attention over knowledge base entries to make the dialogue system differentiable. In various examples the processing unit can apply the end-to-end dialogue agent model to a source of input, fill slots for output from the knowledge base entries, induce a posterior distribution over the entities in a knowledge base or induce a posterior distribution of a target of the requesting user over entities from a knowledge base, develop an end-to-end differentiable model of a dialogue agent, use supervised and/or imitation learning to initialize network parameters, calculate a modified version of an episodic algorithm, e.g., the REINFORCE algorithm, for training an end-to-end differentiable model based on user feedback.Type: ApplicationFiled: January 13, 2017Publication date: March 1, 2018Inventors: Lihong Li, Bhuwan Dhingra, Jianfeng Gao, Xiujun Li, Yun-Nung Chen, Li Deng, Faisal Ahmed
-
Publication number: 20170372199Abstract: A processing unit can train a model as a joint multi-domain recurrent neural network (JRNN), such as a bi-directional recurrent neural network (bRNN) and/or a recurrent neural network with long-short term memory (RNN-LSTM) for spoken language understanding (SLU). The processing unit can use the trained model to, e.g., jointly model slot filling, intent determination, and domain classification. The joint multi-domain model described herein can estimate a complete semantic frame per query, and the joint multi-domain model enables multi-task deep learning leveraging the data from multiple domains. The joint multi-domain recurrent neural network (JRNN) can leverage semantic intents (such as, finding or identifying, e.g., a domain specific goal) and slots (such as, dates, times, locations, subjects, etc.) across multiple domains.Type: ApplicationFiled: August 4, 2016Publication date: December 28, 2017Inventors: Dilek Z Hakkani-Tur, Asli Celikyilmaz, Yun-Nung Chen, Li Deng, Jianfeng Gao, Gokhan Tur, Ye-Yi Wang
-
Publication number: 20170372200Abstract: A processing unit can extract salient semantics to model knowledge carryover, from one turn to the next, in multi-turn conversations. Architecture described herein can use the end-to-end memory networks to encode inputs, e.g., utterances, with intents and slots, which can be stored as embeddings in memory, and in decoding the architecture can exploit latent contextual information from memory, e.g., demographic context, visual context, semantic context, etc. e.g., via an attention model, to leverage previously stored semantics for semantic parsing, e.g., for joint intent prediction and slot tagging. In examples, architecture is configured to build an end-to-end memory network model for contextual, e.g., multi-turn, language understanding, to apply the end-to-end memory network model to multiple turns of conversational input; and to fill slots for output of contextual, e.g., multi-turn, language understanding of the conversational input.Type: ApplicationFiled: August 4, 2016Publication date: December 28, 2017Inventors: Yun-Nung Chen, Dilek Z. Hakkani-Tur, Gokhan Tur, Li Deng, Jianfeng Gao
-
Publication number: 20170092264Abstract: A computer-implemented technique is described herein for detecting actionable items in speech. In one manner of operation, the technique entails: receiving utterance information that expresses at least one utterance made by one participant of a conversation to at least one other participant of the conversation; converting the utterance information into recognized speech information; using a machine-trained model to recognize at least one actionable item associated with the recognized speech information; and performing at least one computer-implemented action associated the actionable item(s). The machine-trained model may correspond to a deep-structured convolutional neural network. In some implementations, the technique produces the machine-trained model using a source environment corpus that is not optimally suited for a target environment in which the model is intended to be applied.Type: ApplicationFiled: September 24, 2015Publication date: March 30, 2017Inventors: Dilek Zeynep Hakkani-Tur, Xiaodong He, Yun-Nung Chen