Patents by Inventor Guangsen WANG
Guangsen WANG has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20250147887Abstract: Methods and systems are presented for providing multi-tiered cache system that works with an artificial intelligence (AI)-based conversation system for facilitating a conversation with users and processing transactions for the users. The multi-tiered cache system includes multiple tiers of cache modules that use different structures for caching and/or querying data. As a new utterance is received, the cache system uses each of the cache modules in sequence to determine whether a cache hit occurs. If a cache miss occurs at a first cache module, the cache system determines if a cache hit occurs at a second cache module. When a response is obtained from one of the cache modules and/or the AI model, the cache system updates the cache modules using the response.Type: ApplicationFiled: January 7, 2025Publication date: May 8, 2025Inventors: Sharmili Srinivasan, Pranav Ashok Dhakras, Soujanya Lanka, Guangsen Wang, Reyha Verma
-
Publication number: 20250061407Abstract: Methods and systems are presented for providing an artificial intelligence (AI)-based conversation system for facilitating a conversation with users and processing transactions for the users. The AI-based conversation system includes an AI model coupled with different backend modules. Based on an utterance submitted by a user during a chat session, the AI model is configured to generate instructions for a backend module to perform a transaction for the user based on a prompt template. The AI model also communicates the instructions to the backend module using a protocol specified in the prompt template. Upon receiving an output from the backend module, the AI model is configured to generate content for the chat session based on the output, and provide the content to the user.Type: ApplicationFiled: June 28, 2024Publication date: February 20, 2025Inventors: Soujanya Lanka, Guangsen Wang
-
Publication number: 20250061453Abstract: Methods and systems are presented for providing an artificial intelligence (AI)-based conversation system for facilitating a conversation with users and processing transactions for the users. The AI-based conversation system includes an AI model coupled with different backend modules. Based on an utterance submitted by a user during a chat session, the AI model is configured to generate instructions for a backend module to perform a transaction for the user based on a prompt template. The AI model also communicates the instructions to the backend module using a protocol specified in the prompt template. Upon receiving an output from the backend module, the AI model is configured to generate content for the chat session based on the output, and provide the content to the user.Type: ApplicationFiled: June 28, 2024Publication date: February 20, 2025Inventors: Soujanya Lanka, Guangsen Wang, Reyha Verma
-
Publication number: 20250063003Abstract: Methods and systems are presented for providing an artificial intelligence (AI)-based conversation system for facilitating a conversation with users and processing transactions for the users. The AI-based conversation system includes an AI model coupled with different backend modules. Based on an utterance submitted by a user during a chat session, the AI model is configured to generate instructions for a backend module to perform a transaction for the user based on a prompt template. The AI model also communicates the instructions to the backend module using a protocol specified in the prompt template. Upon receiving an output from the backend module, the AI model is configured to generate content for the chat session based on the output, and provide the content to the user.Type: ApplicationFiled: June 28, 2024Publication date: February 20, 2025Inventors: Soujanya Lanka, Guangsen Wang, Reyha Verma
-
Patent number: 12112138Abstract: Embodiments provide a software framework for evaluating and troubleshooting real-world task-oriented bot systems. Specifically, the evaluation framework includes a generator that infers dialog acts and entities from bot definitions and generates test cases for the system via model-based paraphrasing. The framework may also include a simulator for task-oriented dialog user simulation that supports both regression testing and end-to-end evaluation. The framework may also include a remediator to analyze and visualize the simulation results, remedy some of the identified issues, and provide actionable suggestions for improving the task-oriented dialog system.Type: GrantFiled: June 2, 2022Date of Patent: October 8, 2024Assignee: Salesforce, Inc.Inventors: Guangsen Wang, Samson Min Rong Tan, Shafiq Rayhan Joty, Gang Wu, Chu Hong Hoi, Ka Chun Au
-
Patent number: 11972754Abstract: Methods and apparatuses are provided for performing sequence to sequence (Seq2Seq) speech recognition training performed by at least one processor. The method includes acquiring a training set comprising a plurality of pairs of input data and target data corresponding to the input data, encoding the input data into a sequence of hidden states, performing a connectionist temporal classification (CTC) model training based on the sequence of hidden states, performing an attention model training based on the sequence of hidden states, and decoding the sequence of hidden states to generate target labels by independently performing the CTC model training and the attention model training.Type: GrantFiled: December 22, 2021Date of Patent: April 30, 2024Assignee: TENCENT AMERICA LLCInventors: Jia Cui, Chao Weng, Guangsen Wang, Jun Wang, Chengzhu Yu, Dan Su, Dong Yu
-
Patent number: 11803618Abstract: A method and apparatus are provided that analyzing sequence-to-sequence data, such as sequence-to-sequence speech data or sequence-to-sequence machine translation data for example, by minimum Bayes risk (MBR) training a sequence-to-sequence model and within introduction of applications of softmax smoothing to an N-best generation of the MBR training of the sequence-to-sequence model.Type: GrantFiled: November 17, 2022Date of Patent: October 31, 2023Assignee: TENCENT AMERICA LLCInventors: Chao Weng, Jia Cui, Guangsen Wang, Jun Wang, Chengzhu Yu, Dan Su, Dong Yu
-
Patent number: 11798534Abstract: Embodiments described herein provide an Adapt-and-Adjust (A2) mechanism for multilingual speech recognition model that combines both adaptation and adjustment methods as an integrated end-to-end training to improve the models' generalization and mitigate the long-tailed issue. Specifically, a multilingual language model mBERT is utilized, and converted into an autoregressive transformer decoder. In addition, a cross-attention module is added to the encoder on top of the mBERT's self-attention layer in order to explore the acoustic space in addition to the text space. The joint training of the encoder and mBERT decoder can bridge the semantic gap between the speech and the text.Type: GrantFiled: January 29, 2021Date of Patent: October 24, 2023Assignee: salesforce.com, inc.Inventors: Guangsen Wang, Chu Hong Hoi, Genta Indra Winata
-
Publication number: 20230237275Abstract: Embodiments provide a software framework for evaluating and troubleshooting real-world task-oriented bot systems. Specifically, the evaluation framework includes a generator that infers dialog acts and entities from bot definitions and generates test cases for the system via model-based paraphrasing. The framework may also include a simulator for task-oriented dialog user simulation that supports both regression testing and end-to-end evaluation. The framework may also include a remediator to analyze and visualize the simulation results, remedy some of the identified issues, and provide actionable suggestions for improving the task-oriented dialog system.Type: ApplicationFiled: June 2, 2022Publication date: July 27, 2023Inventors: Guangsen Wang, Samson Min Rong Tan, Shafiq Rayhan Joty, Gang Wu, Chu Hong Hoi, Ka Chun Au
-
Publication number: 20230092440Abstract: A method and apparatus are provided that analyzing sequence-to-sequence data, such as sequence-to-sequence speech data or sequence-to-sequence machine translation data for example, by minimum Bayes risk (MBR) training a sequence-to-sequence model and within introduction of applications of softmax smoothing to an N-best generation of the MBR training of the sequence-to-sequence model.Type: ApplicationFiled: November 17, 2022Publication date: March 23, 2023Applicant: TENCENT AMERICA LLCInventors: Chao WENG, Jia CUI, Guangsen WANG, Jun WANG, Chengzhu YU, Dan SU, Dong YU
-
Patent number: 11551136Abstract: A method and apparatus are provided that analyzing sequence-to-sequence data, such as sequence-to-sequence speech data or sequence-to-sequence machine translation data for example, by minimum Bayes risk (MBR) training a sequence-to-sequence model and within introduction of applications of softmax smoothing to an N-best generation of the MBR training of the sequence-to-sequence model.Type: GrantFiled: November 14, 2018Date of Patent: January 10, 2023Assignee: TENCENT AMERICA LLCInventors: Chao Weng, Jia Cui, Guangsen Wang, Jun Wang, Chengzhu Yu, Dan Su, Dong Yu
-
Publication number: 20220115005Abstract: Methods and apparatuses are provided for performing sequence to sequence (Seq2Seq) speech recognition training performed by at least one processor. The method includes acquiring a training set comprising a plurality of pairs of input data and target data corresponding to the input data, encoding the input data into a sequence of hidden states, performing a connectionist temporal classification (CTC) model training based on the sequence of hidden states, performing an attention model training based on the sequence of hidden states, and decoding the sequence of hidden states to generate target labels by independently performing the CTC model training and the attention model training.Type: ApplicationFiled: December 22, 2021Publication date: April 14, 2022Applicant: TENCENT AMERICA LLCInventors: Jia CUI, Chao WENG, Guangsen WANG, Jun WANG, Chengzhu YU, Dan SU, Dong YU
-
Publication number: 20220108688Abstract: Embodiments described herein provide an Adapt-and-Adjust (A2) mechanism for multilingual speech recognition model that combines both adaptation and adjustment methods as an integrated end-to-end training to improve the models' generalization and mitigate the long-tailed issue. Specifically, a multilingual language model mBERT is utilized, and converted into an autoregressive transformer decoder. In addition, a cross-attention module is added to the encoder on top of the mBERT's self-attention layer in order to explore the acoustic space in addition to the text space. The joint training of the encoder and mBERT decoder can bridge the semantic gap between the speech and the text.Type: ApplicationFiled: January 29, 2021Publication date: April 7, 2022Inventors: Guangsen Wang, Chu Hong Hoi, Genta Indra Winata
-
Patent number: 11257481Abstract: Methods and apparatuses are provided for performing sequence to sequence (Seq2Seq) speech recognition training performed by at least one processor. The method includes acquiring a training set comprising a plurality of pairs of input data and target data corresponding to the input data, encoding the input data into a sequence of hidden states, performing a connectionist temporal classification (CTC) model training based on the sequence of hidden states, performing an attention model training based on the sequence of hidden states, and decoding the sequence of hidden states to generate target labels by independently performing the CTC model training and the attention model training.Type: GrantFiled: October 24, 2018Date of Patent: February 22, 2022Assignee: TENCENT AMERICA LLCInventors: Jia Cui, Chao Weng, Guangsen Wang, Jun Wang, Chengzhu Yu, Dan Su, Dong Yu
-
Patent number: 10672382Abstract: Methods and apparatuses are provided for performing end-to-end speech recognition training performed by at least one processor. The method includes receiving, by the at least one processor, one or more input speech frames, generating, by the at least one processor, a sequence of encoder hidden states by transforming the input speech frames, computing, by the at least one processor, attention weights based on each of the sequence of encoder hidden states and a current decoder hidden state, performing, by the at least one processor, a decoding operation based on a previous embedded label prediction information and a previous attentional hidden state information generated based on the attention weights; and generating a current embedded label prediction information based on a result of the decoding operation and the attention weights.Type: GrantFiled: October 15, 2018Date of Patent: June 2, 2020Assignee: TENCENT AMERICA LLCInventors: Chao Weng, Jia Cui, Guangsen Wang, Jun Wang, Chengzhu Yu, Dan Su, Dong Yu
-
Publication number: 20200151623Abstract: A method and apparatus are provided that analyzing sequence-to-sequence data, such as sequence-to-sequence speech data or sequence-to-sequence machine translation data for example, by minimum Bayes risk (MBR) training a sequence-to-sequence model and within introduction of applications of softmax smoothing to an N-best generation of the MBR training of the sequence-to-sequence model.Type: ApplicationFiled: November 14, 2018Publication date: May 14, 2020Applicant: TENCENT America LLCInventors: Chao WENG, Jia CUI, Guangsen WANG, Jun WANG, Chengzhu YU, Dan SU, Dong YU
-
Publication number: 20200135174Abstract: Methods and apparatuses are provided for performing sequence to sequence (Seq2Seq) speech recognition training performed by at least one processor. The method includes acquiring a training set comprising a plurality of pairs of input data and target data corresponding to the input data, encoding the input data into a sequence of hidden states, performing a connectionist temporal classification (CTC) model training based on the sequence of hidden states, performing an attention model training based on the sequence of hidden states, and decoding the sequence of hidden states to generate target labels by independently performing the CTC model training and the attention model training.Type: ApplicationFiled: October 24, 2018Publication date: April 30, 2020Applicant: TENCENT AMERICA LLCInventors: Jia CUI, Chao WENG, Guangsen WANG, Jun WANG, Chengzhu YU, Dan SU, Dong YU
-
Publication number: 20200118547Abstract: Methods and apparatuses are provided for performing end-to-end speech recognition training performed by at least one processor. The method includes receiving, by the at least one processor, one or more input speech frames, generating, by the at least one processor, a sequence of encoder hidden states by transforming the input speech frames, computing, by the at least one processor, attention weights based on each of the sequence of encoder hidden states and a current decoder hidden state, performing, by the at least one processor, a decoding operation based on a previous embedded label prediction information and a previous attentional hidden state information generated based on the attention weights; and generating a current embedded label prediction information based on a result of the decoding operation and the attention weights.Type: ApplicationFiled: October 15, 2018Publication date: April 16, 2020Applicant: TENCENT AMERICA LLCInventors: Chao WENG, Jia Cui, Guangsen WANG, Jun Wang, Chengzhu Yu, Dan Su, Dong Yu