Patents by Inventor Erik T. Mueller
Erik T. Mueller has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20250124288Abstract: The disclosed technology involves autonomously identifying goals and sub-goals from a user utterance and generating responses to the user based on the goals and sub-goals.Type: ApplicationFiled: December 23, 2024Publication date: April 17, 2025Inventors: Alexandra Coman, Zachary Kulis, Rui Zhang, Liwei Dai, Erik T. Mueller, Vinay Igure
-
Patent number: 12223423Abstract: The disclosed technology involves autonomously identifying goals and sub-goals from a user utterance and generating responses to the user based on the goals and sub-goals.Type: GrantFiled: October 2, 2020Date of Patent: February 11, 2025Assignee: CAPITAL ONE SERVICES, LLCInventors: Alexandra Coman, Zachary Kulis, Rui Zhang, Liwei Dai, Erik T. Mueller, Vinay Igure
-
Patent number: 12106058Abstract: In a variety of embodiments, machine classifiers may model multi-turn dialogue as a one-to-many prediction task. The machine classifier may be trained using adversarial bootstrapping between a generator and a discriminator with multi-turn capabilities. The machine classifiers may be trained in both auto-regressive and traditional teacher-forcing modes, with the generator including a hierarchical recurrent encoder-decoder network and the discriminator including a bi-directional recurrent neural network. The discriminator input may include a mixture of ground truth labels, the teacher-forcing outputs of the generator, and/or noise data. This mixture of input data may allow for richer feedback on the autoregressive outputs of the generator. The outputs can be ranked based on the discriminator feedback and a response selected from the ranked outputs.Type: GrantFiled: October 27, 2023Date of Patent: October 1, 2024Assignee: Capital One Services, LLCInventors: Oluwatobi Olabiyi, Erik T. Mueller
-
Patent number: 12067981Abstract: Systems and methods for generating responses to user input such as dialogues, and images are discussed. The system may generate, by a response generation module of at least one server, an optimal generated response to the user communication by applying an generative adversarial network. In some embodiments, the generative adversarial network may include a hierarchical recurrent encoder decoder generative adversarial network including a generator and a discriminator component.Type: GrantFiled: June 28, 2021Date of Patent: August 20, 2024Inventors: Oluwatobi Olabiyi, Erik T. Mueller
-
Patent number: 12051409Abstract: Systems and methods are provided herein for autonomously determining and resolving a customer's perceived discrepancy during a customer service interaction. The method can include receiving an incoming communication from a customer; extracting a perceived state and an expected state (possibly of a product or service) based on the incoming communication; determining a perceived discrepancy between the perceived and expected states of the customer; retrieving customer information; extracting a current state of the customer from the retrieved customer information, verifying, by a rule-based platform, the discrepancy; generating a response based on the discrepancy after comparing the perceived stated with the current state, where the response may include a confirmation or a correction related to the discrepancy and a personalized explanation describing the current state of the customer; and outputting, for presentation to the customer, the response.Type: GrantFiled: July 28, 2020Date of Patent: July 30, 2024Assignee: CAPITAL ONE SERVICES, LLCInventors: Alexandra Coman, Liwei Dai, Erik T. Mueller, Rui Zhang
-
Patent number: 12045711Abstract: Memory augmented neural networks may use one or more neural encoders to transform input data into distributed representations and a memory module to store the representations with individual addresses. Memory augmented neural networks allow for few-shot learning capabilities because latent representations are persistent between training examples and gradient-based updates affect only certain memory locations via content-based lookups. When a query vector is not found in memory and the memory is full, existing memories that are positively associated with a particular representation may be identified, redundant memories may be aged, and updated memories may be generated. These updated memories retain relevant information acquired during training and reduce redundancy in the memories stored using the memory module, thereby improving the efficiency of data storage and reducing overfitting of data typically encountered with existing neural networks using memory modules.Type: GrantFiled: June 16, 2021Date of Patent: July 23, 2024Assignee: Capital One Services, LLCInventors: Omar Florez Choque, Anish Khazane, Erik T. Mueller
-
Patent number: 12039280Abstract: Machine classifiers in accordance with embodiments of the invention capture long-term temporal dependencies in particular tasks, such as turn-based dialogues. Machine classifiers may be used to help users to perform tasks indicated by the user. When a user utterance is received, natural language processing techniques may be used to understand the user's intent. Templates may be determined based on the user's intent in the generation of responses to solicit information from the user. A variety of persona attributes may be determined for a user. The persona attributes may be determined based on the user's utterances and/or provided as metadata included with the user's utterances. A response persona may be used to generate responses to the user's utterances such that the generated responses match a tone appropriate to the task. A response persona may be used to generate templates to solicit additional information and/or generate responses appropriate to the task.Type: GrantFiled: April 17, 2023Date of Patent: July 16, 2024Assignee: Capital One Services, LLCInventors: Oluwatobi Olabiyi, Erik T. Mueller, Rui Zhang, Zachary Kulis, Varun Singh
-
Patent number: 12032917Abstract: A computing device may execute a conversational agent that may receive language input. The conversational agent may analyze the language input based on configured goals to determine conclusions regarding the language input. The conversational agent may determine whether to modify the truth of one or more of the conclusions, and whether to include or omit the one or more conclusions or modified conclusions in an output response. The conversational agent may also store justifications for including or omitting each conclusion or modified conclusion. The conversational agent may output a response that indicates the conclusions and/or modified conclusions that were selected for output. A user may request that the conversational agent output the justifications for generating the output response. The conversational agent may output the justifications based on receiving the request.Type: GrantFiled: September 27, 2021Date of Patent: July 9, 2024Assignee: Capital One Services, LLCInventors: Alexandra Coman, Erik T. Mueller
-
Patent number: 12032910Abstract: Systems described herein may use transformer-based machine classifiers to perform a variety of natural language understanding tasks including, but not limited to sentence classification, named entity recognition, sentence similarity, and question answering. The exceptional performance of transformer-based language models is due to their ability to capture long-term temporal dependencies in input sequences. Machine classifiers may be trained using training data sets for multiple tasks, such as but not limited to sentence classification tasks and sequence labeling tasks. Loss masking may be employed in the machine classifier to jointly train the machine classifier on multiple tasks simultaneously. The user of transformer encoders in the machine classifiers, which treat each output sequence independently of other output sequences, in accordance with aspects of the invention do not require joint labeling to model tasks.Type: GrantFiled: September 26, 2022Date of Patent: July 9, 2024Assignee: Capital One Services, LLCInventors: Oluwatobi Olabiyi, Erik T. Mueller, Zachary Kulis, Varun Singh
-
Publication number: 20240185063Abstract: Aspects described herein may allow for the application of stochastic gradient boosting techniques to the training of deep neural networks by disallowing gradient back propagation from examples that are correctly classified by the neural network model while still keeping correctly classified examples in the gradient averaging. Removing the gradient contribution from correctly classified examples may regularize the deep neural network and prevent the model from overfitting. Further aspects described herein may provide for scheduled boosting during the training of the deep neural network model conditioned on a mini-batch accuracy and/or a number of training iterations. The model training process may start un-boosted, using maximum likelihood objectives or another first loss function.Type: ApplicationFiled: February 15, 2024Publication date: June 6, 2024Inventors: Oluwatobi Olabiyi, Erik T. Mueller, Christopher Larson
-
Patent number: 11995523Abstract: A method for determining machine learning training parameters is disclosed. The method can include a processor receiving a first input. The processor may receive a first response to the first input, determine a first intent, and identify a first action. The processor can then determine first trainable parameter(s) and determine whether the first trainable parameter(s) is negative or positive. Further, the processor can update a training algorithm based on the first trainable parameter(s). The processor can then receive a second input and determine a second intent for the second input. The processor can also determine a second action for the second intent and transmit the second action to a user. The processor can then determine second trainable parameter(s) and determine whether the second trainable parameter(s) is positive or negative. Finally, the processor can further update the training algorithm based on the second trainable parameter(s).Type: GrantFiled: January 27, 2021Date of Patent: May 28, 2024Assignee: CAPITAL ONE SERVICES, LLCInventors: Omar Florez Choque, Erik T. Mueller, Zachary Kulis
-
Publication number: 20240119233Abstract: Machine classifiers in accordance with embodiments of the invention capture long-term temporal dependencies in the dialogue data better than the existing RNN-based architectures. Additionally, machine classifiers may model the joint distribution of the context and response as opposed to the conditional distribution of the response given the context as employed in sequence-to-sequence frameworks. Machine classifiers in accordance with embodiments further append random paddings before and/or after the input data to reduce the syntactic redundancy in the input data, thereby improving the performance of the machine classifiers for a variety of dialogue-related tasks. The random padding of the input data may further provide regularization during the training of the machine classifier and/or reduce exposure bias. In a variety of embodiments, the input data may be encoded based on subword tokenization.Type: ApplicationFiled: October 6, 2023Publication date: April 11, 2024Inventors: Oluwatobi Olabiyi, Erik T. Mueller, Rui Zhang
-
Patent number: 11941523Abstract: Aspects described herein may allow for the application of stochastic gradient boosting techniques to the training of deep neural networks by disallowing gradient back propagation from examples that are correctly classified by the neural network model while still keeping correctly classified examples in the gradient averaging. Removing the gradient contribution from correctly classified examples may regularize the deep neural network and prevent the model from overfitting. Further aspects described herein may provide for scheduled boosting during the training of the deep neural network model conditioned on a mini-batch accuracy and/or a number of training iterations. The model training process may start un-boosted, using maximum likelihood objectives or another first loss function.Type: GrantFiled: April 16, 2021Date of Patent: March 26, 2024Assignee: Capital One Services, LLCInventors: Oluwatobi Olabiyi, Erik T. Mueller, Christopher Larson
-
Publication number: 20240054293Abstract: In a variety of embodiments, machine classifiers may model multi-turn dialogue as a one-to-many prediction task. The machine classifier may be trained using adversarial bootstrapping between a generator and a discriminator with multi-turn capabilities. The machine classifiers may be trained in both auto-regressive and traditional teacher-forcing modes, with the generator including a hierarchical recurrent encoder-decoder network and the discriminator including a bi-directional recurrent neural network. The discriminator input may include a mixture of ground truth labels, the teacher-forcing outputs of the generator, and/or noise data. This mixture of input data may allow for richer feedback on the autoregressive outputs of the generator. The outputs can be ranked based on the discriminator feedback and a response selected from the ranked outputs.Type: ApplicationFiled: October 27, 2023Publication date: February 15, 2024Inventors: Oluwatobi Olabiyi, Erik T. Mueller
-
Publication number: 20240046043Abstract: Machine classifiers in accordance with embodiments of the invention capture long-term temporal dependencies in particular tasks, such as turn-based dialogues. Machine classifiers may be used to help users to perform tasks indicated by the user. When a user utterance is received, natural language processing techniques may be used to understand the user's intent. Templates may be determined based on the user's intent in the generation of responses to solicit information from the user. A variety of persona attributes may be determined for a user. The persona attributes may be determined based on the user's utterances and/or provided as metadata included with the user's utterances. A response persona may be used to generate responses to the user's utterances such that the generated responses match a tone appropriate to the task. A response persona may be used to generate templates to solicit additional information and/or generate responses appropriate to the task.Type: ApplicationFiled: October 5, 2023Publication date: February 8, 2024Inventors: Oluwatobi Olabiyi, Erik T. Mueller, Diana Mingels, Zachary Kulis
-
Publication number: 20230394245Abstract: Systems described herein may use machine classifiers to perform a variety of natural language understanding tasks including, but not limited to multi-turn dialogue generation. Machine classifiers in accordance with aspects of the disclosure may model multi-turn dialogue as a one-to-many prediction task. The machine classifier may be trained using adversarial bootstrapping between a generator and a discriminator with multi-turn capabilities. The machine classifiers may be trained in both auto-regressive and traditional teacher-forcing modes, with the maximum likelihood loss of the auto-regressive outputs being weighted by the score from a metric-based discriminator model. The discriminators input may include a mixture of ground truth labels, the teacher-forcing outputs of the generator, and/or negative examples from the dataset. This mixture of input may allow for richer feedback on the autoregressive outputs of the generator.Type: ApplicationFiled: August 16, 2023Publication date: December 7, 2023Inventors: Oluwatobi Olabiyi, Erik T. Mueller
-
Patent number: 11836452Abstract: In a variety of embodiments, machine classifiers may model multi-turn dialogue as a one-to-many prediction task. The machine classifier may be trained using adversarial bootstrapping between a generator and a discriminator with multi-turn capabilities. The machine classifiers may be trained in both auto-regressive and traditional teacher-forcing modes, with the generator including a hierarchical recurrent encoder-decoder network and the discriminator including a bi-directional recurrent neural network. The discriminator input may include a mixture of ground truth labels, the teacher-forcing outputs of the generator, and/or noise data. This mixture of input data may allow for richer feedback on the autoregressive outputs of the generator. The outputs can be ranked based on the discriminator feedback and a response selected from the ranked outputs.Type: GrantFiled: February 27, 2023Date of Patent: December 5, 2023Assignee: Capital One Services, LLCInventors: Oluwatobi Olabiyi, Erik T. Mueller
-
Patent number: 11816439Abstract: Machine classifiers in accordance with embodiments of the invention capture long-term temporal dependencies in particular tasks, such as turn-based dialogues. Machine classifiers may be used to help users to perform tasks indicated by the user. When a user utterance is received, natural language processing techniques may be used to understand the user's intent. Templates may be determined based on the user's intent in the generation of responses to solicit information from the user. A variety of persona attributes may be determined for a user. The persona attributes may be determined based on the user's utterances and/or provided as metadata included with the user's utterances. A response persona may be used to generate responses to the user's utterances such that the generated responses match a tone appropriate to the task. A response persona may be used to generate templates to solicit additional information and/or generate responses appropriate to the task.Type: GrantFiled: September 22, 2022Date of Patent: November 14, 2023Assignee: Capital One Services, LLCInventors: Oluwatobi Olabiyi, Erik T. Mueller, Diana Mingels, Zachary Kulis
-
Patent number: 11816442Abstract: Machine classifiers in accordance with embodiments of the invention capture long-term temporal dependencies in the dialogue data better than the existing RNN-based architectures. Additionally, machine classifiers may model the joint distribution of the context and response as opposed to the conditional distribution of the response given the context as employed in sequence-to-sequence frameworks. Machine classifiers in accordance with embodiments further append random paddings before and/or after the input data to reduce the syntactic redundancy in the input data, thereby improving the performance of the machine classifiers for a variety of dialogue-related tasks. The random padding of the input data may further provide regularization during the training of the machine classifier and/or reduce exposure bias. In a variety of embodiments, the input data may be encoded based on subword tokenization.Type: GrantFiled: March 1, 2023Date of Patent: November 14, 2023Assignee: Capital One Services, LLCInventors: Oluwatobi Olabiyi, Erik T. Mueller, Rui Zhang
-
Patent number: 11775770Abstract: Systems described herein may use machine classifiers to perform a variety of natural language understanding tasks including, but not limited to multi-turn dialogue generation. Machine classifiers in accordance with aspects of the disclosure may model multi-turn dialogue as a one-to-many prediction task. The machine classifier may be trained using adversarial bootstrapping between a generator and a discriminator with multi-turn capabilities. The machine classifiers may be trained in both auto-regressive and traditional teacher-forcing modes, with the maximum likelihood loss of the auto-regressive outputs being weighted by the score from a metric-based discriminator model. The discriminators input may include a mixture of ground truth labels, the teacher-forcing outputs of the generator, and/or negative examples from the dataset. This mixture of input may allow for richer feedback on the autoregressive outputs of the generator.Type: GrantFiled: May 21, 2020Date of Patent: October 3, 2023Assignee: Capital One Services, LLCInventors: Oluwatobi Olabiyi, Erik T. Mueller