Patents by Inventor Erik T. Mueller

Erik T. Mueller has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20250124288
    Abstract: The disclosed technology involves autonomously identifying goals and sub-goals from a user utterance and generating responses to the user based on the goals and sub-goals.
    Type: Application
    Filed: December 23, 2024
    Publication date: April 17, 2025
    Inventors: Alexandra Coman, Zachary Kulis, Rui Zhang, Liwei Dai, Erik T. Mueller, Vinay Igure
  • Patent number: 12223423
    Abstract: The disclosed technology involves autonomously identifying goals and sub-goals from a user utterance and generating responses to the user based on the goals and sub-goals.
    Type: Grant
    Filed: October 2, 2020
    Date of Patent: February 11, 2025
    Assignee: CAPITAL ONE SERVICES, LLC
    Inventors: Alexandra Coman, Zachary Kulis, Rui Zhang, Liwei Dai, Erik T. Mueller, Vinay Igure
  • Patent number: 12106058
    Abstract: In a variety of embodiments, machine classifiers may model multi-turn dialogue as a one-to-many prediction task. The machine classifier may be trained using adversarial bootstrapping between a generator and a discriminator with multi-turn capabilities. The machine classifiers may be trained in both auto-regressive and traditional teacher-forcing modes, with the generator including a hierarchical recurrent encoder-decoder network and the discriminator including a bi-directional recurrent neural network. The discriminator input may include a mixture of ground truth labels, the teacher-forcing outputs of the generator, and/or noise data. This mixture of input data may allow for richer feedback on the autoregressive outputs of the generator. The outputs can be ranked based on the discriminator feedback and a response selected from the ranked outputs.
    Type: Grant
    Filed: October 27, 2023
    Date of Patent: October 1, 2024
    Assignee: Capital One Services, LLC
    Inventors: Oluwatobi Olabiyi, Erik T. Mueller
  • Patent number: 12067981
    Abstract: Systems and methods for generating responses to user input such as dialogues, and images are discussed. The system may generate, by a response generation module of at least one server, an optimal generated response to the user communication by applying an generative adversarial network. In some embodiments, the generative adversarial network may include a hierarchical recurrent encoder decoder generative adversarial network including a generator and a discriminator component.
    Type: Grant
    Filed: June 28, 2021
    Date of Patent: August 20, 2024
    Inventors: Oluwatobi Olabiyi, Erik T. Mueller
  • Patent number: 12051409
    Abstract: Systems and methods are provided herein for autonomously determining and resolving a customer's perceived discrepancy during a customer service interaction. The method can include receiving an incoming communication from a customer; extracting a perceived state and an expected state (possibly of a product or service) based on the incoming communication; determining a perceived discrepancy between the perceived and expected states of the customer; retrieving customer information; extracting a current state of the customer from the retrieved customer information, verifying, by a rule-based platform, the discrepancy; generating a response based on the discrepancy after comparing the perceived stated with the current state, where the response may include a confirmation or a correction related to the discrepancy and a personalized explanation describing the current state of the customer; and outputting, for presentation to the customer, the response.
    Type: Grant
    Filed: July 28, 2020
    Date of Patent: July 30, 2024
    Assignee: CAPITAL ONE SERVICES, LLC
    Inventors: Alexandra Coman, Liwei Dai, Erik T. Mueller, Rui Zhang
  • Patent number: 12045711
    Abstract: Memory augmented neural networks may use one or more neural encoders to transform input data into distributed representations and a memory module to store the representations with individual addresses. Memory augmented neural networks allow for few-shot learning capabilities because latent representations are persistent between training examples and gradient-based updates affect only certain memory locations via content-based lookups. When a query vector is not found in memory and the memory is full, existing memories that are positively associated with a particular representation may be identified, redundant memories may be aged, and updated memories may be generated. These updated memories retain relevant information acquired during training and reduce redundancy in the memories stored using the memory module, thereby improving the efficiency of data storage and reducing overfitting of data typically encountered with existing neural networks using memory modules.
    Type: Grant
    Filed: June 16, 2021
    Date of Patent: July 23, 2024
    Assignee: Capital One Services, LLC
    Inventors: Omar Florez Choque, Anish Khazane, Erik T. Mueller
  • Patent number: 12039280
    Abstract: Machine classifiers in accordance with embodiments of the invention capture long-term temporal dependencies in particular tasks, such as turn-based dialogues. Machine classifiers may be used to help users to perform tasks indicated by the user. When a user utterance is received, natural language processing techniques may be used to understand the user's intent. Templates may be determined based on the user's intent in the generation of responses to solicit information from the user. A variety of persona attributes may be determined for a user. The persona attributes may be determined based on the user's utterances and/or provided as metadata included with the user's utterances. A response persona may be used to generate responses to the user's utterances such that the generated responses match a tone appropriate to the task. A response persona may be used to generate templates to solicit additional information and/or generate responses appropriate to the task.
    Type: Grant
    Filed: April 17, 2023
    Date of Patent: July 16, 2024
    Assignee: Capital One Services, LLC
    Inventors: Oluwatobi Olabiyi, Erik T. Mueller, Rui Zhang, Zachary Kulis, Varun Singh
  • Patent number: 12032917
    Abstract: A computing device may execute a conversational agent that may receive language input. The conversational agent may analyze the language input based on configured goals to determine conclusions regarding the language input. The conversational agent may determine whether to modify the truth of one or more of the conclusions, and whether to include or omit the one or more conclusions or modified conclusions in an output response. The conversational agent may also store justifications for including or omitting each conclusion or modified conclusion. The conversational agent may output a response that indicates the conclusions and/or modified conclusions that were selected for output. A user may request that the conversational agent output the justifications for generating the output response. The conversational agent may output the justifications based on receiving the request.
    Type: Grant
    Filed: September 27, 2021
    Date of Patent: July 9, 2024
    Assignee: Capital One Services, LLC
    Inventors: Alexandra Coman, Erik T. Mueller
  • Patent number: 12032910
    Abstract: Systems described herein may use transformer-based machine classifiers to perform a variety of natural language understanding tasks including, but not limited to sentence classification, named entity recognition, sentence similarity, and question answering. The exceptional performance of transformer-based language models is due to their ability to capture long-term temporal dependencies in input sequences. Machine classifiers may be trained using training data sets for multiple tasks, such as but not limited to sentence classification tasks and sequence labeling tasks. Loss masking may be employed in the machine classifier to jointly train the machine classifier on multiple tasks simultaneously. The user of transformer encoders in the machine classifiers, which treat each output sequence independently of other output sequences, in accordance with aspects of the invention do not require joint labeling to model tasks.
    Type: Grant
    Filed: September 26, 2022
    Date of Patent: July 9, 2024
    Assignee: Capital One Services, LLC
    Inventors: Oluwatobi Olabiyi, Erik T. Mueller, Zachary Kulis, Varun Singh
  • Publication number: 20240185063
    Abstract: Aspects described herein may allow for the application of stochastic gradient boosting techniques to the training of deep neural networks by disallowing gradient back propagation from examples that are correctly classified by the neural network model while still keeping correctly classified examples in the gradient averaging. Removing the gradient contribution from correctly classified examples may regularize the deep neural network and prevent the model from overfitting. Further aspects described herein may provide for scheduled boosting during the training of the deep neural network model conditioned on a mini-batch accuracy and/or a number of training iterations. The model training process may start un-boosted, using maximum likelihood objectives or another first loss function.
    Type: Application
    Filed: February 15, 2024
    Publication date: June 6, 2024
    Inventors: Oluwatobi Olabiyi, Erik T. Mueller, Christopher Larson
  • Patent number: 11995523
    Abstract: A method for determining machine learning training parameters is disclosed. The method can include a processor receiving a first input. The processor may receive a first response to the first input, determine a first intent, and identify a first action. The processor can then determine first trainable parameter(s) and determine whether the first trainable parameter(s) is negative or positive. Further, the processor can update a training algorithm based on the first trainable parameter(s). The processor can then receive a second input and determine a second intent for the second input. The processor can also determine a second action for the second intent and transmit the second action to a user. The processor can then determine second trainable parameter(s) and determine whether the second trainable parameter(s) is positive or negative. Finally, the processor can further update the training algorithm based on the second trainable parameter(s).
    Type: Grant
    Filed: January 27, 2021
    Date of Patent: May 28, 2024
    Assignee: CAPITAL ONE SERVICES, LLC
    Inventors: Omar Florez Choque, Erik T. Mueller, Zachary Kulis
  • Publication number: 20240119233
    Abstract: Machine classifiers in accordance with embodiments of the invention capture long-term temporal dependencies in the dialogue data better than the existing RNN-based architectures. Additionally, machine classifiers may model the joint distribution of the context and response as opposed to the conditional distribution of the response given the context as employed in sequence-to-sequence frameworks. Machine classifiers in accordance with embodiments further append random paddings before and/or after the input data to reduce the syntactic redundancy in the input data, thereby improving the performance of the machine classifiers for a variety of dialogue-related tasks. The random padding of the input data may further provide regularization during the training of the machine classifier and/or reduce exposure bias. In a variety of embodiments, the input data may be encoded based on subword tokenization.
    Type: Application
    Filed: October 6, 2023
    Publication date: April 11, 2024
    Inventors: Oluwatobi Olabiyi, Erik T. Mueller, Rui Zhang
  • Patent number: 11941523
    Abstract: Aspects described herein may allow for the application of stochastic gradient boosting techniques to the training of deep neural networks by disallowing gradient back propagation from examples that are correctly classified by the neural network model while still keeping correctly classified examples in the gradient averaging. Removing the gradient contribution from correctly classified examples may regularize the deep neural network and prevent the model from overfitting. Further aspects described herein may provide for scheduled boosting during the training of the deep neural network model conditioned on a mini-batch accuracy and/or a number of training iterations. The model training process may start un-boosted, using maximum likelihood objectives or another first loss function.
    Type: Grant
    Filed: April 16, 2021
    Date of Patent: March 26, 2024
    Assignee: Capital One Services, LLC
    Inventors: Oluwatobi Olabiyi, Erik T. Mueller, Christopher Larson
  • Publication number: 20240054293
    Abstract: In a variety of embodiments, machine classifiers may model multi-turn dialogue as a one-to-many prediction task. The machine classifier may be trained using adversarial bootstrapping between a generator and a discriminator with multi-turn capabilities. The machine classifiers may be trained in both auto-regressive and traditional teacher-forcing modes, with the generator including a hierarchical recurrent encoder-decoder network and the discriminator including a bi-directional recurrent neural network. The discriminator input may include a mixture of ground truth labels, the teacher-forcing outputs of the generator, and/or noise data. This mixture of input data may allow for richer feedback on the autoregressive outputs of the generator. The outputs can be ranked based on the discriminator feedback and a response selected from the ranked outputs.
    Type: Application
    Filed: October 27, 2023
    Publication date: February 15, 2024
    Inventors: Oluwatobi Olabiyi, Erik T. Mueller
  • Publication number: 20240046043
    Abstract: Machine classifiers in accordance with embodiments of the invention capture long-term temporal dependencies in particular tasks, such as turn-based dialogues. Machine classifiers may be used to help users to perform tasks indicated by the user. When a user utterance is received, natural language processing techniques may be used to understand the user's intent. Templates may be determined based on the user's intent in the generation of responses to solicit information from the user. A variety of persona attributes may be determined for a user. The persona attributes may be determined based on the user's utterances and/or provided as metadata included with the user's utterances. A response persona may be used to generate responses to the user's utterances such that the generated responses match a tone appropriate to the task. A response persona may be used to generate templates to solicit additional information and/or generate responses appropriate to the task.
    Type: Application
    Filed: October 5, 2023
    Publication date: February 8, 2024
    Inventors: Oluwatobi Olabiyi, Erik T. Mueller, Diana Mingels, Zachary Kulis
  • Publication number: 20230394245
    Abstract: Systems described herein may use machine classifiers to perform a variety of natural language understanding tasks including, but not limited to multi-turn dialogue generation. Machine classifiers in accordance with aspects of the disclosure may model multi-turn dialogue as a one-to-many prediction task. The machine classifier may be trained using adversarial bootstrapping between a generator and a discriminator with multi-turn capabilities. The machine classifiers may be trained in both auto-regressive and traditional teacher-forcing modes, with the maximum likelihood loss of the auto-regressive outputs being weighted by the score from a metric-based discriminator model. The discriminators input may include a mixture of ground truth labels, the teacher-forcing outputs of the generator, and/or negative examples from the dataset. This mixture of input may allow for richer feedback on the autoregressive outputs of the generator.
    Type: Application
    Filed: August 16, 2023
    Publication date: December 7, 2023
    Inventors: Oluwatobi Olabiyi, Erik T. Mueller
  • Patent number: 11836452
    Abstract: In a variety of embodiments, machine classifiers may model multi-turn dialogue as a one-to-many prediction task. The machine classifier may be trained using adversarial bootstrapping between a generator and a discriminator with multi-turn capabilities. The machine classifiers may be trained in both auto-regressive and traditional teacher-forcing modes, with the generator including a hierarchical recurrent encoder-decoder network and the discriminator including a bi-directional recurrent neural network. The discriminator input may include a mixture of ground truth labels, the teacher-forcing outputs of the generator, and/or noise data. This mixture of input data may allow for richer feedback on the autoregressive outputs of the generator. The outputs can be ranked based on the discriminator feedback and a response selected from the ranked outputs.
    Type: Grant
    Filed: February 27, 2023
    Date of Patent: December 5, 2023
    Assignee: Capital One Services, LLC
    Inventors: Oluwatobi Olabiyi, Erik T. Mueller
  • Patent number: 11816439
    Abstract: Machine classifiers in accordance with embodiments of the invention capture long-term temporal dependencies in particular tasks, such as turn-based dialogues. Machine classifiers may be used to help users to perform tasks indicated by the user. When a user utterance is received, natural language processing techniques may be used to understand the user's intent. Templates may be determined based on the user's intent in the generation of responses to solicit information from the user. A variety of persona attributes may be determined for a user. The persona attributes may be determined based on the user's utterances and/or provided as metadata included with the user's utterances. A response persona may be used to generate responses to the user's utterances such that the generated responses match a tone appropriate to the task. A response persona may be used to generate templates to solicit additional information and/or generate responses appropriate to the task.
    Type: Grant
    Filed: September 22, 2022
    Date of Patent: November 14, 2023
    Assignee: Capital One Services, LLC
    Inventors: Oluwatobi Olabiyi, Erik T. Mueller, Diana Mingels, Zachary Kulis
  • Patent number: 11816442
    Abstract: Machine classifiers in accordance with embodiments of the invention capture long-term temporal dependencies in the dialogue data better than the existing RNN-based architectures. Additionally, machine classifiers may model the joint distribution of the context and response as opposed to the conditional distribution of the response given the context as employed in sequence-to-sequence frameworks. Machine classifiers in accordance with embodiments further append random paddings before and/or after the input data to reduce the syntactic redundancy in the input data, thereby improving the performance of the machine classifiers for a variety of dialogue-related tasks. The random padding of the input data may further provide regularization during the training of the machine classifier and/or reduce exposure bias. In a variety of embodiments, the input data may be encoded based on subword tokenization.
    Type: Grant
    Filed: March 1, 2023
    Date of Patent: November 14, 2023
    Assignee: Capital One Services, LLC
    Inventors: Oluwatobi Olabiyi, Erik T. Mueller, Rui Zhang
  • Patent number: 11775770
    Abstract: Systems described herein may use machine classifiers to perform a variety of natural language understanding tasks including, but not limited to multi-turn dialogue generation. Machine classifiers in accordance with aspects of the disclosure may model multi-turn dialogue as a one-to-many prediction task. The machine classifier may be trained using adversarial bootstrapping between a generator and a discriminator with multi-turn capabilities. The machine classifiers may be trained in both auto-regressive and traditional teacher-forcing modes, with the maximum likelihood loss of the auto-regressive outputs being weighted by the score from a metric-based discriminator model. The discriminators input may include a mixture of ground truth labels, the teacher-forcing outputs of the generator, and/or negative examples from the dataset. This mixture of input may allow for richer feedback on the autoregressive outputs of the generator.
    Type: Grant
    Filed: May 21, 2020
    Date of Patent: October 3, 2023
    Assignee: Capital One Services, LLC
    Inventors: Oluwatobi Olabiyi, Erik T. Mueller