Patents by Inventor Tripti Saxena

Tripti Saxena has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11675977
    Abstract: Systems, methods, and apparatuses are presented for a novel natural language tokenizer and tagger. In some embodiments, a method for tokenizing text for natural language processing comprises: generating from a pool of documents, a set of statistical models comprising one or more entries each indicating a likelihood of appearance of a character/letter sequence in the pool of documents; receiving a set of rules comprising rules that identify character/letter sequences as valid tokens; transforming one or more entries in the statistical models into new rules that are added to the set of rules when the entries indicate a high likelihood; receiving a document to be processed; dividing the document to be processed into tokens based on the set of statistical models and the set of rules, wherein the statistical models are applied where the rules fail to unambiguously tokenize the document; and outputting the divided tokens for natural language processing.
    Type: Grant
    Filed: March 27, 2020
    Date of Patent: June 13, 2023
    Assignee: Daash Intelligence, Inc.
    Inventors: Robert J. Munro, Rob Voigt, Schuyler D. Erle, Brendan D. Callahan, Gary C. King, Jessica D. Long, Jason Brenier, Tripti Saxena, Stefan Krawczyk
  • Patent number: 11593458
    Abstract: Implementations are directed to receiving a set of training data including a plurality of data points, at least a portion of which are to be labeled for subsequent supervised training of a computer-executable machine learning (ML) model, providing at least one visualization based on the set of training data, the at least one visualization including a graphical representation of at least a portion of the set of training data, receiving user input associated with the at least one visualization, the user input indicating an action associated with a label assigned to a respective data point in the set of training data, executing a transformation on data points of the set of training data based on one or more heuristics representing the user input to provide labeled training data in a set of labeled training data, and transmitting the set of labeled training data for training the ML model.
    Type: Grant
    Filed: May 21, 2020
    Date of Patent: February 28, 2023
    Assignee: Accenture Global Solutions Limited
    Inventors: Phillip Henry Rogers, Andrew E. Fano, Joshua Neland, Allan Enemark, Tripti Saxena, Jana A. Thompson, David William Vinson
  • Patent number: 11288444
    Abstract: Methods, apparatuses and computer readable medium are presented for generating a natural language model. A method for generating a natural language model comprises: selecting from a pool of documents, a first set of documents to be annotated; receiving annotations of the first set of documents elicited by first human readable prompts; training a natural language model using the annotated first set of documents; determining documents in the pool having uncertain natural language processing results according to the trained natural language model and/or the received annotations; selecting from the pool of documents, a second set of documents to be annotated comprising documents having uncertain natural language processing results; receiving annotations of the second set of documents elicited by second human readable prompts; and retraining a natural language model using the annotated second set of documents.
    Type: Grant
    Filed: December 11, 2020
    Date of Patent: March 29, 2022
    Assignee: 100.co, LLC
    Inventors: Robert J. Munro, Schuyler D. Erle, Jason Brenier, Paul A. Tepper, Tripti Saxena, Gary C. King, Jessica D. Long, Brendan D. Callahan, Tyler J. Schnoebelen, Stefan Krawczyk, Veena Basavaraj
  • Publication number: 20210232762
    Abstract: Systems are presented for generating a natural language model. The system may comprise a database module, an application program interface (API) module, a background processing module, and an applications module, each stored on the at least one memory and executable by the at least one processor. The system may be configured to generate the natural language model by: ingesting training data, generating a hierarchical data structure, selecting a plurality of documents among the training data to be annotated, generating an annotation prompt for each document configured to elicit an annotation about said document, receiving the annotation based on the annotation prompt, and generating the natural language model using an adaptive machine learning process configured to determine patterns among the annotations for how the documents in the training data are to be subdivided according to the at least two topical nodes of the hierarchical data structure.
    Type: Application
    Filed: February 3, 2021
    Publication date: July 29, 2021
    Applicant: Al IP INVESTMENTS LTD
    Inventors: Robert J. Munro, Schuyler D. Erie, Christopher Walker, Sarah K. Luger, Jason Brenier, Gary C. King, Paul A. Tepper, Ross Mechanic, Andrew Gilchrist-Scott, Jessica D. Long, James B. Robinson, Brendan D. Callahan, Michelle Casbon, Ujjwal Sarin, Aneesh Nair, Veena Basavaraj, Tripti Saxena, Edgar Nunez, Martha G. Hinrichs, Haley Most, Tyler J. Schnoebelen
  • Publication number: 20210232760
    Abstract: Methods, apparatuses and computer readable medium are presented for generating a natural language model. A method for generating a natural language model comprises: selecting from a pool of documents, a first set of documents to be annotated; receiving annotations of the first set of documents elicited by first human readable prompts; training a natural language model using the annotated first set of documents; determining documents in the pool having uncertain natural language processing results according to the trained natural language model and/or the received annotations; selecting from the pool of documents, a second set of documents to be annotated comprising documents having uncertain natural language processing results; receiving annotations of the second set of documents elicited by second human readable prompts; and retraining a natural language model using the annotated second set of documents.
    Type: Application
    Filed: December 11, 2020
    Publication date: July 29, 2021
    Inventors: Robert J. Munro, Schuyler D. Erle, Jason Brenier, Paul A. Tepper, Tripti Saxena, Gary C. King, Jessica D. Long, Brendan D. Callahan, Tyler J. Schnoebelen, Stefan Krawczyk, Veena Basavaraj
  • Publication number: 20210157984
    Abstract: Systems, methods, and apparatuses are presented for a novel natural language tokenizer and tagger. In some embodiments, a method for tokenizing text for natural language processing comprises: generating from a pool of documents, a set of statistical models comprising one or more entries each indicating a likelihood of appearance of a character/letter sequence in the pool of documents; receiving a set of rules comprising rules that identify character/letter sequences as valid tokens; transforming one or more entries in the statistical models into new rules that are added to the set of rules when the entries indicate a high likelihood; receiving a document to be processed; dividing the document to be processed into tokens based on the set of statistical models and the set of rules, wherein the statistical models are applied where the rules fail to unambiguously tokenize the document; and outputting the divided tokens for natural language processing.
    Type: Application
    Filed: March 27, 2020
    Publication date: May 27, 2021
    Inventors: Robert J. Munro, Rob Voigt, Schuyler D. Erle, Brendan D. Callahan, Gary C. King, Jessica D. Long, Jason Brenier, Tripti Saxena, Stefan Krawczyk
  • Publication number: 20210150130
    Abstract: Methods are presented for generating a natural language model. The method may comprise: ingesting training data representative of documents to be analyzed by the natural language model, generating a hierarchical data structure comprising at least two topical nodes within which the training data is to be subdivided into by the natural language model, selecting a plurality of documents among the training data to be annotated, generating an annotation prompt for each document configured to elicit an annotation about said document indicating which node among the at least two topical nodes said document is to be classified into, receiving the annotation based on the annotation prompt; and generating the natural language model using an adaptive machine learning process configured to determine patterns among the annotations for how the documents in the training data are to be subdivided according to the at least two topical nodes of the hierarchical data structure.
    Type: Application
    Filed: February 20, 2020
    Publication date: May 20, 2021
    Inventors: Robert J. Munro, Schuyler D. Erle, Christopher Walker, Sarah K. Luger, Jason Brenier, Gary C. King, Paul A. Tepper, Ross Mechanic, Andrew Gilchrist-Scott, Jessica D. Long, James B. Robinson, Brendan D. Callahan, Michelle Casban, Ujjwal Sarin, Aneesh Nair, Veena Basavaraj, Tripti Saxena, Edgar Nunez, Martha G. Hinrichs, Haley Most, Tyler Schnoebelen
  • Publication number: 20200285903
    Abstract: Implementations are directed to receiving a set of training data including a plurality of data points, at least a portion of which are to be labeled for subsequent supervised training of a computer-executable machine learning (ML) model, providing at least one visualization based on the set of training data, the at least one visualization including a graphical representation of at least a portion of the set of training data, receiving user input associated with the at least one visualization, the user input indicating an action associated with a label assigned to a respective data point in the set of training data, executing a transformation on data points of the set of training data based on one or more heuristics representing the user input to provide labeled training data in a set of labeled training data, and transmitting the set of labeled training data for training the ML model.
    Type: Application
    Filed: May 21, 2020
    Publication date: September 10, 2020
    Inventors: Phillip Henry Rogers, Andrew E. Fano, Joshua Neland, Allan Enemark, Tripti Saxena, Jana A. Thompson, David William Vinson
  • Publication number: 20200234002
    Abstract: Methods, apparatuses and computer readable medium are presented for generating a natural language model. A method for generating a natural language model comprises: selecting from a pool of documents, a first set of documents to be annotated; receiving annotations of the first set of documents elicited by first human readable prompts; training a natural language model using the annotated first set of documents; determining documents in the pool having uncertain natural language processing results according to the trained natural language model and/or the received annotations; selecting from the pool of documents, a second set of documents to be annotated comprising documents having uncertain natural language processing results; receiving annotations of the second set of documents elicited by second human readable prompts; and retraining a natural language model using the annotated second set of documents.
    Type: Application
    Filed: November 21, 2018
    Publication date: July 23, 2020
    Inventors: Robert J. Munro, Schuyler D. Erle, Jason Brenier, Paul A. Tepper, Tripti Saxena, Gary C. King, Jessica D. Long, Brendan D. Callahan, Tyler J. Schnoebelen, Stefan Krawczyk, Veena Basavaraj
  • Patent number: 10691976
    Abstract: Implementations are directed to receiving a set of training data including a plurality of data points, at least a portion of which are to be labeled for subsequent supervised training of a computer-executable machine learning (ML) model, providing at least one visualization based on the set of training data, the at least one visualization including a graphical representation of at least a portion of the set of training data, receiving user input associated with the at least one visualization, the user input indicating an action associated with a label assigned to a respective data point in the set of training data, executing a transformation on data points of the set of training data based on one or more heuristics representing the user input to provide labeled training data in a set of labeled training data, and transmitting the set of labeled training data for training the ML model.
    Type: Grant
    Filed: November 16, 2017
    Date of Patent: June 23, 2020
    Assignee: Accenture Global Solutions Limited
    Inventors: Phillip Henry Rogers, Andrew E. Fano, Joshua Neland, Allan Enemark, Tripti Saxena, Jana A. Thompson, David William Vinson
  • Publication number: 20200034737
    Abstract: Systems are presented for generating a natural language model. The system may comprise a database module, an application program interface (API) module, a background processing module, and an applications module, each stored on the at least one memory and executable by the at least one processor. The system may be configured to generate the natural language model by: ingesting training data, generating a hierarchical data structure, selecting a plurality of documents among the training data to be annotated, generating an annotation prompt for each document configured to elicit an annotation about said document, receiving the annotation based on the annotation prompt, and generating the natural language model using an adaptive machine learning process configured to determine patterns among the annotations for how the documents in the training data are to be subdivided according to the at least two topical nodes of the hierarchical data structure.
    Type: Application
    Filed: February 28, 2019
    Publication date: January 30, 2020
    Applicant: AIPARC HOLDINGS PTE. LTD. `
    Inventors: Robert J. Munro, Schuyler D. Erle, Christopher Walker, Sarah K. Luger, Jason Brenier, Gary C. King, Paul A. Tepper, Ross Mechanic, Andrew Gilchrist-Scott, Jessica D. Long, James B. Robinson, Brendan D. Callahan, Michelle Casbon, Ujjwal Sarin, Aneesh Nair, Veena Basavaraj, Tripti Saxena, Edgar Nunez, Martha G. Hinrichs, Haley Most, Tyler J. Schnoebelen
  • Publication number: 20190303428
    Abstract: Methods are presented for generating a natural language model. The method may comprise: ingesting training data representative of documents to be analyzed by the natural language model, generating a hierarchical data structure comprising at least two topical nodes within which the training data is to be subdivided into by the natural language model, selecting a plurality of documents among the training data to be annotated, generating an annotation prompt for each document configured to elicit an annotation about said document indicating which node among the at least two topical nodes said document is to be classified into, receiving the annotation based on the annotation prompt; and generating the natural language model using an adaptive machine learning process configured to determine patterns among the annotations for how the documents in the training data are to be subdivided according to the at least two topical nodes of the hierarchical data structure.
    Type: Application
    Filed: November 5, 2018
    Publication date: October 3, 2019
    Inventors: Robert J. Munro, Schuyler D. Erle, Christopher Walker, Sarah K. Luger, Jason Brenier, Gary C. King, Paul A. Tepper, Ross Mechanic, Andrew Gilchrist-Scott, Jessica D. Long, James B. Robinson, Brendan D. Callahan, Michelle Casbon, Ujjwal Sarin, Aneesh Nair, Veena Basavaraj, Tripti Saxena, Edgar Nunez, Martha G. Hinrichs, Haley Most, Tyler Schnoebelen
  • Publication number: 20190205377
    Abstract: Systems, methods, and apparatuses are presented for a novel natural language tokenizer and tagger. In some embodiments, a method for tokenizing text for natural language processing comprises: generating from a pool of documents, a set of statistical models comprising one or more entries each indicating a likelihood of appearance of a character/letter sequence in the pool of documents; receiving a set of rules comprising rules that identify character/letter sequences as valid tokens; transforming one or more entries in the statistical models into new rules that are added to the set of rules when the entries indicate a high likelihood; receiving a document to be processed; dividing the document to be processed into tokens based on the set of statistical models and the set of rules, wherein the statistical models are applied where the rules fail to unambiguously tokenize the document; and outputting the divided tokens for natural language processing.
    Type: Application
    Filed: August 6, 2018
    Publication date: July 4, 2019
    Applicant: Idibon, Inc.
    Inventors: Robert J. Munro, Rob Voigt, Schuyler D. Erle, Brendan D. Callahan, Gary C. King, Jessica D. Long, Jason Brenier, Tripti Saxena, Stefan Krawczyk
  • Publication number: 20190147297
    Abstract: Implementations are directed to receiving a set of training data including a plurality of data points, at least a portion of which are to be labeled for subsequent supervised training of a computer-executable machine learning (ML) model, providing at least one visualization based on the set of training data, the at least one visualization including a graphical representation of at least a portion of the set of training data, receiving user input associated with the at least one visualization, the user input indicating an action associated with a label assigned to a respective data point in the set of training data, executing a transformation on data points of the set of training data based on one or more heuristics representing the user input to provide labeled training data in a set of labeled training data, and transmitting the set of labeled training data for training the ML model.
    Type: Application
    Filed: November 16, 2017
    Publication date: May 16, 2019
    Inventors: Phillip Henry Rogers, Andrew E. Fano, Joshua Neland, Allan Enemark, Tripti Saxena, Jana A. Thompson, David William Vinson
  • Patent number: 10127214
    Abstract: Methods are presented for generating a natural language model. The method may comprise: ingesting training data representative of documents to be analyzed by the natural language model, generating a hierarchical data structure comprising at least two topical nodes within which the training data is to be subdivided into by the natural language model, selecting a plurality of documents among the training data to be annotated, generating an annotation prompt for each document configured to elicit an annotation about said document indicating which node among the at least two topical nodes said document is to be classified into, receiving the annotation based on the annotation prompt; and generating the natural language model using an adaptive machine learning process configured to determine patterns among the annotations for how the documents in the training data are to be subdivided according to the at least two topical nodes of the hierarchical data structure.
    Type: Grant
    Filed: December 9, 2015
    Date of Patent: November 13, 2018
    Assignee: Sansa Al Inc.
    Inventors: Robert J. Munro, Schuyler D. Erle, Christopher Walker, Sarah K. Luger, Jason Brenier, Gary C. King, Paul A. Tepper, Ross Mechanic, Andrew Gilchrist-Scott, Jessica D. Long, James B. Robinson, Brendan D. Callahan, Michelle Casbon, Ujjwal Sarin, Aneesh Nair, Veena Basavaraj, Tripti Saxena, Edgar Nunez, Martha G. Hinrichs, Haley Most, Tyler J. Schnoebelen
  • Patent number: 9965458
    Abstract: Systems, methods, and apparatuses are presented for a novel natural language tokenizer and tagger. In some embodiments, a method for tokenizing text for natural language processing comprises: generating from a pool of documents, a set of statistical models comprising one or more entries each indicating a likelihood of appearance of a character/letter sequence in the pool of documents; receiving a set of rules comprising rules that identify character/letter sequences as valid tokens; transforming one or more entries in the statistical models into new rules that are added to the set of rules when the entries indicate a high likelihood; receiving a document to be processed; dividing the document to be processed into tokens based on the set of statistical models and the set of rules, wherein the statistical models are applied where the rules fail to unambiguously tokenize the document; and outputting the divided tokens for natural language processing.
    Type: Grant
    Filed: December 9, 2015
    Date of Patent: May 8, 2018
    Assignee: Sansa AI Inc.
    Inventors: Robert J. Munro, Rob Voigt, Schuyler D. Erle, Brendan D. Callahan, Gary C. King, Jessica D. Long, Jason Brenier, Tripti Saxena, Stefan Krawczyk
  • Publication number: 20180095946
    Abstract: Systems, methods, and apparatuses are presented for a novel natural language tokenizer and tagger. In some embodiments, a method for tokenizing text for natural language processing comprises: generating from a pool of documents, a set of statistical models comprising one or more entries each indicating a likelihood of appearance of a character/letter sequence in the pool of documents; receiving a set of rules comprising rules that identify character/letter sequences as valid tokens; transforming one or more entries in the statistical models into new rules that are added to the set of rules when the entries indicate a high likelihood; receiving a document to be processed; dividing the document to be processed into tokens based on the set of statistical models and the set of rules, wherein the statistical models are applied where the rules fail to unambiguously tokenize the document; and outputting the divided tokens for natural language processing.
    Type: Application
    Filed: May 16, 2017
    Publication date: April 5, 2018
    Applicant: Idibon, Inc.
    Inventors: Robert Munro, Rob Voigt, Schuyler D. Erle, Brendan D. Callahan, Gary C. King, Jessica D. Long, Jason Brenier, Tripti Saxena, Stefan Krawczyk
  • Publication number: 20160162456
    Abstract: Methods are presented for generating a natural language model. The method may comprise: ingesting training data representative of documents to be analyzed by the natural language model, generating a hierarchical data structure comprising at least two topical nodes within which the training data is to be subdivided into by the natural language model, selecting a plurality of documents among the training data to be annotated, generating an annotation prompt for each document configured to elicit an annotation about said document indicating which node among the at least two topical nodes said document is to be classified into, receiving the annotation based on the annotation prompt; and generating the natural language model using an adaptive machine learning process configured to determine patterns among the annotations for how the documents in the training data are to be subdivided according to the at least two topical nodes of the hierarchical data structure.
    Type: Application
    Filed: December 9, 2015
    Publication date: June 9, 2016
    Applicant: Idibon, Inc.
    Inventors: Robert J. Munro, Schuyler D. Erle, Christopher Walker, Sarah K. Luger, Jason Brenier, Gary C. King, Paul A. Tepper, Ross Mechanic, Andrew Gilchrist-Scott, Jessica D. Long, James B. Robinson, Brendan D. Callahan, Michelle Casbon, Ujjwal Sarin, Aneesh Nair, Veena Basavaraj, Tripti Saxena, Edgar Nunez, Martha G. Hinrichs, Haley Most, Tyler J. Schnoebelen
  • Publication number: 20160162466
    Abstract: Systems, methods, and apparatuses are presented for a novel natural language tokenizer and tagger. In some embodiments, a method for tokenizing text for natural language processing comprises: generating from a pool of documents, a set of statistical models comprising one or more entries each indicating a likelihood of appearance of a character/letter sequence in the pool of documents; receiving a set of rules comprising rules that identify character/letter sequences as valid tokens; transforming one or more entries in the statistical models into new rules that are added to the set of rules when the entries indicate a high likelihood; receiving a document to be processed; dividing the document to be processed into tokens based on the set of statistical models and the set of rules, wherein the statistical models are applied where the rules fail to unambiguously tokenize the document; and outputting the divided tokens for natural language processing.
    Type: Application
    Filed: December 9, 2015
    Publication date: June 9, 2016
    Applicant: Idibon, Inc.
    Inventors: Robert J. Munro, Rob Voigt, Schuyler D. Erle, Brendan D. Callahan, Gary C. King, Jessica D. Long, Jason Brenier, Tripti Saxena, Stefan Krawczyk
  • Publication number: 20160162457
    Abstract: Methods, apparatuses and computer readable medium are presented for generating a natural language model. A method for generating a natural language model comprises: selecting from a pool of documents, a first set of documents to be annotated; receiving annotations of the first set of documents elicited by first human readable prompts; training a natural language model using the annotated first set of documents; determining documents in the pool having uncertain natural language processing results according to the trained natural language model and/or the received annotations; selecting from the pool of documents, a second set of documents to be annotated comprising documents having uncertain natural language processing results; receiving annotations of the second set of documents elicited by second human readable prompts; and retraining a natural language model using the annotated second set of documents.
    Type: Application
    Filed: December 9, 2015
    Publication date: June 9, 2016
    Applicant: Idibon, Inc.
    Inventors: Robert J. Munro, Schuyler D. Erle, Jason Brenier, Paul A. Tepper, Tripti Saxena, Gary C. King, Jessica D. Long, Brendan D. Callahan, Tyler J. Schnoebelen, Stefan Krawczyk, Veena Basavaraj