Patents by Inventor Michael R. Glass
Michael R. Glass has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11961513Abstract: A decoder includes a feature extraction circuit for calculating one or more feature vectors. An acoustic model circuit is coupled to receive one or more feature vectors from and assign one or more likelihood values to the one or more feature vectors. A memory architecture that utilizes on-chip state lattices and an off-chip memory for storing states of transition of the decoder is used to reduce reading and writing to the off-chip memory. The on-chip state lattice is populated with at least one of the states of transition stored in the off-chip memory. An on-chip word is generated from a snapshot from the on-chip state lattice. The on-chip state lattice and the on-chip word lattice act as an on-chip cache to reduce reading and writing to the off-chip memory.Type: GrantFiled: July 29, 2021Date of Patent: April 16, 2024Assignee: Massachusetts Institute of TechnologyInventors: Michael R. Price, James R. Glass, Anantha P. Chandrakasan
-
Patent number: 11556712Abstract: Methods and systems for natural language processing include pretraining a machine learning model that is based on a bidirectional encoder representations from transformers model, using a span selection training data set that associates a masked word with a passage. A natural language processing task is performed using the span selection pretrained machine learning model.Type: GrantFiled: October 8, 2019Date of Patent: January 17, 2023Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Michael R. Glass, Alfio Massimiliano Gliozzo
-
Patent number: 11151410Abstract: A computer-implemented method for data labeling is provided. The computer-implemented method assigns pseudo-labels to unlabeled examples of data using a similarity metric on an embedding space to produce pseudo-labeled examples. A curriculum learning model is trained using the pseudo-labeled examples. The curriculum learning model trained with the pseudo-labeled examples is employed in in a fine-tuning task to enhance classification accuracy of the data.Type: GrantFiled: September 7, 2018Date of Patent: October 19, 2021Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Patrick Watson, Bishwaranjan Bhattacharjee, Siyu Huo, Noel C. Codella, Brian M. Belgodere, Parijat Dube, Michael R. Glass, John R. Kender, Matthew L. Hill
-
Publication number: 20210103775Abstract: Methods and systems for natural language processing include pretraining a machine learning model that is based on a bidirectional encoder representations from transformers model, using a span selection training data set that associates a masked word with a passage. A natural language processing task is performed using the span selection pretrained machine learning model.Type: ApplicationFiled: October 8, 2019Publication date: April 8, 2021Inventors: Michael R. Glass, Alfio Massimiliano Gliozzo
-
Publication number: 20200082210Abstract: A computer-implemented method for data labeling is provided. The computer-implemented method assigns pseudo-labels to unlabeled examples of data using a similarity metric on an embedding space to produce pseudo-labeled examples. A curriculum learning model is trained using the pseudo-labeled examples. The curriculum learning model trained with the pseudo-labeled examples is employed in in a fine-tuning task to enhance classification accuracy of the data.Type: ApplicationFiled: September 7, 2018Publication date: March 12, 2020Inventors: Patrick Watson, Bishwaranjan Bhattacharjee, Siyu Huo, Noel C. Codella, Brian M. Belgodere, Parijat Dube, Michael R. Glass, John R. Kender, Matthew L. Hill
-
Patent number: 10445654Abstract: According to an aspect, learning parameters in a feed forward probabilistic graphical model includes creating an inference model via a computer processor. The creation of the inference model includes receiving a training set that includes multiple scenarios, each scenario comprised of one or more natural language statements, and each scenario corresponding to a plurality of candidate answers. The creation also includes constructing evidence graphs for each of the multiple scenarios based on the training set, and calculating weights for common features across the evidence graphs that will maximize a probability of the inference model locating correct answers from corresponding candidate answers across all of the multiple scenarios. In response to an inquiry from a user that includes a scenario, the inference model constructs an evidence graph and recursively constructs formulas to express a confidence of each node in the evidence graph in terms of its parents in the evidence graph.Type: GrantFiled: September 1, 2015Date of Patent: October 15, 2019Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Michael R. Glass, James W. Murdock, IV
-
Patent number: 10430722Abstract: According to an aspect, learning parameters in a feed forward probabilistic graphical model includes creating an inference model via a computer processor. The creation of the inference model includes receiving a training set that includes multiple scenarios, each scenario comprised of one or more natural language statements, and each scenario corresponding to a plurality of candidate answers. The creation also includes constructing evidence graphs for each of the multiple scenarios based on the training set, and calculating weights for common features across the evidence graphs that will maximize a probability of the inference model locating correct answers from corresponding candidate answers across all of the multiple scenarios. In response to an inquiry from a user that includes a scenario, the inference model constructs an evidence graph and recursively constructs formulas to express a confidence of each node in the evidence graph in terms of its parents in the evidence graph.Type: GrantFiled: November 23, 2015Date of Patent: October 1, 2019Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Michael R. Glass, James W. Murdock, IV
-
Patent number: 10339214Abstract: A method, system and computer program product for recognizing terms in a specified corpus. In one embodiment, the method comprises providing a set of known terms t?T, each of the known terms t belonging to a set of types ? (t)={?1, . . . }, wherein each of the terms is comprised of a list of words, t=w1, w2, . . . , wn, and the union of all the words for all the terms is a word set W. The method further comprises using the set of terms T and the set of types to determine a set of pattern-to-type mappings p??; and using the set of pattern-to-type mappings to recognize terms in the specified corpus and, for each of the recognized terms in the specified corpus, to recognize one or more of the types ? for said each recognized term.Type: GrantFiled: November 2, 2012Date of Patent: July 2, 2019Assignee: International Business Machines CorporationInventors: Michael R. Glass, Alfio M. Gliozzo
-
Patent number: 10282421Abstract: Embodiments provide a system and method for short form and long form detection. Using a language-independent process, the detection system can ingest a corpus of documents, pre-process those documents by tokenizing the documents and performing a part-of-speech analysis, and can filter one or more candidate short forms using one or more filters that select for semantic criteria. Semantic criteria can include the part of speech of a token, whether the token contains more than a pre-determined amount of symbols or digits, whether the token appears too frequently in the corpus of documents, and whether the token has at least one uppercase letter. The detection system can detect short forms independent of case and punctuation, and independent of language-specific metaphone variants.Type: GrantFiled: June 29, 2018Date of Patent: May 7, 2019Assignee: International Business Machines CorporationInventors: Md Faisal M. Chowdhury, Michael R. Glass, Alfio M. Gliozzo
-
Patent number: 10261990Abstract: Embodiments provide a system and method for short form and long form detection. Using a language-independent process, the detection system can ingest a corpus of documents, pre-process those documents by tokenizing the documents and performing a part-of-speech analysis, and can filter one or more candidate short forms using one or more filters that select for semantic criteria. Semantic criteria can include the part of speech of a token, whether the token contains more than a pre-determined amount of symbols or digits, whether the token appears too frequently in the corpus of documents, and whether the token has at least one uppercase letter. The detection system can detect short forms independent of case and punctuation, and independent of language-specific metaphone variants.Type: GrantFiled: June 28, 2016Date of Patent: April 16, 2019Assignee: International Business Machines CorporationInventors: Md Faisal M. Chowdhury, Michael R. Glass, Alfio M. Gliozzo
-
Publication number: 20180365210Abstract: Embodiments provide a system and method for short form and long form detection. Given candidate short forms, the system can generate one or more n-gram combinations, resulting in one or more candidate short form and n-gram combination pairs. For each candidate short form and n-gram combination pair, the system can calculate an approximate string matching distance, calculate a best possible alignment score, calculate a confidence score, calculate a topic similarity score, and calculate a semantic similarity score. The system can determine the validity, through a meta learner, of the one or more valid candidate short form and n-gram combination pairs based upon each short form and n-gram combination pair's confidence score, topic similarity score, and semantic similarity score, and store the valid short form and n-gram combination pairs in a repository. The system has no language specific constraints and can extract short form and long form pairs from documents written in various languages.Type: ApplicationFiled: August 22, 2018Publication date: December 20, 2018Inventors: Md Faisal M. Chowdhury, Michael R. Glass, Alfio M. Gliozzo
-
Publication number: 20180307681Abstract: Embodiments provide a system and method for short form and long form detection. Using a language-independent process, the detection system can ingest a corpus of documents, pre-process those documents by tokenizing the documents and performing a part-of-speech analysis, and can filter one or more candidate short forms using one or more filters that select for semantic criteria. Semantic criteria can include the part of speech of a token, whether the token contains more than a pre-determined amount of symbols or digits, whether the token appears too frequently in the corpus of documents, and whether the token has at least one uppercase letter. The detection system can detect short forms independent of case and punctuation, and independent of language-specific metaphone variants.Type: ApplicationFiled: June 29, 2018Publication date: October 25, 2018Inventors: Md Faisal M. Chowdhury, Michael R. Glass, Alfio M. Gliozzo
-
Patent number: 10083170Abstract: Embodiments provide a system and method for short form and long form detection. Given candidate short forms, the system can generate one or more n-gram combinations, resulting in one or more candidate short form and n-gram combination pairs. For each candidate short form and n-gram combination pair, the system can calculate an approximate string matching distance, calculate a best possible alignment score, calculate a confidence score, calculate a topic similarity score, and calculate a semantic similarity score. The system can determine the validity, through a meta learner, of the one or more valid candidate short form and n-gram combination pairs based upon each short form and n-gram combination pair's confidence score, topic similarity score, and semantic similarity score, and store the valid short form and n-gram combination pairs in a repository. The system has no language specific constraints and can extract short form and long form pairs from documents written in various languages.Type: GrantFiled: June 28, 2016Date of Patent: September 25, 2018Assignee: International Business Machines CorporationInventors: Md Faisal M. Chowdhury, Michael R. Glass, Alfio M. Gliozzo
-
Patent number: 9946764Abstract: According to an aspect, a processing system of a question answering computer system determines a first set of relations between one or more pairs of terms in a question. The processing system also determines a second set of relations between one or more pairs of terms in a candidate passage including a candidate answer to the question. The processing system matches the first set of relations to the second set of relations. A plurality of scores is determined by the processing system based on the matching. The processing system aggregates the scores to produce an answer score indicative of a level of support that the candidate answer correctly answers the question.Type: GrantFiled: March 6, 2015Date of Patent: April 17, 2018Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Michael A. Barborak, James J. Fan, Michael R. Glass, Aditya A. Kalyanpur, Adam P. Lally, James W. Murdock, IV, Benjamin P. Segal
-
Patent number: 9946763Abstract: According to an aspect, a processing system of a question answering computer system determines a first set of relations between one or more pairs of terms in a question. The processing system also determines a second set of relations between one or more pairs of terms in a candidate passage including a candidate answer to the question. The processing system matches the first set of relations to the second set of relations. A plurality of scores is determined by the processing system based on the matching. The processing system aggregates the scores to produce an answer score indicative of a level of support that the candidate answer correctly answers the question.Type: GrantFiled: November 5, 2014Date of Patent: April 17, 2018Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Michael A. Barborak, James J. Fan, Michael R. Glass, Aditya A. Kalyanpur, Adam P. Lally, James W. Murdock, IV, Benjamin P. Segal
-
Publication number: 20170371862Abstract: Embodiments provide a system and method for short form and long form detection. Using a language-independent process, the detection system can ingest a corpus of documents, pre-process those documents by tokenizing the documents and performing a part-of-speech analysis, and can filter one or more candidate short forms using one or more filters that select for semantic criteria. Semantic criteria can include the part of speech of a token, whether the token contains more than a pre-determined amount of symbols or digits, whether the token appears too frequently in the corpus of documents, and whether the token has at least one uppercase letter. The detection system can detect short forms independent of case and punctuation, and independent of language-specific metaphone variants.Type: ApplicationFiled: June 28, 2016Publication date: December 28, 2017Inventors: Md Faisal M. Chowdhury, Michael R. Glass, Alfio M. Gliozzo
-
Publication number: 20170371857Abstract: Embodiments provide a system and method for short form and long form detection. Given candidate short forms, the system can generate one or more n-gram combinations, resulting in one or more candidate short form and n-gram combination pairs. For each candidate short form and n-gram combination pair, the system can calculate an approximate string matching distance, calculate a best possible alignment score, calculate a confidence score, calculate a topic similarity score, and calculate a semantic similarity score. The system can determine the validity, through a meta learner, of the one or more valid candidate short form and n-gram combination pairs based upon each short form and n-gram combination pair's confidence score, topic similarity score, and semantic similarity score, and store the valid short form and n-gram combination pairs in a repository. The system has no language specific constraints and can extract short form and long form pairs from documents written in various languages.Type: ApplicationFiled: June 28, 2016Publication date: December 28, 2017Inventors: Md Faisal M. Chowdhury, Michael R. Glass, Alfio M. Gliozzo
-
Publication number: 20170061301Abstract: According to an aspect, learning parameters in a feed forward probabilistic graphical model includes creating an inference model via a computer processor. The creation of the inference model includes receiving a training set that includes multiple scenarios, each scenario comprised of one or more natural language statements, and each scenario corresponding to a plurality of candidate answers. The creation also includes constructing evidence graphs for each of the multiple scenarios based on the training set, and calculating weights for common features across the evidence graphs that will maximize a probability of the inference model locating correct answers from corresponding candidate answers across all of the multiple scenarios. In response to an inquiry from a user that includes a scenario, the inference model constructs an evidence graph and recursively constructs formulas to express a confidence of each node in the evidence graph in terms of its parents in the evidence graph.Type: ApplicationFiled: November 23, 2015Publication date: March 2, 2017Inventors: Michael R. Glass, James W. Murdock, IV
-
Publication number: 20170061324Abstract: According to an aspect, learning parameters in a feed forward probabilistic graphical model includes creating an inference model via a computer processor. The creation of the inference model includes receiving a training set that includes multiple scenarios, each scenario comprised of one or more natural language statements, and each scenario corresponding to a plurality of candidate answers. The creation also includes constructing evidence graphs for each of the multiple scenarios based on the training set, and calculating weights for common features across the evidence graphs that will maximize a probability of the inference model locating correct answers from corresponding candidate answers across all of the multiple scenarios. In response to an inquiry from a user that includes a scenario, the inference model constructs an evidence graph and recursively constructs formulas to express a confidence of each node in the evidence graph in terms of its parents in the evidence graph.Type: ApplicationFiled: September 1, 2015Publication date: March 2, 2017Inventors: Michael R. Glass, James W. Murdock, IV
-
Publication number: 20160124962Abstract: According to an aspect, a processing system of a question answering computer system determines a first set of relations between one or more pairs of terms in a question. The processing system also determines a second set of relations between one or more pairs of terms in a candidate passage including a candidate answer to the question. The processing system matches the first set of relations to the second set of relations. A plurality of scores is determined by the processing system based on the matching. The processing system aggregates the scores to produce an answer score indicative of a level of support that the candidate answer correctly answers the question.Type: ApplicationFiled: November 5, 2014Publication date: May 5, 2016Inventors: Michael A. Barborak, James J. Fan, Michael R. Glass, Aditya A. Kalyanpur, Adam P. Lally, James W. Murdock, IV, Benjamin P. Segal