Patents by Inventor Jimmie Goode
Jimmie Goode has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 12271413Abstract: A recurrent neural network (RNN) method implemented on a computer system is used to produce summaries of unstructured text generated by multiple networks of individuals interacting over time by encoding the unstructured text into intermediate representations and decoding the intermediate representations into summaries of each network. Parameter data for the RNN is obtained by using multiple different versions of the same source texts to train the computer system. The method and computer system can be used to identify which of the networks match a query by determining which network generates the query with low or lowest cost.Type: GrantFiled: February 28, 2022Date of Patent: April 8, 2025Assignee: Pulselight Holdings, Inc.Inventors: Jonathan William Mugan, Laura Hitt, Jimmie Goode, Russ Gregory, Yuan Qu
-
Patent number: 12051002Abstract: A method includes receiving, by a processor, an input data set. The input data set includes a plurality of features. The method includes determining, by the processor, one or more characteristics of the input data set. The method includes, based on the one or more characteristics, adjusting, by the processor, one or more architectural parameters of an automated model generation process. The automated model generation process is configured to generate a plurality of models using a weighted randomization process. The one or more architectural parameters weight the weighted randomization process to adjust a probability of generation of models having particular architectural features. The method further includes executing, by the processor, the automated model generation process to output a mode, the model including data representative of a neural network.Type: GrantFiled: April 14, 2020Date of Patent: July 30, 2024Assignee: SPARKCOGNITION, INC.Inventors: Tyler S. McDonnell, Sari Andoni, Junhwan Choi, Jimmie Goode, Yiyun Lan, Keith D. Moore, Gavin Sellers
-
Publication number: 20230186053Abstract: A device includes one or more processors configured to process a portion of time-series data using a trained encoder network to generate a dimensionally reduced encoding of the portion of the time-series data. The one or more processors are further configured to process the dimensionally reduced encoding using a trained decoder network to determine decoder output data. The one or more processors are also configured to set parameters of a predictive machine-learning model based on the decoder output data, wherein the predictive machine-learning model is configured to, based on the parameters, determine a predicted future value of the time-series data.Type: ApplicationFiled: December 9, 2021Publication date: June 15, 2023Inventors: Sari Andoni, Jimmie Goode
-
Publication number: 20230071667Abstract: A device includes one or more processors configured to process first input time-series data associated with a first time range using an embedding generator to generate an input embedding. The input embedding includes a positional embedding and a temporal embedding. The positional embedding indicates a position of an input value within the first input time-series data. The temporal embedding indicates that a first time associated with the input value is included in a particular day, a particular week, a particular month, a particular year, a particular holiday, or a combination thereof. The processors are configured to process the input embedding using a predictor to generate second predicted time-series data associated with a second time range. The second time range is subsequent to at least a portion of the first time range. The processors are configured to provide, to a second device, an output based on the second predicted time-series data.Type: ApplicationFiled: September 7, 2022Publication date: March 9, 2023Inventors: Tyler S. McDonnell, Jimmie Goode, William Jurayj, Nikolai Warner, Udaivir Yadav
-
Publication number: 20220269707Abstract: A recurrent neural network (RNN) method implemented on a computer system is used to produce summaries of unstructured text generated by multiple networks of individuals interacting over time by encoding the unstructured text into intermediate representations and decoding the intermediate representations into summaries of each network. Parameter data for the RNN is obtained by using multiple different versions of the same source texts to train the computer system. The method and computer system can be used to identify which of the networks match a query by determining which network generates the query with low or lowest cost.Type: ApplicationFiled: February 28, 2022Publication date: August 25, 2022Applicant: PULSELIGHT HOLDINGS, INC.Inventors: JONATHAN WILLIAM MUGAN, LAURA HITT, JIMMIE GOODE, RUSS GREGORY, YUAN QU
-
Patent number: 11263250Abstract: A recurrent neural network (RNN) method implemented on a computer system is used to produce summaries of unstructured text generated by multiple networks of individuals interacting over time by encoding the unstructured text into intermediate representations and decoding the intermediate representations into summaries of each network. Parameter data for the RNN is obtained by using multiple different versions of the same source texts to train the computer system. The method and computer system can be used to identify which of the networks match a query by determining which network generates the query with low or lowest cost.Type: GrantFiled: October 14, 2019Date of Patent: March 1, 2022Assignee: Pulselight Holdings, Inc.Inventors: Jonathan William Mugan, Laura Hitt, Jimmie Goode, Russ Gregory, Yuan Qu
-
Publication number: 20200364253Abstract: A recurrent neural network (RNN) method implemented on a computer system is used to produce summaries of unstructured text generated by multiple networks of individuals interacting over time by encoding the unstructured text into intermediate representations and decoding the intermediate representations into summaries of each network. Parameter data for the RNN is obtained by using multiple different versions of the same source texts to train the computer system. The method and computer system can be used to identify which of the networks match a query by determining which network generates the query with low or lowest cost.Type: ApplicationFiled: October 14, 2019Publication date: November 19, 2020Inventors: JONATHAN WILLIAM MUGAN, LAURA HITT, JIMMIE GOODE, RUSS GREGORY, YUAN QU
-
Publication number: 20200242480Abstract: A method includes receiving, by a processor, an input data set. The input data set includes a plurality of features. The method includes determining, by the processor, one or more characteristics of the input data set. The method includes, based on the one or more characteristics, adjusting, by the processor, one or more architectural parameters of an automated model generation process. The automated model generation process is configured to generate a plurality of models using a weighted randomization process. The one or more architectural parameters weight the weighted randomization process to adjust a probability of generation of models having particular architectural features. The method further includes executing, by the processor, the automated model generation process to output a mode, the model including data representative of a neural network.Type: ApplicationFiled: April 14, 2020Publication date: July 30, 2020Inventors: Tyler S. McDonnell, Sari Andoni, Junhwan Choi, Jimmie Goode, Yiyun Lan, Keith D. Moore, Gavin Sellers
-
Publication number: 20200175378Abstract: A method includes receiving, by a processor, an input data set. The input data set includes a plurality of features. The method includes determining, by the processor, one or more characteristics of the input data set. The method includes, based on the one or more characteristics, adjusting, by the processor, one or more architectural parameters of an automated model generation process. The automated model generation process is configured to generate a plurality of models using a weighted randomization process. The one or more architectural parameters weight the weighted randomization process to adjust a probability of generation of models having particular architectural features. The method further includes executing, by the processor, the automated model generation process to output a mode, the model including data representative of a neural network.Type: ApplicationFiled: November 29, 2018Publication date: June 4, 2020Inventors: Tyler S. McDonnell, Sari Andoni, Junhwan Choi, Jimmie Goode, Yiyun Lan, Keith D. Moore, Gavin Sellers
-
Patent number: 10657447Abstract: A method includes receiving, by a processor, an input data set. The input data set includes a plurality of features. The method includes determining, by the processor, one or more characteristics of the input data set. The method includes, based on the one or more characteristics, adjusting, by the processor, one or more architectural parameters of an automated model generation process. The automated model generation process is configured to generate a plurality of models using a weighted randomization process. The one or more architectural parameters weight the weighted randomization process to adjust a probability of generation of models having particular architectural features. The method further includes executing, by the processor, the automated model generation process to output a mode, the model including data representative of a neural network.Type: GrantFiled: November 29, 2018Date of Patent: May 19, 2020Assignee: SPARKCOGNITION, INC.Inventors: Tyler S. McDonnell, Sari Andoni, Junhwan Choi, Jimmie Goode, Yiyun Lan, Keith D. Moore, Gavin Sellers
-
Patent number: 10445356Abstract: A recurrent neural network (RNN) method implemented on a computer system is used to produce summaries of unstructured text generated by multiple networks of individuals interacting over time by encoding the unstructured text into intermediate representations and decoding the intermediate representations into summaries of each network. Parameter data for the RNN is obtained by using multiple different versions of the same source texts to train the computer system. The method and computer system can be used to identify which of the networks match a query by determining which network generates the query with low or lowest cost.Type: GrantFiled: June 23, 2017Date of Patent: October 15, 2019Assignee: Pulselight Holdings, Inc.Inventors: Jonathan William Mugan, Laura Hitt, Jimmie Goode, Russ Gregory, Yuan Qu
-
Publication number: 20140214722Abstract: A computer-implemented method for forecasting losses in a financial portfolio includes estimating parameters of an autoregressive-moving-average generalized-autoregressive-conditional-heteroscedastic (ARMA-GARCH) model for each individual asset in a financial portfolio by performing a parallel maximum likelihood estimation, estimating parameters of a copula dependence structure for standardized residuals of the ARMA-GARCH model, and estimating a Value-at-Risk (VaR) for the financial portfolio from the ARMA-GARCH model parameters and the copula dependence structure parameters.Type: ApplicationFiled: January 24, 2014Publication date: July 31, 2014Applicant: THE RESEARCH FOUNDATION OF THE STATE UNIVERSITY OF NEW YORKInventors: JAMES GLIMM, Jimmie Goode, Beiyu Lin, Nicholas Pezolano, Svetlozar Rachev