Patents by Inventor Heiko H. Ludwig
Heiko H. Ludwig has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240089081Abstract: An example system includes a processor to compute a tensor of indicators indicating a presence of partial sums in an encrypted vector of indicators. The processor can also securely reorder an encrypted array based on the computed tensor of indicators to generate a reordered encrypted array.Type: ApplicationFiled: August 25, 2022Publication date: March 14, 2024Inventors: Eyal KUSHNIR, Hayim SHAUL, Omri SOCEANU, Ehud AHARONI, Nathalie BARACALDO ANGEL, Runhua XU, Heiko H. LUDWIG
-
Patent number: 11856021Abstract: Computer-implemented methods, program products, and systems for provenance-based defense against poison attacks are disclosed. In one approach, a method includes: receiving observations and corresponding provenance data from data sources; determining whether the observations are poisoned based on the corresponding provenance data; and removing the poisoned observation(s) from a final training dataset used to train a final prediction model. Another implementation involves provenance-based defense against poison attacks in a fully untrusted data environment. Untrusted data points are grouped according to provenance signature, and the groups are used to train learning algorithms and generate complete and filtered prediction models. The results of applying the prediction models to an evaluation dataset are compared, and poisoned data points identified where the performance of the filtered prediction model exceeds the performance of the complete prediction model.Type: GrantFiled: March 22, 2023Date of Patent: December 26, 2023Assignee: International Business Machines CorporationInventors: Nathalie Baracaldo-Angel, Bryant Chen, Evelyn Duesterwald, Heiko H. Ludwig
-
Publication number: 20230409959Abstract: According to one embodiment, a method, computer system, and computer program product for grouped federated learning is provided. The embodiment may include initializing a plurality of aggregation groups including a plurality of parties and a plurality of local aggregators. The embodiment may also include submitting a query to a first party from the plurality of parties. The embodiment may further include submitting an initial response to the query from the first party or a second party from the plurality of parties to a first local aggregator from the plurality of local aggregators. The embodiment may also include submitting a final response from the first local aggregator or a second local aggregator from the plurality of local aggregators to a global aggregator. The embodiment may further include building a machine learning model based on the final response.Type: ApplicationFiled: June 21, 2022Publication date: December 21, 2023Inventors: Ali Anwar, Yi Zhou, NATHALIE BARACALDO ANGEL, Runhua Xu, YUYA JEREMY ONG, Annie K Abay, Heiko H. Ludwig, Gegi Thomas, Jayaram Kallapalayam Radhakrishnan, Laura Wynter
-
Patent number: 11824968Abstract: Techniques regarding privacy preservation in a federated learning environment are provided. For example, one or more embodiments described herein can comprise a system, which can comprise a memory that can store computer executable components. The system can also comprise a processor, operably coupled to the memory, and that can execute the computer executable components stored in the memory. The computer executable components can comprise a plurality of machine learning components that can execute a machine learning algorithm to generate a plurality of model parameters. The computer executable components can also comprise an aggregator component that can synthesize a machine learning model based on an aggregate of the plurality of model parameters. The aggregator component can communicate with the plurality of machine learning components via a data privacy scheme that comprises a privacy process and a homomorphic encryption process in a federated learning environment.Type: GrantFiled: September 13, 2021Date of Patent: November 21, 2023Inventors: Nathalie Baracaldo Angel, Stacey Truex, Heiko H. Ludwig, Ali Anwar, Thomas Steinke, Rui Zhang
-
Publication number: 20230231875Abstract: Computer-implemented methods, program products, and systems for provenance-based defense against poison attacks are disclosed. In one approach, a method includes: receiving observations and corresponding provenance data from data sources; determining whether the observations are poisoned based on the corresponding provenance data; and removing the poisoned observation(s) from a final training dataset used to train a final prediction model. Another implementation involves provenance-based defense against poison attacks in a fully untrusted data environment. Untrusted data points are grouped according to provenance signature, and the groups are used to train learning algorithms and generate complete and filtered prediction models. The results of applying the prediction models to an evaluation dataset are compared, and poisoned data points identified where the performance of the filtered prediction model exceeds the performance of the complete prediction model.Type: ApplicationFiled: March 22, 2023Publication date: July 20, 2023Inventors: Nathalie Baracaldo-Angel, Bryant Chen, Evelyn Duesterwald, Heiko H. Ludwig
-
Patent number: 11689566Abstract: Computer-implemented methods, program products, and systems for provenance-based defense against poison attacks are disclosed. In one approach, a method includes: receiving observations and corresponding provenance data from data sources; determining whether the observations are poisoned based on the corresponding provenance data; and removing the poisoned observation(s) from a final training dataset used to train a final prediction model. Another implementation involves provenance-based defense against poison attacks in a fully untrusted data environment. Untrusted data points are grouped according to provenance signature, and the groups are used to train learning algorithms and generate complete and filtered prediction models. The results of applying the prediction models to an evaluation dataset are compared, and poisoned data points identified where the performance of the filtered prediction model exceeds the performance of the complete prediction model.Type: GrantFiled: July 10, 2018Date of Patent: June 27, 2023Assignee: International Business Machines CorporationInventors: Nathalie Baracaldo-Angel, Bryant Chen, Evelyn Duesterwald, Heiko H. Ludwig
-
Publication number: 20230186168Abstract: A computer-implemented method according to one embodiment includes issuing a hyperparameter optimization (HPO) query to a plurality of computing devices; receiving HPO results from each of the plurality of computing devices; generating a unified performance metric surface utilizing the HPO results from each of the plurality of computing devices; and determining optimal global hyperparameters, utilizing the unified performance metric surface.Type: ApplicationFiled: December 9, 2021Publication date: June 15, 2023Inventors: Yi Zhou, Parikshit Ram, Nathalie Baracaldo Angel, Theodoros Salonidis, Horst Cornelius Samulowitz, Martin Wistuba, Heiko H. Ludwig
-
Patent number: 11645582Abstract: One embodiment provides a method for federated learning across a plurality of data parties, comprising assigning each data party with a corresponding namespace in an object store, assigning a shared namespace in the object store, and triggering a round of federated learning by issuing a customized learning request to at least one data party. Each customized learning request issued to a data party triggers the data party to locally train a model based on training data owned by the data party and model parameters stored in the shared namespace, and upload a local model resulting from the local training to a corresponding namespace in the object store the data party is assigned with. The method further comprises retrieving, from the object store, local models uploaded to the object store during the round of federated learning, and aggregating the local models to obtain a shared model.Type: GrantFiled: March 27, 2020Date of Patent: May 9, 2023Assignee: International Business Machines CorporationInventors: Shashank Rajamoni, Ali Anwar, Yi Zhou, Heiko H. Ludwig, Nathalie Baracaldo Angel
-
Patent number: 11645515Abstract: Embodiments relate to a system, program product, and method for automatically determining which activation data points in a neural model have been poisoned to erroneously indicate association with a particular label or labels. A neural network is trained using potentially poisoned training data. Each of the training data points is classified using the network to retain the activations of the last hidden layer, and segment those activations by the label of corresponding training data. Clustering is applied to the retained activations of each segment, and a cluster assessment is conducted for each cluster associated with each label to distinguish clusters with potentially poisoned activations from clusters populated with legitimate activations. The assessment includes executing a set of analyses and integrating the results of the analyses into a determination as to whether a training data set is poisonous based on determining if resultant activation clusters are poisoned.Type: GrantFiled: September 16, 2019Date of Patent: May 9, 2023Assignee: International Business Machines CorporationInventors: Nathalie Baracaldo Angel, Bryant Chen, Biplav Srivastava, Heiko H. Ludwig
-
Patent number: 11601468Abstract: Systems, computer-implemented methods, and computer program products that can facilitate detection of an adversarial backdoor attack on a trained model at inference time are provided. According to an embodiment, a system can comprise a memory that stores computer executable components and a processor that executes the computer executable components stored in the memory. The computer executable components can comprise a log component that records predictions and corresponding activation values generated by a trained model based on inference requests. The computer executable components can further comprise an analysis component that employs a model at an inference time to detect a backdoor trigger request based on the predictions and the corresponding activation values. In some embodiments, the log component records the predictions and the corresponding activation values from one or more layers of the trained model.Type: GrantFiled: June 25, 2019Date of Patent: March 7, 2023Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Nathalie Baracaldo Angel, Yi Zhou, Bryant Chen, Ali Anwar, Heiko H. Ludwig
-
Patent number: 11588621Abstract: Systems and techniques that facilitate universal and efficient privacy-preserving vertical federated learning are provided. In various embodiments, a key distribution component can distribute respective feature-dimension public keys and respective sample-dimension public keys to respective participants in a vertical federated learning framework governed by a coordinator, wherein the respective participants can send to the coordinator respective local model updates encrypted by the respective feature-dimension public keys and respective local datasets encrypted by the respective sample-dimension public keys. In various embodiments, an inference prevention component can verify a participant-related weight vector generated by the coordinator, based on which the key distribution component can distribute to the coordinator a functional feature-dimension secret key that can aggregate the encrypted respective local model updates into a sample-related weight vector.Type: GrantFiled: December 6, 2019Date of Patent: February 21, 2023Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Nathalie Baracaldo Angel, Runhua Xu, Yi Zhou, Ali Anwar, Heiko H. Ludwig
-
Publication number: 20230017500Abstract: One embodiment of the invention provides a method for federated learning (FL) comprising training a machine learning (ML) model collaboratively by initiating a round of FL across data parties. Each data party is allocated tokens to utilize during the training. The method further comprises maintaining, for each data party, a corresponding data usage profile indicative of an amount of data the data party consumed during the training and a corresponding participation profile indicative of an amount of data the data party provided during the training. The method further comprises selectively allocating new tokens to the data parties based on each participation profile maintained, selectively allocating additional new tokens to the data parties based on each data usage profile maintained, and reimbursing one or more tokens utilized during the training to the data parties based on one or more measurements of accuracy of the ML model.Type: ApplicationFiled: July 12, 2021Publication date: January 19, 2023Inventors: Ali Anwar, Syed Amer Zawad, Yi Zhou, Nathalie Baracaldo Angel, Kamala Micaela Noelle Varma, Annie Abay, Ebube Chuba, Yuya Jeremy Ong, Heiko H. Ludwig
-
Patent number: 11538236Abstract: Embodiments relate to a system, program product, and method for processing an untrusted data set to automatically determine which data points there are poisonous. A neural network is trained network using potentially poisoned training data. Each of the training data points is classified using the network to retain the activations of at least one hidden layer, and segment those activations by the label of corresponding training data. Clustering is applied to the retained activations of each segment, and a clustering assessment is conducted to remove an identified cluster from the data set, form a new training set, and train a second neural model with the new training set. The removed cluster and corresponding data are applied to the trained second neural model to analyze and classify data in the removed cluster as either legitimate or poisonous.Type: GrantFiled: September 16, 2019Date of Patent: December 27, 2022Assignee: International Business Machines CorporationInventors: Nathalie Baracaldo Angel, Bryant Chen, Heiko H. Ludwig
-
Patent number: 11487963Abstract: Embodiments relate to a system, program product, and method for automatically determining which activation data points in a neural model have been poisoned to erroneously indicate association with a particular label or labels. A neural network is trained network using potentially poisoned training data. Each of the training data points is classified using the network to retain the activations of the last hidden layer, and segment those activations by the label of corresponding training data. Clustering is applied to the retained activations of each segment, and a cluster assessment is conducted for each cluster associated with each label to distinguish clusters with potentially poisoned activations from clusters populated with legitimate activations. The assessment includes analyzing, for each cluster, a distance of a median of the activations therein to medians of the activations in the labels.Type: GrantFiled: September 16, 2019Date of Patent: November 1, 2022Assignee: International Business Machines CorporationInventors: Nathalie Baracaldo Angel, Bryant Chen, Biplav Srivastava, Heiko H. Ludwig
-
Patent number: 11341394Abstract: Embodiments relate to systematic explanation of neural model behavior and effective deduction of its vulnerabilities. Input data is received for the neural model and applied to the model to generate output data. Accuracy of the output data is evaluated with respect to the neural model, and one or more neural model vulnerabilities are identified that correspond to the output data accuracy. An explanation of the output data and the identified one or more vulnerabilities is generated, wherein the explanation serves as an indicator of alignment of the input data with the output data.Type: GrantFiled: July 24, 2019Date of Patent: May 24, 2022Assignee: International Business Machines CorporationInventors: Heiko H. Ludwig, Hogun Park, Mu Qiao, Peifeng Yin, Shubhi Asthana, Shun Jiang, Sunhwan Lee
-
Patent number: 11334709Abstract: A computer-implemented method according to one embodiment includes identifying a topic associated with a received notification, determining a plurality of policies associated with the topic, determining a current environmental context, determining a generalization level, utilizing the plurality of policies and the current environmental context, modifying the notification, based on the generalization level, and presenting the modified notification.Type: GrantFiled: November 13, 2018Date of Patent: May 17, 2022Assignee: International Business Machines CorporationInventors: Nathalie Baracaldo-Angel, Margaret H. Szymanski, Eric K. Butler, Heiko H. Ludwig
-
Publication number: 20220114259Abstract: One or more computer processors determine a tolerance value, and a norm value associated with an untrusted model and an adversarial training method. The one or more computer processors generate a plurality of interpolated adversarial images ranging between a pair of images utilizing the adversarial training method, wherein each image in the pair of images is from a different class. The one or more computer processors detect a backdoor associated with the untrusted model utilizing the generated plurality of interpolated adversarial images. The one or more computer processors harden the untrusted model by training the untrusted model with the generated plurality of interpolated adversarial images.Type: ApplicationFiled: October 13, 2020Publication date: April 14, 2022Inventors: Heiko H. Ludwig, Ebube Chuba, Bryant Chen, Benjamin James Edwards, Taesung Lee, Ian Michael Molloy
-
Patent number: 11240243Abstract: According to one embodiment, a method, computer system, and computer program product for preventing statistical inference attacks is provided. The present invention may include splitting records into items, and classifying these items into shared items and private items; grouping the private items according to privacy and confidentiality requirements; restricting access of the private items to stakeholders based on the confidentiality requirements using cryptographic keys; generating and encrypting one or more placeholders for both existent and non-existent stakeholders; storing private items in private storage as indicated by links; creating shared records comprising links, placeholders, and shared items; adding integrity signatures to the shared records; and publishing the shared records to a shared medium.Type: GrantFiled: September 13, 2017Date of Patent: February 1, 2022Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATIONInventors: Nathalie Baracaldo Angel, Robert Engel, Heiko H. Ludwig
-
Publication number: 20210409197Abstract: Techniques regarding privacy preservation in a federated learning environment are provided. For example, one or more embodiments described herein can comprise a system, which can comprise a memory that can store computer executable components. The system can also comprise a processor, operably coupled to the memory, and that can execute the computer executable components stored in the memory. The computer executable components can comprise a plurality of machine learning components that can execute a machine learning algorithm to generate a plurality of model parameters. The computer executable components can also comprise an aggregator component that can synthesize a machine learning model based on an aggregate of the plurality of model parameters. The aggregator component can communicate with the plurality of machine learning components via a data privacy scheme that comprises a privacy process and a homomorphic encryption process in a federated learning environment.Type: ApplicationFiled: September 13, 2021Publication date: December 30, 2021Inventors: Nathalie Baracaldo Angel, Stacey Truex, Heiko H. Ludwig, Ali Anwar, Thomas Steinke, Rui Zhang
-
Patent number: 11188789Abstract: One embodiment provides a method comprising receiving a training set comprising a plurality of data points, where a neural network is trained as a classifier based on the training set. The method further comprises, for each data point of the training set, classifying the data point with one of a plurality of classification labels using the trained neural network, and recording neuronal activations of a portion of the trained neural network in response to the data point. The method further comprises, for each classification label that a portion of the training set has been classified with, clustering a portion of all recorded neuronal activations that are in response to the portion of the training set, and detecting one or more poisonous data points in the portion of the training set based on the clustering.Type: GrantFiled: August 7, 2018Date of Patent: November 30, 2021Assignee: International Business Machines CorporationInventors: Bryant Chen, Wilka Carvalho, Heiko H. Ludwig, Ian Michael Molloy, Taesung Lee, Jialong Zhang, Benjamin J. Edwards