Patents by Inventor Hans Peter Graf

Hans Peter Graf has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11741712
    Abstract: A method for using a multi-hop reasoning framework to perform multi-step compositional long-term reasoning is presented. The method includes extracting feature maps and frame-level representations from a video stream by using a convolutional neural network (CNN), performing object representation learning and detection, linking objects through time via tracking to generate object tracks and image feature tracks, feeding the object tracks and the image feature tracks to a multi-hop transformer that hops over frames in the video stream while concurrently attending to one or more of the objects in the video stream until the multi-hop transformer arrives at a correct answer, and employing video representation learning and recognition from the objects and image context to locate a target object within the video stream.
    Type: Grant
    Filed: September 1, 2021
    Date of Patent: August 29, 2023
    Inventors: Asim Kadav, Farley Lai, Hans Peter Graf, Alexandru Niculescu-Mizil, Renqiang Min, Honglu Zhou
  • Publication number: 20230253068
    Abstract: A method for implementing deep reinforcement learning with T-cell receptor (TCR) mutation policies to generate binding TCRs recognizing target peptides for immunotherapy is presented. The method includes extracting peptides to identify a virus or tumor cells, collecting a library of TCRs from target patients, predicting, by a deep neural network, interaction scores between the extracted peptides and the TCRs from the target patients, developing a deep reinforcement learning (DRL) framework with TCR mutation policies to generate TCRs with maximum binding scores, defining reward functions based on a reconstruction-based score and a density estimation-based score, randomly sampling batches of TCRs and following a policy network to mutate the TCRs, outputting mutated TCRs, and ranking the outputted TCRs to utilize top-ranked TCR candidates to target the virus or the tumor cells for immunotherapy.
    Type: Application
    Filed: January 9, 2023
    Publication date: August 10, 2023
    Inventors: Renqiang Min, Hans Peter Graf, Ziqi Chen
  • Publication number: 20230154167
    Abstract: A method for implementing source-free domain adaptive detection is presented. The method includes, in a pretraining phase, applying strong data augmentation to labeled source images to produce perturbed labeled source images and training an object detection model by using the perturbed labeled source images to generate a source-only model. The method further includes, in an adaptation phase, training a self-trained mean teacher model by generating a weakly augmented image and multiple strongly augmented images from unlabeled target images, generating a plurality of region proposals from the weakly augmented image, selecting a region proposal from the plurality of region proposals as a pseudo ground truth, detecting, by the self-trained mean teacher model, object boxes and selecting pseudo ground truth boxes by employing a confidence constraint and a consistency constraint, and training a student model by using one of the multiple strongly augmented images jointly with an object detection loss.
    Type: Application
    Filed: October 14, 2022
    Publication date: May 18, 2023
    Inventors: Kai Li, Renqiang Min, Hans Peter Graf
  • Publication number: 20230153606
    Abstract: A method is provided that includes training a CLIP model to learn embeddings of images and text from matched image-text pairs. The text represents image attributes. The method trains a StyleGAN on images in a training dataset of matched image-text pairs. The method also trains, using a CLIP model guided contrastive loss which attracts matched text embedding pairs and repels unmatched pairs, a text-to-direction model to predict a text direction that is semantically aligned with an input text responsive to the input text and a random latent code. A triplet loss is used to learn text directions using the embeddings learned by the trained CLIP model. The method generates, by the trained StyleGAN, positive and negative synthesized images by respectively adding and subtracting the text direction in the latent space of the trained StyleGAN corresponding to a word for each of the words in the training dataset.
    Type: Application
    Filed: October 19, 2022
    Publication date: May 18, 2023
    Inventors: Renqiang Min, Kai Li, Hans Peter Graf, Zhiheng Li
  • Publication number: 20230148017
    Abstract: A method for compositional reasoning of group activity in videos with keypoint-only modality is presented. The method includes obtaining video frames from a video stream received from a plurality of video image capturing devices, extracting keypoints all of persons detected in the video frames to define keypoint data, tokenizing the keypoint data with time and segment information, clustering groups of keypoint persons in the video frames and passing the clustering groups through multi-scale prediction, and performing a prediction to provide a group activity prediction of a scene in the video frames.
    Type: Application
    Filed: October 5, 2022
    Publication date: May 11, 2023
    Inventors: Asim Kadav, Farley Lai, Hans Peter Graf, Honglu Zhou
  • Publication number: 20230129568
    Abstract: Systems and methods for predicting T-Cell receptor (TCR)-peptide interaction, including training a deep learning model for the prediction of TCR-peptide interaction by determining a multiple sequence alignment (MSA) for TCR-peptide pair sequences from a dataset of TCR-peptide pair sequences using a sequence analyzer, building TCR structures and peptide structures using the MSA and corresponding structures from a Protein Data Bank (PDB) using a MODELLER, and generating an extended TCR-peptide training dataset based on docking energy scores determined by docking peptides to TCRs using physical modeling based on the TCR structures and peptide structures built using the MODELLER. TCR-peptide pairs are classified and labeled as positive or negative pairs using pseudo-labels based on the docking energy scores, and the deep learning model is iteratively retrained based on the extended TCR-peptide training dataset and the pseudo-labels until convergence.
    Type: Application
    Filed: October 20, 2022
    Publication date: April 27, 2023
    Inventors: Renqiang Min, Hans Peter Graf, Erik Kruus, Yiren Jian
  • Publication number: 20230085160
    Abstract: A method for generating binding peptides presented by any given Major Histocompatibility Complex (MHC) protein is presented. The method includes, given a peptide and an MHC protein pair, enabling a Reinforcement Learning (RL) agent to interact with and exploit a peptide mutation environment by repeatedly mutating the peptide and observing an observation score of the peptide, learning to form a mutation policy, via a mutation policy network, to iteratively mutate amino acids of the peptide to obtain desired presentation scores, and generating, based on the desired presentation scores, qualified peptides and binding motifs of MHC Class I proteins.
    Type: Application
    Filed: August 30, 2022
    Publication date: March 16, 2023
    Inventors: Renqiang Min, Hans Peter Graf, Ziqi Chen
  • Publication number: 20230083313
    Abstract: A system for binding peptide search for immunotherapy is presented. The system includes employing a deep neural network to predict a peptide presentation given Major Histocompatibility Complex allele sequences and peptide sequences, training a Variational Autoencoder (VAE) to reconstruct peptides by converting the peptide sequences into continuous embedding vectors, running a Monte Carlo Tree Search to generate a first set of positive peptide vaccine candidates, running a Bayesian Optimization search with the trained VAE and a Backpropagation search with the trained VAE to generate a second set of positive peptide vaccine candidates, using a sampling from a Position Weight Matrix (sPWM) to generate a third set of positive peptide vaccine candidates, screening and merging the first, second, and third sets of positive peptide vaccine candidates, and outputting qualified peptides for immunotherapy from the screened and merged sets of positive peptide vaccine candidates.
    Type: Application
    Filed: August 30, 2022
    Publication date: March 16, 2023
    Inventors: Renqiang Min, Hans Peter Graf, Ziqi Chen
  • Patent number: 11568247
    Abstract: A computer-implemented method executed by at least one processor for performing mini-batching in deep learning by improving cache utilization is presented. The method includes temporally localizing a candidate clip in a video stream based on a natural language query, encoding a state, via a state processing module, into a joint visual and linguistic representation, feeding the joint visual and linguistic representation into a policy learning module, wherein the policy learning module employs a deep learning network to selectively extract features for select frames for video-text analysis and includes a fully connected linear layer and a long short-term memory (LSTM), outputting a value function from the LSTM, generating an action policy based on the encoded state, wherein the action policy is a probabilistic distribution over a plurality of possible actions given the encoded state, and rewarding policy actions that return clips matching the natural language query.
    Type: Grant
    Filed: March 16, 2020
    Date of Patent: January 31, 2023
    Inventors: Asim Kadav, Iain Melvin, Hans Peter Graf, Meera Hahn
  • Patent number: 11475590
    Abstract: Aspects of the present disclosure describe systems, methods and structures for an efficient multi-person posetracking method that advantageously achieves state-of-the-art performance on PoseTrack datasets by only using keypoint information in a tracking step without optical flow or convolution routines. As a consequence, our method has fewer parameters and FLOPs and achieves faster FPS. Our method benefits from our parameter-free tracking method that outperforms commonly used bounding box propagation in top-down methods. Finally, we disclose tokenization and embedding multi-person pose keypoint information in the transformer architecture that can be re-used for other pose tasks such as pose-based action recognition.
    Type: Grant
    Filed: September 9, 2020
    Date of Patent: October 18, 2022
    Inventors: Asim Kadav, Farley Lai, Hans Peter Graf, Michael Snower
  • Publication number: 20220327489
    Abstract: Systems and methods for matching job descriptions with job applicants is provided. The method includes allocating each of one or more job applicants' curriculum vitae (CV) into sections; applying max pooled word embedding to each section of the job applicants' CVs; using concatenated max-pooling and average-pooling to compose the section embeddings into an applicant's CV representation; allocating each of one or more job position descriptions into specified sections; applying max pooled word embedding to each section of the job position descriptions; using concatenated max-pooling and average-pooling to compose the section embeddings into a job representation; calculating a cosine similarity between each of the job representations and each of the CV representations to perform job-to-applicant matching; and presenting an ordered list of the one or more job applicants or an ordered list of the one or more job position descriptions to a user.
    Type: Application
    Filed: April 6, 2022
    Publication date: October 13, 2022
    Inventors: Renqiang Min, Iain Melvin, Christopher A White, Christopher Malon, Hans Peter Graf
  • Publication number: 20220328127
    Abstract: A computer-implemented method is provided for generating new binding peptides to Major Histocompatibility Complex (MHC) proteins. The method includes training, by a processor device, a Generative Adversarial Network GAN having a generator and a discriminator only on a set of binding peptide sequences given training data comprising the set of binding peptide sequences and a set of non-binding peptide sequences. A GAN training objective includes the discriminator being iteratively updated to distinguish generated peptide sequences from sampled binding peptide sequences as fake or real and the generator being iteratively updated to fool the discriminator. The training includes optimizing the GAN training objective while learning two projection vectors for a binding class with two cross-entropy losses. A first loss discriminating binding peptide sequences in the training data from non-binding peptide sequences in the training data.
    Type: Application
    Filed: April 1, 2022
    Publication date: October 13, 2022
    Inventors: Renqiang Min, Hans Peter Graf, Ligong Han
  • Publication number: 20220327425
    Abstract: Methods and systems for training a machine learning model include embedding a state, including a peptide sequence and a protein, as a vector. An action, including a modification to an amino acid in the peptide sequence, is predicted using a presentation score of the peptide sequence by the protein as a reward. A mutation policy model is trained, using the state and the reward, to generate modifications that increase the presentation score.
    Type: Application
    Filed: April 1, 2022
    Publication date: October 13, 2022
    Inventors: Renqiang Min, Hans Peter Graf, Ligong Han
  • Publication number: 20220319635
    Abstract: Methods and systems for training a model include encoding training peptide sequences using an encoder model. A new peptide sequence is generated using a generator model. The encoder model, the generator model, and the discriminator model are trained to cause the generator model to generate new peptides that the discriminator mistakes for the training peptide sequences, including learning projection vectors with respective cross-entropy losses for binding sequences and non-binding sequences.
    Type: Application
    Filed: April 1, 2022
    Publication date: October 6, 2022
    Inventors: Renqiang Min, Hans Peter Graf, Ligong Han
  • Patent number: 11423655
    Abstract: A computer-implemented method is provided for disentangled data generation. The method includes accessing, by a variational autoencoder, a plurality of supervision signals. The method further includes accessing, by the variational autoencoder, a plurality of auxiliary tasks that utilize the supervision signals as reward signals to learn a disentangled representation. The method also includes training the variational autoencoder to disentangle a sequential data input into a time-invariant factor and a time-varying factor using a self-supervised training approach which is based on outputs of the auxiliary tasks obtained by using the supervision signals to accomplish the plurality of auxiliary tasks.
    Type: Grant
    Filed: November 3, 2020
    Date of Patent: August 23, 2022
    Inventors: Renqiang Min, Yizhe Zhu, Asim Kadav, Hans Peter Graf
  • Publication number: 20220254152
    Abstract: A method for learning disentangled representations of videos is presented. The method includes feeding each frame of video data into an encoder to produce a sequence of visual features, passing the sequence of visual features through a deep convolutional network to obtain a posterior of a dynamic latent variable and a posterior of a static latent variable, sampling static and dynamic representations from the posterior of the static latent variable and the posterior of the dynamic latent variable, respectively, concatenating the static and dynamic representations to be fed into a decoder to generate reconstructed sequences, and applying three regularizers to the dynamic and static latent variables to trigger representation disentanglement. To facilitate the disentangled sequential representation learning, orthogonal factorization in generative adversarial network (GAN) latent space is leveraged to pre-train a generator as a decoder in the method.
    Type: Application
    Filed: January 27, 2022
    Publication date: August 11, 2022
    Inventors: Renqiang Min, Hans Peter Graf, Ligong Han
  • Publication number: 20220237884
    Abstract: A computer-implemented method is provided for action localization. The method includes converting one or more video frames into person keypoints and object keypoints. The method further includes embedding position, timestamp, instance, and type information with the person keypoints and object keypoints to obtain keypoint embeddings. The method also includes predicting, by a hierarchical transformer encoder using the keypoint embeddings, human actions and bounding box information of when and where the human actions occur in the one or more video frames.
    Type: Application
    Filed: January 27, 2022
    Publication date: July 28, 2022
    Inventors: Asim Kadav, Farley Lai, Hans Peter Graf, Yi Huang
  • Publication number: 20220171989
    Abstract: A computer-implemented method for representation disentanglement is provided. The method includes encoding an input vector into an embedding. The method further includes learning, by a hardware processor, disentangled representations of the input vector including a style embedding and a content embedding by performing sample-based mutual information minimization on the embedding under a Wasserstein distance regularization and a Kullback-Leibler (KL) divergence. The method also includes decoding the style and content embeddings to obtain a reconstructed vector.
    Type: Application
    Filed: November 18, 2021
    Publication date: June 2, 2022
    Inventors: Renqiang Min, Asim Kadav, Hans Peter Graf, Ligong Han
  • Publication number: 20220130490
    Abstract: Methods and systems for generating a peptide sequence include transforming an input peptide sequence into disentangled representations, including a structural representation and an attribute representation, using an autoencoder model. One of the disentangled representations is modified. The disentangled representations, including the modified disentangled representation, are transformed to generate a new peptide sequence using the autoencoder model.
    Type: Application
    Filed: October 26, 2021
    Publication date: April 28, 2022
    Inventors: Renqiang Min, Igor Durdanovic, Hans Peter Graf
  • Publication number: 20220101007
    Abstract: A method for using a multi-hop reasoning framework to perform multi-step compositional long-term reasoning is presented. The method includes extracting feature maps and frame-level representations from a video stream by using a convolutional neural network (CNN), performing object representation learning and detection, linking objects through time via tracking to generate object tracks and image feature tracks, feeding the object tracks and the image feature tracks to a multi-hop transformer that hops over frames in the video stream while concurrently attending to one or more of the objects in the video stream until the multi-hop transformer arrives at a correct answer, and employing video representation learning and recognition from the objects and image context to locate a target object within the video stream.
    Type: Application
    Filed: September 1, 2021
    Publication date: March 31, 2022
    Inventors: Asim Kadav, Farley Lai, Hans Peter Graf, Alexandru Niculescu-Mizil, Renqiang Min, Honglu Zhou