Patents by Inventor Samuel Schulter
Samuel Schulter has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 12373484Abstract: Systems and methods are provided for identifying and retrieving semantically similar images from a database. Semantic analysis is performed on an input query utilizing a vision language model to identify semantic concepts associated with the input query. A preliminary set of images is retrieved from the database for semantic concepts identified. Relevant concepts are extracted for images with a tokenizer by comparing images against a predefined label space to identify relevant concepts. A ranked list of relevant concepts is generated based on occurrence frequency within the set. The preliminary set of images is refined based on selecting specific relevant concepts from the ranked list by the user by combining the input query with the specific relevant concepts. Additional semantic analysis is iteratively performed to retrieve additional sets of images semantically similar to the combined input query and selection of the specific relevant concepts until a threshold condition is met.Type: GrantFiled: April 18, 2024Date of Patent: July 29, 2025Assignee: NEC CorporationInventors: Vijay Kumar Baikampady Gopalkrishna, Samuel Schulter, Manmohan Chandraker, Xiang Yu
-
Publication number: 20250218162Abstract: Systems and methods receive an annotated driving dataset including images capturing driving scenes and annotations including bounding boxes locating objects in the images. An image-caption dataset is obtained including images from common scenes and captions describing the images. A specialized dataset includes data of specific rare or unseen categories. Problem-specific knowledge generates a list of rare or unseen categories. Dataset tuning is performed by applying vision language model (VLM) sub-categorization, cut and paste, image generation, or caption filtering to the annotated driving dataset, the image-caption dataset, and the specialized dataset based on the problem-specific knowledge. A combined dataset includes outputs from the dataset tuning and the annotated driving dataset. A machine learning model is trained using the combined dataset.Type: ApplicationFiled: November 6, 2024Publication date: July 3, 2025Inventors: Samuel Schulter, Abhishek Aich, Manmohan Chandraker
-
Publication number: 20250182294Abstract: Methods and systems for image segmentation include generating features at multiple scales from an input image using a backbone model. The features are encoded using a transformer encoder that creates a per-pixel embedding map from a high-resolution scale of the multiple scales using deformable attention layers that operate on progressively higher-resolution scales of the multiple scales. The features are decoded using a transformer decoder to generate a segmentation mask.Type: ApplicationFiled: December 4, 2024Publication date: June 5, 2025Inventors: Abhishek Aich, Yumin Suh, Samuel Schulter
-
Publication number: 20250148757Abstract: Systems and methods for a self-improving data engine for autonomous vehicles is presented. To train the self-improving data engine for autonomous vehicles (SIDE), multi-modality dense captioning (MMDC) models can detect unrecognized classes from diversified descriptions for input images. A vision-language-model (VLM) can generate textual features from the diversified descriptions and image features from corresponding images to the diversified descriptions. Curated features, including curated textual features and curated image features, can be obtained by comparing similarity scores between the textual features and top-ranked image features based on their likelihood scores. Generate annotations, including bounding boxes and labels, can be generated for the curated features by comparing the similarity scores of labels generated by a zero-shot classifier and the curated textual features. The SIDE can be trained using the curated features, annotations, and feedback.Type: ApplicationFiled: October 30, 2024Publication date: May 8, 2025Inventors: Jong-Chyi Su, Sparsh Garg, Samuel Schulter, Manmohan Chandraker, Mingfu Liang
-
Publication number: 20250139527Abstract: Systems and methods for a self-improving model for agentic visual program synthesis. An agent can be continuously trained using an optimal training tuple to perform a corrective action to a monitored entity which in turn generates new input data for the training. To train the agent, an input question can be decomposed into vision model tasks to generate task outputs. The task outputs can be corrected based on feedback to obtain corrected task outputs. The optimal training tuple can be generated by comparing an optimal tuple threshold with a similarity score of the input image, the input question, and the corrected task outputs.Type: ApplicationFiled: October 29, 2024Publication date: May 1, 2025Inventors: Vijay Kumar Baikampady Gopalkrishna, Samuel Schulter, Manmohan Chandraker, Zaid Khan
-
Publication number: 20250118096Abstract: Methods and systems for object detection include generating a negative description for an input image based on a positive description of the input image using a language model. A negative image is generated based on the input image and the negative description by replacing a portion of the input image that is described by the positive description with content that is described by the negative description using a generative image model. An object detection model is trained with the input image, the positive description, the negative description, and the negative image.Type: ApplicationFiled: October 2, 2024Publication date: April 10, 2025Inventors: Samuel Schulter, Abhishek Aich, Vijay Kumar Baikampady Gopalkrishna
-
Publication number: 20250117947Abstract: Methods and systems for segmentation include encoding an image using a backbone model to generate feature maps. An exit point based on one of the feature maps. The feature maps are processed with a dynamic transformer encoder that includes layers, exiting the dynamic transformer encoder at a layer identified by the exit point. An output of the dynamic transformer encoder is decoded to output a segmentation of the image.Type: ApplicationFiled: September 23, 2024Publication date: April 10, 2025Inventors: Abhishek Aich, Yumin Suh, Samuel Schulter, Manyi Yao
-
Publication number: 20250118053Abstract: Systems and methods for visual object detection using explicit negatives. To train an artificial intelligence model with explicit negatives, a data sampler can sample input data from a language-based dataset to select images with annotations. A negative generation engine can generate explicit negatives representing sentences that include contradicting words that are semantically related to the annotations by using an external knowledgebase. A model trainer can minimize the classification loss of positive labels while decreasing the confidence score of the explicit negatives for the artificial intelligence model. The negative generation engine can be optimized to generate next explicit negatives. The artificial intelligence model can backpropagate using positive labels and the next explicit negatives to generate supervisory loss corresponding to the net explicit negatives. The artificial intelligence model can detect objects from an input image.Type: ApplicationFiled: October 3, 2024Publication date: April 10, 2025Inventors: Samuel Schulter, Vijay Kumar Baikampady Gopalkrishna, Yumin Suh
-
Publication number: 20250118063Abstract: Systems and methods include detecting one or more objects in an image and generating one or more captions for the image. One or more predicted categories of the one or more objects detected in the image and the one or more captions are matched. From the one or more predicted categories, a category that is not successfully predicted in the image is identified. Data is curated to improve the category that is not successfully predicted in the image. A perception model is finetuned using data curated.Type: ApplicationFiled: September 20, 2024Publication date: April 10, 2025Inventors: Jong-Chyi Su, Samuel Schulter, Sparsh Garg, Manmohan Chandraker, Mingfu Liang
-
Publication number: 20250118067Abstract: Systems and methods include generating a detection output for an image over multiple iterations by applying a dropout randomly to a different convolutional layer of a learning model for each iteration. The detection outputs are clustered, on labels, for each iteration. A total surface area for the clusters is computed over the iteration. A confidence is computed for the image using the total surface area for the clusters as an uncertainty score. A system is disabled if the confidence is below a threshold.Type: ApplicationFiled: September 17, 2024Publication date: April 10, 2025Inventors: Sparsh Garg, Samuel Schulter, Yumin Suh
-
Publication number: 20250118044Abstract: Systems and methods for identifying novel objects in an image include detecting one or more objects in an image and generating one or more captions for the image. One or more predicted categories of the one or more objects detected in the image and the one or more captions are matched to identify, from the one or more predicted categories, a category of a novel object in the image. An image feature and a text description feature are generated using a description of the novel object. A relevant image is selected using a similarity score between the image feature and the text description feature. A model is updated using the relevant image and associated description of the novel object.Type: ApplicationFiled: September 20, 2024Publication date: April 10, 2025Inventors: Jong-Chyi Su, Samuel Schulter, Sparsh Garg, Manmohan Chandraker, Mingfu Liang
-
Publication number: 20250115276Abstract: Methods and systems for object detection include generating a negative description for an input image of a road scene, based on a positive description of the input image, using a language model. A negative image is generated based on the input image and the negative description by replacing a portion of the input image that is described by the positive description with content that is described by the negative description using a generative image model. An object detection model is trained with the input image, the positive description, the negative description, and the negative image. An object is identified within a driving scene using the trained object detection model. A driving action is performed in a self-driving vehicle responsive to the identified object.Type: ApplicationFiled: October 2, 2024Publication date: April 10, 2025Inventors: Samuel Schulter, Vijay Kumar Baikampady Gopalkrishna, Yumin Suh
-
Patent number: 12254681Abstract: Systems and methods are provided for multi-modal test-time adaptation. The method includes inputting a digital image into a pre-trained Camera Intra-modal Pseudo-label Generator, and inputting a point cloud set into a pre-trained Lidar Intra-modal Pseudo-label Generator. The method further includes applying a fast 2-dimension (2D) model, and a slow 2D model, to the inputted digital image to apply pseudo-labels, and applying a fast 3-dimension (3D) model, and a slow 3D model, to the inputted point cloud set to apply pseudo-labels. The method further includes fusing pseudo-label predictions from the fast models and the slow models through an Inter-modal Pseudo-label Refinement module to obtain robust pseudo labels, and measuring a prediction consistency for the pseudo-labels.Type: GrantFiled: September 6, 2022Date of Patent: March 18, 2025Assignee: NEC CorporationInventors: Yi-Hsuan Tsai, Bingbing Zhuang, Samuel Schulter, Buyu Liu, Sparsh Garg, Ramin Moslemi, Inkyu Shin
-
Patent number: 12205356Abstract: Methods and systems for detecting faults include capturing an image of a scene using a camera. The image is embedded using a segmentation model that includes an image branch having an image embedding layer that embeds images into a joint latent space and a text branch having a text embedding layer that embeds text into the joint latent space. Semantic information is generated for a region of the image corresponding to a predetermined static object using the embedded image. A fault of the camera is identified based on a discrepancy between the semantic information and semantic information of the predetermined static image. The fault of the camera is corrected.Type: GrantFiled: March 23, 2023Date of Patent: January 21, 2025Assignee: NEC CorporationInventors: Samuel Schulter, Sparsh Garg, Manmohan Chandraker
-
Publication number: 20240379234Abstract: Methods and systems for visual question answering include decomposing an initial question to generate a sub-question. The initial question and an image are applied to a visual question answering model to generate an answer and a confidence score. It is determined that the confidence score is below a threshold value. The sub-question is applied to the visual question answering model, responsive to the determination that the confidence score is below a threshold value, to generate a final answer.Type: ApplicationFiled: May 9, 2024Publication date: November 14, 2024Inventors: Vijay Kumar Baikampady Gopalkrishnan, Samuel Schulter, Manmohan Chandraker
-
Publication number: 20240378874Abstract: Systems and methods are provided for multi-dataset panoptic segmentation, including processing received images from multiple datasets to extract multi-scale features using a backbone network, each of the multiple datasets including a unique label space, generating text-embeddings for class names from the unique label space for each of the multiple datasets, and integrating the text-embeddings with visual features extracted from the received images to create a unified semantic space. A transformer-based segmentation model is trained using the unified semantic space to predict segmentation masks and classes for the received images, and a unified panoptic segmentation map is generated from the predicted segmentation masks and classes by performing inference using a panoptic interference algorithm.Type: ApplicationFiled: May 9, 2024Publication date: November 14, 2024Inventors: Samuel Schulter, Abhishek Aich
-
Publication number: 20240378454Abstract: Systems and methods for optimizing models for open-vocabulary detection. Region proposals can be obtained by employing a pre-trained vision-language model and a pre-trained region proposal network. Object feature predictions can be obtained by employing a trained teacher neural network with the region proposals. Object feature predictions can be filtered above a threshold to obtain pseudo labels. A student neural network with a split-and-fusion detection head can be trained by utilizing the region proposals, base ground truth class labels and the pseudo labels. The pseudo labels can be optimized by reducing the noise from the pseudo labels by employing the trained split-and-fusion detection head of the trained student neural network to obtain optimized object detections. An action can be performed relative to a scene layout based on the optimized object detections.Type: ApplicationFiled: May 9, 2024Publication date: November 14, 2024Inventors: Samuel Schulter, Yumin Suh, Manmohan Chandraker, Vijay Kumar Baikampady Gopalkrishna
-
Patent number: 12131422Abstract: A method for achieving high-fidelity novel view synthesis and 3D reconstruction for large-scale scenes is presented. The method includes obtaining images from a video stream received from a plurality of video image capturing devices, grouping the images into different image clusters representing a large-scale 3D scene, training a neural radiance field (NeRF) and an uncertainty multilayer perceptron (MLP) for each of the image clusters to generate a plurality of NeRFs and a plurality of uncertainty MLPs for the large-scale 3D scene, applying a rendering loss and an entropy loss to the plurality of NeRFs, performing uncertainty-based fusion to the plurality of NeRFs to define a fused NeRF, and jointly fine-tuning the plurality of NeRFs and the plurality of uncertainty MLPs, and during inference, applying the fused NeRF for novel view synthesis of the large-scale 3D scene.Type: GrantFiled: October 11, 2022Date of Patent: October 29, 2024Assignee: NEC CorporationInventors: Bingbing Zhuang, Samuel Schulter, Yi-Hsuan Tsai, Buyu Liu, Nanbo Li
-
Publication number: 20240354583Abstract: Methods and systems for training a model include annotating a subset of an unlabeled training dataset, that includes images of road scenes, with labels. A road defect detection model is iteratively trained, including adding pseudo-labels to a remainder of examples from the unlabeled training dataset and training the road defect detection model based on the labels and the pseudo-labels.Type: ApplicationFiled: March 25, 2024Publication date: October 24, 2024Inventors: Sparsh Garg, Samuel Schulter, Bingbing Zhuang, Manmohan Chandraker
-
Publication number: 20240354336Abstract: Systems and methods are provided for identifying and retrieving semantically similar images from a database. Semantic analysis is performed on an input query utilizing a vision language model to identify semantic concepts associated with the input query. A preliminary set of images is retrieved from the database for semantic concepts identified. Relevant concepts are extracted for images with a tokenizer by comparing images against a predefined label space to identify relevant concepts. A ranked list of relevant concepts is generated based on occurrence frequency within the set. The preliminary set of images is refined based on selecting specific relevant concepts from the ranked list by the user by combining the input query with the specific relevant concepts. Additional semantic analysis is iteratively performed to retrieve additional sets of images semantically similar to the combined input query and selection of the specific relevant concepts until a threshold condition is met.Type: ApplicationFiled: April 18, 2024Publication date: October 24, 2024Inventors: Vijay Kumar Baikampady Gopalkrishna, Samuel Schulter, Manmohan Chandraker, Xiang Yu