Patents by Inventor Senthil Kumar Yogamani

Senthil Kumar Yogamani has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20250139882
    Abstract: In some aspects of the disclosure, an apparatus includes a processing system that includes one or more processors and one or more memories coupled to the one or more processors. The processing system is configured to receive sensor data associated with a scene and to generate a cylindrical representation associated with the scene. The processing system is further configured to modify the cylindrical representation based on detecting a feature of the cylindrical representation being included in a first region of the cylindrical representation. Modifying the cylindrical representation includes relocating the feature from the first region to a second region that is different than the first region. The processing system is further configured to perform, based on the modified cylindrical representation, one or more three-dimensional (3D) perception operations associated with the scene.
    Type: Application
    Filed: October 31, 2023
    Publication date: May 1, 2025
    Inventors: Behnaz Rezaei, Varun Ravi Kumar, Senthil Kumar Yogamani
  • Publication number: 20250131742
    Abstract: Aspects presented herein may improve the accuracy and reliability of object detections performed by multiple object detection models. In one aspect, a UE detects (1) a set of polylines from at least one of a set of bird's eye view (BEV) features or a set of perspective view (PV) features associated with a set of images and (2) a set of three-dimensional (3D) objects in the set of BEV features. The UE associates the set of polylines with the set of 3D objects. The UE updates the set of polylines based on a set of nearby 3D objects or updates the set of 3D objects based on a set of nearby polylines. The UE outputs an indication of the updated set of polylines or the updated set of 3D objects.
    Type: Application
    Filed: October 23, 2023
    Publication date: April 24, 2025
    Inventors: Varun RAVI KUMAR, Senthil Kumar YOGAMANI, Heesoo MYEONG
  • Publication number: 20250095354
    Abstract: An apparatus includes a memory and processing circuitry in communication with the memory. The processing circuitry is configured to process a joint graph representation using a graph neural network (GNN) to form an enhanced graph representation. The joint graph representation includes first features from a voxelized point cloud, and second features from a plurality of camera images. The enhanced graph representation includes enhanced first features and enhanced second features. The processing circuitry is further configured to perform a diffusion processes on the enhanced first features and the enhanced second features of the enhanced graph representation to form a denoised graph representation having denoised first features and denoised second features, and fuse the denoised first features and the denoised second features of the denoised graph representation using a graph attention network (GAT) to form a fused point cloud having fused features.
    Type: Application
    Filed: September 14, 2023
    Publication date: March 20, 2025
    Inventors: Varun Ravi Kumar, Debasmit Das, Senthil Kumar Yogamani
  • Publication number: 20250094535
    Abstract: According to aspects described herein, a device can extract first features from frames of first sensor data and second features from frames of second sensor data (captured after the first sensor data). The device can obtain first weighted features based on the first features and second weighted features based on the second features. The device can aggregate the first weighted features to determine a first feature vector and the second weighted features to determine a second feature vector. The device can obtain a first transformed feature vector (based on transforming the first feature vector into a coordinate space) and a second transformed feature vector (based on transforming the second feature vector into the coordinate space). The device can aggregate first transformed weighted features (based on the first transformed feature vector) and second transformed weighted features (based on the second transformed feature vector) to determine a fused feature vector.
    Type: Application
    Filed: September 18, 2023
    Publication date: March 20, 2025
    Inventors: Shivansh RAO, Sweta PRIYADARSHI, Varun RAVI KUMAR, Senthil Kumar YOGAMANI, Arunkumar NEHRUR RAVI, Vasudev BHASKARAN
  • Publication number: 20250095173
    Abstract: An example device for training a neural network includes a memory configured to store a neural network model for the neural network; and a processing system comprising one or more processors implemented in circuitry, the processing system being configured to: extract image features from an image of an area, the image features representing objects in the area; extract point cloud features from a point cloud representation of the area, the point cloud features representing the objects in the area; add Gaussian noise to a ground truth depth map for the area to generate a noisy ground truth depth map, the ground truth depth map representing accurate positions of the objects in the area; and train the neural network using the image features, the point cloud features, and the noisy ground truth depth map to generate a depth map.
    Type: Application
    Filed: September 14, 2023
    Publication date: March 20, 2025
    Inventors: Savitha Srinivasan, Varun Ravi Kumar, Senthil Kumar Yogamani
  • Publication number: 20250095379
    Abstract: This disclosure provides systems, methods, and devices for vehicle driving assistance systems that support image processing. In a first aspect, a method includes receiving a first image frame from the first camera and a second image frame from the second camera. The method may also include determining a first set of optical flows between the first image frame and the second image frame and determining a second set of optical flows based on the first set of optical flows and positions of the first camera and the second camera on the vehicle. Position data may be determined for objects in an area surrounding the vehicle based on the second set of optical flows. Other aspects and features are also claimed and described.
    Type: Application
    Filed: September 15, 2023
    Publication date: March 20, 2025
    Inventors: Madhumitha Sakthi, Louis Joseph Kerofsky, Senthil Kumar Yogamani
  • Publication number: 20250095168
    Abstract: Systems and techniques are described herein for processing data. For instance, a method for processing data is provided. The method may include obtaining source features generated based on first sensor data captured using a first set of sensors; obtaining source semantic attributes related to the source features; obtaining target features generated based on second sensor data captured using a second set of sensors; obtaining map information; obtaining location information of a device comprising the second set of sensors; obtaining target semantic attributes from the map information based on the location information; aligning the target features with a set of the source features, based on the source semantic attributes and the target semantic attributes, to generate aligned target features; and processing the aligned target features to generate an output.
    Type: Application
    Filed: September 15, 2023
    Publication date: March 20, 2025
    Inventors: Julia KABALAR, Kiran BANGALORE RAVI, Nirnai ACH, Mireille Lucette Laure GREGOIRE, Varun RAVI KUMAR, Senthil Kumar YOGAMANI
  • Publication number: 20250095329
    Abstract: A device for processing sensor data is configured to receive the first frames from a first sensor; receive the second frames from a second sensor; perform a first feature extraction on the first frames using a first dynamic neural network to determine first features; perform a second feature extraction on the second frames using a second dynamic neural network to determine second features; determine a first delay associated with the first features; determine a second delay associated with the second features; modify a topology of the second dynamic neural network based on the first delay and the second delay; and use the second dynamic neural network with the modified topology to generate an output.
    Type: Application
    Filed: September 19, 2023
    Publication date: March 20, 2025
    Inventors: Kiran Bangalore Ravi, Julia Kabalar, Nirnai Ach, Mireille Lucette Laure Gregoire, Senthil Kumar Yogamani
  • Publication number: 20250086978
    Abstract: An apparatus includes a memory for storing image data and position data, wherein the image data comprises a set of two-dimensional (2D) camera images, and wherein the position data comprises a set of three-dimensional (3D) point cloud frames. The apparatus also includes processing circuitry in communication with the memory, wherein the processing circuitry is configured to convert the set of 2D camera images into a first 3D representation of a 3D environment corresponding to the image data and the position data, wherein the set of 3D point cloud frames comprises a second 3D representation of the 3D environment. The processing circuitry is also configured to generate, based on the first 3D representation and the second 3D representation, a set of bird's eye view (BEV) feature kernels in a continuous space; and generate, based on the set of BEV feature kernels, an output.
    Type: Application
    Filed: September 13, 2023
    Publication date: March 13, 2025
    Inventors: Kiran Bangalore Ravi, Varun Ravi Kumar, Senthil Kumar Yogamani
  • Publication number: 20250085413
    Abstract: This disclosure provides systems, methods, and devices for vehicle driving assistance systems that support image processing. In a first aspect, a method of image processing includes receiving image BEV features and receiving first radio detection and ranging (RADAR) BEV features. The first RADAR BEV features that are received are determined based on first RADAR data associated with a first data type. First normalized RADAR BEV features are determined, which includes rescaling the first RADAR BEV features using a first attention mechanism based on the image BEV features and the first RADAR BEV features. Fused data is determined that combines the first normalized RADAR BEV features and the image BEV features. Other aspects and features are also claimed and described.
    Type: Application
    Filed: September 7, 2023
    Publication date: March 13, 2025
    Inventors: Senthil Kumar Yogamani, Varun Ravi Kumar
  • Publication number: 20250086977
    Abstract: This disclosure provides systems, methods, and devices for processing and aligning sensor data features for navigation. In a first aspect, a method is provided that includes determining, based on received sensor data, a first set of features for an area surrounding the vehicle. A second set of features for the area surrounding the vehicle may be determined based on an occupancy map for the area surrounding the vehicle. A third set of features may be determined that align the first set of features with the second set of features. The third set of features may align each of at least a subset of the second set of features with at least one corresponding feature from the first set of features. Other aspects and features are also claimed and described.
    Type: Application
    Filed: September 8, 2023
    Publication date: March 13, 2025
    Inventors: Venkatraman Narayanan, Varun Ravi Kumar, Senthil Kumar Yogamani
  • Publication number: 20250086979
    Abstract: Systems that support graph neural network (GNN) implemented multi-modal spatiotemporal fusion are provided. Identifying and tracking an object in images captured by an imaging system is facilitated by generating a graph based on multimodal data received from a plurality of sensors. The graph encodes spatial components and spatial data associated with the images and encodes temporal data associated with the images. Pooled features are generated, through application of a first graph attention network (GAT), by pooling spatial features and temporal features. The spatial features are based on the spatial component and on the spatial relationship, and the temporal features are based on the temporal relationship. A three dimensional bounding box associated with the object is decoded by propagating the pooled features through a fully connected layer.
    Type: Application
    Filed: September 7, 2023
    Publication date: March 13, 2025
    Inventors: Venkatraman Narayanan, Varun Ravi Kumar, Senthil Kumar Yogamani
  • Publication number: 20250085407
    Abstract: A method includes receiving a plurality of images, wherein a first image of the one or more images comprises a range image and a second image comprises a camera image and filtering the first image to generate a filtered first image. The method also includes generating a plurality of depth estimates based on the second image and generating an attention map by combining the filtered first image and the plurality of depth estimates. Additionally, the method includes generating a consistency score indicative of a consistency of depth estimates between the first image and the second image based on the attention map, modulating one or more features extracted from the second image based on the consistency score using a gating mechanism to generate modulated one or more features, and generating a classification of one or more soiled regions in the second image based on the modulated one or more features.
    Type: Application
    Filed: September 11, 2023
    Publication date: March 13, 2025
    Inventors: Varun Ravi Kumar, Senthil Kumar Yogamani, Shivansh Rao
  • Publication number: 20250076487
    Abstract: This disclosure provides systems, methods, and devices for vehicle driving assistance systems that support image processing. In a first aspect, a method includes receiving a first image frame from a first camera; receiving second image frame from a second camera. The fields-of-view of the cameras partially overlap. A set of coordinates associated with pixel values of the first image frame and pixel values of the second image frame are determined. The set of coordinates correspond to an overlap of the first field-of-view and the second field-of-view. A first uncertainty window metric is determined based the set of coordinates and first uncertainty values. A second uncertainty window metric is determined based on the first uncertainty values and second uncertainty values associated with RADAR. Fused depth data is determined based on the set of coordinates, the RADAR data, and the first and second uncertainty window metrics. Other aspects and features are also claimed and described.
    Type: Application
    Filed: September 6, 2023
    Publication date: March 6, 2025
    Inventors: Madhumitha Sakthi, Louis Joseph Kerofsky, Senthil Kumar Yogamani
  • Publication number: 20250078437
    Abstract: A method for selecting one or more Regions of Interest (RoIs) for human annotations includes obtaining sensor data generated by one or more sensors of a vehicle; applying at least one class-agnostic heuristic function to the sensor data to determine a presence and an approximate position of one or more objects in an RoI of the sensor data; selecting one or more RoIs having proposed annotations for the one or more objects for refinement by an annotator; and outputting the one or more selected RoIs.
    Type: Application
    Filed: August 14, 2024
    Publication date: March 6, 2025
    Inventors: Julia Kabalar, Mireille Lucette Laure Gregoire, Hazem Ahmed Mohamed Mohamed Rashed, Dorel Mircea Coman, Nirnai Ach, Kiran Bangalore Ravi, Senthil Kumar Yogamani
  • Publication number: 20250080685
    Abstract: A method of image processing includes receiving first feature data from image content captured with a sensor, the first feature data having a first set of states with values that change non-linearly over time, generating second feature data based at least in part on the first feature data, the second feature data having a second set of states with values that change approximately linearly over time relative to a linear operator, wherein the second set of states is greater than the first set of states, and predicting movement of one or more objects in the image content based at least in part on the second feature data.
    Type: Application
    Filed: September 6, 2023
    Publication date: March 6, 2025
    Inventors: Ming-Yuan Yu, Varun Ravi Kumar, Senthil Kumar Yogamani
  • Publication number: 20250078294
    Abstract: A method includes receiving one or more images, wherein at least one of the one or more images depicts a water region and analyzing, by one or more processors, the one or more images using a first machine learning model to determine a depth of the water region. The method also includes analyzing, by the one or more processors, the one or more images using a second machine learning model to determine a surface normal of the water region and performing, by the one or more processors, using a third machine learning model, multi-class segmentation of the one or more images. Additionally, the method includes performing one or more fusion operations on outputs of at least two of the first machine learning model, the second machine learning model and the third machine learning model to generate a classification of the water region.
    Type: Application
    Filed: August 30, 2023
    Publication date: March 6, 2025
    Inventors: Varun Ravi Kumar, Debasmit Das, Senthil Kumar Yogamani
  • Publication number: 20250069184
    Abstract: A method of processing image content includes constructing a first graph representation having a first level of point sparsity from a first point cloud data, and performing diffusion-based upsampling on the first graph representation to generate a second graph representation having a second level of point sparsity. Performing diffusion-based upsampling includes inputting the first graph representation into a diffusion-based trained model to generate a first intermediate graph representation having a first intermediate level of point sparsity, inputting the first intermediate graph representation into the diffusion-based trained model to generate a second intermediate graph representation having a second intermediate level of point sparsity, and generating the second graph representation based on at least on the second intermediate graph representation.
    Type: Application
    Filed: August 24, 2023
    Publication date: February 27, 2025
    Inventors: Varun Ravi Kumar, Risheek Garrepalli, Senthil Kumar Yogamani
  • Publication number: 20250065907
    Abstract: Certain aspects of the present disclosure provide techniques and apparatus for improved machine learning. A set of object detections, each respective object detection in the set of object detections corresponding to a respective object detected in an environment, is accessed. Based on the set of object detections, a graph representation comprising a plurality of nodes is generated, where each respective node in the plurality of nodes corresponds to a respective object detection in the set of object detections. A set of output features is generated based on processing the graph representation using a trained message passing network. A predicted object relationship graph is generated based on processing the set of output features using a layer of a trained machine learning model.
    Type: Application
    Filed: August 25, 2023
    Publication date: February 27, 2025
    Inventors: Venkatraman NARAYANAN, Varun RAVI KUMAR, Senthil Kumar YOGAMANI
  • Publication number: 20250060481
    Abstract: An apparatus includes a memory and processing circuitry in communication with the memory. The processing circuitry is configured to apply, based on a positional encoding model, a first feature conditioning module to a set of bird's eye view (BEV) position data features corresponding to position data to generate a set of conditioned BEV position data features, and apply, based on the position encoding model, a second feature conditioning module to a set of perspective image data features corresponding to image data to generate a set of conditioned perspective image data features. The processing circuitry is also configured to generate, based on the positional encoding model, the set of conditioned BEV position data features, and the set of conditioned perspective image data features, a weighted summation. Additionally, the processing circuitry is configured to generate, based on the weighted summation, a set of BEV image data features.
    Type: Application
    Filed: August 18, 2023
    Publication date: February 20, 2025
    Inventors: Meysam Sadeghigooghari, Varun Ravi Kumar, Senthil Kumar Yogamani