Patents by Inventor Sakthivel Sivaraman

Sakthivel Sivaraman has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20250156717
    Abstract: Apparatuses, systems, and techniques are described to determine locations of objects using images including digital representations of those objects. In at least one embodiment, a gaze of one or more occupants of a vehicle is determined independently of a location of one or more sensors used to detect those occupants.
    Type: Application
    Filed: January 15, 2025
    Publication date: May 15, 2025
    Inventors: Feng Hu, Niranjan Avadhanam, Yuzhuo Ren, Sujay Yadawadkar, Sakthivel Sivaraman, Hairong Jiang, Siyue Wu
  • Publication number: 20250142208
    Abstract: In various examples, an image processing pipeline may switch between different operating or switching modes based on speed of ego-motion and/or the active gear (e.g., park vs. drive) of a vehicle or other ego-machine in which an RGB/IR camera is being used. For example, a first operating or switching mode that toggles between IR and RGB imaging modes at a fixed frame rate or interval may be used when the vehicle is in motion, in a particular gear (e.g., drive), and/or traveling above a threshold speed. In another example, a second operating or switching mode that toggles between IR and RGB imaging modes based on detected light intensity may be used when the vehicle is in stationary, in park (or out of gear), and/or traveling below a threshold speed.
    Type: Application
    Filed: October 25, 2023
    Publication date: May 1, 2025
    Inventors: Sakthivel SIVARAMAN, Rajath SHETTY, Animesh KHEMKA, Niranjan AVADHANAM
  • Publication number: 20250124734
    Abstract: Interactions with virtual systems may be difficult when users inadvertently fail to provide sufficient information to proceed with their requests. Certain types of inputs, such as auditory inputs, may lack sufficient information to properly provide a response to the user. Additional information, such as image data, may enable user gestures or poses to supplement the auditory inputs to enable response generation without requesting additional information from users.
    Type: Application
    Filed: December 23, 2024
    Publication date: April 17, 2025
    Inventors: Sakthivel Sivaraman, Nishant Puri, Yuzhuo Ren, Atousa Torabi, Shubhadeep Das, Niranjan Avadhanam, Sumit Kumar Bhattacharya, Jason Roche
  • Patent number: 12236351
    Abstract: Apparatuses, systems, and techniques are described to determine locations of objects using images including digital representations of those objects. In at least one embodiment, a gaze of one or more occupants of a vehicle is determined independently of a location of one or more sensors used to detect those occupants.
    Type: Grant
    Filed: October 30, 2023
    Date of Patent: February 25, 2025
    Assignee: Nvidia Corporation
    Inventors: Feng Hu, Niranjan Avadhanam, Yuzhuo Ren, Sujay Yadawadkar, Sakthivel Sivaraman, Hairong Jiang, Siyue Wu
  • Publication number: 20250050831
    Abstract: In various examples, systems and methods are disclosed that accurately identify driver and passenger in-cabin activities that may indicate a biomechanical distraction that prevents a driver from being fully engaged in driving a vehicle. In particular, image data representative of an image of an occupant of a vehicle may be applied to one or more deep neural networks (DNNs). Using the DNNs, data indicative of key point locations corresponding to the occupant may be computed, a shape and/or a volume corresponding to the occupant may be reconstructed, a position and size of the occupant may be estimated, hand gesture activities may be classified, and/or body postures or poses may be classified. These determinations may be used to determine operations or settings for the vehicle to increase not only the safety of the occupants, but also of surrounding motorists, bicyclists, and pedestrians.
    Type: Application
    Filed: October 30, 2024
    Publication date: February 13, 2025
    Inventors: Atousa Torabi, Sakthivel Sivaraman, Niranjan Avadhanam, Shagan Sah
  • Patent number: 12211308
    Abstract: Interactions with virtual systems may be difficult when users inadvertently fail to provide sufficient information to proceed with their requests. Certain types of inputs, such as auditory inputs, may lack sufficient information to properly provide a response to the user. Additional information, such as image data, may enable user gestures or poses to supplement the auditory inputs to enable response generation without requesting additional information from users.
    Type: Grant
    Filed: August 31, 2021
    Date of Patent: January 28, 2025
    Assignee: Nvidia Corporation
    Inventors: Sakthivel Sivaraman, Nishant Puri, Yuzhuo Ren, Atousa Torabi, Shubhadeep Das, Niranjan Avadhanam, Sumit Kumar Bhattacharya, Jason Roche
  • Publication number: 20250022288
    Abstract: In various examples, sensor data (e.g., image and/or RADAR data) may be used to detect occupants and classify them (e.g., as children or adults) using one or more predictions that represent estimated age (e.g., based on detected limb length, a detected face) and/or detected child presence (e.g., based on detecting an occupied child seat). In some embodiments, multiple predictions generated using multiple machine learning models (and optionally one or more corresponding confidence values) may be combined using a state machine and/or one or more machine learning models to generate a combined assessment of occupant presence and/or age for each occupant and/or supported occupant slot. As such, the techniques described herein may be utilized to detect child presence, detect unattended child presence, determine age or size of a particular occupant, and/or take some responsive action (e.g., trigger an alarm, control temperature, unlock door(s), permit or disable airbag deployment, etc.).
    Type: Application
    Filed: July 10, 2023
    Publication date: January 16, 2025
    Inventors: Sakthivel SIVARAMAN, Arjun GURU, Rajath SHETTY, Shagan SAH, Varsha HEDAU
  • Publication number: 20250022290
    Abstract: In various examples, image-based three-dimensional occupant assessment for in-cabin monitoring systems and applications are provided. An evaluation function may determine a 3D representation of an occupant of a machine by evaluating sensor data comprising an image frame from an optical image sensor. The 3D representation may comprise at least one characteristic representative of a size of the occupant, (e.g., a 3D pose and/or 3D shape), which may be used to derive other characteristics such as, but not limited to weight, height, and/or age. A first processing path may generate a representation of one or more features corresponding to at least a portion of the occupant based on optical image data, and a second processing path may determine a depth corresponding to the one or more features based on depth data derived from the optical image data and ground truth depth data corresponding to the interior of the machine.
    Type: Application
    Filed: July 10, 2023
    Publication date: January 16, 2025
    Inventors: Sakthivel SIVARAMAN, Arjun Guru, Rajath Shetty, Umar Iqbal, Orazio Gallo, Hang Su, Abhishek Badki, Varsha Hedau
  • Publication number: 20250022289
    Abstract: In various examples, occupant assessment using multi-modal sensor fusion for monitoring systems and applications are provided. In some embodiments, an occupant monitoring system comprises an occupant evaluation function that may predict at least one characteristic representative of a size of the occupant. The occupant evaluation function may include a first processing path that generates a representation of features corresponding to the occupant based on optical image data, and a second processing path that performs operations to determine a depth corresponding to the one or more features based on depth data derived from the optical image data and the point cloud depth data. In some embodiments, a three-dimensional pose detection model generates a three-dimensional pose estimate of the occupant using the optical image data, and the three-dimensional pose estimate is scaled to an absolute pose based on the point cloud depth data.
    Type: Application
    Filed: July 10, 2023
    Publication date: January 16, 2025
    Inventors: Sakthivel SIVARAMAN, Rajath SHETTY, Hairong JIANG, Arjun GURU, Yuzhuo REN
  • Publication number: 20250022155
    Abstract: In various examples, systems and methods for pose detection model training for predicting three-dimensional pose estimates using two-dimensional image data are provided. The occupant pose detection model may be trained using multi-view image sensor training data that includes image frames that capture a pose of a training subject within a machine interior using multiple synchronized optical image sensors placed around the machine interior that produce a set of captured image frames of the training subject from different viewpoints. Based on the multi-view image sensor training data, the occupant pose detection model may generate a set of individual, predicted 3D pose estimates for the training subject from a captured image frame from each of the respective optical image sensors. To adjust the occupant pose detection model during training, a loss feedback may be generated that comprises a pose alignment loss, a pose depth loss, and/or a ground truth kinematic loss.
    Type: Application
    Filed: July 10, 2023
    Publication date: January 16, 2025
    Inventors: Sakthivel SIVARAMAN, Arjun Guru, Rajath Shetty
  • Patent number: 12162418
    Abstract: In various examples, systems and methods are disclosed that accurately identify driver and passenger in-cabin activities that may indicate a biomechanical distraction that prevents a driver from being fully engaged in driving a vehicle. In particular, image data representative of an image of an occupant of a vehicle may be applied to one or more deep neural networks (DNNs). Using the DNNs, data indicative of key point locations corresponding to the occupant may be computed, a shape and/or a volume corresponding to the occupant may be reconstructed, a position and size of the occupant may be estimated, hand gesture activities may be classified, and/or body postures or poses may be classified. These determinations may be used to determine operations or settings for the vehicle to increase not only the safety of the occupants, but also of surrounding motorists, bicyclists, and pedestrians.
    Type: Grant
    Filed: October 5, 2023
    Date of Patent: December 10, 2024
    Assignee: NVIDIA Corporation
    Inventors: Atousa Torabi, Sakthivel Sivaraman, Niranjan Avadhanam, Shagan Sah
  • Publication number: 20240371136
    Abstract: In various examples, the present disclosure relates to using temporal filters for automated real-time classification. The technology described herein improves the performance of a multiclass classifier that may be used to classify a temporal sequence of input signals-such as input signals representative of video frames. A performance improvement may be achieved, at least in part, by applying a temporal filter to an output of the multiclass classifier. For example, the temporal filter may leverage classifications associated with preceding input signals to improve the final classification given to a subsequent signal. In some embodiments, the temporal filter may also use data from a confusion matrix to correct for the probable occurrence of certain types of classification errors. The temporal filter may be a linear filter, a nonlinear filter, an adaptive filter, and/or a statistical filter.
    Type: Application
    Filed: July 12, 2024
    Publication date: November 7, 2024
    Inventors: Sakthivel Sivaraman, Shagan Sah, Niranjan Avadhanam
  • Patent number: 12073604
    Abstract: In various examples, the present disclosure relates to using temporal filters for automated real-time classification. The technology described herein improves the performance of a multiclass classifier that may be used to classify a temporal sequence of input signals—such as input signals representative of video frames. A performance improvement may be achieved, at least in part, by applying a temporal filter to an output of the multiclass classifier. For example, the temporal filter may leverage classifications associated with preceding input signals to improve the final classification given to a subsequent signal. In some embodiments, the temporal filter may also use data from a confusion matrix to correct for the probable occurrence of certain types of classification errors. The temporal filter may be a linear filter, a nonlinear filter, an adaptive filter, and/or a statistical filter.
    Type: Grant
    Filed: June 12, 2023
    Date of Patent: August 27, 2024
    Assignee: NVIDIA Corporation
    Inventors: Sakthivel Sivaraman, Shagan Sah, Niranjan Avadhanam
  • Publication number: 20240062067
    Abstract: Apparatuses, systems, and techniques are described to determine locations of objects using images including digital representations of those objects. In at least one embodiment, a gaze of one or more occupants of a vehicle is determined independently of a location of one or more sensors used to detect those occupants.
    Type: Application
    Filed: October 30, 2023
    Publication date: February 22, 2024
    Inventors: Feng Hu, Niranjan Avadhanam, Yuzhuo Ren, Sujay Yadawadkar, Sakthivel Sivaraman, Hairong Jiang, Siyue Wu
  • Publication number: 20240034260
    Abstract: In various examples, systems and methods are disclosed that accurately identify driver and passenger in-cabin activities that may indicate a biomechanical distraction that prevents a driver from being fully engaged in driving a vehicle. In particular, image data representative of an image of an occupant of a vehicle may be applied to one or more deep neural networks (DNNs). Using the DNNs, data indicative of key point locations corresponding to the occupant may be computed, a shape and/or a volume corresponding to the occupant may be reconstructed, a position and size of the occupant may be estimated, hand gesture activities may be classified, and/or body postures or poses may be classified. These determinations may be used to determine operations or settings for the vehicle to increase not only the safety of the occupants, but also of surrounding motorists, bicyclists, and pedestrians.
    Type: Application
    Filed: October 5, 2023
    Publication date: February 1, 2024
    Inventors: Atousa Torabi, Sakthivel Sivaraman, Niranjan Avadhanam, Shagan Sah
  • Patent number: 11851015
    Abstract: In various examples, systems and methods are disclosed that accurately identify driver and passenger in-cabin activities that may indicate a biomechanical distraction that prevents a driver from being fully engaged in driving a vehicle. In particular, image data representative of an image of an occupant of a vehicle may be applied to one or more deep neural networks (DNNs). Using the DNNs, data indicative of key point locations corresponding to the occupant may be computed, a shape and/or a volume corresponding to the occupant may be reconstructed, a position and size of the occupant may be estimated, hand gesture activities may be classified, and/or body postures or poses may be classified. These determinations may be used to determine operations or settings for the vehicle to increase not only the safety of the occupants, but also of surrounding motorists, bicyclists, and pedestrians.
    Type: Grant
    Filed: September 7, 2022
    Date of Patent: December 26, 2023
    Assignee: NVIDIA Corporation
    Inventors: Atousa Torabi, Sakthivel Sivaraman, Niranjan Avadhanam, Shagan Sah
  • Patent number: 11851014
    Abstract: In various examples, systems and methods are disclosed that accurately identify driver and passenger in-cabin activities that may indicate a biomechanical distraction that prevents a driver from being fully engaged in driving a vehicle. In particular, image data representative of an image of an occupant of a vehicle may be applied to one or more deep neural networks (DNNs). Using the DNNs, data indicative of key point locations corresponding to the occupant may be computed, a shape and/or a volume corresponding to the occupant may be reconstructed, a position and size of the occupant may be estimated, hand gesture activities may be classified, and/or body postures or poses may be classified. These determinations may be used to determine operations or settings for the vehicle to increase not only the safety of the occupants, but also of surrounding motorists, bicyclists, and pedestrians.
    Type: Grant
    Filed: September 7, 2022
    Date of Patent: December 26, 2023
    Assignee: NVIDIA Corporation
    Inventors: Atousa Torabi, Sakthivel Sivaraman, Niranjan Avadhanam, Shagan Sah
  • Patent number: 11803759
    Abstract: Apparatuses, systems, and techniques are described to determine locations of objects using images including digital representations of those objects. In at least one embodiment, a gaze of one or more occupants of a vehicle is determined independently of a location of one or more sensors used to detect those occupants.
    Type: Grant
    Filed: October 11, 2021
    Date of Patent: October 31, 2023
    Assignee: Nvidia Corporation
    Inventors: Feng Hu, Niranjan Avadhanam, Yuzhuo Ren, Sujay Yadawadkar, Sakthivel Sivaraman, Hairong Jiang, Siyue Wu
  • Publication number: 20230326182
    Abstract: In various examples, the present disclosure relates to using temporal filters for automated real-time classification. The technology described herein improves the performance of a multiclass classifier that may be used to classify a temporal sequence of input signals—such as input signals representative of video frames. A performance improvement may be achieved, at least in part, by applying a temporal filter to an output of the multiclass classifier. For example, the temporal filter may leverage classifications associated with preceding input signals to improve the final classification given to a subsequent signal. In some embodiments, the temporal filter may also use data from a confusion matrix to correct for the probable occurrence of certain types of classification errors. The temporal filter may be a linear filter, a nonlinear filter, an adaptive filter, and/or a statistical filter.
    Type: Application
    Filed: June 12, 2023
    Publication date: October 12, 2023
    Inventors: Sakthivel Sivaraman, Shagan Sah, Niranjan Avadhanam
  • Patent number: 11721089
    Abstract: In various examples, the present disclosure relates to using temporal filters for automated real-time classification. The technology described herein improves the performance of a multiclass classifier that may be used to classify a temporal sequence of input signals—such as input signals representative of video frames. A performance improvement may be achieved, at least in part, by applying a temporal filter to an output of the multiclass classifier. For example, the temporal filter may leverage classifications associated with preceding input signals to improve the final classification given to a subsequent signal. In some embodiments, the temporal filter may also use data from a confusion matrix to correct for the probable occurrence of certain types of classification errors. The temporal filter may be a linear filter, a nonlinear filter, an adaptive filter, and/or a statistical filter.
    Type: Grant
    Filed: January 7, 2022
    Date of Patent: August 8, 2023
    Assignee: NVIDIA Corporation
    Inventors: Sakthivel Sivaraman, Shagan Sah, Niranjan Avadhanam