Patents by Inventor Sakthivel Sivaraman
Sakthivel Sivaraman has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20250156717Abstract: Apparatuses, systems, and techniques are described to determine locations of objects using images including digital representations of those objects. In at least one embodiment, a gaze of one or more occupants of a vehicle is determined independently of a location of one or more sensors used to detect those occupants.Type: ApplicationFiled: January 15, 2025Publication date: May 15, 2025Inventors: Feng Hu, Niranjan Avadhanam, Yuzhuo Ren, Sujay Yadawadkar, Sakthivel Sivaraman, Hairong Jiang, Siyue Wu
-
Publication number: 20250142208Abstract: In various examples, an image processing pipeline may switch between different operating or switching modes based on speed of ego-motion and/or the active gear (e.g., park vs. drive) of a vehicle or other ego-machine in which an RGB/IR camera is being used. For example, a first operating or switching mode that toggles between IR and RGB imaging modes at a fixed frame rate or interval may be used when the vehicle is in motion, in a particular gear (e.g., drive), and/or traveling above a threshold speed. In another example, a second operating or switching mode that toggles between IR and RGB imaging modes based on detected light intensity may be used when the vehicle is in stationary, in park (or out of gear), and/or traveling below a threshold speed.Type: ApplicationFiled: October 25, 2023Publication date: May 1, 2025Inventors: Sakthivel SIVARAMAN, Rajath SHETTY, Animesh KHEMKA, Niranjan AVADHANAM
-
Publication number: 20250124734Abstract: Interactions with virtual systems may be difficult when users inadvertently fail to provide sufficient information to proceed with their requests. Certain types of inputs, such as auditory inputs, may lack sufficient information to properly provide a response to the user. Additional information, such as image data, may enable user gestures or poses to supplement the auditory inputs to enable response generation without requesting additional information from users.Type: ApplicationFiled: December 23, 2024Publication date: April 17, 2025Inventors: Sakthivel Sivaraman, Nishant Puri, Yuzhuo Ren, Atousa Torabi, Shubhadeep Das, Niranjan Avadhanam, Sumit Kumar Bhattacharya, Jason Roche
-
Patent number: 12236351Abstract: Apparatuses, systems, and techniques are described to determine locations of objects using images including digital representations of those objects. In at least one embodiment, a gaze of one or more occupants of a vehicle is determined independently of a location of one or more sensors used to detect those occupants.Type: GrantFiled: October 30, 2023Date of Patent: February 25, 2025Assignee: Nvidia CorporationInventors: Feng Hu, Niranjan Avadhanam, Yuzhuo Ren, Sujay Yadawadkar, Sakthivel Sivaraman, Hairong Jiang, Siyue Wu
-
Publication number: 20250050831Abstract: In various examples, systems and methods are disclosed that accurately identify driver and passenger in-cabin activities that may indicate a biomechanical distraction that prevents a driver from being fully engaged in driving a vehicle. In particular, image data representative of an image of an occupant of a vehicle may be applied to one or more deep neural networks (DNNs). Using the DNNs, data indicative of key point locations corresponding to the occupant may be computed, a shape and/or a volume corresponding to the occupant may be reconstructed, a position and size of the occupant may be estimated, hand gesture activities may be classified, and/or body postures or poses may be classified. These determinations may be used to determine operations or settings for the vehicle to increase not only the safety of the occupants, but also of surrounding motorists, bicyclists, and pedestrians.Type: ApplicationFiled: October 30, 2024Publication date: February 13, 2025Inventors: Atousa Torabi, Sakthivel Sivaraman, Niranjan Avadhanam, Shagan Sah
-
Patent number: 12211308Abstract: Interactions with virtual systems may be difficult when users inadvertently fail to provide sufficient information to proceed with their requests. Certain types of inputs, such as auditory inputs, may lack sufficient information to properly provide a response to the user. Additional information, such as image data, may enable user gestures or poses to supplement the auditory inputs to enable response generation without requesting additional information from users.Type: GrantFiled: August 31, 2021Date of Patent: January 28, 2025Assignee: Nvidia CorporationInventors: Sakthivel Sivaraman, Nishant Puri, Yuzhuo Ren, Atousa Torabi, Shubhadeep Das, Niranjan Avadhanam, Sumit Kumar Bhattacharya, Jason Roche
-
Publication number: 20250022288Abstract: In various examples, sensor data (e.g., image and/or RADAR data) may be used to detect occupants and classify them (e.g., as children or adults) using one or more predictions that represent estimated age (e.g., based on detected limb length, a detected face) and/or detected child presence (e.g., based on detecting an occupied child seat). In some embodiments, multiple predictions generated using multiple machine learning models (and optionally one or more corresponding confidence values) may be combined using a state machine and/or one or more machine learning models to generate a combined assessment of occupant presence and/or age for each occupant and/or supported occupant slot. As such, the techniques described herein may be utilized to detect child presence, detect unattended child presence, determine age or size of a particular occupant, and/or take some responsive action (e.g., trigger an alarm, control temperature, unlock door(s), permit or disable airbag deployment, etc.).Type: ApplicationFiled: July 10, 2023Publication date: January 16, 2025Inventors: Sakthivel SIVARAMAN, Arjun GURU, Rajath SHETTY, Shagan SAH, Varsha HEDAU
-
Publication number: 20250022290Abstract: In various examples, image-based three-dimensional occupant assessment for in-cabin monitoring systems and applications are provided. An evaluation function may determine a 3D representation of an occupant of a machine by evaluating sensor data comprising an image frame from an optical image sensor. The 3D representation may comprise at least one characteristic representative of a size of the occupant, (e.g., a 3D pose and/or 3D shape), which may be used to derive other characteristics such as, but not limited to weight, height, and/or age. A first processing path may generate a representation of one or more features corresponding to at least a portion of the occupant based on optical image data, and a second processing path may determine a depth corresponding to the one or more features based on depth data derived from the optical image data and ground truth depth data corresponding to the interior of the machine.Type: ApplicationFiled: July 10, 2023Publication date: January 16, 2025Inventors: Sakthivel SIVARAMAN, Arjun Guru, Rajath Shetty, Umar Iqbal, Orazio Gallo, Hang Su, Abhishek Badki, Varsha Hedau
-
OCCUPANT EVALUATION USING MULTI-MODAL SENSOR FUSION FOR IN-CABIN MONITORING SYSTEMS AND APPLICATIONS
Publication number: 20250022289Abstract: In various examples, occupant assessment using multi-modal sensor fusion for monitoring systems and applications are provided. In some embodiments, an occupant monitoring system comprises an occupant evaluation function that may predict at least one characteristic representative of a size of the occupant. The occupant evaluation function may include a first processing path that generates a representation of features corresponding to the occupant based on optical image data, and a second processing path that performs operations to determine a depth corresponding to the one or more features based on depth data derived from the optical image data and the point cloud depth data. In some embodiments, a three-dimensional pose detection model generates a three-dimensional pose estimate of the occupant using the optical image data, and the three-dimensional pose estimate is scaled to an absolute pose based on the point cloud depth data.Type: ApplicationFiled: July 10, 2023Publication date: January 16, 2025Inventors: Sakthivel SIVARAMAN, Rajath SHETTY, Hairong JIANG, Arjun GURU, Yuzhuo REN -
Publication number: 20250022155Abstract: In various examples, systems and methods for pose detection model training for predicting three-dimensional pose estimates using two-dimensional image data are provided. The occupant pose detection model may be trained using multi-view image sensor training data that includes image frames that capture a pose of a training subject within a machine interior using multiple synchronized optical image sensors placed around the machine interior that produce a set of captured image frames of the training subject from different viewpoints. Based on the multi-view image sensor training data, the occupant pose detection model may generate a set of individual, predicted 3D pose estimates for the training subject from a captured image frame from each of the respective optical image sensors. To adjust the occupant pose detection model during training, a loss feedback may be generated that comprises a pose alignment loss, a pose depth loss, and/or a ground truth kinematic loss.Type: ApplicationFiled: July 10, 2023Publication date: January 16, 2025Inventors: Sakthivel SIVARAMAN, Arjun Guru, Rajath Shetty
-
Patent number: 12162418Abstract: In various examples, systems and methods are disclosed that accurately identify driver and passenger in-cabin activities that may indicate a biomechanical distraction that prevents a driver from being fully engaged in driving a vehicle. In particular, image data representative of an image of an occupant of a vehicle may be applied to one or more deep neural networks (DNNs). Using the DNNs, data indicative of key point locations corresponding to the occupant may be computed, a shape and/or a volume corresponding to the occupant may be reconstructed, a position and size of the occupant may be estimated, hand gesture activities may be classified, and/or body postures or poses may be classified. These determinations may be used to determine operations or settings for the vehicle to increase not only the safety of the occupants, but also of surrounding motorists, bicyclists, and pedestrians.Type: GrantFiled: October 5, 2023Date of Patent: December 10, 2024Assignee: NVIDIA CorporationInventors: Atousa Torabi, Sakthivel Sivaraman, Niranjan Avadhanam, Shagan Sah
-
Publication number: 20240371136Abstract: In various examples, the present disclosure relates to using temporal filters for automated real-time classification. The technology described herein improves the performance of a multiclass classifier that may be used to classify a temporal sequence of input signals-such as input signals representative of video frames. A performance improvement may be achieved, at least in part, by applying a temporal filter to an output of the multiclass classifier. For example, the temporal filter may leverage classifications associated with preceding input signals to improve the final classification given to a subsequent signal. In some embodiments, the temporal filter may also use data from a confusion matrix to correct for the probable occurrence of certain types of classification errors. The temporal filter may be a linear filter, a nonlinear filter, an adaptive filter, and/or a statistical filter.Type: ApplicationFiled: July 12, 2024Publication date: November 7, 2024Inventors: Sakthivel Sivaraman, Shagan Sah, Niranjan Avadhanam
-
Patent number: 12073604Abstract: In various examples, the present disclosure relates to using temporal filters for automated real-time classification. The technology described herein improves the performance of a multiclass classifier that may be used to classify a temporal sequence of input signals—such as input signals representative of video frames. A performance improvement may be achieved, at least in part, by applying a temporal filter to an output of the multiclass classifier. For example, the temporal filter may leverage classifications associated with preceding input signals to improve the final classification given to a subsequent signal. In some embodiments, the temporal filter may also use data from a confusion matrix to correct for the probable occurrence of certain types of classification errors. The temporal filter may be a linear filter, a nonlinear filter, an adaptive filter, and/or a statistical filter.Type: GrantFiled: June 12, 2023Date of Patent: August 27, 2024Assignee: NVIDIA CorporationInventors: Sakthivel Sivaraman, Shagan Sah, Niranjan Avadhanam
-
Publication number: 20240062067Abstract: Apparatuses, systems, and techniques are described to determine locations of objects using images including digital representations of those objects. In at least one embodiment, a gaze of one or more occupants of a vehicle is determined independently of a location of one or more sensors used to detect those occupants.Type: ApplicationFiled: October 30, 2023Publication date: February 22, 2024Inventors: Feng Hu, Niranjan Avadhanam, Yuzhuo Ren, Sujay Yadawadkar, Sakthivel Sivaraman, Hairong Jiang, Siyue Wu
-
Publication number: 20240034260Abstract: In various examples, systems and methods are disclosed that accurately identify driver and passenger in-cabin activities that may indicate a biomechanical distraction that prevents a driver from being fully engaged in driving a vehicle. In particular, image data representative of an image of an occupant of a vehicle may be applied to one or more deep neural networks (DNNs). Using the DNNs, data indicative of key point locations corresponding to the occupant may be computed, a shape and/or a volume corresponding to the occupant may be reconstructed, a position and size of the occupant may be estimated, hand gesture activities may be classified, and/or body postures or poses may be classified. These determinations may be used to determine operations or settings for the vehicle to increase not only the safety of the occupants, but also of surrounding motorists, bicyclists, and pedestrians.Type: ApplicationFiled: October 5, 2023Publication date: February 1, 2024Inventors: Atousa Torabi, Sakthivel Sivaraman, Niranjan Avadhanam, Shagan Sah
-
Patent number: 11851015Abstract: In various examples, systems and methods are disclosed that accurately identify driver and passenger in-cabin activities that may indicate a biomechanical distraction that prevents a driver from being fully engaged in driving a vehicle. In particular, image data representative of an image of an occupant of a vehicle may be applied to one or more deep neural networks (DNNs). Using the DNNs, data indicative of key point locations corresponding to the occupant may be computed, a shape and/or a volume corresponding to the occupant may be reconstructed, a position and size of the occupant may be estimated, hand gesture activities may be classified, and/or body postures or poses may be classified. These determinations may be used to determine operations or settings for the vehicle to increase not only the safety of the occupants, but also of surrounding motorists, bicyclists, and pedestrians.Type: GrantFiled: September 7, 2022Date of Patent: December 26, 2023Assignee: NVIDIA CorporationInventors: Atousa Torabi, Sakthivel Sivaraman, Niranjan Avadhanam, Shagan Sah
-
Patent number: 11851014Abstract: In various examples, systems and methods are disclosed that accurately identify driver and passenger in-cabin activities that may indicate a biomechanical distraction that prevents a driver from being fully engaged in driving a vehicle. In particular, image data representative of an image of an occupant of a vehicle may be applied to one or more deep neural networks (DNNs). Using the DNNs, data indicative of key point locations corresponding to the occupant may be computed, a shape and/or a volume corresponding to the occupant may be reconstructed, a position and size of the occupant may be estimated, hand gesture activities may be classified, and/or body postures or poses may be classified. These determinations may be used to determine operations or settings for the vehicle to increase not only the safety of the occupants, but also of surrounding motorists, bicyclists, and pedestrians.Type: GrantFiled: September 7, 2022Date of Patent: December 26, 2023Assignee: NVIDIA CorporationInventors: Atousa Torabi, Sakthivel Sivaraman, Niranjan Avadhanam, Shagan Sah
-
Patent number: 11803759Abstract: Apparatuses, systems, and techniques are described to determine locations of objects using images including digital representations of those objects. In at least one embodiment, a gaze of one or more occupants of a vehicle is determined independently of a location of one or more sensors used to detect those occupants.Type: GrantFiled: October 11, 2021Date of Patent: October 31, 2023Assignee: Nvidia CorporationInventors: Feng Hu, Niranjan Avadhanam, Yuzhuo Ren, Sujay Yadawadkar, Sakthivel Sivaraman, Hairong Jiang, Siyue Wu
-
Publication number: 20230326182Abstract: In various examples, the present disclosure relates to using temporal filters for automated real-time classification. The technology described herein improves the performance of a multiclass classifier that may be used to classify a temporal sequence of input signals—such as input signals representative of video frames. A performance improvement may be achieved, at least in part, by applying a temporal filter to an output of the multiclass classifier. For example, the temporal filter may leverage classifications associated with preceding input signals to improve the final classification given to a subsequent signal. In some embodiments, the temporal filter may also use data from a confusion matrix to correct for the probable occurrence of certain types of classification errors. The temporal filter may be a linear filter, a nonlinear filter, an adaptive filter, and/or a statistical filter.Type: ApplicationFiled: June 12, 2023Publication date: October 12, 2023Inventors: Sakthivel Sivaraman, Shagan Sah, Niranjan Avadhanam
-
Patent number: 11721089Abstract: In various examples, the present disclosure relates to using temporal filters for automated real-time classification. The technology described herein improves the performance of a multiclass classifier that may be used to classify a temporal sequence of input signals—such as input signals representative of video frames. A performance improvement may be achieved, at least in part, by applying a temporal filter to an output of the multiclass classifier. For example, the temporal filter may leverage classifications associated with preceding input signals to improve the final classification given to a subsequent signal. In some embodiments, the temporal filter may also use data from a confusion matrix to correct for the probable occurrence of certain types of classification errors. The temporal filter may be a linear filter, a nonlinear filter, an adaptive filter, and/or a statistical filter.Type: GrantFiled: January 7, 2022Date of Patent: August 8, 2023Assignee: NVIDIA CorporationInventors: Sakthivel Sivaraman, Shagan Sah, Niranjan Avadhanam