Patents by Inventor Atousa Torabi

Atousa Torabi has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240034260
    Abstract: In various examples, systems and methods are disclosed that accurately identify driver and passenger in-cabin activities that may indicate a biomechanical distraction that prevents a driver from being fully engaged in driving a vehicle. In particular, image data representative of an image of an occupant of a vehicle may be applied to one or more deep neural networks (DNNs). Using the DNNs, data indicative of key point locations corresponding to the occupant may be computed, a shape and/or a volume corresponding to the occupant may be reconstructed, a position and size of the occupant may be estimated, hand gesture activities may be classified, and/or body postures or poses may be classified. These determinations may be used to determine operations or settings for the vehicle to increase not only the safety of the occupants, but also of surrounding motorists, bicyclists, and pedestrians.
    Type: Application
    Filed: October 5, 2023
    Publication date: February 1, 2024
    Inventors: Atousa Torabi, Sakthivel Sivaraman, Niranjan Avadhanam, Shagan Sah
  • Patent number: 11851014
    Abstract: In various examples, systems and methods are disclosed that accurately identify driver and passenger in-cabin activities that may indicate a biomechanical distraction that prevents a driver from being fully engaged in driving a vehicle. In particular, image data representative of an image of an occupant of a vehicle may be applied to one or more deep neural networks (DNNs). Using the DNNs, data indicative of key point locations corresponding to the occupant may be computed, a shape and/or a volume corresponding to the occupant may be reconstructed, a position and size of the occupant may be estimated, hand gesture activities may be classified, and/or body postures or poses may be classified. These determinations may be used to determine operations or settings for the vehicle to increase not only the safety of the occupants, but also of surrounding motorists, bicyclists, and pedestrians.
    Type: Grant
    Filed: September 7, 2022
    Date of Patent: December 26, 2023
    Assignee: NVIDIA Corporation
    Inventors: Atousa Torabi, Sakthivel Sivaraman, Niranjan Avadhanam, Shagan Sah
  • Patent number: 11851015
    Abstract: In various examples, systems and methods are disclosed that accurately identify driver and passenger in-cabin activities that may indicate a biomechanical distraction that prevents a driver from being fully engaged in driving a vehicle. In particular, image data representative of an image of an occupant of a vehicle may be applied to one or more deep neural networks (DNNs). Using the DNNs, data indicative of key point locations corresponding to the occupant may be computed, a shape and/or a volume corresponding to the occupant may be reconstructed, a position and size of the occupant may be estimated, hand gesture activities may be classified, and/or body postures or poses may be classified. These determinations may be used to determine operations or settings for the vehicle to increase not only the safety of the occupants, but also of surrounding motorists, bicyclists, and pedestrians.
    Type: Grant
    Filed: September 7, 2022
    Date of Patent: December 26, 2023
    Assignee: NVIDIA Corporation
    Inventors: Atousa Torabi, Sakthivel Sivaraman, Niranjan Avadhanam, Shagan Sah
  • Publication number: 20230410650
    Abstract: In various examples, audio alerts of emergency response vehicles may be detected and classified using audio captured by microphones of an autonomous or semi-autonomous machine in order to identify travel directions, locations, and/or types of emergency response vehicles in the environment. For example, a plurality of microphone arrays may be disposed on an autonomous or semi-autonomous machine and used to generate audio signals corresponding to sounds in the environment. These audio signals may be processed to determine a location and/or direction of travel of an emergency response vehicle (e.g., using triangulation). Additionally, to identify siren types—and thus emergency response vehicle types corresponding thereto—the audio signals may be used to generate representations of a frequency spectrum that may be processed using a deep neural network (DNN) that outputs probabilities of alert types being represented by the audio data.
    Type: Application
    Filed: September 6, 2023
    Publication date: December 21, 2023
    Inventors: Ambrish Dantrey, Atousa Torabi, Anshul Jain, Ram Ganapathi, Abhijit Patait, Revanth Reddy Nalla, Niranjan Avadhanam
  • Patent number: 11816987
    Abstract: In various examples, audio alerts of emergency response vehicles may be detected and classified using audio captured by microphones of an autonomous or semi-autonomous machine in order to identify travel directions, locations, and/or types of emergency response vehicles in the environment. For example, a plurality of microphone arrays may be disposed on an autonomous or semi-autonomous machine and used to generate audio signals corresponding to sounds in the environment. These audio signals may be processed to determine a location and/or direction of travel of an emergency response vehicle (e.g., using triangulation). Additionally, to identify siren types—and thus emergency response vehicle types corresponding thereto—the audio signals may be used to generate representations of a frequency spectrum that may be processed using a deep neural network (DNN) that outputs probabilities of alert types being represented by the audio data.
    Type: Grant
    Filed: November 18, 2020
    Date of Patent: November 14, 2023
    Assignee: NVIDIA Corporation
    Inventors: Ambrish Dantrey, Atousa Torabi, Anshul Jain, Ram Ganapathi, Abhijit Patait, Revanth Reddy Nalla, Niranjan Avadhanam
  • Publication number: 20230356728
    Abstract: Approaches for an advanced AI-assisted vehicle can utilize an extensive suite of sensors inside and outside the vehicle, providing information to a computing platform running one or more neural networks. The neural networks can perform functions such as facial recognition, eye tracking, gesture recognition, head position, and gaze tracking to monitor the condition and safety of the driver and passengers. The system also identifies and tracks body pose and signals of people inside and outside the vehicle to understand their intent and actions. The system can track driver gaze to identify objects the driver might not see, such as cross-traffic and approaching cyclists. The system can provide notification of potential hazards, advice, and warnings. The system can also take corrective action, which may include controlling one or more vehicle subsystems, or when necessary, autonomously controlling the entire vehicle. The system can work with vehicle systems for enhanced analytics and recommendations.
    Type: Application
    Filed: May 8, 2023
    Publication date: November 9, 2023
    Inventors: Anshul Jain, Ratin Kumar, Feng Hu, Niranjan Avadhanam, Atousa Torabi, Hairong Jiang, Ram Ganapathi, Taek Kim
  • Patent number: 11682272
    Abstract: Systems and methods are disclosed herein for a pedestrian crossing warning system that may use multi-modal technology to determine attributes of a person and provide a warning to the person in response to a calculated risk level to effect a reduction of the risk level. The system may utilize sensors to receive data indicative of a trajectory of a person external to the vehicle. Specific attributes of the person such as age or walking aids may be determined. Based on the trajectory data and the specific attributes, a risk level may be determined by the system using a machine learning model. The system may cause emission of a warning to the person in response to the risk level.
    Type: Grant
    Filed: July 7, 2020
    Date of Patent: June 20, 2023
    Assignee: NVIDIA Corporation
    Inventors: Niranjan Avadhanam, Sumit Bhattacharya, Atousa Torabi, Jason Conrad Roche
  • Publication number: 20230064049
    Abstract: Interactions with virtual systems may be difficult when users inadvertently fail to provide sufficient information to proceed with their requests. Certain types of inputs, such as auditory inputs, may lack sufficient information to properly provide a response to the user. Additional information, such as image data, may enable user gestures or poses to supplement the auditory inputs to enable response generation without requesting additional information from users.
    Type: Application
    Filed: August 31, 2021
    Publication date: March 2, 2023
    Inventors: Sakthivel Sivaraman, Nishant Puri, Yuzhuo Ren, Atousa Torabi, Shubhadeep Das, Niranjan Avadhanam, Sumit Kumar Bhattacharya, Jason Roche
  • Publication number: 20230001872
    Abstract: In various examples, systems and methods are disclosed that accurately identify driver and passenger in-cabin activities that may indicate a biomechanical distraction that prevents a driver from being fully engaged in driving a vehicle. In particular, image data representative of an image of an occupant of a vehicle may be applied to one or more deep neural networks (DNNs). Using the DNNs, data indicative of key point locations corresponding to the occupant may be computed, a shape and/or a volume corresponding to the occupant may be reconstructed, a position and size of the occupant may be estimated, hand gesture activities may be classified, and/or body postures or poses may be classified. These determinations may be used to determine operations or settings for the vehicle to increase not only the safety of the occupants, but also of surrounding motorists, bicyclists, and pedestrians.
    Type: Application
    Filed: September 7, 2022
    Publication date: January 5, 2023
    Inventors: Atousa Torabi, Sakthivel Sivaraman, Niranjan Avadhanam, Shagan Sah
  • Publication number: 20220410830
    Abstract: In various examples, systems and methods are disclosed that accurately identify driver and passenger in-cabin activities that may indicate a biomechanical distraction that prevents a driver from being fully engaged in driving a vehicle. In particular, image data representative of an image of an occupant of a vehicle may be applied to one or more deep neural networks (DNNs). Using the DNNs, data indicative of key point locations corresponding to the occupant may be computed, a shape and/or a volume corresponding to the occupant may be reconstructed, a position and size of the occupant may be estimated, hand gesture activities may be classified, and/or body postures or poses may be classified. These determinations may be used to determine operations or settings for the vehicle to increase not only the safety of the occupants, but also of surrounding motorists, bicyclists, and pedestrians.
    Type: Application
    Filed: September 7, 2022
    Publication date: December 29, 2022
    Inventors: Atousa Torabi, Sakthivel Sivaraman, Niranjan Avadhanam, Shagan Sah
  • Patent number: 11485308
    Abstract: In various examples, systems and methods are disclosed that accurately identify driver and passenger in-cabin activities that may indicate a biomechanical distraction that prevents a driver from being fully engaged in driving a vehicle. In particular, image data representative of an image of an occupant of a vehicle may be applied to one or more deep neural networks (DNNs). Using the DNNs, data indicative of key point locations corresponding to the occupant may be computed, a shape and/or a volume corresponding to the occupant may be reconstructed, a position and size of the occupant may be estimated, hand gesture activities may be classified, and/or body postures or poses may be classified. These determinations may be used to determine operations or settings for the vehicle to increase not only the safety of the occupants, but also of surrounding motorists, bicyclists, and pedestrians.
    Type: Grant
    Filed: June 29, 2020
    Date of Patent: November 1, 2022
    Assignee: NVIDIA Corporation
    Inventors: Atousa Torabi, Sakthivel Sivaraman, Niranjan Avadhanam, Shagan Sah
  • Patent number: 11409791
    Abstract: Systems, methods and articles of manufacture for modeling a joint language-visual space. A textual query to be evaluated relative to a video library is received from a requesting entity. The video library contains a plurality of instances of video content. One or more instances of video content from the video library that correspond to the textual query are determined, by analyzing the textual query using a data model that includes a soft-attention neural network module that is jointly trained with a language Long Short-term Memory (LSTM) neural network module and a video LSTM neural network module. At least an indication of the one or more instances of video content is returned to the requesting entity.
    Type: Grant
    Filed: June 12, 2017
    Date of Patent: August 9, 2022
    Assignee: Disney Enterprises, Inc.
    Inventors: Atousa Torabi, Leonid Sigal
  • Publication number: 20220157165
    Abstract: In various examples, audio alerts of emergency response vehicles may be detected and classified using audio captured by microphones of an autonomous or semi-autonomous machine in order to identify travel directions, locations, and/or types of emergency response vehicles in the environment. For example, a plurality of microphone arrays may be disposed on an autonomous or semi-autonomous machine and used to generate audio signals corresponding to sounds in the environment. These audio signals may be processed to determine a location and/or direction of travel of an emergency response vehicle (e.g., using triangulation). Additionally, to identify siren types—and thus emergency response vehicle types corresponding thereto—the audio signals may be used to generate representations of a frequency spectrum that may be processed using a deep neural network (DNN) that outputs probabilities of alert types being represented by the audio data.
    Type: Application
    Filed: November 18, 2020
    Publication date: May 19, 2022
    Inventors: Ambrish Dantrey, Atousa Torabi, Anshul Jain, Ram Ganapathi, Abhijit Patait, Revanth Reddy Nalla, Niranjan Avadhanam
  • Publication number: 20220012988
    Abstract: Systems and methods are disclosed herein for a pedestrian crossing warning system that may use multi-modal technology to determine attributes of a person and provide a warning to the person in response to a calculated risk level to effect a reduction of the risk level. The system may utilize sensors to receive data indicative of a trajectory of a person external to the vehicle. Specific attributes of the person such as age or walking aids may be determined. Based on the trajectory data and the specific attributes, a risk level may be determined by the system using a machine learning model. The system may cause emission of a warning to the person in response to the risk level.
    Type: Application
    Filed: July 7, 2020
    Publication date: January 13, 2022
    Inventors: Niranjan Avadhanam, Sumit Bhattacharya, Atousa Torabi, Jason Conrad Roche
  • Publication number: 20210402942
    Abstract: In various examples, systems and methods are disclosed that accurately identify driver and passenger in-cabin activities that may indicate a biomechanical distraction that prevents a driver from being fully engaged in driving a vehicle. In particular, image data representative of an image of an occupant of a vehicle may be applied to one or more deep neural networks (DNNs). Using the DNNs, data indicative of key point locations corresponding to the occupant may be computed, a shape and/or a volume corresponding to the occupant may be reconstructed, a position and size of the occupant may be estimated, hand gesture activities may be classified, and/or body postures or poses may be classified. These determinations may be used to determine operations or settings for the vehicle to increase not only the safety of the occupants, but also of surrounding motorists, bicyclists, and pedestrians.
    Type: Application
    Filed: June 29, 2020
    Publication date: December 30, 2021
    Inventors: Atousa Torabi, Sakthivel Sivaraman, Niranjan Avadhanam, Shagan Sah
  • Patent number: 11055537
    Abstract: There is provided a system comprising a label database including a plurality of label, a non-transitory memory storing an executable code, and a hardware processor executing the executable code to receive a media content including a plurality of segments, each segment including a plurality of frames, extract a first plurality of features from a segment, extract a second plurality of features from each frame of the segment, determine an attention weight for each frame of the segment based on the first plurality of features extracted from the segment and the second plurality of features extracted from the segment, and determine that the segment depicts one of the plurality of labels in a label database based on the first plurality of features, the second plurality of features, and the attention weight of each frame of the plurality of frames of the segment.
    Type: Grant
    Filed: July 5, 2016
    Date of Patent: July 6, 2021
    Assignee: Disney Enterprises, Inc.
    Inventors: Atousa Torabi, Leonid Sigal
  • Publication number: 20170357720
    Abstract: Systems, methods and articles of manufacture for modeling a joint language-visual space. A textual query to be evaluated relative to a video library is received from a requesting entity. The video library contains a plurality of instances of video content. One or more instances of video content from the video library that correspond to the textual query are determined, by analyzing the textual query using a data model that includes a soft-attention neural network module that is jointly trained with a language Long Short-term Memory (LSTM) neural network module and a video LSTM neural network module. At least an indication of the one or more instances of video content is returned to the requesting entity.
    Type: Application
    Filed: June 12, 2017
    Publication date: December 14, 2017
    Inventors: Atousa TORABI, Leonid SIGAL
  • Publication number: 20170308754
    Abstract: There is provided a system comprising a label database including a plurality of label, a non-transitory memory storing an executable code, and a hardware processor executing the executable code to receive a media content including a plurality of segments, each segment including a plurality of frames, extract a first plurality of features from a segment, extract a second plurality of features from each frame of the segment, determine an attention weight for each frame of the segment based on the first plurality of features extracted from the segment and the second plurality of features extracted from the segment, and determine that the segment depicts one of the plurality of labels in a label database based on the first plurality of features, the second plurality of features, and the attention weight of each frame of the plurality of frames of the segment.
    Type: Application
    Filed: July 5, 2016
    Publication date: October 26, 2017
    Inventors: Atousa Torabi, Leonid Sigal