Patents by Inventor Deva Ramanan

Deva Ramanan has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11794731
    Abstract: Systems and methods of determining trajectories of an actor in an environment in which a vehicle is operating are provided. The method includes, by an object detection system of a vehicle in an environment, detecting an actor that may move within a scene in the environment. The method further includes using context of the scene to determine a reference polyline for the actor and determining a kinematic history of the actor. The method additionally includes using the kinematic history to predict a waypoint, which is a predicted position of the actor at a conclusion of a waypoint time period, and identifying a segment of the reference polyline, the segment extending from a current location to a point along the reference polyline that is closest to the waypoint and determining a trajectory for the actor conditioned by the segment of the reference polyline.
    Type: Grant
    Filed: May 5, 2021
    Date of Patent: October 24, 2023
    Assignee: Ford Global Technologies, LLC
    Inventors: Siddhesh Shyam Khandelwal, William Junbo Qi, Jagjeet Singh, Andrew T. Hartnett, Deva Ramanan
  • Patent number: 11704912
    Abstract: A method is disclosed for evaluating a classifier used to determine a traffic light signal state in images. The method includes, by a computer vision system of a vehicle, receiving at least one image of a traffic signal device of an imminent intersection. The traffic signal device includes a traffic signal face including one or more traffic signal elements. The method includes classifying, by a traffic light classifier (TLC), a classification state of the traffic signal face using labeled images correlated to the received at least one image. The classification state controls an operation of the vehicle at the intersection. The method includes evaluating a performance of the classifying of the classification state generated by the TLC. The evaluation is a label-free performance evaluation based on unlabeled images. The method includes training the TLC based on the evaluated performance.
    Type: Grant
    Filed: June 16, 2020
    Date of Patent: July 18, 2023
    Assignee: Ford Global Technologies, LLC
    Inventors: Guy Hotson, Richard L. Kwant, Brett Browning, Deva Ramanan
  • Patent number: 11568545
    Abstract: Various embodiments of a framework which allow, as an alternative to resource-taxing decompression, efficient computation of feature maps using a compressed content data subset, such as video, by exploiting the motion information, such as a motion vector, present in the compressed video. This framework allows frame-specific object recognition and action detection algorithms to be applied to compressed video and other media files by executing only on I-frames in a Group of Pictures and linearly interpolating the results. Training and machine learning increases recognition accuracy. Yielding significant computational gains, this approach accelerates frame-wise feature extraction I-frame/P-frame/P-frame videos as well as I-frame/P-frame/B-frame videos. The present techniques may also be used for segmentation to identify and label respective regions for objects in a video.
    Type: Grant
    Filed: December 27, 2019
    Date of Patent: January 31, 2023
    Assignee: A9.com, Inc.
    Inventors: R. Manmatha, Hexiang Hu, Deva Ramanan
  • Publication number: 20220122620
    Abstract: Systems and methods for siren detection in a vehicle are provided. A method includes recording an audio segment, using a first audio recording device coupled to an autonomous vehicle, separating, using a computing device coupled to the autonomous vehicle, the audio segment into one or more audio clips, generating a spectrogram of the one or more audio clips, and inputting each spectrogram into a Convolutional Neural Network (CNN) run on the computing device. The CNN may be pretrained to detect one or more sirens present in spectrographic data. The method further includes determining, using the CNN, whether a siren is present in the audio segment, and if the siren is determined to be present in the audio segment, determining a course of action of the autonomous vehicle.
    Type: Application
    Filed: October 19, 2020
    Publication date: April 21, 2022
    Inventors: Olivia Watkins, Nathan Pendleton, Guy Hotson, Chao Fang, Richard L. Kwant, Weihua Gao, Deva Ramanan, Nicolas Cebron, Brett Browning
  • Publication number: 20220048498
    Abstract: Systems and methods of determining trajectories of an actor in an environment in which a vehicle is operating are provided. The method includes, by an object detection system of a vehicle in an environment, detecting an actor that may move within a scene in the environment. The method further includes using context of the scene to determine a reference polyline for the actor and determining a kinematic history of the actor. The method additionally includes using the kinematic history to predict a waypoint, which is a predicted position of the actor at a conclusion of a waypoint time period, and identifying a segment of the reference polyline, the segment extending from a current location to a point along the reference polyline that is closest to the waypoint and determining a trajectory for the actor conditioned by the segment of the reference polyline.
    Type: Application
    Filed: May 5, 2021
    Publication date: February 17, 2022
    Inventors: Siddhesh Shyam Khandelwal, William Junbo Qi, Jagjeet Singh, Andrew T. Hartnett, Deva Ramanan
  • Publication number: 20220048503
    Abstract: Systems and methods of determining trajectories of an actor in an environment in which a vehicle is operating are provided. The method includes detecting an actor that may move within a scene in the environment by an object detection system of a vehicle in the environment, determining a kinematic history of the actor, and using context of the scene and the kinematic history of the actor to determine a plurality of reference polylines for the actor. The method further includes generating a contextual embedding of the kinematic history of the actor to generate a plurality of predicted trajectories of the actor, in which the generating conditions each of the predicted trajectories to correspond to one of the reference polylines. The method additionally includes using, by the vehicle, the plurality of predicted trajectories to plan movement of the vehicle.
    Type: Application
    Filed: May 5, 2021
    Publication date: February 17, 2022
    Inventors: Siddhesh Shyam Khandelwal, William Junbo Qi, Jagjeet Singh, Andrew T. Hartnett, Deva Ramanan
  • Publication number: 20210390349
    Abstract: A method is disclosed for evaluating a classifier used to determine a traffic light signal state in images. The method includes, by a computer vision system of a vehicle, receiving at least one image of a traffic signal device of an imminent intersection. The traffic signal device includes a traffic signal face including one or more traffic signal elements. The method includes classifying, by a traffic light classifier (TLC), a classification state of the traffic signal face using labeled images correlated to the received at least one image. The classification state controls an operation of the vehicle at the intersection. The method includes evaluating a performance of the classifying of the classification state generated by the TLC. The evaluation is a label-free performance evaluation based on unlabeled images. The method includes training the TLC based on the evaluated performance.
    Type: Application
    Filed: June 16, 2020
    Publication date: December 16, 2021
    Inventors: Guy Hotson, Richard L. Kwant, Brett Browning, Deva Ramanan
  • Publication number: 20210342924
    Abstract: Various embodiments of a framework which allow, as an alternative to resource-taxing decompression, efficient computation of feature maps using a compressed content data subset, such as video, by exploiting the motion information, such as a motion vector, present in the compressed video. This framework allows frame-specific object recognition and action detection algorithms to be applied to compressed video and other media files by executing only on I-frames in a Group of Pictures and linearly interpolating the results. Training and machine learning increases recognition accuracy. Yielding significant computational gains, this approach accelerates frame-wise feature extraction I-frame/P-frame/P-frame videos as well as I-frame/P-frame/B-frame videos. The present techniques may also be used for segmentation to identify and label respective regions for objects in a video.
    Type: Application
    Filed: December 27, 2019
    Publication date: November 4, 2021
    Inventors: R. Manmatha, Hexiang Hu, Deva Ramanan
  • Publication number: 20200143457
    Abstract: Various embodiments of a framework which allow, as an alternative to resource-taxing decompression, efficient computation of feature maps using a compressed content data subset, such as video, by exploiting the motion information, such as a motion vector, present in the compressed video. This framework allows frame-specific object recognition and action detection algorithms to be applied to compressed video and other media files by executing only on I-frames in a Group of Pictures and linearly interpolating the results. Training and machine learning increases recognition accuracy. Yielding significant computational gains, this approach accelerates frame-wise feature extraction I-frame/P-frame/P-frame videos as well as I-frame/P-frame/B-frame videos. The present techniques may also be used for segmentation to identify and label respective regions for objects in a video.
    Type: Application
    Filed: December 27, 2019
    Publication date: May 7, 2020
    Inventors: R. Manmatha, Hexiang Hu, Deva Ramanan
  • Patent number: 10528819
    Abstract: Various embodiments of a framework which allow, as an alternative to resource-taxing decompression, efficient computation of feature maps using a compressed content data subset, such as video, by exploiting the motion information, such as a motion vector, present in the compressed video. This framework allows frame-specific object recognition and action detection algorithms to be applied to compressed video and other media files by executing only on I-frames in a Group of Pictures and linearly interpolating the results. Training and machine learning increases recognition accuracy. Yielding significant computational gains, this approach accelerates frame-wise feature extraction I-frame/P-frame/P-frame videos as well as I-frame/P-frame/B-frame videos. The present techniques may also be used for segmentation to identify and label respective regions for objects in a video.
    Type: Grant
    Filed: November 20, 2017
    Date of Patent: January 7, 2020
    Assignee: A9.COM, INC.
    Inventors: R. Manmatha, Hexiang Hu, Deva Ramanan