Patents by Inventor Deva Ramanan
Deva Ramanan has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11794731Abstract: Systems and methods of determining trajectories of an actor in an environment in which a vehicle is operating are provided. The method includes, by an object detection system of a vehicle in an environment, detecting an actor that may move within a scene in the environment. The method further includes using context of the scene to determine a reference polyline for the actor and determining a kinematic history of the actor. The method additionally includes using the kinematic history to predict a waypoint, which is a predicted position of the actor at a conclusion of a waypoint time period, and identifying a segment of the reference polyline, the segment extending from a current location to a point along the reference polyline that is closest to the waypoint and determining a trajectory for the actor conditioned by the segment of the reference polyline.Type: GrantFiled: May 5, 2021Date of Patent: October 24, 2023Assignee: Ford Global Technologies, LLCInventors: Siddhesh Shyam Khandelwal, William Junbo Qi, Jagjeet Singh, Andrew T. Hartnett, Deva Ramanan
-
Patent number: 11704912Abstract: A method is disclosed for evaluating a classifier used to determine a traffic light signal state in images. The method includes, by a computer vision system of a vehicle, receiving at least one image of a traffic signal device of an imminent intersection. The traffic signal device includes a traffic signal face including one or more traffic signal elements. The method includes classifying, by a traffic light classifier (TLC), a classification state of the traffic signal face using labeled images correlated to the received at least one image. The classification state controls an operation of the vehicle at the intersection. The method includes evaluating a performance of the classifying of the classification state generated by the TLC. The evaluation is a label-free performance evaluation based on unlabeled images. The method includes training the TLC based on the evaluated performance.Type: GrantFiled: June 16, 2020Date of Patent: July 18, 2023Assignee: Ford Global Technologies, LLCInventors: Guy Hotson, Richard L. Kwant, Brett Browning, Deva Ramanan
-
Patent number: 11568545Abstract: Various embodiments of a framework which allow, as an alternative to resource-taxing decompression, efficient computation of feature maps using a compressed content data subset, such as video, by exploiting the motion information, such as a motion vector, present in the compressed video. This framework allows frame-specific object recognition and action detection algorithms to be applied to compressed video and other media files by executing only on I-frames in a Group of Pictures and linearly interpolating the results. Training and machine learning increases recognition accuracy. Yielding significant computational gains, this approach accelerates frame-wise feature extraction I-frame/P-frame/P-frame videos as well as I-frame/P-frame/B-frame videos. The present techniques may also be used for segmentation to identify and label respective regions for objects in a video.Type: GrantFiled: December 27, 2019Date of Patent: January 31, 2023Assignee: A9.com, Inc.Inventors: R. Manmatha, Hexiang Hu, Deva Ramanan
-
Publication number: 20220122620Abstract: Systems and methods for siren detection in a vehicle are provided. A method includes recording an audio segment, using a first audio recording device coupled to an autonomous vehicle, separating, using a computing device coupled to the autonomous vehicle, the audio segment into one or more audio clips, generating a spectrogram of the one or more audio clips, and inputting each spectrogram into a Convolutional Neural Network (CNN) run on the computing device. The CNN may be pretrained to detect one or more sirens present in spectrographic data. The method further includes determining, using the CNN, whether a siren is present in the audio segment, and if the siren is determined to be present in the audio segment, determining a course of action of the autonomous vehicle.Type: ApplicationFiled: October 19, 2020Publication date: April 21, 2022Inventors: Olivia Watkins, Nathan Pendleton, Guy Hotson, Chao Fang, Richard L. Kwant, Weihua Gao, Deva Ramanan, Nicolas Cebron, Brett Browning
-
Publication number: 20220048498Abstract: Systems and methods of determining trajectories of an actor in an environment in which a vehicle is operating are provided. The method includes, by an object detection system of a vehicle in an environment, detecting an actor that may move within a scene in the environment. The method further includes using context of the scene to determine a reference polyline for the actor and determining a kinematic history of the actor. The method additionally includes using the kinematic history to predict a waypoint, which is a predicted position of the actor at a conclusion of a waypoint time period, and identifying a segment of the reference polyline, the segment extending from a current location to a point along the reference polyline that is closest to the waypoint and determining a trajectory for the actor conditioned by the segment of the reference polyline.Type: ApplicationFiled: May 5, 2021Publication date: February 17, 2022Inventors: Siddhesh Shyam Khandelwal, William Junbo Qi, Jagjeet Singh, Andrew T. Hartnett, Deva Ramanan
-
Publication number: 20220048503Abstract: Systems and methods of determining trajectories of an actor in an environment in which a vehicle is operating are provided. The method includes detecting an actor that may move within a scene in the environment by an object detection system of a vehicle in the environment, determining a kinematic history of the actor, and using context of the scene and the kinematic history of the actor to determine a plurality of reference polylines for the actor. The method further includes generating a contextual embedding of the kinematic history of the actor to generate a plurality of predicted trajectories of the actor, in which the generating conditions each of the predicted trajectories to correspond to one of the reference polylines. The method additionally includes using, by the vehicle, the plurality of predicted trajectories to plan movement of the vehicle.Type: ApplicationFiled: May 5, 2021Publication date: February 17, 2022Inventors: Siddhesh Shyam Khandelwal, William Junbo Qi, Jagjeet Singh, Andrew T. Hartnett, Deva Ramanan
-
Publication number: 20210390349Abstract: A method is disclosed for evaluating a classifier used to determine a traffic light signal state in images. The method includes, by a computer vision system of a vehicle, receiving at least one image of a traffic signal device of an imminent intersection. The traffic signal device includes a traffic signal face including one or more traffic signal elements. The method includes classifying, by a traffic light classifier (TLC), a classification state of the traffic signal face using labeled images correlated to the received at least one image. The classification state controls an operation of the vehicle at the intersection. The method includes evaluating a performance of the classifying of the classification state generated by the TLC. The evaluation is a label-free performance evaluation based on unlabeled images. The method includes training the TLC based on the evaluated performance.Type: ApplicationFiled: June 16, 2020Publication date: December 16, 2021Inventors: Guy Hotson, Richard L. Kwant, Brett Browning, Deva Ramanan
-
Publication number: 20210342924Abstract: Various embodiments of a framework which allow, as an alternative to resource-taxing decompression, efficient computation of feature maps using a compressed content data subset, such as video, by exploiting the motion information, such as a motion vector, present in the compressed video. This framework allows frame-specific object recognition and action detection algorithms to be applied to compressed video and other media files by executing only on I-frames in a Group of Pictures and linearly interpolating the results. Training and machine learning increases recognition accuracy. Yielding significant computational gains, this approach accelerates frame-wise feature extraction I-frame/P-frame/P-frame videos as well as I-frame/P-frame/B-frame videos. The present techniques may also be used for segmentation to identify and label respective regions for objects in a video.Type: ApplicationFiled: December 27, 2019Publication date: November 4, 2021Inventors: R. Manmatha, Hexiang Hu, Deva Ramanan
-
Publication number: 20200143457Abstract: Various embodiments of a framework which allow, as an alternative to resource-taxing decompression, efficient computation of feature maps using a compressed content data subset, such as video, by exploiting the motion information, such as a motion vector, present in the compressed video. This framework allows frame-specific object recognition and action detection algorithms to be applied to compressed video and other media files by executing only on I-frames in a Group of Pictures and linearly interpolating the results. Training and machine learning increases recognition accuracy. Yielding significant computational gains, this approach accelerates frame-wise feature extraction I-frame/P-frame/P-frame videos as well as I-frame/P-frame/B-frame videos. The present techniques may also be used for segmentation to identify and label respective regions for objects in a video.Type: ApplicationFiled: December 27, 2019Publication date: May 7, 2020Inventors: R. Manmatha, Hexiang Hu, Deva Ramanan
-
Patent number: 10528819Abstract: Various embodiments of a framework which allow, as an alternative to resource-taxing decompression, efficient computation of feature maps using a compressed content data subset, such as video, by exploiting the motion information, such as a motion vector, present in the compressed video. This framework allows frame-specific object recognition and action detection algorithms to be applied to compressed video and other media files by executing only on I-frames in a Group of Pictures and linearly interpolating the results. Training and machine learning increases recognition accuracy. Yielding significant computational gains, this approach accelerates frame-wise feature extraction I-frame/P-frame/P-frame videos as well as I-frame/P-frame/B-frame videos. The present techniques may also be used for segmentation to identify and label respective regions for objects in a video.Type: GrantFiled: November 20, 2017Date of Patent: January 7, 2020Assignee: A9.COM, INC.Inventors: R. Manmatha, Hexiang Hu, Deva Ramanan