Patents by Inventor Ajay Divakaran

Ajay Divakaran has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20200394499
    Abstract: Techniques are disclosed for identifying multimodal subevents within an event having spatially-related and temporally-related features. In one example, a system receives a Spatio-Temporal Graph (STG) comprising (1) a plurality of nodes, each node having a feature descriptor that describes a feature present in the event, (2) a plurality of spatial edges, each spatial edge describing a spatial relationship between two of the plurality of nodes, and (3) a plurality of temporal edges, each temporal edge describing a temporal relationship between two of the plurality of nodes. Furthermore, the STG comprises at least one of: (1) variable-length descriptors for the feature descriptors or (2) temporal edges that span multiple time steps for the event. A machine learning system processes the STG to identify the multimodal subevents for the event. In some examples, the machine learning system comprises stacked Spatio-Temporal Graph Convolutional Networks (STGCNs), each comprising a plurality of STGCN layers.
    Type: Application
    Filed: June 12, 2019
    Publication date: December 17, 2020
    Inventors: Yi Yao, Ajay Divakaran, Pallabi Ghosh
  • Patent number: 10824916
    Abstract: Systems and methods for improving the accuracy of a computer system for object identification/classification through the use of weakly supervised learning are provided herein. In some embodiments, the method includes (a) receiving at least one set of curated data, wherein the curated data includes labeled images, (b) using the curated data to train a deep network model for identifying objects within images, wherein the trained deep network model has a first accuracy level for identifying objects, receiving a first target accuracy level for object identification of the deep network model, determining, automatically via the computer system, an amount of weakly labeled data needed to train the deep network model to achieve the first target accuracy level, and augmenting the deep network model using weakly supervised learning and the weakly labeled data to achieve the first target accuracy level for object identification by the deep network model.
    Type: Grant
    Filed: September 10, 2018
    Date of Patent: November 3, 2020
    Assignee: SRI International
    Inventors: Karan Sikka, Ajay Divakaran, Parneet Kaur
  • Publication number: 20200193245
    Abstract: A method, apparatus and system for understanding visual content includes determining at least one region proposal for an image, attending at least one symbol of the proposed image region, attending a portion of the proposed image region using information regarding the attended symbol, extracting appearance features of the attended portion of the proposed image region, fusing the appearance features of the attended image region and features of the attended symbol, projecting the fused features into a semantic embedding space having been trained using fused attended appearance features and attended symbol features of images having known descriptive messages, computing a similarity measure between the projected, fused features and fused attended appearance features and attended symbol features embedded in the semantic embedding space having at least one associated descriptive message and predicting a descriptive message for an image associated with the projected, fused features.
    Type: Application
    Filed: December 17, 2019
    Publication date: June 18, 2020
    Inventors: Ajay Divakaran, Karan Sikka, Karuna Ahuja, Anirban Roy
  • Patent number: 10679063
    Abstract: A computing system for recognizing salient events depicted in a video utilizes learning algorithms to detect audio and visual features of the video. The computing system identifies one or more salient events depicted in the video based on the audio and visual features.
    Type: Grant
    Filed: September 4, 2015
    Date of Patent: June 9, 2020
    Assignee: SRI International
    Inventors: Hui Cheng, Ajay Divakaran, Elizabeth Shriberg, Harpreet Singh Sawhney, Jingen Liu, Ishani Chakraborty, Omar Javed, David Chisolm, Behjat Siddiquie, Steven S. Weiner
  • Publication number: 20200134398
    Abstract: Inferring multimodal content intent in a common geometric space in order to improve recognition of influential impacts of content includes mapping the multimodal content in a common geometric space by embedding a multimodal feature vector representing a first modality of the multimodal content and a second modality of the multimodal content and inferring intent of the multimodal content mapped into the common geometric space such that connections between multimodal content result in an improvement in recognition of the influential impact of the multimodal content.
    Type: Application
    Filed: April 12, 2019
    Publication date: April 30, 2020
    Inventors: Julia Kruk, Jonah M. Lubin, Karan Sikka, Xiao Lin, Ajay Divakaran
  • Publication number: 20200082224
    Abstract: Systems and methods for improving the accuracy of a computer system for object identification/classification through the use of weakly supervised learning are provided herein. In some embodiments, the method includes (a) receiving at least one set of curated data, wherein the curated data includes labeled images, (b) using the curated data to train a deep network model for identifying objects within images, wherein the trained deep network model has a first accuracy level for identifying objects, receiving a first target accuracy level for object identification of the deep network model, determining, automatically via the computer system, an amount of weakly labeled data needed to train the deep network model to achieve the first target accuracy level, and augmenting the deep network model using weakly supervised learning and the weakly labeled data to achieve the first target accuracy level for object identification by the deep network model.
    Type: Application
    Filed: September 10, 2018
    Publication date: March 12, 2020
    Inventors: Karan Sikka, Ajay Divakaran, Parneet Kaur
  • Publication number: 20190325243
    Abstract: A method, apparatus and system for zero shot object detection includes, in a semantic embedding space having embedded object class labels, training the space by embedding extracted features of bounding boxes and object class labels of labeled bounding boxes of known object classes into the space, determining regions in an image having unknown object classes on which to perform object detection as proposed bounding boxes, extracting features of the proposed bounding boxes, projecting the extracted features of the proposed bounding boxes into the space, computing a similarity measure between the projected features of the proposed bounding boxes and the embedded, extracted features of the bounding boxes of the known object classes in the space, and predicting an object class label for proposed bounding boxes by determining a nearest embedded object class label to the projected features of the proposed bounding boxes in the space based on the similarity measures.
    Type: Application
    Filed: April 12, 2019
    Publication date: October 24, 2019
    Inventors: Karan Sikka, Ajay Divakaran, Ankan Bansal
  • Publication number: 20190325342
    Abstract: Embedding multimodal content in a common geometric space includes for each of a plurality of content of the multimodal content, creating a respective, first modality feature vector representative of content of the multimodal content having a first modality using a first machine learning model; for each of a plurality of content of the multimodal content, creating a respective, second modality feature vector representative of content of the multimodal content having a second modality using a second machine learning model; and semantically embedding the respective, first modality feature vectors and the respective, second modality feature vectors in a common geometric space that provides logarithm-like warping of distance space in the geometric space to capture hierarchical relationships between seemingly disparate, embedded modality feature vectors of content in the geometric space; wherein embedded modality feature vectors that are related, across modalities, are closer together in the geometric space than un
    Type: Application
    Filed: April 12, 2019
    Publication date: October 24, 2019
    Inventors: Karan Sikka, Ajay Divakaran, Julia Kruk
  • Patent number: 10303768
    Abstract: Technologies to detect persuasive multimedia content by using affective and semantic concepts extracted from the audio-visual content as well as the sentiment of associated comments are disclosed. The multimedia content is analyzed and compared with a persuasiveness model.
    Type: Grant
    Filed: October 2, 2015
    Date of Patent: May 28, 2019
    Assignee: SRI International
    Inventors: Ajay Divakaran, Behjat Siddiquie, David Chisholm, Elizabeth Shriberg
  • Patent number: 10268900
    Abstract: A system for object detection and tracking includes technologies to, among other things, detect and track moving objects, such as pedestrians and/or vehicles, in a real-world environment, handle static and dynamic occlusions, and continue tracking moving objects across the fields of view of multiple different cameras.
    Type: Grant
    Filed: February 27, 2018
    Date of Patent: April 23, 2019
    Assignee: SRI International
    Inventors: Ajay Divakaran, Qian Yu, Amir Tamrakar, Harpreet Singh Sawhney, Jiejie Zhu, Omar Javed, Jingen Liu, Hui Cheng, Jayakrishnan Eledath
  • Patent number: 10198509
    Abstract: A complex video event classification, search and retrieval system can generate a semantic representation of a video or of segments within the video, based on one or more complex events that are depicted in the video, without the need for manual tagging. The system can use the semantic representations to, among other things, provide enhanced video search and retrieval capabilities.
    Type: Grant
    Filed: January 25, 2016
    Date of Patent: February 5, 2019
    Assignee: SRI International
    Inventors: Hui Cheng, Harpreet Singh Sawhney, Ajay Divakaran, Qian Yu, Jingen Liu, Amir Tamrakar, Saad Ali, Omar Javed
  • Patent number: 10068024
    Abstract: Methods and apparatuses of the present invention generally relate to generating actionable data based on multimodal data from unsynchronized data sources. In an exemplary embodiment, the method comprises receiving multimodal data from one or more unsynchronized data sources, extracting concepts from the multimodal data, the concepts comprising at least one of objects, actions, scenes and emotions, indexing the concepts for searchability; and generating actionable data based on the concepts.
    Type: Grant
    Filed: December 18, 2015
    Date of Patent: September 4, 2018
    Assignee: SRI International
    Inventors: Harpreet Singh Sawhney, Jayakrishnan Eledath, Ajay Divakaran, Mayank Bansal, Hui Cheng
  • Publication number: 20180189573
    Abstract: A system for object detection and tracking includes technologies to, among other things, detect and track moving objects, such as pedestrians and/or vehicles, in a real-world environment, handle static and dynamic occlusions, and continue tracking moving objects across the fields of view of multiple different cameras.
    Type: Application
    Filed: February 27, 2018
    Publication date: July 5, 2018
    Inventors: Ajay Divakaran, Qian Yu, Amir Tamrakar, Harpreet Singh Sawhney, Jiejie Zhu, Omar Javed, Jingen Liu, Hui Cheng, Jayakrishnan Eledath
  • Patent number: 9977972
    Abstract: A computer implemented method for determining a vehicle type of a vehicle detected in an image is disclosed. An image having a detected vehicle is received. A number of vehicle models having salient feature points is projected on the detected vehicle. A first set of features derived from each of the salient feature locations of the vehicle models is compared to a second set of features derived from corresponding salient feature locations of the detected vehicle to form a set of positive match scores (p-scores) and a set of negative match scores (n-scores). The detected vehicle is classified as one of the vehicle models based at least in part on the set of p-scores and the set of n-scores.
    Type: Grant
    Filed: October 21, 2014
    Date of Patent: May 22, 2018
    Assignee: SRI International
    Inventors: Saad Masood Khan, Hui Cheng, Dennis Lee Matthies, Harpreet Singh Sawhney, Sang-Hack Jung, Chris Broaddus, Bogdan Calin Mihai Matei, Ajay Divakaran
  • Publication number: 20180075774
    Abstract: A method and system for analyzing at least one food item on a food plate is disclosed. A plurality of images of the food plate is received by an image capturing device. A description of the at least one food item on the food plate is received by a recognition device. The description is at least one of a voice description and a text description. At least one processor extracts a list of food items from the description; classifies and segments the at least one food item from the list using color and texture features derived from the plurality of images; and estimates the volume of the classified and segmented at least one food item. The processor is also configured to estimate the caloric content of the at least one food item.
    Type: Application
    Filed: November 20, 2017
    Publication date: March 15, 2018
    Inventors: Manika PURI, Zhiwei Zhu, Jeffrey Lubin, Tom Pschar, Ajay Divakaran, Harpreet Sawhney
  • Patent number: 9916520
    Abstract: A food recognition assistant system includes technologies to recognize foods and combinations of foods depicted in a digital picture of food. Some embodiments include technologies to estimate portion size and calories, and to estimate nutritional value of the foods. In some embodiments, data identifying recognized foods and related information are generated in an automated fashion without relying on human assistance to identify the foods. In some embodiments, the system includes technologies for achieving automatic food detection and recognition in a real-life setting with a cluttered background, without the images being taken in a controlled lab setting, and without requiring additional user input (such as user-defined bounding boxes). Some embodiments of the system include technologies for personalizing the food classification based on user-specific habits, location and/or other criteria.
    Type: Grant
    Filed: December 11, 2014
    Date of Patent: March 13, 2018
    Assignee: SRI International
    Inventors: Ajay Divakaran, Weiyu Zhang, Qian Yu, Harpreet S. Sawhney
  • Patent number: 9904852
    Abstract: A system for object detection and tracking includes technologies to, among other things, detect and track moving objects, such as pedestrians and/or vehicles, in a real-world environment, handle static and dynamic occlusions, and continue tracking moving objects across the fields of view of multiple different cameras.
    Type: Grant
    Filed: May 23, 2014
    Date of Patent: February 27, 2018
    Assignee: SRI International
    Inventors: Ajay Divakaran, Qian Yu, Amir Tamrakar, Harpreet Singh Sawhney, Jiejie Zhu, Omar Javed, Jingen Liu, Hui Cheng, Jayakrishnan Eledath
  • Patent number: 9875445
    Abstract: Technologies for analyzing temporal components of multimodal data to detect short-term multimodal events, determine relationships between short-term multimodal events, and recognize long-term multimodal events, using a deep learning architecture, are disclosed.
    Type: Grant
    Filed: February 25, 2015
    Date of Patent: January 23, 2018
    Assignee: SRI International
    Inventors: Mohamed R. Amer, Behjat Siddiquie, Ajay Divakaran, Colleen Richey, Saad Khan, Hapreet S. Sawhney, Timothy J. Shields
  • Patent number: 9734426
    Abstract: A food recognition assistant system includes technologies to recognize foods and combinations of foods depicted in a digital picture of food. Some embodiments include technologies to estimate portion size and calories, and to estimate nutritional value of the foods. In some embodiments, data identifying recognized foods and related information are generated in an automated fashion without relying on human assistance to identify the foods. In some embodiments, the system includes technologies for achieving automatic food detection and recognition in a real-life setting with a cluttered background, without the images being taken in a controlled lab setting, and without requiring additional user input (such as user-defined bounding boxes). Some embodiments of the system include technologies for personalizing the food classification based on user-specific habits, location and/or other criteria.
    Type: Grant
    Filed: December 11, 2014
    Date of Patent: August 15, 2017
    Assignee: SRI International
    Inventors: Ajay Divakaran, Weiyu Zhang, Qian Yu, Harpreet S. Sawhney
  • Patent number: 9734730
    Abstract: A multi-modal interaction modeling system can model a number of different aspects of a human interaction across one or more temporal interaction sequences. Some versions of the system can generate assessments of the nature or quality of the interaction or portions thereof, which can be used to, among other things, provide assistance to one or more of the participants in the interaction.
    Type: Grant
    Filed: January 31, 2013
    Date of Patent: August 15, 2017
    Assignee: SRI International
    Inventors: Ajay Divakaran, Behjat Siddiquie, Saad Khan, Jeffrey Lubin, Harpreet S. Sawhney