Patents by Inventor Ajay Divakaran

Ajay Divakaran has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11934793
    Abstract: A method, apparatus and system for training an embedding space for content comprehension and response includes, for each layer of a hierarchical taxonomy having at least two layers including respective words resulting in layers of varying complexity, determining a set of words associated with a layer of the hierarchical taxonomy, determining a question answer pair based on a question generated using at least one word of the set of words and at least one content domain, determining a vector representation for the generated question and for content related to the at least one content domain of the question answer pair, and embedding the question vector representation and the content vector representations into a common embedding space where vector representations that are related, are closer in the embedding space than unrelated embedded vector representations. Requests for content can then be fulfilled using the trained, common embedding space.
    Type: Grant
    Filed: November 1, 2021
    Date of Patent: March 19, 2024
    Assignee: SRI International
    Inventors: Ajay Divakaran, Karan Sikka, Yi Yao, Yunye Gong, Stephanie Nunn, Pritish Sahu, Michael A. Cogswell, Jesse Hostetler, Sara Rutherford-Quach
  • Publication number: 20240062042
    Abstract: In general, the disclosure describes techniques for implementing an MI-based attack detector. In an example, a method includes training a neural network using training data, applying stochastic quantization to one or more layers of the neural network, generating, using the trained neural network, an ensemble of neural networks having a plurality of quantized members, wherein at least one of weights or activations of each of the plurality of quantized members have different bit precision, and combining predictions of the plurality of quantized members of the ensemble to detect one or more adversarial attacks and/or determine performance of the ensemble of neural networks.
    Type: Application
    Filed: August 17, 2023
    Publication date: February 22, 2024
    Inventors: Aswin Nadamuni Raghavan, Saurabh Farkya, Jesse Albert Hostetler, Avraham Joshua Ziskind, Michael Piacentino, Ajay Divakaran, Zhengyu Chen
  • Publication number: 20240054294
    Abstract: A method, apparatus and system for moderating multilingual content data, for example, presented during a communication session include receiving or pulling content data that can include multilingual content, classifying, using a first machine learning system, the content data by projecting the content data into a trained embedding space to determine at least one English-language classification for the content data, and determining, using a second machine learning system, if the content data violates at least one predetermined moderation rule, wherein the second machine learning system is trained to determine from English-language classifications determined by the first machine learning system if the content data violates moderation rules. In some embodiments, the method apparatus and system can further include prohibiting a presentation of the content data related to the at least one English-language classification determined to violate the at least one predetermined moderation rule.
    Type: Application
    Filed: August 14, 2023
    Publication date: February 15, 2024
    Inventors: Karan SIKKA, Meng YE, Ajay DIVAKARAN
  • Publication number: 20240005654
    Abstract: A computing system comprising a memory configured to store an artificial intelligence (AI) model and an image, and a computation engine executing one or more processors may be configured to perform the techniques for error-based explanations for AI behavior. The computation engine may execute the AI model to analyze the image to output a result. The AI model may, when analyzing the image to output the result, process, based on data indicative of the result, the image to assign an error score to each image feature extracted from the image, and obtain, based on the error scores, an error map. The AI model may next update, based on the error map and to obtain a first updated image, the image to visually indicate the error score assigned to each of the image features, and output one or more of the error scores, the error map, and the first updated image.
    Type: Application
    Filed: March 24, 2022
    Publication date: January 4, 2024
    Inventors: Arijit Ray, Michael A. Cogswell, Ajay Divakaran, Yi Yao, Giedrius T. Burachas, Kamran Alipour
  • Patent number: 11790213
    Abstract: Techniques are disclosed for identifying multimodal subevents within an event having spatially-related and temporally-related features. In one example, a system receives a Spatio-Temporal Graph (STG) comprising (1) a plurality of nodes, each node having a feature descriptor that describes a feature present in the event, (2) a plurality of spatial edges, each spatial edge describing a spatial relationship between two of the plurality of nodes, and (3) a plurality of temporal edges, each temporal edge describing a temporal relationship between two of the plurality of nodes. Furthermore, the STG comprises at least one of: (1) variable-length descriptors for the feature descriptors or (2) temporal edges that span multiple time steps for the event. A machine learning system processes the STG to identify the multimodal subevents for the event. In some examples, the machine learning system comprises stacked Spatio-Temporal Graph Convolutional Networks (STGCNs), each comprising a plurality of STGCN layers.
    Type: Grant
    Filed: June 12, 2019
    Date of Patent: October 17, 2023
    Assignee: SRI INTERNATIONAL
    Inventors: Yi Yao, Ajay Divakaran, Pallabi Ghosh
  • Patent number: 11610384
    Abstract: A method, apparatus and system for zero shot object detection includes, in a semantic embedding space having embedded object class labels, training the space by embedding extracted features of bounding boxes and object class labels of labeled bounding boxes of known object classes into the space, determining regions in an image having unknown object classes on which to perform object detection as proposed bounding boxes, extracting features of the proposed bounding boxes, projecting the extracted features of the proposed bounding boxes into the space, computing a similarity measure between the projected features of the proposed bounding boxes and the embedded, extracted features of the bounding boxes of the known object classes in the space, and predicting an object class label for proposed bounding boxes by determining a nearest embedded object class label to the projected features of the proposed bounding boxes in the space based on the similarity measures.
    Type: Grant
    Filed: June 2, 2021
    Date of Patent: March 21, 2023
    Assignee: SRI International
    Inventors: Karan Sikka, Ajay Divakaran, Ankan Bansal
  • Publication number: 20230031449
    Abstract: A method, apparatus and system for comprehension-based question answering using a hierarchical taxonomy include receiving a word-based question, associating the word-based question with a layer of the hierarchical taxonomy, in which the hierarchical taxonomy includes at least two layers, each of the at least two layers including respective words resulting in the at least two layers having varying levels complexity, determining which layer of the at least two layers of the hierarchical taxonomy comprises a layer of complexity one level less than the layer of the hierarchical taxonomy associated with the word-based question, and using a pre-trained language model, answering the word-based question using only words associated with the layer of the at least two layers of the hierarchical taxonomy having the one less level of complexity.
    Type: Application
    Filed: July 20, 2022
    Publication date: February 2, 2023
    Inventors: Ajay DIVAKARAN, Michael A. COGSWELL, Pritish SAHU
  • Publication number: 20220414476
    Abstract: A method, apparatus and system for adapting a pre-trained network for application to a different dataset includes arranging at least two different types of active adaptation modules in a pipeline configuration, wherein an output of a previous active adaptation module produces an input for a next active adaptation module in the pipeline in the form of adapted network data until a last active adaptation module, and wherein each of the at least two different types of adaptation modules can be switched on or off, determining at least one respective hyperparameter for each of the at least two different types of active adaptation modules, and applying the at least one respective determined hyperparameter to each of the at least two different types of active adaptation modules for processing received data from the pretrained network to determine an adapted network.
    Type: Application
    Filed: June 20, 2022
    Publication date: December 29, 2022
    Inventors: Xiao LIN, Meng YE, Yunye GONG, Giedrius T. BURACHAS, Ajay DIVAKARAN, Yi YAO, Nikoletta BASIOU
  • Publication number: 20220138433
    Abstract: A method, apparatus and system for training an embedding space for content comprehension and response includes, for each layer of a hierarchical taxonomy having at least two layers including respective words resulting in layers of varying complexity, determining a set of words associated with a layer of the hierarchical taxonomy, determining a question answer pair based on a question generated using at least one word of the set of words and at least one content domain, determining a vector representation for the generated question and for content related to the at least one content domain of the question answer pair, and embedding the question vector representation and the content vector representations into a common embedding space where vector representations that are related, are closer in the embedding space than unrelated embedded vector representations. Requests for content can then be fulfilled using the trained, common embedding space.
    Type: Application
    Filed: November 1, 2021
    Publication date: May 5, 2022
    Inventors: Ajay DIVAKARAN, Karan SIKKA, Yi YAO, Yunye GONG, Stephanie NUNN, Pritish SAHU, Michael A. COGSWELL, Jesse HOSTETLER, Sara RUTHERFORD-QUACH
  • Patent number: 11238631
    Abstract: A method, apparatus and system for visual grounding of a caption in an image include projecting at least two parsed phrases of the caption into a trained semantic embedding space, projecting extracted region proposals of the image into the trained semantic embedding space, aligning the extracted region proposals and the at least two parsed phrases, aggregating the aligned region proposals and the at least two parsed phrases to determine a caption-conditioned image representation and projecting the caption-conditioned image representation and the caption into a semantic embedding space to align the caption-conditioned image representation and the caption. The method, apparatus and system can further include a parser for parsing the caption into the at least two parsed phrases and a region proposal module for extracting the region proposals from the image.
    Type: Grant
    Filed: April 22, 2020
    Date of Patent: February 1, 2022
    Assignee: SRI International
    Inventors: Karan Sikka, Ajay Divakaran, Samyak Datta
  • Patent number: 11210572
    Abstract: A method, apparatus and system for understanding visual content includes determining at least one region proposal for an image, attending at least one symbol of the proposed image region, attending a portion of the proposed image region using information regarding the attended symbol, extracting appearance features of the attended portion of the proposed image region, fusing the appearance features of the attended image region and features of the attended symbol, projecting the fused features into a semantic embedding space having been trained using fused attended appearance features and attended symbol features of images having known descriptive messages, computing a similarity measure between the projected, fused features and fused attended appearance features and attended symbol features embedded in the semantic embedding space having at least one associated descriptive message and predicting a descriptive message for an image associated with the projected, fused features.
    Type: Grant
    Filed: December 17, 2019
    Date of Patent: December 28, 2021
    Assignee: SRI International
    Inventors: Ajay Divakaran, Karan Sikka, Karuna Ahuja, Anirban Roy
  • Publication number: 20210390492
    Abstract: In some examples, a computer-implemented collaboration assessment model identifies actions of each of two or more individuals depicted in video data, identify, based at least on the identified actions of each of the two or more individuals depicted in the video data, first behaviors at a first collaboration assessment level, identify, based at least on the identified actions of each of the two or more individuals depicted in the video data, second behaviors at a second collaboration assessment level different from the first collaboration assessment level, and generate and output, based at least on the first behaviors at the first collaboration assessment level and the second behaviors at the second collaboration assessment level, an indication of at least one of an assessment of a collaboration effort of the two or more individuals or respective assessments of individual contributions of the two or more individuals to the collaboration effort.
    Type: Application
    Filed: June 15, 2021
    Publication date: December 16, 2021
    Inventors: Swati Dhamija, Amir Tamrakar, Nonye M. Alozie, Elizabeth McBride, Ajay Divakaran, Anirudh Som, Sujeong Kim, Bladimir Lopez-Prado
  • Publication number: 20210390400
    Abstract: Techniques are described for neural networks based on Progressive Neural ODEs (PODEs). In an example, a method to progressively train a neural ordinary differential equation (NODE) model comprises processing, by a machine learning system executed by a computing system, first training data, the first training data having a first complexity, to perform training of a first layer for the NODE model; and after performing the first training, processing second training data, the second training data having a second complexity that is higher than the first complexity, to perform training of a second layer for the NODE model.
    Type: Application
    Filed: June 15, 2021
    Publication date: December 16, 2021
    Inventors: Yi Yao, Ajay Divakaran, Hammad A. Ayyubi
  • Publication number: 20210374531
    Abstract: In general, the disclosure describes techniques for characterizing a dynamical system and a neural ordinary differential equation (NODE)-based controller for the dynamical system. An example analysis system is configured to: obtain a set of parameters of a NODE model used to implement the NODE-based controller, the NODE model trained to control the dynamical system; determine, based on the set of parameters, a system property of a combined system comprising the dynamical system and the NODE-based controller, the system property comprising one or more of an accuracy, safety, reliability, reachability, or controllability of the combined system; and output the system property to modify one or more of the dynamical system or the NODE-based controller to meet a required specification for the combined system.
    Type: Application
    Filed: May 26, 2021
    Publication date: December 2, 2021
    Inventors: Ajay Divakaran, Anirban Roy, Susmit Jha
  • Publication number: 20210295082
    Abstract: A method, apparatus and system for zero shot object detection includes, in a semantic embedding space having embedded object class labels, training the space by embedding extracted features of bounding boxes and object class labels of labeled bounding boxes of known object classes into the space, determining regions in an image having unknown object classes on which to perform object detection as proposed bounding boxes, extracting features of the proposed bounding boxes, projecting the extracted features of the proposed bounding boxes into the space, computing a similarity measure between the projected features of the proposed bounding boxes and the embedded, extracted features of the bounding boxes of the known object classes in the space, and predicting an object class label for proposed bounding boxes by determining a nearest embedded object class label to the projected features of the proposed bounding boxes in the space based on the similarity measures.
    Type: Application
    Filed: June 2, 2021
    Publication date: September 23, 2021
    Inventors: Karan Sikka, Ajay Divakaran, Ankan Bansal
  • Publication number: 20210297498
    Abstract: A method, apparatus and system for determining user-content associations for determining and providing user-preferred content using multimodal embeddings include creating an embedding space for multimodal content by creating a first modality vector representation of the multimodal content having a first modality, creating a second modality vector representation of the multimodal content having a second modality, creating a user vector representation, as a third modality, for each user associated with at least a portion of the multimodal content, and embedding the first and the second modality vector representations and the user vector representations in the common embedding space using at least a mixture of loss functions for each modality pair of the first, the at least second and the third modalities that pushes closer co-occurring pairs of multimodal content.
    Type: Application
    Filed: March 4, 2021
    Publication date: September 23, 2021
    Inventors: Ajay Divakaran, Karan Sikka, Arijit Ray, Xiao Lin, Yi Yao
  • Patent number: 11055555
    Abstract: A method, apparatus and system for zero shot object detection includes, in a semantic embedding space having embedded object class labels, training the space by embedding extracted features of bounding boxes and object class labels of labeled bounding boxes of known object classes into the space, determining regions in an image having unknown object classes on which to perform object detection as proposed bounding boxes, extracting features of the proposed bounding boxes, projecting the extracted features of the proposed bounding boxes into the space, computing a similarity measure between the projected features of the proposed bounding boxes and the embedded, extracted features of the bounding boxes of the known object classes in the space, and predicting an object class label for proposed bounding boxes by determining a nearest embedded object class label to the projected features of the proposed bounding boxes in the space based on the similarity measures.
    Type: Grant
    Filed: April 12, 2019
    Date of Patent: July 6, 2021
    Assignee: SRI International
    Inventors: Karan Sikka, Ajay Divakaran, Ankan Bansal
  • Publication number: 20210081056
    Abstract: Methods, computing devices, and computer-program products are provided for implementing a virtual personal assistant. In various implementations, a virtual personal assistant can be configured to receive sensory input, including at least two different types of information. The virtual personal assistant can further be configured to determine semantic information from the sensory input, and to identify a context-specific framework. The virtual personal assistant can further be configured to determine a current intent. Determining the current intent can include using the semantic information and the context-specific framework. The virtual personal assistant can further be configured to determine a current input state. Determining the current input state can include using the semantic information and one or more behavioral models. The behavioral models can include one or more interpretations of previously-provided semantic information.
    Type: Application
    Filed: December 1, 2020
    Publication date: March 18, 2021
    Inventors: Ajay Divakaran, Amir Tamrakar, Girish Acharya, William Mark, Greg Ho, Jihua Huang, David Salter, Edgar Kalns, Michael Wessel, Min Yin, James Carpenter, Brent Mombourquette, Kenneth Nitz, Elizabeth Shriberg, Eric Law, Michael Frandsen, Hyong-Gyun Kim, Cory Albright, Andreas Tsiartas
  • Publication number: 20210056742
    Abstract: A method, apparatus and system for visual grounding of a caption in an image include projecting at least two parsed phrases of the caption into a trained semantic embedding space, projecting extracted region proposals of the image into the trained semantic embedding space, aligning the extracted region proposals and the at least two parsed phrases, aggregating the aligned region proposals and the at least two parsed phrases to determine a caption-conditioned image representation and projecting the caption-conditioned image representation and the caption into a semantic embedding space to align the caption-conditioned image representation and the caption. The method, apparatus and system can further include a parser for parsing the caption into the at least two parsed phrases and a region proposal module for extracting the region proposals from the image.
    Type: Application
    Filed: April 22, 2020
    Publication date: February 25, 2021
    Inventors: Karan Sikka, Ajay Divakaran, Samyak Datta
  • Patent number: 10884503
    Abstract: Methods, computing devices, and computer-program products are provided for implementing a virtual personal assistant. In various implementations, a virtual personal assistant can be configured to receive sensory input, including at least two different types of information. The virtual personal assistant can further be configured to determine semantic information from the sensory input, and to identify a context-specific framework. The virtual personal assistant can further be configured to determine a current intent. Determining the current intent can include using the semantic information and the context-specific framework. The virtual personal assistant can further be configured to determine a current input state. Determining the current input state can include using the semantic information and one or more behavioral models. The behavioral models can include one or more interpretations of previously-provided semantic information.
    Type: Grant
    Filed: October 24, 2016
    Date of Patent: January 5, 2021
    Assignee: SRI International
    Inventors: Ajay Divakaran, Amir Tamrakar, Girish Acharya, William Mark, Greg Ho, Jihua Huang, David Salter, Edgar Kalns, Michael Wessel, Min Yin, James Carpenter, Brent Mombourquette, Kenneth Nitz, Elizabeth Shriberg, Eric Law, Michael Frandsen, Hyong-Gyun Kim, Cory Albright, Andreas Tsiartas