Patents by Inventor Hrituraj Singh

Hrituraj Singh has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 12198048
    Abstract: In some embodiments, a multimodal computing system receives a query and identifies, from source documents, text passages and images that are relevant to the query. The multimodal computing system accesses a multimodal question-answering model that includes a textual stream of language models and a visual stream of language models. Each of the textual stream and the visual stream contains a set of transformer-based models and each transformer-based model includes a cross-attention layer using data generated by both the textual stream and visual stream of language models as an input. The multimodal computing system identifies text relevant to the query by applying the textual stream to the text passages and computes, using the visual stream, relevance scores of the images to the query, respectively. The multimodal computing system further generates a response to the query by including the text and/or an image according to the relevance scores.
    Type: Grant
    Filed: January 20, 2021
    Date of Patent: January 14, 2025
    Assignee: Adobe Inc.
    Inventors: Hrituraj Singh, Jatin Lamba, Denil Pareshbhai Mehta, Balaji Vasan Srinivasan, Anshul Nasery, Aishwarya Agarwal
  • Publication number: 20240331102
    Abstract: Certain aspects and features of this disclosure relate to semantically-aware image extrapolation. In one example, an input image is segmented to produce an input segmentation map of object instances in the input image. An object generation network is used to generate an extrapolated semantic label map for an extrapolated image. The extrapolated semantic label map includes instances in the original image and instances that will appear in an outpainted region of the extrapolated image. A panoptic label map is derived from coordinates of output instances in the extrapolated image and used to identify partial instances and boundaries. Instance-aware context normalization is used to apply one or more characteristics from the input image to the outpainted region to maintain semantic continuity. The extrapolated image includes the original image and the outpainted region and can be rendered or stored for future use.
    Type: Application
    Filed: June 7, 2024
    Publication date: October 3, 2024
    Inventors: Kuldeep Kulkarni, Soumya Dash, Hrituraj Singh, Bholeshwar Khurana, Aniruddha Mahapatra, Abhishek Bhatia
  • Patent number: 12020403
    Abstract: Certain aspects and features of this disclosure relate to semantically-aware image extrapolation. In one example, an input image is segmented to produce an input segmentation map of object instances in the input image. An object generation network is used to generate an extrapolated semantic label map for an extrapolated image. The extrapolated semantic label map includes instances in the original image and instances that will appear in an outpainted region of the extrapolated image. A panoptic label map is derived from coordinates of output instances in the extrapolated image and used to identify partial instances and boundaries. Instance-aware context normalization is used to apply one or more characteristics from the input image to the outpainted region to maintain semantic continuity. The extrapolated image includes the original image and the outpainted region and can be rendered or stored for future use.
    Type: Grant
    Filed: November 8, 2021
    Date of Patent: June 25, 2024
    Assignee: Adobe Inc.
    Inventors: Kuldeep Kulkarni, Soumya Dash, Hrituraj Singh, Bholeshwar Khurana, Aniruddha Mahapatra, Abhishek Bhatia
  • Publication number: 20230169632
    Abstract: Certain aspects and features of this disclosure relate to semantically-aware image extrapolation. In one example, an input image is segmented to produce an input segmentation map of object instances in the input image. An object generation network is used to generate an extrapolated semantic label map for an extrapolated image. The extrapolated semantic label map includes instances in the original image and instances that will appear in an outpainted region of the extrapolated image. A panoptic label map is derived from coordinates of output instances in the extrapolated image and used to identify partial instances and boundaries. Instance-aware context normalization is used to apply one or more characteristics from the input image to the outpainted region to maintain semantic continuity. The extrapolated image includes the original image and the outpainted region and can be rendered or stored for future use.
    Type: Application
    Filed: November 8, 2021
    Publication date: June 1, 2023
    Inventors: Kuldeep Kulkarni, Soumya Dash, Hrituraj Singh, Bholeshwar Khurana, Aniruddha Mahapatra, Abhishek Bhatia
  • Patent number: 11657225
    Abstract: Systems and methods for generating a tuned summary using a word generation model. An example method includes receiving, at a decoder of the word generation model, a training data learned subspace representation of training data. The method also includes identifying tunable linguistic characteristics of the word generation model and training the decoder to output a training tuned summary of the training data learned subspace representation based on at least one of the tunable linguistic characteristics. The method further includes receiving an input text and a target characteristic token, and generating, by the trained decoder of the word generation model, each word of a tuned summary of the input text from a learned subspace representation and from feedback about preceding words of the tuned summary, wherein the tuned summary is tuned to target characteristics represented by the target characteristic token.
    Type: Grant
    Filed: June 15, 2021
    Date of Patent: May 23, 2023
    Assignee: ADOBE INC.
    Inventors: Balaji Vasan Srinivasan, Kushal Chawla, Mithlesh Kumar, Hrituraj Singh, Arijit Pramanik
  • Publication number: 20220230061
    Abstract: In some embodiments, a multimodal computing system receives a query and identifies, from source documents, text passages and images that are relevant to the query. The multimodal computing system accesses a multimodal question-answering model that includes a textual stream of language models and a visual stream of language models. Each of the textual stream and the visual stream contains a set of transformer-based models and each transformer-based model includes a cross-attention layer using data generated by both the textual stream and visual stream of language models as an input. The multimodal computing system identifies text relevant to the query by applying the textual stream to the text passages and computes, using the visual stream, relevance scores of the images to the query, respectively. The multimodal computing system further generates a response to the query by including the text and/or an image according to the relevance scores.
    Type: Application
    Filed: January 20, 2021
    Publication date: July 21, 2022
    Inventors: Hrituraj Singh, Jatin Lamba, Denil Pareshbhai Mehta, Balaji Vasan Srinivasan, Anshul Nasery, Aishwarya Agarwal
  • Patent number: 11386114
    Abstract: Embodiments are disclosed for determining an answer to a query associated with a graphical representation of data. In particular, in one or more embodiments, the disclosed systems and methods comprise obtaining a visual embedding for a graphical representation of data, the visual embedding representing a plurality of graphical elements. The one or more embodiment further include obtaining a query embedding for a query associated with the graphical representation of data, the query embedding representing a plurality of textual elements of the query with at least one textual element substituted with an identifier for at least one graphical element of the set of graphical elements. The one or more embodiment further include generating a chart sequence from the visual embedding and a query sequence from the query embedding, generating an output sequence based on the graph and the query sequences, and determining an answer to the query from the output sequence.
    Type: Grant
    Filed: October 21, 2020
    Date of Patent: July 12, 2022
    Assignee: Adobe Inc.
    Inventors: Hrituraj Singh, Sumit Shekhar
  • Publication number: 20220121679
    Abstract: Embodiments are disclosed for determining an answer to a query associated with a graphical representation of data. In particular, in one or more embodiments, the disclosed systems and methods comprise obtaining a visual embedding for a graphical representation of data, the visual embedding representing a plurality of graphical elements. The one or more embodiment further include obtaining a query embedding for a query associated with the graphical representation of data, the query embedding representing a plurality of textual elements of the query with at least one textual element substituted with an identifier for at least one graphical element of the set of graphical elements. The one or more embodiment further include generating a chart sequence from the visual embedding and a query sequence from the query embedding, generating an output sequence based on the graph and the query sequences, and determining an answer to the query from the output sequence.
    Type: Application
    Filed: October 21, 2020
    Publication date: April 21, 2022
    Inventors: Hrituraj SINGH, Sumit SHEKHAR
  • Publication number: 20210312129
    Abstract: Certain embodiments involve tuning summaries of input text to a target characteristic using a word generation model. For example, a method for generating a tuned summary using a word generation model includes generating a learned subspace representation of input text and a target characteristic token associated with the input text by applying an encoder to the input text and the target characteristic token. The method also includes generating, by a decoder, each word of a tuned summary of the input text from the learned subspace representation and from a feedback about preceding words of the tuned summary. The tuned summary is tuned to target characteristics represented by the target characteristic token.
    Type: Application
    Filed: June 15, 2021
    Publication date: October 7, 2021
    Inventors: Balaji Vasan Srinivasan, Kushal Chawla, Mithlesh Kumar, Hrituraj Singh, Arijit Pramanik
  • Patent number: 11062087
    Abstract: Certain embodiments involve tuning summaries of input text to a target characteristic using a word generation model. For example, a method for generating a tuned summary using a word generation model includes generating a learned subspace representation of input text and a target characteristic token associated with the input text by applying an encoder to the input text and the target characteristic token. The method also includes generating, by a decoder, each word of a tuned summary of the input text from the learned subspace representation and from a feedback about preceding words of the tuned summary. The tuned summary is tuned to target characteristics represented by the target characteristic token.
    Type: Grant
    Filed: January 30, 2019
    Date of Patent: July 13, 2021
    Assignee: ADOBE INC.
    Inventors: Balaji Vasan Srinivasan, Kushal Chawla, Mithlesh Kumar, Hrituraj Singh, Arijit Pramanik
  • Publication number: 20200242197
    Abstract: Certain embodiments involve tuning summaries of input text to a target characteristic using a word generation model. For example, a method for generating a tuned summary using a word generation model includes generating a learned subspace representation of input text and a target characteristic token associated with the input text by applying an encoder to the input text and the target characteristic token. The method also includes generating, by a decoder, each word of a tuned summary of the input text from the learned subspace representation and from a feedback about preceding words of the tuned summary. The tuned summary is tuned to target characteristics represented by the target characteristic token.
    Type: Application
    Filed: January 30, 2019
    Publication date: July 30, 2020
    Inventors: Balaji Vasan Srinivasan, Kushal Chawla, Mithlesh Kumar, Hrituraj Singh, Arijit Pramanik