Patents by Inventor Aniruddha Mahapatra
Aniruddha Mahapatra has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 12373923Abstract: Certain aspects and features of this disclosure relate to semantically-aware image extrapolation. In one example, an input image is segmented to produce an input segmentation map of object instances in the input image. An object generation network is used to generate an extrapolated semantic label map for an extrapolated image. The extrapolated semantic label map includes instances in the original image and instances that will appear in an outpainted region of the extrapolated image. A panoptic label map is derived from coordinates of output instances in the extrapolated image and used to identify partial instances and boundaries. Instance-aware context normalization is used to apply one or more characteristics from the input image to the outpainted region to maintain semantic continuity. The extrapolated image includes the original image and the outpainted region and can be rendered or stored for future use.Type: GrantFiled: June 7, 2024Date of Patent: July 29, 2025Assignee: Adobe Inc.Inventors: Kuldeep Kulkarni, Soumya Dash, Hrituraj Singh, Bholeshwar Khurana, Aniruddha Mahapatra, Abhishek Bhatia
-
Patent number: 12282992Abstract: Systems and methods for machine learning based controllable animation of still images is provided. In one embodiment, a still image including a fluid element is obtained. Using a flow refinement machine learning model, a refined dense optical flow is generated for the still image based on a selection mask that includes the fluid element and a dense optical flow generated from a motion hint that indicates a direction of animation. The refined dense optical flow indicates a pattern of apparent motion for the at least one fluid element. Thereafter, a plurality of video frames is generated by projecting a plurality of pixels of the still image using the refined dense optical flow.Type: GrantFiled: July 1, 2022Date of Patent: April 22, 2025Assignee: Adobe Inc.Inventors: Kuldeep Kulkarni, Aniruddha Mahapatra
-
Patent number: 12159109Abstract: Embodiments of the present invention provide systems, methods, and computer storage media for pre-training entity extraction models to facilitate domain adaptation in resource-constrained domains. In an example embodiment, a first machine learning model is used to encode sentences of a source domain corpus and a target domain corpus into sentence embeddings. The sentence embeddings of the target domain corpus are combined into a target corpus embedding. Training sentences from the source domain corpus within a threshold of similarity to the target corpus embedding are selected. A second machine learning model is trained on the training sentences selected from the source domain corpus.Type: GrantFiled: November 12, 2021Date of Patent: December 3, 2024Assignee: Adobe Inc.Inventors: Aniruddha Mahapatra, Sharmila Reddy Nangi, Aparna Garimella, Anandha velu Natarajan
-
Publication number: 20240331102Abstract: Certain aspects and features of this disclosure relate to semantically-aware image extrapolation. In one example, an input image is segmented to produce an input segmentation map of object instances in the input image. An object generation network is used to generate an extrapolated semantic label map for an extrapolated image. The extrapolated semantic label map includes instances in the original image and instances that will appear in an outpainted region of the extrapolated image. A panoptic label map is derived from coordinates of output instances in the extrapolated image and used to identify partial instances and boundaries. Instance-aware context normalization is used to apply one or more characteristics from the input image to the outpainted region to maintain semantic continuity. The extrapolated image includes the original image and the outpainted region and can be rendered or stored for future use.Type: ApplicationFiled: June 7, 2024Publication date: October 3, 2024Inventors: Kuldeep Kulkarni, Soumya Dash, Hrituraj Singh, Bholeshwar Khurana, Aniruddha Mahapatra, Abhishek Bhatia
-
Patent number: 12020403Abstract: Certain aspects and features of this disclosure relate to semantically-aware image extrapolation. In one example, an input image is segmented to produce an input segmentation map of object instances in the input image. An object generation network is used to generate an extrapolated semantic label map for an extrapolated image. The extrapolated semantic label map includes instances in the original image and instances that will appear in an outpainted region of the extrapolated image. A panoptic label map is derived from coordinates of output instances in the extrapolated image and used to identify partial instances and boundaries. Instance-aware context normalization is used to apply one or more characteristics from the input image to the outpainted region to maintain semantic continuity. The extrapolated image includes the original image and the outpainted region and can be rendered or stored for future use.Type: GrantFiled: November 8, 2021Date of Patent: June 25, 2024Assignee: Adobe Inc.Inventors: Kuldeep Kulkarni, Soumya Dash, Hrituraj Singh, Bholeshwar Khurana, Aniruddha Mahapatra, Abhishek Bhatia
-
Publication number: 20240135197Abstract: Embodiments are disclosed for expanding a seed scene using proposals from a generative model of scene graphs. The method may include clustering subgraphs according to respective one or more maximal connected subgraphs of a scene graph. The scene graph includes a plurality of nodes and edges. The method also includes generating a scene sequence for the scene graph based on the clustered subgraphs. A first machine learning model determines a predicted node in response to receiving the scene sequence. A second machine learning model determines a predicted edge in response to receiving the scene sequence and the predicted node. A scene graph is output according to the predicted node and the predicted edge.Type: ApplicationFiled: October 10, 2022Publication date: April 25, 2024Applicant: Adobe Inc.Inventors: Vishwa VINAY, Tirupati Saketh CHANDRA, Rishi AGARWAL, Kuldeep KULKARNI, Hiransh GUPTA, Aniruddha MAHAPATRA, Vaidehi Ramesh PATIL
-
Publication number: 20240005587Abstract: Systems and methods for machine learning based controllable animation of still images is provided. In one embodiment, a still image including a fluid element is obtained. Using a flow refinement machine learning model, a refined dense optical flow is generated for the still image based on a selection mask that includes the fluid element and a dense optical flow generated from a motion hint that indicates a direction of animation. The refined dense optical flow indicates a pattern of apparent motion for the at least one fluid element. Thereafter, a plurality of video frames is generated by projecting a plurality of pixels of the still image using the refined dense optical flow.Type: ApplicationFiled: July 1, 2022Publication date: January 4, 2024Inventors: Kuldeep KULKARNI, Aniruddha MAHAPATRA
-
Publication number: 20230262237Abstract: Systems and methods for image processing are described. The systems and methods include receiving a plurality of frames of a video at an edge device, wherein the video depicts an action that spans the plurality of frames, compressing, using an encoder network, each of the plurality of frames to obtain compressed frame features, wherein the compressed frame features include fewer data bits than the plurality of frames of the video, classifying, using a classification network, the compressed frame features at the edge device to obtain action classification information corresponding to the action in the video, and transmitting the action classification information from the edge device to a central server.Type: ApplicationFiled: February 15, 2022Publication date: August 17, 2023Inventors: Subrata Mitra, Aniruddha Mahapatra, Kuldeep Sharad Kulkarni, Abhishek Yadav, Abhijith Kuruba, Manoj Kilaru
-
Publication number: 20230169632Abstract: Certain aspects and features of this disclosure relate to semantically-aware image extrapolation. In one example, an input image is segmented to produce an input segmentation map of object instances in the input image. An object generation network is used to generate an extrapolated semantic label map for an extrapolated image. The extrapolated semantic label map includes instances in the original image and instances that will appear in an outpainted region of the extrapolated image. A panoptic label map is derived from coordinates of output instances in the extrapolated image and used to identify partial instances and boundaries. Instance-aware context normalization is used to apply one or more characteristics from the input image to the outpainted region to maintain semantic continuity. The extrapolated image includes the original image and the outpainted region and can be rendered or stored for future use.Type: ApplicationFiled: November 8, 2021Publication date: June 1, 2023Inventors: Kuldeep Kulkarni, Soumya Dash, Hrituraj Singh, Bholeshwar Khurana, Aniruddha Mahapatra, Abhishek Bhatia
-
Publication number: 20230153533Abstract: Embodiments of the present invention provide systems, methods, and computer storage media for pre-training entity extraction models to facilitate domain adaptation in resource-constrained domains. In an example embodiment, a first machine learning model is used to encode sentences of a source domain corpus and a target domain corpus into sentence embeddings. The sentence embeddings of the target domain corpus are combined into a target corpus embedding. Training sentences from the source domain corpus within a threshold of similarity to the target corpus embedding are selected. A second machine learning model is trained on the training sentences selected from the source domain corpus.Type: ApplicationFiled: November 12, 2021Publication date: May 18, 2023Inventors: Aniruddha Mahapatra, Sharmila Reddy Nangi, Aparna Garimella, Anandha velu Natarajan