Patents by Inventor Amir GHODRATI

Amir GHODRATI has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 12266179
    Abstract: A processor-implemented method for processing a video includes receiving the video as an input at an artificial neural network (ANN). The video includes a sequence of frames. A set of features of a current frame of the video and a prior frame of the video are extracted. The set of features including a set of support features for a set of pixels of the prior frame to be aligned with a set of reference features of the current frame. A similarity between a support feature for each pixel in the set of pixels of the set of support features of the prior frame and a corresponding reference feature of the current frame is computed. An attention map is generated based on the similarity. An output including a reconstruction of the current frame is generated based on the attention map.
    Type: Grant
    Filed: March 16, 2022
    Date of Patent: April 1, 2025
    Assignee: QUALCOMM INCORPORATED
    Inventors: Davide Abati, Amirhossein Habibian, Amir Ghodrati
  • Publication number: 20240330662
    Abstract: Certain aspects of the present disclosure provide techniques and apparatus for processing multidimensional content using neural networks. An example method generally includes decomposing a multidimensional input into a plurality of two-dimensional subspaces, wherein the plurality of two-dimensional subspaces share a common dimension. A first attention matrix is generated based on a projection of tokens in a first two-dimensional subspace of the plurality of two-dimensional subspaces via an attention block of a transformer neural network, and a second attention matrix is generated based on a projection of tokens in a second two-dimensional subspace of the plurality of two-dimensional subspaces via the attention block of the transformer neural network. An output of the transformer neural network is generated based on a combination of the first attention matrix and the second attention matrix.
    Type: Application
    Filed: March 30, 2023
    Publication date: October 3, 2024
    Inventors: Amir GHODRATI, Amirhossein HABIBIAN
  • Publication number: 20240160896
    Abstract: Certain aspects of the present disclosure provide techniques and apparatus for improved attention-based machine learning. A first attention propagation output is generated using a first transformer block of a plurality of transformer blocks, this generation including processing input data for the first transformer block using a first self-attention sub-block of the first transformer block. The first attention propagation output is propagated to a second transformer block of the plurality of transformer blocks. An output for the second transformer block is generated, this generation including generating output features for the second transformer block based on the first attention propagation output.
    Type: Application
    Filed: June 15, 2023
    Publication date: May 16, 2024
    Inventors: Shashanka VENKATARAMANAN, Amir GHODRATI, Amirhossein HABIBIAN
  • Patent number: 11842540
    Abstract: Systems and techniques are provided for performing holistic video understanding. For example a process can include obtaining a first video and determining, using a machine learning model decision engine, a first machine learning model from a set of machine learning models to use for processing at least a portion of the first video. The first machine learning model can be determined based on one or more characteristics of at least the portion of the first video. The process can include processing at least the portion of the first video using the first machine learning model.
    Type: Grant
    Filed: March 31, 2021
    Date of Patent: December 12, 2023
    Assignee: QUALCOMM Incorporated
    Inventors: Haitam Ben Yahia, Amir Ghodrati, Mihir Jain, Amirhossein Habibian
  • Publication number: 20230154157
    Abstract: A processor-implemented method of video processing using includes receiving, via an artificial neural network (ANN), a video including a first frame and a second frame. A saliency map is generated based on the first frame of the video. The second frame of the video is sampled based on the saliency map. A first portion of the second frame is sampled at a first resolution and a second portion of the second frame is sampled at a second resolution. The first resolution is different than the second resolution. A resampled second frame is generated based on the sampling of the second frame. The resampled second frame is processed to determine an inference associated with the video.
    Type: Application
    Filed: October 25, 2022
    Publication date: May 18, 2023
    Inventors: Babak EHTESHAMI BEJNORDI, Amir GHODRATI, Fatih Murat PORIKLI, Amirhossein HABIBIAN
  • Publication number: 20220318553
    Abstract: Systems and techniques are provided for performing holistic video understanding. For example a process can include obtaining a first video and determining, using a machine learning model decision engine, a first machine learning model from a set of machine learning models to use for processing at least a portion of the first video. The first machine learning model can be determined based on one or more characteristics of at least the portion of the first video. The process can include processing at least the portion of the first video using the first machine learning model.
    Type: Application
    Filed: March 31, 2021
    Publication date: October 6, 2022
    Inventors: Haitam BEN YAHIA, Amir GHODRATI, Mihir JAIN, Amirhossein HABIBIAN
  • Publication number: 20220301311
    Abstract: A processor-implemented method for processing a video includes receiving the video as an input at an artificial neural network (ANN). The video includes a sequence of frames. A set of features of a current frame of the video and a prior frame of the video are extracted. The set of features including a set of support features for a set of pixels of the prior frame to be aligned with a set of reference features of the current frame. A similarity between a support feature for each pixel in the set of pixels of the set of support features of the prior frame and a corresponding reference feature of the current frame is computed. An attention map is generated based on the similarity. An output including a reconstruction of the current frame is generated based on the attention map.
    Type: Application
    Filed: March 16, 2022
    Publication date: September 22, 2022
    Inventors: Davide ABATI, Amirhossein HABIBIAN, Amir GHODRATI
  • Publication number: 20220157045
    Abstract: Certain aspects of the present disclosure provide techniques for processing with an auto exiting machine learning model architecture, including processing input data in a first portion of a classification model to generate first intermediate activation data; providing the first intermediate activation data to a first gate; making a determination by the first gate whether or not to exit processing by the classification model; and generating a classification result from one of a plurality of classifiers of the classification model.
    Type: Application
    Filed: November 15, 2021
    Publication date: May 19, 2022
    Inventors: Babak EHTESHAMI BEJNORDI, Amirhossein HABIBIAN, Fatih Murat PORIKLI, Amir GHODRATI
  • Patent number: 10896342
    Abstract: A method of pixel-wise localization of an actor and an action in a sequence of frames includes receiving a natural language query describing the action and the actor. The method also includes receiving the sequence of frames. The method further includes localizing the action and the actor in the sequence of frames based on the natural language query.
    Type: Grant
    Filed: November 13, 2018
    Date of Patent: January 19, 2021
    Assignee: Qualcomm Incorporated
    Inventors: Kirill Gavrilyuk, Amir Ghodrati, Zhenyang Li, Cornelis Gerardus Maria Snoek
  • Publication number: 20190147284
    Abstract: A method of pixel-wise localization of an actor and an action in a sequence of frames includes receiving a natural language query describing the action and the actor. The method also includes receiving the sequence of frames. The method further includes localizing the action and the actor in the sequence of frames based on the natural language query.
    Type: Application
    Filed: November 13, 2018
    Publication date: May 16, 2019
    Inventors: Kirill GAVRILYUK, Amir GHODRATI, Zhenyang LI, Cornelis Gerardus Maria SNOEK