Patents by Inventor Imran Saleemi

Imran Saleemi has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20230044969
    Abstract: The present disclosure describes techniques of improving video matting. The techniques comprise extracting features from each frame of a video by an encoder of a model, wherein the video comprises a plurality of frames; incorporating, by a decoder of the model, into any particular frame temporal information extracted from one or more frames previous to the particular frame, wherein the particular frame and the one or more previous frames are among the plurality of frames of the video, and the decoder is a recurrent decoder; and generating a representation of a foreground object included in the particular frame by the model, wherein the model is trained using segmentation dataset and matting dataset.
    Type: Application
    Filed: August 6, 2021
    Publication date: February 9, 2023
    Inventors: Linjie YANG, Peter LIN, Imran SALEEMI
  • Patent number: 11087130
    Abstract: Various embodiments are disclosed for simultaneous object localization and attribute classification using multitask deep neural networks. In an embodiment, a method comprises: obtaining, by a processing circuit, an image from an image capture device in an environment, the image including a target object in the environment; generating, by the processing circuit, predictions from the image for the target object using a multitask deep neural network, the multitask deep neural network including a network trunk and side branches, the network trunk configured for multi-scale feature extraction guided by supervision information provided by the side branches during training of the multitask deep neural network, the side branches configured as learning task-specific classifiers; and using, by the processing circuit, the predictions to localize the target object in the environment and to classify the target object and at least one attribute of the target object.
    Type: Grant
    Filed: December 26, 2018
    Date of Patent: August 10, 2021
    Assignee: RetailNext, Inc.
    Inventors: Dong Liu, Imran Saleemi, Mark Jamtgaard
  • Patent number: 10891741
    Abstract: Various embodiments are disclosed for detecting, tracking and counting objects of interest in video frames using fusion of image and depth modalities. In an embodiment, a method comprises: obtaining multiple frames of stereo image pairs from an image capturing device; rectifying each frame of the stereo image pairs; computing a stereo disparity for each frame of the stereo image pairs; determining a first set of object detections in each frame using the computed stereo disparity; determining a second set of object detections in each left or right frame of the stereo image pair using one or more machine learning models; fusing the first and second sets of object detections; and creating tracks based on the fused object detections.
    Type: Grant
    Filed: December 26, 2018
    Date of Patent: January 12, 2021
    Assignee: RetailNext, Inc.
    Inventors: Imran Saleemi, Mark Jamtgaard, Dong Liu
  • Publication number: 20200242784
    Abstract: Various embodiments are disclosed for detecting, tracking and counting objects of interest in video. In an embodiment, a method of detecting and tracking objects of interest comprises: obtaining, by a computing device, multiple frames of images from an image capturing device; detecting, by the computing device, objects of interest in each frame; accumulating, by the computing device, multiple frames of object detections; creating, by the computing device, object tracks based on a batch of object detections over multiple frames; and associating, by the computing device, the object tracks over consecutive batches.
    Type: Application
    Filed: April 13, 2020
    Publication date: July 30, 2020
    Applicant: RetailNext, Inc.
    Inventors: Imran Saleemi, Mark Jamtgaard, Arun Nair
  • Patent number: 10621735
    Abstract: Various embodiments are disclosed for detecting, tracking and counting objects of interest in video. In an embodiment, a method of detecting and tracking objects of interest comprises: obtaining, by a computing device, multiple frames of images from an image capturing device; detecting, by the computing device, objects of interest in each frame; accumulating, by the computing device, multiple frames of object detections; creating, by the computing device, object tracks based on a batch of object detections over multiple frames; and associating, by the computing device, the object tracks over consecutive batches.
    Type: Grant
    Filed: November 15, 2018
    Date of Patent: April 14, 2020
    Assignee: RetailNext, Inc.
    Inventors: Imran Saleemi, Mark Jamtgaard, Arun Nair
  • Publication number: 20190325588
    Abstract: Various embodiments are disclosed for detecting, tracking and counting objects of interest in video. In an embodiment, a method of detecting and tracking objects of interest comprises: obtaining, by a computing device, multiple frames of images from an image capturing device; detecting, by the computing device, objects of interest in each frame; accumulating, by the computing device, multiple frames of object detections; creating, by the computing device, object tracks based on a batch of object detections over multiple frames; and associating, by the computing device, the object tracks over consecutive batches.
    Type: Application
    Filed: November 15, 2018
    Publication date: October 24, 2019
    Inventors: Imran Saleemi, Mark Jamtgaard, Arun Nair
  • Publication number: 20190205643
    Abstract: Various embodiments are disclosed for simultaneous object localization and attribute classification using multitask deep neural networks. In an embodiment, a method comprises: obtaining, by a processing circuit, an image from an image capture device in an environment, the image including a target object in the environment; generating, by the processing circuit, predictions from the image for the target object using a multitask deep neural network, the multitask deep neural network including a network trunk and side branches, the network trunk configured for multi-scale feature extraction guided by supervision information provided by the side branches during training of the multitask deep neural network, the side branches configured as learning task-specific classifiers; and using, by the processing circuit, the predictions to localize the target object in the environment and to classify the target object and at least one attribute of the target object.
    Type: Application
    Filed: December 26, 2018
    Publication date: July 4, 2019
    Inventors: Dong Liu, Imran Saleemi, Mark Jamtgaard
  • Publication number: 20190206066
    Abstract: Various embodiments are disclosed for detecting, tracking and counting objects of interest in video frames using fusion of image and depth modalities. In an embodiment, a method comprises: obtaining multiple frames of stereo image pairs from an image capturing device; rectifying each frame of the stereo image pairs; computing a stereo disparity for each frame of the stereo image pairs; determining a first set of object detections in each frame using the computed stereo disparity; determining a second set of object detections in each left or right frame of the stereo image pair using one or more machine learning models; fusing the first and second sets of object detections; and creating tracks based on the fused object detections.
    Type: Application
    Filed: December 26, 2018
    Publication date: July 4, 2019
    Inventors: Imran Saleemi, Mark Jamtgaard, Dong Liu
  • Patent number: 10134146
    Abstract: Various embodiments are disclosed for detecting, tracking and counting objects of interest in video. In an embodiment, a method of detecting and tracking objects of interest comprises: obtaining, by a computing device, multiple frames of images from an image capturing device; detecting, by the computing device, objects of interest in each frame; accumulating, by the computing device, multiple frames of object detections; creating, by the computing device, object tracks based on a batch of object detections over multiple frames; and associating, by the computing device, the object tracks over consecutive batches.
    Type: Grant
    Filed: January 13, 2017
    Date of Patent: November 20, 2018
    Assignee: RetailNext, Inc.
    Inventors: Imran Saleemi, Mark Jamtgaard, Arun Nair
  • Patent number: 9946952
    Abstract: A method for counting individuals in an image containing a dense, uniform or non-uniform crowd. The current invention leverages multiple sources of information to compute an estimate of the number of individuals present in a dense crowd visible in a single image. This approach relies on multiple sources, such as low confidence head detections, repetition of texture elements (using SIFT), and frequency-domain analysis to estimate counts, along with confidence associated with observing individuals in an image region. Additionally, a global consistency constraint can be employed on counts using Markov Random Field. This caters for disparity in counts in local neighborhoods and across scales. The methodology was tested on a new dataset of fifty (50) crowd images containing over 64,000 annotated humans, with the head counts ranging from 94 to 4,543. Efficient and accurate results were attained.
    Type: Grant
    Filed: June 25, 2014
    Date of Patent: April 17, 2018
    Assignee: University of Central Florida Research Foundation, Inc.
    Inventors: Haroon Idrees, Imran Saleemi, Mubarak Shah
  • Publication number: 20180005071
    Abstract: A method for counting individuals in an image containing a dense, uniform or non-uniform crowd. The current invention leverages multiple sources of information to compute an estimate of the number of individuals present in a dense crowd visible in a single image. This approach relies on multiple sources, such as low confidence head detections, repetition of texture elements (using SIFT), and frequency-domain analysis to estimate counts, along with confidence associated with observing individuals in an image region. Additionally, a global consistency constraint can be employed on counts using Markov Random Field. This caters for disparity in counts in local neighborhoods and across scales. The methodology was tested on a new dataset of fifty (50) crowd images containing over 64,000 annotated humans, with the head counts ranging from 94 to 4,543. Efficient and accurate results were attained.
    Type: Application
    Filed: June 25, 2014
    Publication date: January 4, 2018
    Inventors: Haroon Idrees, Imran Saleemi, Mubarak Shah
  • Publication number: 20170206669
    Abstract: Various embodiments are disclosed for detecting, tracking and counting objects of interest in video. In an embodiment, a method of detecting and tracking objects of interest comprises: obtaining, by a computing device, multiple frames of images from an image capturing device; detecting, by the computing device, objects of interest in each frame; accumulating, by the computing device, multiple frames of object detections; creating, by the computing device, object tracks based on a batch of object detections over multiple frames; and associating, by the computing device, the object tracks over consecutive batches.
    Type: Application
    Filed: January 13, 2017
    Publication date: July 20, 2017
    Inventors: Imran Saleemi, Mark Jamtgaard, Arun Nair