Patents by Inventor Kristen Grauman
Kristen Grauman has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11823392Abstract: A method, system and computer program product for segmenting generic foreground objects in images and videos. For segmenting generic foreground objects in videos, an appearance stream of an image in a video frame is processed using a first deep neural network. Furthermore, a motion stream of an optical flow image in the video frame is processed using a second deep neural network. The appearance and motion streams are then joined to combine complementary appearance and motion information to perform segmentation of generic objects in the video frame. Generic foreground objects are segmented in images by training a convolutional deep neural network to estimate a likelihood that a pixel in an image belongs to a foreground object. After receiving the image, the likelihood that the pixel in the image is part of the foreground object as opposed to background is then determined using the trained convolutional deep neural network.Type: GrantFiled: August 2, 2022Date of Patent: November 21, 2023Assignee: Board of Regents, The University of Texas SystemInventors: Kristen Grauman, Suyog Dutt Jain, Bo Xiong
-
Publication number: 20220375102Abstract: A method, system and computer program product for segmenting generic foreground objects in images and videos. For segmenting generic foreground objects in videos, an appearance stream of an image in a video frame is processed using a first deep neural network. Furthermore, a motion stream of an optical flow image in the video frame is processed using a second deep neural network. The appearance and motion streams are then joined to combine complementary appearance and motion information to perform segmentation of generic objects in the video frame. Generic foreground objects are segmented in images by training a convolutional deep neural network to estimate a likelihood that a pixel in an image belongs to a foreground object. After receiving the image, the likelihood that the pixel in the image is part of the foreground object as opposed to background is then determined using the trained convolutional deep neural network.Type: ApplicationFiled: August 2, 2022Publication date: November 24, 2022Inventors: Kristen Grauman, Suyog Dutt Jain, Bo Xiong
-
Patent number: 11423548Abstract: A method, system and computer program product for segmenting generic foreground objects in images and videos. For segmenting generic foreground objects in videos, an appearance stream of an image in a video frame is processed using a first deep neural network. Furthermore, a motion stream of an optical flow image in the video frame is processed using a second deep neural network. The appearance and motion streams are then joined to combine complementary appearance and motion information to perform segmentation of generic objects in the video frame. Generic foreground objects are segmented in images by training a convolutional deep neural network to estimate a likelihood that a pixel in an image belongs to a foreground object. After receiving the image, the likelihood that the pixel in the image is part of the foreground object as opposed to background is then determined using the trained convolutional deep neural network.Type: GrantFiled: December 5, 2017Date of Patent: August 23, 2022Assignee: Board of Regents, The University of Texas SystemInventors: Kristen Grauman, Suyog Dutt Jain, Bo Xiong
-
Publication number: 20190355128Abstract: A method, system and computer program product for segmenting generic foreground objects in images and videos. For segmenting generic foreground objects in videos, an appearance stream of an image in a video frame is processed using a first deep neural network. Furthermore, a motion stream of an optical flow image in the video frame is processed using a second deep neural network. The appearance and motion streams are then joined to combine complementary appearance and motion information to perform segmentation of generic objects in the video frame. Generic foreground objects are segmented in images by training a convolutional deep neural network to estimate a likelihood that a pixel in an image belongs to a foreground object. After receiving the image, the likelihood that the pixel in the image is part of the foreground object as opposed to background is then determined using the trained convolutional deep neural network.Type: ApplicationFiled: December 5, 2017Publication date: November 21, 2019Inventors: Kristen Grauman, Suyog Dutt Jain, Bo Xiong
-
Patent number: 9292517Abstract: A method, system and computer program product for efficiently identifying images, videos, audio files or documents relevant to a user. Using either manual annotations or learned functions, the method predicts the relative strength of an attribute in an image, video, audio file or document from a pool of images, videos, audio files or documents. At query time, the system presents an initial set of reference images, videos, audio files or documents, and the user selects among them to provide relative attribute feedback. Using the resulting constraints in the multi-dimensional attribute space, the relevance function for the pool of images, videos, audio files or documents is updated and the relevance of the pool of images, videos, audio files or documents is re-computed. This procedure iterates using the accumulated constraints until the top-ranked images, videos, audio files or documents are acceptably close to the user's envisioned image, video, audio file or document.Type: GrantFiled: August 13, 2013Date of Patent: March 22, 2016Assignee: Board of Regents, The University of Texas SystemInventors: Kristen Grauman, Adriana Kovashka, Devi Parikh
-
Patent number: 9176993Abstract: A method, system and computer program product for efficiently identifying images, videos, audio files or documents relevant to a user using binary search trees in attribute space for guiding relevance feedback. A binary tree is constructed for each relative attribute of interest. A “pivot exemplar” (at a node of the binary tree) is set for each relative attribute's binary tree as corresponding to the database image, video, audio file or document with a median relative attribute value among that subtree's child examples. A pivot exemplar out of the available current pivot exemplars that has the highest expected information gain is selected to be provided to the user. Comparative attribute feedback is then received from the user regarding whether a degree of the attribute in the user's target image, video, audio file or document is more, less or equal with the attribute displayed in the selected pivot exemplar.Type: GrantFiled: August 13, 2013Date of Patent: November 3, 2015Assignee: Board of Regents, The University of Texas SystemInventors: Kristen Grauman, Adriana Kovashka
-
Publication number: 20140188863Abstract: A method, system and computer program product for efficiently identifying images, videos, audio files or documents relevant to a user. Using either manual annotations or learned functions, the method predicts the relative strength of an attribute in an image, video, audio file or document from a pool of images, videos, audio files or documents. At query time, the system presents an initial set of reference images, videos, audio files or documents, and the user selects among them to provide relative attribute feedback. Using the resulting constraints in the multi-dimensional attribute space, the relevance function for the pool of images, videos, audio files or documents is updated and the relevance of the pool of images, videos, audio files or documents is re-computed. This procedure iterates using the accumulated constraints until the top-ranked images, videos, audio files or documents are acceptably close to the user's envisioned image, video, audio file or document.Type: ApplicationFiled: August 13, 2013Publication date: July 3, 2014Applicant: Board of Regents, The University of Texas SystemInventors: Kristen Grauman, Adriana Kovashka, Devi Parikh
-
Publication number: 20140188901Abstract: A method, system and computer program product for efficiently identifying images, videos, audio files or documents relevant to a user using binary search trees in attribute space for guiding relevance feedback. A binary tree is constructed for each relative attribute of interest. A “pivot exemplar” (at a node of the binary tree) is set for each relative attribute's binary tree as corresponding to the database image, video, audio file or document with a median relative attribute value among that subtree's child examples. A pivot exemplar out of the available current pivot exemplars that has the highest expected information gain is selected to be provided to the user. Comparative attribute feedback is then received from the user regarding whether a degree of the attribute in the user's target image, video, audio file or document is more, less or equal with the attribute displayed in the selected pivot exemplar.Type: ApplicationFiled: August 13, 2013Publication date: July 3, 2014Applicant: Board of Regents, The University of Texas SystemInventors: Kristen Grauman, Adriana Kovashka
-
Patent number: 7949186Abstract: A method for classifying or comparing objects includes detecting points of interest within two objects, computing feature descriptors at said points of interest, forming a multi-resolution histogram over feature descriptors for each object and computing a weighted intersection of multi-resolution histogram for each object. An alternative embodiment includes a method for matching objects by defining a plurality of bins for multi-resolution histograms having various levels and a plurality of cluster groups, each group having a center, for each point of interest, calculating a bin index, a bin count and a maximal distance to the bin center and providing a path vector indicative of the bins chosen at each level. Still another embodiment includes a method for matching objects comprising creating a set of feature vectors for each object of interest, mapping each set of feature vectors to a single high-dimensional vector to create an embedding vector and encoding each embedding vector with a binary hash string.Type: GrantFiled: March 15, 2007Date of Patent: May 24, 2011Assignee: Massachusetts Institute of TechnologyInventors: Kristen Grauman, Trevor Darrell
-
Publication number: 20070217676Abstract: A method for classifying or comparing objects includes detecting points of interest within two objects, computing feature descriptors at said points of interest, forming a multi-resolution histogram over feature descriptors for each object and computing a weighted intersection of multi-resolution histogram for each object. An alternative embodiment includes a method for matching objects by defining a plurality of bins for multi-resolution histograms having various levels and a plurality of cluster groups, each group having a center, for each point of interest, calculating a bin index, a bin count and a maximal distance to the bin center and providing a path vector indicative of the bins chosen at each level. Still another embodiment includes a method for matching objects comprising creating a set of feature vectors for each object of interest, mapping each set of feature vectors to a single high-dimensional vector to create an embedding vector and encoding each embedding vector with a binary hash string.Type: ApplicationFiled: March 15, 2007Publication date: September 20, 2007Inventors: Kristen Grauman, Trevor Darrell