Patents by Inventor Diane Larlus-Larrondo
Diane Larlus-Larrondo has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11263753Abstract: A method and system pre-trains a convolutional neural network for image recognition based upon masked language modeling by inputting, to the convolutional neural network, an image; outputting, from the convolutional neural network, a visual embedding tensor of visual embedding vectors; tokenizing a caption to create a list of tokens, at least one token having visual correspondence to the image received by the convolutional neural network; randomly selecting one of the tokens in the list of tokens to be masked, the selected token being taken as ground truth; computing, using a language model neural network, hidden representations of the tokens; using the hidden representation of the masked token, as a query vector, to pool the visual embedding vectors in the visual embedding tensor, attentively; predicting the masked token by mapping the pooled visual embedding vectors to the tokens; determining a prediction loss associated with the masked token; and back-propagating the prediction loss to the convolutional neType: GrantFiled: April 7, 2020Date of Patent: March 1, 2022Inventors: Diane Larlus-Larrondo, Julien Perez, Mert Bulent Sariyildiz
-
Publication number: 20210312628Abstract: A method and system pre-trains a convolutional neural network for image recognition based upon masked language modeling by inputting, to the convolutional neural network, an image; outputting, from the convolutional neural network, a visual embedding tensor of visual embedding vectors; tokenizing a caption to create a list of tokens, at least one token having visual correspondence to the image received by the convolutional neural network; randomly selecting one of the tokens in the list of tokens to be masked, the selected token being taken as ground truth; computing, using a language model neural network, hidden representations of the tokens; using the hidden representation of the masked token, as a query vector, to pool the visual embedding vectors in the visual embedding tensor, attentively; predicting the masked token by mapping the pooled visual embedding vectors to the tokens; determining a prediction loss associated with the masked token; and back-propagating the prediction loss to the convolutional neType: ApplicationFiled: April 7, 2020Publication date: October 7, 2021Applicant: Naver CorporationInventors: Diane Larlus-Larrondo, Julien Perez, Mert Bulent Sariyildiz
-
Patent number: 11138469Abstract: A method for re-identification of a subject in an image by pre-training a convolutional neural network to recognize individuals within a closed set of possible identifications and further pre-training the convolutional neural network using classification loss; training the pre-trained convolutional neural network by sequentially processing a plurality of triplet of images, each triplet containing a query image degraded by adding random noise to a region of the query image, a positive image corresponding to an image of a same subject as in the query image, and a negative image corresponding to an image of a different subject as in the query image by (a) ranking the triplets by the triplet loss computed, (b) selecting a subset of triplets among the plurality of triplets, and (c) retraining the pre-trained convolutional neural network on each of the triplets of the subset of triplets.Type: GrantFiled: November 6, 2019Date of Patent: October 5, 2021Inventors: Jon Almazan, Bojana Gajic, Naila Murray, Diane Larlus-Larrondo
-
Publication number: 20200226421Abstract: A method for re-identification of a subject in an image from a set of images, by pre-training, by a data processor in a first server, a convolutional neural network with ImageNet to recognize individuals within a closed set of possible identifications and further pre-training the convolutional neural network using classification loss to realize person identification; training, by the data processor in the first server, the pre-trained convolutional neural network by sequentially processing a plurality of triplet of images and allowing a different size input for each image, each triplet containing a query image degraded by adding random noise to a region of the query image, a positive image corresponding to an image of a same subject as in the query image, and a negative image corresponding to an image of a different subject as in the query image by (a) ranking the triplets by the triplet loss computed (b) selecting a subset of triplets among the plurality of triplets, the subset of triplets having the largestType: ApplicationFiled: November 6, 2019Publication date: July 16, 2020Applicant: Naver CorporationInventors: Jon Almazan, Bojana Gajic, Naila Murray, Diane Larlus-Larrondo
-
Patent number: 10678846Abstract: In a method for detecting an object in an input image, an input image vector representing the input image is generated by performing a regional maximum activations of convolutions (R-MAC) using a convolutional neural network (CNN) applied to the input image and using regions for the R-MAC defined by applying a region proposal network (RPN) to the output of the CNN applied to the input image. Likewise, a reference image vector representing a reference image depicting the object is generated by performing the R-MAC using the CNN applied to the reference image and using regions for the R MAC defined by applying the RPN to the output of the CNN applied to the reference image. A similarity metric between the input image vector and the reference image vector is computed, and the object is detected as present in the input image if the similarity metric satisfies a detection criterion.Type: GrantFiled: March 10, 2017Date of Patent: June 9, 2020Assignee: Xerox CorporationInventors: Albert Gordo Soldevila, Jon Almazan, Jerome Revaud, Diane Larlus-Larrondo
-
Publication number: 20180373955Abstract: Similar images are identified by semantically matching human-supplied text captions accompanying training images. An image representation function is trained to produce similar vectors for similar images according to this similarity. The trained function is applied to non-training second images in a different database to produce second vectors. This trained function does not require the second images to contain captions. A query image is matched to the second images by applying the trained function to the query image to produce a query vector, and the second images are ranked based on how closely the second vectors match the query vector, and the top ranking ones of the second images are output as a response to the query image.Type: ApplicationFiled: June 27, 2017Publication date: December 27, 2018Applicant: Xerox CorporationInventors: Albert Gordo Soldevila, Diane Larlus-Larrondo
-
Publication number: 20180260415Abstract: In a method for detecting an object in an input image, an input image vector representing the input image is generated by performing a regional maximum activations of convolutions (R-MAC) using a convolutional neural network (CNN) applied to the input image and using regions for the R-MAC defined by applying a region proposal network (RPN) to the output of the CNN applied to the input image. Likewise, a reference image vector representing a reference image depicting the object is generated by performing the R-MAC using the CNN applied to the reference image and using regions for the R MAC defined by applying the RPN to the output of the CNN applied to the reference image. A similarity metric between the input image vector and the reference image vector is computed, and the object is detected as present in the input image if the similarity metric satisfies a detection criterion.Type: ApplicationFiled: March 10, 2017Publication date: September 13, 2018Applicant: Xerox CorporationInventors: Albert Gordo Soldevila, Jon Almazan, Jerome Revaud, Diane Larlus-Larrondo
-
Publication number: 20170330059Abstract: A method for generating object and part detectors includes accessing a collection of training images. The collection of training images includes images annotated with an object label and images annotated with a respective part label for each of a plurality of parts of the object. Joint appearance-geometric embeddings for regions of a set of the training images are generated. At least one detector for the object and its parts is learnt using annotations of the training images and respective joint appearance-geometric embeddings, e.g., using multi-instance learning for generating parameters of scoring functions which are used to identify high scoring regions for learning the object and its parts. The detectors may be output or used to label regions of a new image with object and part labels.Type: ApplicationFiled: May 11, 2016Publication date: November 16, 2017Applicant: Xerox CorporationInventors: David Novotny, Diane Larlus Larrondo, Andrea Vedaldi
-
Patent number: 9697439Abstract: An object detection method includes for each of a set of patches of an image, encoding features of the patch with a non-linear mapping function, and computing per-patch statistics based on the encoded features for approximating a window-level non-linear operation by a patch-level operation. Then, windows are extracted from the image, each window comprising a sub-set of the set of patches. Each of the windows is scored based on the computed patch statistics of the respective sub-set of patches. Objects, if any, can then be detected in the image, based on the window scores. The method and system allow the non-linear operations to be performed only at the patch level, reducing the computation time of the method, since there are generally many more windows than patches, while not impacting performance unduly, as compared to a system which performs non-linear operations at the window level.Type: GrantFiled: October 2, 2014Date of Patent: July 4, 2017Assignee: XEROX CORPORATIONInventors: Adrien Gaidon, Diane Larlus-Larrondo, Florent C. Perronnin
-
Patent number: 9639806Abstract: A system and method for evaluating iconicity of an image are provided. In the method, at least one test image is received, each test image including an object in a selected class. Properties related to iconicity are computed for each test image. The properties may include one or more of: a) a direct measure of iconicity, which is computed with a direct iconicity prediction model which has been learned on a set of training images, each training image labeled with an iconicity score; b) one or more class-independent properties; and c) one or more class-dependent properties. A measure of iconicity of each of the test images is computed, based on the computed properties. By combining a set of complementary properties, an iconicity measure which shows good agreement with human evaluations of iconicity can be obtained.Type: GrantFiled: April 15, 2014Date of Patent: May 2, 2017Assignee: XEROX CORPORATIONInventors: Yangmuzi Zhang, Diane Larlus-Larrondo, Florent C. Perronnin
-
Patent number: 9606988Abstract: A system and method predict the translation quality of a translated input document. The method includes receiving an input document pair composed of a plurality of sentence pairs, each sentence pair including a source sentence in a source language and a machine translation of the source language sentence to a target language sentence. For each of the sentence pairs, a representation of the sentence pair is generated, based on a set of features extracted for the sentence pair. Using a generative model, a representation of the input document pair is generated, based on the sentence pair representations. A translation quality of the translated input document is computed, based on the representation of the input document pair.Type: GrantFiled: November 4, 2014Date of Patent: March 28, 2017Assignee: XEROX CORPORATIONInventors: Jean-Marc Andreoli, Diane Larlus-Larrondo, Jean-Luc Meunier
-
Patent number: 9514391Abstract: In an image classification method, a feature vector representing an input image is generated by unsupervised operations including extracting local descriptors from patches distributed over the input image, and a classification value for the input image is generated by applying a neural network (NN) to the feature vector. Extracting the feature vector may include encoding the local descriptors extracted from each patch using a generative model, such as Fisher vector encoding, aggregating the encoded local descriptors to form a vector, projecting the vector into a space of lower dimensionality, for example using Principal Component Analysis (PCA), and normalizing the feature vector of lower dimensionality to produce the feature vector representing the input image. A set of mid-level features representing the input image may be generated as the output of an intermediate layer of the NN.Type: GrantFiled: April 20, 2015Date of Patent: December 6, 2016Assignee: XEROX CORPORATIONInventors: Florent C. Perronnin, Diane Larlus-Larrondo
-
Publication number: 20160307071Abstract: In an image classification method, a feature vector representing an input image is generated by unsupervised operations including extracting local descriptors from patches distributed over the input image, and a classification value for the input image is generated by applying a neural network (NN) to the feature vector. Extracting the feature vector may include encoding the local descriptors extracted from each patch using a generative model, such as Fisher vector encoding, aggregating the encoded local descriptors to form a vector, projecting the vector into a space of lower dimensionality, for example using Principal Component Analysis (PCA), and normalizing the feature vector of lower dimensionality to produce the feature vector representing the input image. A set of mid-level features representing the input image may be generated as the output of an intermediate layer of the NN.Type: ApplicationFiled: April 20, 2015Publication date: October 20, 2016Inventors: Florent C. Perronnin, Diane Larlus-Larrondo
-
Patent number: 9443164Abstract: A system and method for object instance localization in an image are disclosed. In the method, keypoints are detected in a target image and candidate regions are detected by matching the detected keypoints to keypoints detected in a set of reference images. Similarity measures between global descriptors computed for the located candidate regions and global descriptors for the reference images are computed and labels are assigned to at least some of the candidate regions based on the computed similarity measures. Performing the region detection based on keypoint matching while performing the labeling based on global descriptors improves object instance detection.Type: GrantFiled: December 2, 2014Date of Patent: September 13, 2016Inventors: Milan Sulc, Albert Gordo Soldevila, Diane Larlus Larrondo, Florent C. Perronnin
-
Publication number: 20160155011Abstract: A system and method for object instance localization in an image are disclosed. In the method, keypoints are detected in a target image and candidate regions are detected by matching the detected keypoints to keypoints detected in a set of reference images. Similarity measures between global descriptors computed for the located candidate regions and global descriptors for the reference images are computed and labels are assigned to at least some of the candidate regions based on the computed similarity measures. Performing the region detection based on keypoint matching while performing the labeling based on global descriptors improves object instance detection.Type: ApplicationFiled: December 2, 2014Publication date: June 2, 2016Inventors: Milan Sulc, Albert Gordo Soldevila, Diane Larlus Larrondo, Florent C. Perronnin
-
Publication number: 20160124944Abstract: A system and method predict the translation quality of a translated input document. The method includes receiving an input document pair composed of a plurality of sentence pairs, each sentence pair including a source sentence in a source language and a machine translation of the source language sentence to a target language sentence. For each of the sentence pairs, a representation of the sentence pair is generated, based on a set of features extracted for the sentence pair. Using a generative model, a representation of the input document pair is generated, based on the sentence pair representations. A translation quality of the translated input document is computed, based on the representation of the input document pair.Type: ApplicationFiled: November 4, 2014Publication date: May 5, 2016Inventors: Jean-Marc Andreoli, Diane Larlus-Larrondo, Jean-Luc Meunier
-
Publication number: 20160098619Abstract: An object detection method includes for each of a set of patches of an image, encoding features of the patch with a non-linear mapping function, and computing per-patch statistics based on the encoded features for approximating a window-level non-linear operation by a patch-level operation. Then, windows are extracted from the image, each window comprising a sub-set of the set of patches. Each of the windows is scored based on the computed patch statistics of the respective sub-set of patches. Objects, if any, can then be detected in the image, based on the window scores. The method and system allow the non-linear operations to be performed only at the patch level, reducing the computation time of the method, since there are generally many more windows than patches, while not impacting performance unduly, as compared to a system which performs non-linear operations at the window level.Type: ApplicationFiled: October 2, 2014Publication date: April 7, 2016Inventors: Adrien Gaidon, Diane Larlus-Larrondo, Florent C. Perronnin
-
Publication number: 20150294191Abstract: A system and method for evaluating iconicity of an image are provided. In the method, at least one test image is received, each test image including an object in a selected class. Properties related to iconicity are computed for each test image. The properties may include one or more of: a) a direct measure of iconicity, which is computed with a direct iconicity prediction model which has been learned on a set of training images, each training image labeled with an iconicity score; b) one or more class-independent properties; and c) one or more class-dependent properties. A measure of iconicity of each of the test images is computed, based on the computed properties. By combining a set of complementary properties, an iconicity measure which shows good agreement with human evaluations of iconicity can be obtained.Type: ApplicationFiled: April 15, 2014Publication date: October 15, 2015Applicant: Xerox CorporationInventors: Yangmuzi Zhang, Diane Larlus-Larrondo, Florent C. Perronnin
-
Patent number: 9158995Abstract: A computer implemented method for localization of an object, such as a license plate, in an input image includes generating a task-dependent representation of the input image based on relevance scores for the object to be localized. The relevance scores are output by a classifier for a plurality of locations in the input image, such as patches. The classifier is trained on patches extracted from training images and their respective relevance labels. One or more similar images are identified from a set of images, based on a comparison of the task-dependent representation of the input image and task-dependent representations of images in the set of images. A location of the object in the input image is identified based on object location annotations for the similar images.Type: GrantFiled: March 14, 2013Date of Patent: October 13, 2015Assignee: XEROX CORPORATIONInventors: Jose Antonio Rodriguez-Serrano, Diane Larlus-Larrondo
-
Publication number: 20150235160Abstract: A system and method for generating gold questions for labeling tasks are disclosed. The method includes sampling a positive class from a predefined set of classes to be used in labeling documents, based on a computed measure of class popularity. A set of negative classes is identified from the set of classes based on a distance measure between the positive class and other classes in the set of classes. A gold question is generated which includes a document representative of the positive class and a set of candidate answers. The candidate answers include a label for the positive class and a label for each of the negative classes in the identified set of negative classes. A task may be generated which includes the gold question and a plurality of standard questions which each include a document to be labeled. A computer processor may implement all or part of the method.Type: ApplicationFiled: February 20, 2014Publication date: August 20, 2015Applicant: Xerox CorporationInventors: Diane Larlus-Larrondo, Vivek Kumar Mishra, Pramod Sankar Kompalli, Florent C. Perronnin