Patents by Inventor Albert Gordo Soldevila
Albert Gordo Soldevila has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 10678846Abstract: In a method for detecting an object in an input image, an input image vector representing the input image is generated by performing a regional maximum activations of convolutions (R-MAC) using a convolutional neural network (CNN) applied to the input image and using regions for the R-MAC defined by applying a region proposal network (RPN) to the output of the CNN applied to the input image. Likewise, a reference image vector representing a reference image depicting the object is generated by performing the R-MAC using the CNN applied to the reference image and using regions for the R MAC defined by applying the RPN to the output of the CNN applied to the reference image. A similarity metric between the input image vector and the reference image vector is computed, and the object is detected as present in the input image if the similarity metric satisfies a detection criterion.Type: GrantFiled: March 10, 2017Date of Patent: June 9, 2020Assignee: Xerox CorporationInventors: Albert Gordo Soldevila, Jon Almazan, Jerome Revaud, Diane Larlus-Larrondo
-
Patent number: 10635949Abstract: A system and method enable semantic comparisons to be made between word images and concepts. Training word images and their concept labels are used to learn parameters of a neural network for embedding word images and concepts in a semantic subspace in which comparisons can be made between word images and concepts without the need for transcribing the text content of the word image. The training of the neural network aims to minimize a ranking loss over the training set where non relevant concepts for an image which are ranked more highly than relevant ones penalize the ranking loss.Type: GrantFiled: July 7, 2015Date of Patent: April 28, 2020Assignee: XEROX CORPORATIONInventors: Albert Gordo Soldevila, Jon Almazán Almazán, Naila Murray, Florent C. Perronnin
-
Publication number: 20180373955Abstract: Similar images are identified by semantically matching human-supplied text captions accompanying training images. An image representation function is trained to produce similar vectors for similar images according to this similarity. The trained function is applied to non-training second images in a different database to produce second vectors. This trained function does not require the second images to contain captions. A query image is matched to the second images by applying the trained function to the query image to produce a query vector, and the second images are ranked based on how closely the second vectors match the query vector, and the top ranking ones of the second images are output as a response to the query image.Type: ApplicationFiled: June 27, 2017Publication date: December 27, 2018Applicant: Xerox CorporationInventors: Albert Gordo Soldevila, Diane Larlus-Larrondo
-
Publication number: 20180260414Abstract: A method for query expansion uses a representation of an input query object, such as an image, to retrieve representations of similar objects retrieved using the query object representation as a query. Given the set of image representations, a weight is predicted for each using a prediction model which assigns different weights to the image representations. An expanded query is generated as a weighted aggregation (e.g., sum) of the query object representation and at least a subset of the set of similar object representations in which each object representation is weighted with its predicted weight. A higher weight can thus be given to one of the similar object representations, in the expanded query, than to another.Type: ApplicationFiled: March 10, 2017Publication date: September 13, 2018Applicant: Xerox CorporationInventor: Albert Gordo Soldevila
-
Publication number: 20180260415Abstract: In a method for detecting an object in an input image, an input image vector representing the input image is generated by performing a regional maximum activations of convolutions (R-MAC) using a convolutional neural network (CNN) applied to the input image and using regions for the R-MAC defined by applying a region proposal network (RPN) to the output of the CNN applied to the input image. Likewise, a reference image vector representing a reference image depicting the object is generated by performing the R-MAC using the CNN applied to the reference image and using regions for the R MAC defined by applying the RPN to the output of the CNN applied to the reference image. A similarity metric between the input image vector and the reference image vector is computed, and the object is detected as present in the input image if the similarity metric satisfies a detection criterion.Type: ApplicationFiled: March 10, 2017Publication date: September 13, 2018Applicant: Xerox CorporationInventors: Albert Gordo Soldevila, Jon Almazan, Jerome Revaud, Diane Larlus-Larrondo
-
Publication number: 20180101750Abstract: A method is disclosed for performing multiple classification of an image simultaneously using multiple classifiers, where information between the classifiers is shared explicitly and is achieved with a low-rank decomposition of the classifier weights. The method includes applying an input image to classifiers and, more particularly, multiplying the extracted input image features by |?| embedding matrices ?c to generate a latent representation of d-dimensions for each of the |?| characters. The embedding matrices are uncorrelated with a position of the extracted character. The step of applying the extracted character to the classifiers further includes projecting the latent representation with a decoding matrix shared by all the character embedding matrices to generate scores of every character in an alphabet at every position. At least one of the multiplying the extracted input image features and the projecting the latent representation with the decoding matrix are performed with a processor.Type: ApplicationFiled: October 11, 2016Publication date: April 12, 2018Applicant: Xerox CorporationInventor: Albert Gordo Soldevila
-
Patent number: 9928436Abstract: Methods and systems recognize alphanumeric characters in an image by computing individual representations of every character of an alphabet at every character position within a certain word transcription length. These methods and systems embed the individual representations of each alphabet character in a common vectorial subspace (using a matrix) and embed a received image of an alphanumeric word into the common vectorial subspace (using the matrix). Such methods and systems compute the utility value of the embedded alphabet characters at every one of the character positions with respect to the embedded alphanumeric character image; and compute the best transcription alphabet character of every one of the image characters based on the utility value of each embedded alphabet character at each character position. Such methods and systems then assign the best transcription alphabet character for each of the character positions to produce a recognized alphanumeric word within the received image.Type: GrantFiled: July 8, 2015Date of Patent: March 27, 2018Assignee: Conduent Business Services, LLCInventors: Albert Gordo Soldevila, Jon Almazan
-
Patent number: 9792492Abstract: A method for extracting a representation from an image includes inputting an image to a pre-trained neural network. The gradient of a loss function is computed with respect to parameters of the neural network, for the image. A gradient representation is extracted for the image based on the computed gradients, which can be used, for example, for classification or retrieval.Type: GrantFiled: July 7, 2015Date of Patent: October 17, 2017Assignee: XEROX CORPORATIONInventors: Albert Gordo Soldevila, Adrien Gaidon, Florent C. Perronnin
-
Patent number: 9785855Abstract: Methods and systems for license plate recognition utilizing a trained neural network. In an example embodiment, a neural network can be subject to operations involving iteratively training and adapting the neural network for a particular task such as, for example, text recognition in the context of a license plate recognition application. The neural network can be trained to perform generic text recognition utilizing a plurality of training samples. The neural network can be applied to a cropped image of a license plate in order to recognize text and produce a license plate transcription with respect to the license plate. An example of such a neural network is a CNN (Convolutional Neural. Network).Type: GrantFiled: December 17, 2015Date of Patent: October 10, 2017Assignee: Conduent Business Services, LLCInventors: Albert Gordo Soldevila, Jon Almazan
-
Patent number: 9767381Abstract: A system and method provide object localization in a query image based on a global representation of the image generated with a model derived from a convolutional neural network. Representations of annotated images and a query image are each generated based on activations output by a layer of the model which precedes the fully-connected layers of the neural network. A similarity is computed between the query image representation and each of the annotated image representations to identify a subset of the annotated images having the highest computed similarity. Object location information from at least one of the subset of annotated images is transferred to the query image and information is output, based on the transferred object location information.Type: GrantFiled: September 22, 2015Date of Patent: September 19, 2017Assignee: XEROX CORPORATIONInventors: José A. Rodríguez-Serrano, Albert Gordo Soldevila
-
Patent number: 9762393Abstract: Authentication methods are disclosed for determining whether a person or object to be authenticated is a member of a set of authorized persons or objects. A query signature is acquired comprising a vector whose elements store values of an ordered set of features for the person or object to be authenticated. The query signature is compared with an aggregate signature comprising a vector whose elements store values of the ordered set of features for the set of authorized persons or objects. The individual signatures for the authorized persons or objects are not stored; only the aggregate signature. It is determined whether the person or object to be authenticated is a member of the set of authorized persons or objects based on the comparison. The comparing may comprise computing an inner product of the query signature and the aggregate signature, with the determining being based on the inner product.Type: GrantFiled: March 19, 2015Date of Patent: September 12, 2017Assignee: Conduent Business Services, LLCInventors: Albert Gordo Soldevila, Naila Murray, Florent C. Perronnin
-
Publication number: 20170177965Abstract: Methods and systems for license plate recognition utilizing a trained neural network. In an example embodiment, a neural network can be subject to operations involving iteratively training and adapting the neural network for a particular task such as, for example, text recognition in the context of a license plate recognition application. The neural network can be trained to perform generic text recognition utilizing a plurality of training samples. The neural network can be applied to a cropped image of a license plate in order to recognize text and produce a license plate transcription with respect to the license plate. An example of such a neural network is a CNN (Convolutional Neural Network).Type: ApplicationFiled: December 17, 2015Publication date: June 22, 2017Inventors: Albert Gordo Soldevila, Jon Almazan
-
Patent number: 9626594Abstract: A system and method for comparing a text image with or without a wildcard character and a character string are provided. The method includes embedding a character string into a vectorial space by extracting a set of features from the character string and generating a character string representation based on the extracted features, such as a spatial pyramid bag of characters (SPBOC) representation. A text image is embedded into a vectorial space by extracting a set of features from the text image and generating a text image representation based on the text image extracted features. A similarity between the text image representation and the character string representation is computed, which includes computing a function of the text image representation and character string representation.Type: GrantFiled: January 21, 2015Date of Patent: April 18, 2017Assignee: XEROX CORPORATIONInventors: Albert Gordo Soldevila, José Antonio Rodríguez-Serrano, Florent Perronnin
-
Publication number: 20170083792Abstract: A system and method provide object localization in a query image based on a global representation of the image generated with a model derived from a convolutional neural network. Representations of annotated images and a query image are each generated based on activations output by a layer of the model which precedes the fully-connected layers of the neural network. A similarity is computed between the query image representation and each of the annotated image representations to identify a subset of the annotated images having the highest computed similarity. Object location information from at least one of the subset of annotated images is transferred to the query image and information is output, based on the transferred object location information.Type: ApplicationFiled: September 22, 2015Publication date: March 23, 2017Applicant: Xerox CorporationInventors: José A. Rodríguez-Serrano, Albert Gordo Soldevila
-
Publication number: 20170011279Abstract: A system and method enable semantic comparisons to be made between word images and concepts. Training word images and their concept labels are used to learn parameters of a neural network for embedding word images and concepts in a semantic subspace in which comparisons can be made between word images and concepts without the need for transcribing the text content of the word image. The training of the neural network aims to minimize a ranking loss over the training set where non relevant concepts for an image which are ranked more highly than relevant ones penalize the ranking loss.Type: ApplicationFiled: July 7, 2015Publication date: January 12, 2017Applicant: Xerox CorporationInventors: Albert Gordo Soldevila, Jon Almazán Almazán, Naila Murray, Florent C. Perronnin
-
Publication number: 20170011280Abstract: A method for extracting a representation from an image includes inputting an image to a pre-trained neural network. The gradient of a loss function is computed with respect to parameters of the neural network, for the image. A gradient representation is extracted for the image based on the computed gradients, which can be used, for example, for classification or retrieval.Type: ApplicationFiled: July 7, 2015Publication date: January 12, 2017Applicant: Xerox CorporationInventors: Albert Gordo Soldevila, Adrien Gaidon, Florent C. Perronnin
-
Publication number: 20170011273Abstract: Methods and systems recognize alphanumeric characters in an image by computing individual representations of every character of an alphabet at every character position within a certain word transcription length. These methods and systems embed the individual representations of each alphabet character in a common vectorial subspace (using a matrix) and embed a received image of an alphanumeric word into the common vectorial subspace (using the matrix). Such methods and systems compute the utility value of the embedded alphabet characters at every one of the character positions with respect to the embedded alphanumeric character image; and compute the best transcription alphabet character of every one of the image characters based on the utility value of each embedded alphabet character at each character position. Such methods and systems then assign the best transcription alphabet character for each of the character positions to produce a recognized alphanumeric word within the received image.Type: ApplicationFiled: July 8, 2015Publication date: January 12, 2017Inventors: Albert Gordo Soldevila, Jon Almazan
-
Publication number: 20160277190Abstract: Authentication methods are disclosed for determining whether a person or object to be authenticated is a member of a set of authorized persons or objects. A query signature is acquired comprising a vector whose elements store values of an ordered set of features for the person or object to be authenticated. The query signature is compared with an aggregate signature comprising a vector whose elements store values of the ordered set of features for the set of authorized persons or objects. The individual signatures for the authorized persons or objects are not stored; only the aggregate signature. It is determined whether the person or object to be authenticated is a member of the set of authorized persons or objects based on the comparison. The comparing may comprise computing an inner product of the query signature and the aggregate signature, with the determining being based on the inner product.Type: ApplicationFiled: March 19, 2015Publication date: September 22, 2016Inventors: Albert Gordo Soldevila, Naila Murray, Florent C. Perronnin
-
Patent number: 9443164Abstract: A system and method for object instance localization in an image are disclosed. In the method, keypoints are detected in a target image and candidate regions are detected by matching the detected keypoints to keypoints detected in a set of reference images. Similarity measures between global descriptors computed for the located candidate regions and global descriptors for the reference images are computed and labels are assigned to at least some of the candidate regions based on the computed similarity measures. Performing the region detection based on keypoint matching while performing the labeling based on global descriptors improves object instance detection.Type: GrantFiled: December 2, 2014Date of Patent: September 13, 2016Inventors: Milan Sulc, Albert Gordo Soldevila, Diane Larlus Larrondo, Florent C. Perronnin
-
Publication number: 20160210532Abstract: A system and method for comparing a text image with or without a wildcard character and a character string are provided. The method includes embedding a character string into a vectorial space by extracting a set of features from the character string and generating a character string representation based on the extracted features, such as a spatial pyramid bag of characters (SPBOC) representation. A text image is embedded into a vectorial space by extracting a set of features from the text image and generating a text image representation based on the text image extracted features. A similarity between the text image representation and the character string representation is computed, which includes computing a function of the text image representation and character string representation.Type: ApplicationFiled: January 21, 2015Publication date: July 21, 2016Inventors: Albert Gordo Soldevila, José Antonio Rodríguez-Serrano, Florent Perronnin