Patents by Inventor Albert Gordo Soldevila

Albert Gordo Soldevila has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 10678846
    Abstract: In a method for detecting an object in an input image, an input image vector representing the input image is generated by performing a regional maximum activations of convolutions (R-MAC) using a convolutional neural network (CNN) applied to the input image and using regions for the R-MAC defined by applying a region proposal network (RPN) to the output of the CNN applied to the input image. Likewise, a reference image vector representing a reference image depicting the object is generated by performing the R-MAC using the CNN applied to the reference image and using regions for the R MAC defined by applying the RPN to the output of the CNN applied to the reference image. A similarity metric between the input image vector and the reference image vector is computed, and the object is detected as present in the input image if the similarity metric satisfies a detection criterion.
    Type: Grant
    Filed: March 10, 2017
    Date of Patent: June 9, 2020
    Assignee: Xerox Corporation
    Inventors: Albert Gordo Soldevila, Jon Almazan, Jerome Revaud, Diane Larlus-Larrondo
  • Patent number: 10635949
    Abstract: A system and method enable semantic comparisons to be made between word images and concepts. Training word images and their concept labels are used to learn parameters of a neural network for embedding word images and concepts in a semantic subspace in which comparisons can be made between word images and concepts without the need for transcribing the text content of the word image. The training of the neural network aims to minimize a ranking loss over the training set where non relevant concepts for an image which are ranked more highly than relevant ones penalize the ranking loss.
    Type: Grant
    Filed: July 7, 2015
    Date of Patent: April 28, 2020
    Assignee: XEROX CORPORATION
    Inventors: Albert Gordo Soldevila, Jon Almazán Almazán, Naila Murray, Florent C. Perronnin
  • Publication number: 20180373955
    Abstract: Similar images are identified by semantically matching human-supplied text captions accompanying training images. An image representation function is trained to produce similar vectors for similar images according to this similarity. The trained function is applied to non-training second images in a different database to produce second vectors. This trained function does not require the second images to contain captions. A query image is matched to the second images by applying the trained function to the query image to produce a query vector, and the second images are ranked based on how closely the second vectors match the query vector, and the top ranking ones of the second images are output as a response to the query image.
    Type: Application
    Filed: June 27, 2017
    Publication date: December 27, 2018
    Applicant: Xerox Corporation
    Inventors: Albert Gordo Soldevila, Diane Larlus-Larrondo
  • Publication number: 20180260414
    Abstract: A method for query expansion uses a representation of an input query object, such as an image, to retrieve representations of similar objects retrieved using the query object representation as a query. Given the set of image representations, a weight is predicted for each using a prediction model which assigns different weights to the image representations. An expanded query is generated as a weighted aggregation (e.g., sum) of the query object representation and at least a subset of the set of similar object representations in which each object representation is weighted with its predicted weight. A higher weight can thus be given to one of the similar object representations, in the expanded query, than to another.
    Type: Application
    Filed: March 10, 2017
    Publication date: September 13, 2018
    Applicant: Xerox Corporation
    Inventor: Albert Gordo Soldevila
  • Publication number: 20180260415
    Abstract: In a method for detecting an object in an input image, an input image vector representing the input image is generated by performing a regional maximum activations of convolutions (R-MAC) using a convolutional neural network (CNN) applied to the input image and using regions for the R-MAC defined by applying a region proposal network (RPN) to the output of the CNN applied to the input image. Likewise, a reference image vector representing a reference image depicting the object is generated by performing the R-MAC using the CNN applied to the reference image and using regions for the R MAC defined by applying the RPN to the output of the CNN applied to the reference image. A similarity metric between the input image vector and the reference image vector is computed, and the object is detected as present in the input image if the similarity metric satisfies a detection criterion.
    Type: Application
    Filed: March 10, 2017
    Publication date: September 13, 2018
    Applicant: Xerox Corporation
    Inventors: Albert Gordo Soldevila, Jon Almazan, Jerome Revaud, Diane Larlus-Larrondo
  • Publication number: 20180101750
    Abstract: A method is disclosed for performing multiple classification of an image simultaneously using multiple classifiers, where information between the classifiers is shared explicitly and is achieved with a low-rank decomposition of the classifier weights. The method includes applying an input image to classifiers and, more particularly, multiplying the extracted input image features by |?| embedding matrices ?c to generate a latent representation of d-dimensions for each of the |?| characters. The embedding matrices are uncorrelated with a position of the extracted character. The step of applying the extracted character to the classifiers further includes projecting the latent representation with a decoding matrix shared by all the character embedding matrices to generate scores of every character in an alphabet at every position. At least one of the multiplying the extracted input image features and the projecting the latent representation with the decoding matrix are performed with a processor.
    Type: Application
    Filed: October 11, 2016
    Publication date: April 12, 2018
    Applicant: Xerox Corporation
    Inventor: Albert Gordo Soldevila
  • Patent number: 9928436
    Abstract: Methods and systems recognize alphanumeric characters in an image by computing individual representations of every character of an alphabet at every character position within a certain word transcription length. These methods and systems embed the individual representations of each alphabet character in a common vectorial subspace (using a matrix) and embed a received image of an alphanumeric word into the common vectorial subspace (using the matrix). Such methods and systems compute the utility value of the embedded alphabet characters at every one of the character positions with respect to the embedded alphanumeric character image; and compute the best transcription alphabet character of every one of the image characters based on the utility value of each embedded alphabet character at each character position. Such methods and systems then assign the best transcription alphabet character for each of the character positions to produce a recognized alphanumeric word within the received image.
    Type: Grant
    Filed: July 8, 2015
    Date of Patent: March 27, 2018
    Assignee: Conduent Business Services, LLC
    Inventors: Albert Gordo Soldevila, Jon Almazan
  • Patent number: 9792492
    Abstract: A method for extracting a representation from an image includes inputting an image to a pre-trained neural network. The gradient of a loss function is computed with respect to parameters of the neural network, for the image. A gradient representation is extracted for the image based on the computed gradients, which can be used, for example, for classification or retrieval.
    Type: Grant
    Filed: July 7, 2015
    Date of Patent: October 17, 2017
    Assignee: XEROX CORPORATION
    Inventors: Albert Gordo Soldevila, Adrien Gaidon, Florent C. Perronnin
  • Patent number: 9785855
    Abstract: Methods and systems for license plate recognition utilizing a trained neural network. In an example embodiment, a neural network can be subject to operations involving iteratively training and adapting the neural network for a particular task such as, for example, text recognition in the context of a license plate recognition application. The neural network can be trained to perform generic text recognition utilizing a plurality of training samples. The neural network can be applied to a cropped image of a license plate in order to recognize text and produce a license plate transcription with respect to the license plate. An example of such a neural network is a CNN (Convolutional Neural. Network).
    Type: Grant
    Filed: December 17, 2015
    Date of Patent: October 10, 2017
    Assignee: Conduent Business Services, LLC
    Inventors: Albert Gordo Soldevila, Jon Almazan
  • Patent number: 9767381
    Abstract: A system and method provide object localization in a query image based on a global representation of the image generated with a model derived from a convolutional neural network. Representations of annotated images and a query image are each generated based on activations output by a layer of the model which precedes the fully-connected layers of the neural network. A similarity is computed between the query image representation and each of the annotated image representations to identify a subset of the annotated images having the highest computed similarity. Object location information from at least one of the subset of annotated images is transferred to the query image and information is output, based on the transferred object location information.
    Type: Grant
    Filed: September 22, 2015
    Date of Patent: September 19, 2017
    Assignee: XEROX CORPORATION
    Inventors: José A. Rodríguez-Serrano, Albert Gordo Soldevila
  • Patent number: 9762393
    Abstract: Authentication methods are disclosed for determining whether a person or object to be authenticated is a member of a set of authorized persons or objects. A query signature is acquired comprising a vector whose elements store values of an ordered set of features for the person or object to be authenticated. The query signature is compared with an aggregate signature comprising a vector whose elements store values of the ordered set of features for the set of authorized persons or objects. The individual signatures for the authorized persons or objects are not stored; only the aggregate signature. It is determined whether the person or object to be authenticated is a member of the set of authorized persons or objects based on the comparison. The comparing may comprise computing an inner product of the query signature and the aggregate signature, with the determining being based on the inner product.
    Type: Grant
    Filed: March 19, 2015
    Date of Patent: September 12, 2017
    Assignee: Conduent Business Services, LLC
    Inventors: Albert Gordo Soldevila, Naila Murray, Florent C. Perronnin
  • Publication number: 20170177965
    Abstract: Methods and systems for license plate recognition utilizing a trained neural network. In an example embodiment, a neural network can be subject to operations involving iteratively training and adapting the neural network for a particular task such as, for example, text recognition in the context of a license plate recognition application. The neural network can be trained to perform generic text recognition utilizing a plurality of training samples. The neural network can be applied to a cropped image of a license plate in order to recognize text and produce a license plate transcription with respect to the license plate. An example of such a neural network is a CNN (Convolutional Neural Network).
    Type: Application
    Filed: December 17, 2015
    Publication date: June 22, 2017
    Inventors: Albert Gordo Soldevila, Jon Almazan
  • Patent number: 9626594
    Abstract: A system and method for comparing a text image with or without a wildcard character and a character string are provided. The method includes embedding a character string into a vectorial space by extracting a set of features from the character string and generating a character string representation based on the extracted features, such as a spatial pyramid bag of characters (SPBOC) representation. A text image is embedded into a vectorial space by extracting a set of features from the text image and generating a text image representation based on the text image extracted features. A similarity between the text image representation and the character string representation is computed, which includes computing a function of the text image representation and character string representation.
    Type: Grant
    Filed: January 21, 2015
    Date of Patent: April 18, 2017
    Assignee: XEROX CORPORATION
    Inventors: Albert Gordo Soldevila, José Antonio Rodríguez-Serrano, Florent Perronnin
  • Publication number: 20170083792
    Abstract: A system and method provide object localization in a query image based on a global representation of the image generated with a model derived from a convolutional neural network. Representations of annotated images and a query image are each generated based on activations output by a layer of the model which precedes the fully-connected layers of the neural network. A similarity is computed between the query image representation and each of the annotated image representations to identify a subset of the annotated images having the highest computed similarity. Object location information from at least one of the subset of annotated images is transferred to the query image and information is output, based on the transferred object location information.
    Type: Application
    Filed: September 22, 2015
    Publication date: March 23, 2017
    Applicant: Xerox Corporation
    Inventors: José A. Rodríguez-Serrano, Albert Gordo Soldevila
  • Publication number: 20170011279
    Abstract: A system and method enable semantic comparisons to be made between word images and concepts. Training word images and their concept labels are used to learn parameters of a neural network for embedding word images and concepts in a semantic subspace in which comparisons can be made between word images and concepts without the need for transcribing the text content of the word image. The training of the neural network aims to minimize a ranking loss over the training set where non relevant concepts for an image which are ranked more highly than relevant ones penalize the ranking loss.
    Type: Application
    Filed: July 7, 2015
    Publication date: January 12, 2017
    Applicant: Xerox Corporation
    Inventors: Albert Gordo Soldevila, Jon Almazán Almazán, Naila Murray, Florent C. Perronnin
  • Publication number: 20170011280
    Abstract: A method for extracting a representation from an image includes inputting an image to a pre-trained neural network. The gradient of a loss function is computed with respect to parameters of the neural network, for the image. A gradient representation is extracted for the image based on the computed gradients, which can be used, for example, for classification or retrieval.
    Type: Application
    Filed: July 7, 2015
    Publication date: January 12, 2017
    Applicant: Xerox Corporation
    Inventors: Albert Gordo Soldevila, Adrien Gaidon, Florent C. Perronnin
  • Publication number: 20170011273
    Abstract: Methods and systems recognize alphanumeric characters in an image by computing individual representations of every character of an alphabet at every character position within a certain word transcription length. These methods and systems embed the individual representations of each alphabet character in a common vectorial subspace (using a matrix) and embed a received image of an alphanumeric word into the common vectorial subspace (using the matrix). Such methods and systems compute the utility value of the embedded alphabet characters at every one of the character positions with respect to the embedded alphanumeric character image; and compute the best transcription alphabet character of every one of the image characters based on the utility value of each embedded alphabet character at each character position. Such methods and systems then assign the best transcription alphabet character for each of the character positions to produce a recognized alphanumeric word within the received image.
    Type: Application
    Filed: July 8, 2015
    Publication date: January 12, 2017
    Inventors: Albert Gordo Soldevila, Jon Almazan
  • Publication number: 20160277190
    Abstract: Authentication methods are disclosed for determining whether a person or object to be authenticated is a member of a set of authorized persons or objects. A query signature is acquired comprising a vector whose elements store values of an ordered set of features for the person or object to be authenticated. The query signature is compared with an aggregate signature comprising a vector whose elements store values of the ordered set of features for the set of authorized persons or objects. The individual signatures for the authorized persons or objects are not stored; only the aggregate signature. It is determined whether the person or object to be authenticated is a member of the set of authorized persons or objects based on the comparison. The comparing may comprise computing an inner product of the query signature and the aggregate signature, with the determining being based on the inner product.
    Type: Application
    Filed: March 19, 2015
    Publication date: September 22, 2016
    Inventors: Albert Gordo Soldevila, Naila Murray, Florent C. Perronnin
  • Patent number: 9443164
    Abstract: A system and method for object instance localization in an image are disclosed. In the method, keypoints are detected in a target image and candidate regions are detected by matching the detected keypoints to keypoints detected in a set of reference images. Similarity measures between global descriptors computed for the located candidate regions and global descriptors for the reference images are computed and labels are assigned to at least some of the candidate regions based on the computed similarity measures. Performing the region detection based on keypoint matching while performing the labeling based on global descriptors improves object instance detection.
    Type: Grant
    Filed: December 2, 2014
    Date of Patent: September 13, 2016
    Inventors: Milan Sulc, Albert Gordo Soldevila, Diane Larlus Larrondo, Florent C. Perronnin
  • Publication number: 20160210532
    Abstract: A system and method for comparing a text image with or without a wildcard character and a character string are provided. The method includes embedding a character string into a vectorial space by extracting a set of features from the character string and generating a character string representation based on the extracted features, such as a spatial pyramid bag of characters (SPBOC) representation. A text image is embedded into a vectorial space by extracting a set of features from the text image and generating a text image representation based on the text image extracted features. A similarity between the text image representation and the character string representation is computed, which includes computing a function of the text image representation and character string representation.
    Type: Application
    Filed: January 21, 2015
    Publication date: July 21, 2016
    Inventors: Albert Gordo Soldevila, José Antonio Rodríguez-Serrano, Florent Perronnin