Patents by Inventor Ronan Stéfan Collobert
Ronan Stéfan Collobert has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 11551668Abstract: In one embodiment, a method includes generating audio segments from a speech signal, generating latent representations that respectively correspond to the audio segments, the latent representations comprising a first subset and a second subset, generating quantized representations that respectively correspond to the latent representations, masking the second subset of the latent representations, using a machine-learning model to process the first subset of the latent representations and the masked second subset of the latent representations to generate contextualized representations that respectively correspond to the latent representations, pre-training the machine-learning model based on comparisons between (1) a subset of the contextualized representations that respectively correspond to the masked second subset of the latent representations and (2) a subset of the quantized representations that respectively correspond to the masked second subset of the latent representations, and training the pre-trainedType: GrantFiled: December 30, 2020Date of Patent: January 10, 2023Assignee: Meta Platforms, Inc.Inventors: Alexei Baevski, Yuhao Zhou, Abdelrahman Mohamed, Michael Auli, Ronan Stéfan Collobert, Alexis Conneau
-
Patent number: 11023772Abstract: In one embodiment, a feature map of an image having h×w pixels and a patch having one or more pixels of the image are received. The patch has been processed by a first set of layers of a convolutional neural network and contains an object centered within the patch. The patch is then processed using the feature map and one or more pixel classifiers of a classification layer of a deep-learning model, where the classification layer includes h×w pixel classifiers, with each pixel classifier corresponding to a respective pixel of the patch. Each of the pixel classifiers used to process the patch outputs a respective value indicating whether the corresponding pixel belongs to the object centered in the patch.Type: GrantFiled: October 1, 2019Date of Patent: June 1, 2021Assignee: Facebook, Inc.Inventors: Pedro Henrique Oliveira Pinheiro, Ronan Stéfan Collobert, Piotr Dollar
-
Publication number: 20200034653Abstract: In one embodiment, a feature map of an image having h×w pixels and a patch having one or more pixels of the image are received. The patch has been processed by a first set of layers of a convolutional neural network and contains an object centered within the patch. The patch is then processed using the feature map and one or more pixel classifiers of a classification layer of a deep-learning model, where the classification layer includes h×w pixel classifiers, with each pixel classifier corresponding to a respective pixel of the patch. Each of the pixel classifiers used to process the patch outputs a respective value indicating whether the corresponding pixel belongs to the object centered in the patch.Type: ApplicationFiled: October 1, 2019Publication date: January 30, 2020Inventors: Pedro Henrique Oliveira Pinheiro, Ronan Stéfan Collobert, Piotr Dollar
-
Patent number: 10496895Abstract: In one embodiment a plurality of patches of an image are processed, using a first set of layers of a convolutional neural network, to output a plurality of object proposals associated with the plurality of patches of the image. Each patch includes one or more pixels of the image. Each object proposal includes a prediction as to a location of an object in the respective patch. Using a second set of layers of the convolutional neural network, the plurality of object proposals outputted by the first set of layers are processed to generate a plurality of refined object proposals. Each refined object proposal includes pixel-level information for the respective patch of the image. The first layer in the second set of layers of the convolutional neural network takes as input the plurality of object proposals outputted by the first set of layers.Type: GrantFiled: December 22, 2017Date of Patent: December 3, 2019Assignee: Facebook, Inc.Inventors: Pedro Henrique Oliveira Pinheiro, Ronan Stéfan Collobert, Piotr Dollar
-
Patent number: 10496896Abstract: In one embodiment, a plurality of patches of an image are processed using a first-pass of a first deep-learning model to generate object-level information for each of the patches. Each patch includes one or more pixels of the image. Using a second-pass of the first deep-learning model, a respective object proposal is generated for each of the plurality of patches of the image. The second-pass takes as input the first-pass output, and the generated respective object proposals comprise pixel-level information for each of the patches. Using a second deep-learning model, a respective score is computed for each object proposal. The second deep-learning model takes as input the first-pass output, and the object score includes a likelihood that the respective patch of the object proposal contains an entire object.Type: GrantFiled: March 29, 2019Date of Patent: December 3, 2019Assignee: Facebook, Inc.Inventors: Pedro Henrique Oliveira Pinheiro, Ronan Stéfan Collobert, Piotr Dollar
-
Publication number: 20190228259Abstract: In one embodiment, a plurality of patches of an image are processed using a first-pass of a first deep-learning model to generate object-level information for each of the patches. Each patch includes one or more pixels of the image. Using a second-pass of the first deep-learning model, a respective object proposal is generated for each of the plurality of patches of the image. The second-pass takes as input the first-pass output, and the generated respective object proposals comprise pixel-level information for each of the patches. Using a second deep-learning model, a respective score is computed for each object proposal. The second deep-learning model takes as input the first-pass output, and the object score includes a likelihood that the respective patch of the object proposal contains an entire object.Type: ApplicationFiled: March 29, 2019Publication date: July 25, 2019Inventors: Pedro Henrique Oliveira Pinheiro, Ronan Stéfan Collobert, Piotr Dollar
-
Patent number: 10255522Abstract: In one embodiment, a plurality of patches of an image are processed using a first deep-learning model to detect a plurality of features associated with the first patch of the image. Each patch includes one or more pixels of the image. Using a second deep-learning model, a respective object proposal is generated for each of the plurality of patches of the image. The second deep-learning model takes as input the plurality of detected features associated with the respective patch of the image, and each object proposal includes a prediction as to a location of an object in the patch. Using a third deep-learning model, a respective score is computed for each object proposal generated using the second deep-learning model. The third deep-learning model takes as input the plurality of detected features associated with the respective patch of the image, and the object score may include a likelihood that the patch contains an entire object.Type: GrantFiled: June 15, 2017Date of Patent: April 9, 2019Assignee: Facebook, Inc.Inventors: Pedro Henrique Oliveira Pinheiro, Ronan Stéfan Collobert, Piotr Dollar
-
Publication number: 20180285686Abstract: In one embodiment a plurality of patches of an image are processed, using a first set of layers of a convolutional neural network, to output a plurality of object proposals associated with the plurality of patches of the image. Each patch includes one or more pixels of the image. Each object proposal includes a prediction as to a location of an object in the respective patch. Using a second set of layers of the convolutional neural network, the plurality of object proposals outputted by the first set of layers are processed to generate a plurality of refined object proposals. Each refined object proposal includes pixel-level information for the respective patch of the image. The first layer in the second set of layers of the convolutional neural network takes as input the plurality of object proposals outputted by the first set of layers.Type: ApplicationFiled: December 22, 2017Publication date: October 4, 2018Inventors: Pedro Henrique Oliveira Pinheiro, Ronan Stéfan Collobert, Piotr Dollar
-
Publication number: 20170364771Abstract: In one embodiment, a plurality of patches of an image are processed using a first deep-learning model to detect a plurality of features associated with the first patch of the image. Each patch includes one or more pixels of the image. Using a second deep-learning model, a respective object proposal is generated for each of the plurality of patches of the image. The second deep-learning model takes as input the plurality of detected features associated with the respective patch of the image, and each object proposal includes a prediction as to a location of an object in the patch. Using a third deep-learning model, a respective score is computed for each object proposal generated using the second deep-learning model. The third deep-learning model takes as input the plurality of detected features associated with the respective patch of the image, and the object score may include a likelihood that the patch contains an entire object.Type: ApplicationFiled: June 15, 2017Publication date: December 21, 2017Inventors: Pedro Henrique Oliveira Pinheiro, Ronan Stéfan Collobert, Piotr Dollar
-
Publication number: 20170046613Abstract: Systems, methods, and non-transitory computer-readable media can obtain a content item to be evaluated by a set of cascaded convolutional neural networks, the set of cascaded convolutional neural networks including at least a first convolutional neural network (CNN) and a second CNN. The content item can be provided to the first CNN as input, wherein an output of the first CNN includes data describing at least one region of interest in the content item and at least one first concept corresponding to the region of interest. The output of the first CNN can be provided to the second CNN as input, wherein an output of the second CNN includes data describing at least one second concept corresponding to the region of interest, the second concept being more accurate than the first concept.Type: ApplicationFiled: April 5, 2016Publication date: February 16, 2017Inventors: Balamanohar Paluri, Lubomir Bourdev, Ronan Stéfan Collobert, Chen Sun