Patents by Inventor Anelia Angelova
Anelia Angelova has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20240037926Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for performing instance segmentation by detecting and segmenting individual objects in an image. In one aspect, a method comprises: processing an image to generate data identifying a region of the image that depicts a particular object; obtaining data defining a plurality of example object segmentations; generating a respective weight value for each of the example object segmentations; for each of a plurality of pixels in the region of the image, determining a score characterizing a likelihood that the pixel is included in the particular object depicted in the region of the image using: (i) the example object segmentations, and (ii) the weight values for the example object segmentations; and generating a segmentation of the particular object depicted in the region of the image using the scores for the pixels in the region of the image.Type: ApplicationFiled: October 12, 2023Publication date: February 1, 2024Inventors: Weicheng Kuo, Anelia Angelova, Tsung-Yi Lin
-
Publication number: 20240029413Abstract: A method involves the training of a model by dynamically adjusting the number of examples within each training batch. The dynamic adjustment is accomplished by adjusting the number of examples per task within each training batch according to the performance of the model on the tasks that the model is being trained on. In some embodiments, this method is applied to cross-modal vision-language tasks. This model may also be applied to the pre-training of a model that can be later fine-tuned for a more specific task(s).Type: ApplicationFiled: July 12, 2023Publication date: January 25, 2024Inventors: Anthony Jacob Piergiovanni, Weiching Kuo, Wei Li, Anelia Angelova
-
Publication number: 20230419521Abstract: A system for generating a depth output for an image is described. The system receives input images that depict the same scene, each input image including one or more potential objects. The system generates, for each input image, a respective background image and processes the background images to generate a camera motion output that characterizes the motion of the camera between the input images. For each potential object, the system generates a respective object motion output for the potential object based on the input images and the camera motion output. The system processes a particular input image of the input images using a depth prediction neural network (NN) to generate a depth output for the particular input image, and updates the current values of parameters of the depth prediction NN based on the particular depth output, the camera motion output, and the object motion outputs for the potential objects.Type: ApplicationFiled: September 13, 2023Publication date: December 28, 2023Inventors: Vincent Michael Casser, Soeren Pirk, Reza Mahjourian, Anelia Angelova
-
Publication number: 20230409899Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for processing a network input using a computer vision neural network with learned tokenization.Type: ApplicationFiled: June 21, 2022Publication date: December 21, 2023Inventors: Michael Sahngwon Ryoo, Anthony Jacob Piergiovanni, Anelia Angelova, Anurag Arnab, Mostafa Dehghani
-
Publication number: 20230394306Abstract: Provided is an efficient multi-modal processing model. The multi-modal processing model can process input data from multiple different domains to generate a prediction for a multi-modal processing task. A machine-learned multi-modal processing model can include an adaptive tokenization layer that is configured to adaptively tokenize features generated from the multi-modal inputs into sets of tokens. Specifically, the tokens may have a smaller data size relative to the features from the inputs, thereby enabling a reduced number of processing operations to be performed overall, thereby improving the efficiency of model.Type: ApplicationFiled: June 2, 2023Publication date: December 7, 2023Inventors: Anthony J. Piergiovanni, Wei-Cheng Kuo, Anelia Angelova
-
Patent number: 11823443Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for performing instance segmentation by detecting and segmenting individual objects in an image. In one aspect, a method comprises: processing an image to generate data identifying a region of the image that depicts a particular object; obtaining data defining a plurality of example object segmentations; generating a respective weight value for each of the example object segmentations; for each of a plurality of pixels in the region of the image, determining a score characterizing a likelihood that the pixel is included in the particular object depicted in the region of the image using: (i) the example object segmentations, and (ii) the weight values for the example object segmentations; and generating a segmentation of the particular object depicted in the region of the image using the scores for the pixels in the region of the image.Type: GrantFiled: August 14, 2019Date of Patent: November 21, 2023Assignee: Google LLCInventors: Weicheng Kuo, Anelia Angelova, Tsung-Yi Lin
-
Patent number: 11790549Abstract: A system includes a neural network implemented by one or more computers, in which the neural network includes an image depth prediction neural network and a camera motion estimation neural network. The neural network is configured to receive a sequence of images. The neural network is configured to process each image in the sequence of images using the image depth prediction neural network to generate, for each image, a respective depth output that characterizes a depth of the image, and to process a subset of images in the sequence of images using the camera motion estimation neural network to generate a camera motion output that characterizes the motion of a camera between the images in the subset. The image depth prediction neural network and the camera motion estimation neural network have been jointly trained using an unsupervised learning technique.Type: GrantFiled: May 27, 2022Date of Patent: October 17, 2023Assignee: Google LLCInventors: Reza Mahjourian, Martin Wicke, Anelia Angelova
-
Patent number: 11783500Abstract: A system for generating a depth output for an image is described. The system receives input images that depict the same scene, each input image including one or more potential objects. The system generates, for each input image, a respective background image and processes the background images to generate a camera motion output that characterizes the motion of the camera between the input images. For each potential object, the system generates a respective object motion output for the potential object based on the input images and the camera motion output. The system processes a particular input image of the input images using a depth prediction neural network (NN) to generate a depth output for the particular input image, and updates the current values of parameters of the depth prediction NN based on the particular depth output, the camera motion output, and the object motion outputs for the potential objects.Type: GrantFiled: September 5, 2019Date of Patent: October 10, 2023Assignee: Google LLCInventors: Vincent Michael Casser, Soeren Pirk, Reza Mahjourian, Anelia Angelova
-
Patent number: 11769269Abstract: A method includes receiving a first depth map that includes a plurality of first pixel depths and a second depth map that includes a plurality of second pixel depths. The first depth map corresponds to a reference depth scale and the second depth map corresponds to a relative depth scale. The method includes aligning the second pixel depths with the first pixel depths. The method includes transforming the aligned region of the second pixel depths such that transformed second edge pixel depths of the aligned region are coextensive with first edge pixel depths surrounding the corresponding region of the first pixel depths. The method includes generating a third depth map. The third depth map includes a first region corresponding to the first pixel depths and a second region corresponding to the transformed and aligned region of the second pixel depths.Type: GrantFiled: August 1, 2022Date of Patent: September 26, 2023Assignee: Google LLCInventors: Guy Satat, Michael Quinlan, Sean Kirmani, Anelia Angelova, Ariel Gordon
-
Patent number: 11734847Abstract: A system includes an image depth prediction neural network implemented by one or more computers. The image depth prediction neural network is a recurrent neural network that is configured to receive a sequence of images and, for each image in the sequence: process the image in accordance with a current internal state of the recurrent neural network to (i) update the current internal state and (ii) generate a depth output that characterizes a predicted depth of a future image in the sequence.Type: GrantFiled: January 15, 2021Date of Patent: August 22, 2023Assignee: Google LLCInventors: Anelia Angelova, Martin Wicke, Reza Mahjourian
-
Publication number: 20230114556Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for processing a network input using a neural network to generate a network output.Type: ApplicationFiled: July 14, 2021Publication date: April 13, 2023Inventors: Michael Sahngwon Ryoo, Anthony Jacob Piergiovanni, Anelia Angelova
-
Publication number: 20230035454Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for generating an optical flow label from a lidar point cloud. One of the methods includes obtaining data specifying a training example, including a first image of a scene in an environment captured at a first time point and a second image of the scene in the environment captured at a second time point. For each of a plurality of lidar points, a respective second corresponding pixel in the second image is obtained and a respective velocity estimate for the lidar point at the second time point is obtained. A respective first corresponding pixel in the first image is determined using the velocity estimate for the lidar point. A proxy optical flow ground truth for the training example is generated based on an estimate of optical flow of the pixel between the first and second images.Type: ApplicationFiled: July 23, 2021Publication date: February 2, 2023Inventors: Daniel Rudolf Maurer, Alper Ayvaci, Robert William Anderson, Rico Jonschkowski, Austin Charles Stone, Anelia Angelova, Nichola Abdo, Christopher John Sweeney
-
Patent number: 11544498Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for training a neural network using consistency measures. One of the methods includes processing a particular training example from a mediator training data set using a first neural network to generate a first output for a first machine learning task; processing the particular training example in the mediator training data set using each of one or more second neural networks, wherein each second neural network is configured to generate a second output for a respective second machine learning task; determining, for each second machine learning task, a consistency target output for the first machine learning task; determining, for each second machine learning task, an error between the first output and the consistency target output corresponding to the second machine learning task; and generating a parameter update for the first neural network from the determined errors.Type: GrantFiled: March 5, 2021Date of Patent: January 3, 2023Assignee: Google LLCInventors: Ariel Gordon, Soeren Pirk, Anelia Angelova, Vincent Michael Casser, Yao Lu, Anthony Brohan, Zhao Chen, Jan Dlabal
-
Publication number: 20220366590Abstract: A method includes receiving a first depth map that includes a plurality of first pixel depths and a second depth map that includes a plurality of second pixel depths. The first depth map corresponds to a reference depth scale and the second depth map corresponds to a relative depth scale. The method includes aligning the second pixel depths with the first pixel depths. The method includes transforming the aligned region of the second pixel depths such that transformed second edge pixel depths of the aligned region are coextensive with first edge pixel depths surrounding the corresponding region of the first pixel depths. The method includes generating a third depth map. The third depth map includes a first region corresponding to the first pixel depths and a second region corresponding to the transformed and aligned region of the second pixel depths.Type: ApplicationFiled: August 1, 2022Publication date: November 17, 2022Inventors: Guy Satat, Michael Quinlan, Sean Kirmani, Anelia Angelova, Ariel Gordon
-
Publication number: 20220366257Abstract: Generally, the present disclosure is directed to a neural architecture search process for finding small and fast video processing networks for understanding of video data. The neural architecture search process can automatically design networks that provide comparable video processing performance at a fraction of the computational and storage cost of larger existing models, thereby conserving computing resources such as memory and processor usage.Type: ApplicationFiled: September 16, 2020Publication date: November 17, 2022Inventors: Anthony J. Piergiovanni, Anelia Angelova, Michael Sahngwon Ryoo
-
Publication number: 20220335624Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for training a neural network to predict optical flow. One of the methods includes obtaining a batch of one or more training image pairs; for each of the pairs: processing the first training image and the second training image using the neural network to generate a final optical flow estimate; generating a cropped final optical flow estimate from the final optical flow estimate; and training the neural network using the cropped optical flow estimate.Type: ApplicationFiled: April 14, 2022Publication date: October 20, 2022Inventors: Daniel Rudolf Maurer, Austin Charles Stone, Alper Ayvaci, Anelia Angelova, Rico Jonschkowski
-
Publication number: 20220305647Abstract: Techniques are disclosed that enable the generation of predicted sequences of terminals using a generator model portion of a prediction model. Various implementations include controlling actuators of a robot based on the predicted sequences of terminals. Additional or alternative implementations include jointly training the generator model portion of the prediction model using a discriminator model portion of the prediction model using, for example, stochastic adversarial based sampling.Type: ApplicationFiled: August 27, 2019Publication date: September 29, 2022Inventors: Anthony Jacob Piergiovanni, Anelia Angelova, Alexander Toshev, Michael Ryoo
-
Patent number: 11450018Abstract: A method includes receiving a first depth map that includes a plurality of first pixel depths and a second depth map that includes a plurality of second pixel depths. The first depth map corresponds to a reference depth scale and the second depth map corresponds to a relative depth scale. The method includes aligning the second pixel depths with the first pixel depths. The method includes transforming the aligned region of the second pixel depths such that transformed second edge pixel depths of the aligned region are coextensive with first edge pixel depths surrounding the corresponding region of the first pixel depths. The method includes generating a third depth map. The third depth map includes a first region corresponding to the first pixel depths and a second region corresponding to the transformed and aligned region of the second pixel depths.Type: GrantFiled: December 24, 2019Date of Patent: September 20, 2022Assignee: X Development LLCInventors: Guy Satat, Michael Quinlan, Sean Kirmani, Anelia Angelova, Ariel Gordon
-
Publication number: 20220292701Abstract: A system includes a neural network implemented by one or more computers, in which the neural network includes an image depth prediction neural network and a camera motion estimation neural network. The neural network is configured to receive a sequence of images. The neural network is configured to process each image in the sequence of images using the image depth prediction neural network to generate, for each image, a respective depth output that characterizes a depth of the image, and to process a subset of images in the sequence of images using the camera motion estimation neural network to generate a camera motion output that characterizes the motion of a camera between the images in the subset. The image depth prediction neural network and the camera motion estimation neural network have been jointly trained using an unsupervised learning technique.Type: ApplicationFiled: May 27, 2022Publication date: September 15, 2022Inventors: Reza Mahjourian, Martin Wicke, Anelia Angelova
-
Publication number: 20220189154Abstract: Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for determining one or more neural network architectures of a neural network for performing a video processing neural network task. In one aspect, a method comprises: at each of a plurality of iterations: selecting a parent neural network architecture from a set of neural network architectures; training a neural network having the parent neural network architecture to perform the video processing neural network task, comprising determining trained values of connection weight parameters of the parent neural network architecture; generating a new neural network architecture based at least in part on the trained values of the connection weight parameters of the parent neural network architecture; and adding the new neural network architecture to the set of neural network architectures.Type: ApplicationFiled: May 22, 2020Publication date: June 16, 2022Inventors: Michael Sahngwon Ryoo, Anthony Jacob Piergiovanni, Mingxing Tan, Anelia Angelova