Patents by Inventor Andrei Vakunov

Andrei Vakunov has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20230410329
    Abstract: Example aspects of the present disclosure are directed to computing systems and methods for hand tracking using a machine-learned system for palm detection and key-point localization of hand landmarks. In particular, example aspects of the present disclosure are directed to a multi-model hand tracking system that performs both palm detection and hand landmark detection. Given a sequence of image frames, for example, the hand tracking system can detect one or more palms depicted in each image frame. For each palm detected within an image frame, the machine-learned system can determine a plurality of hand landmark positions of a hand associated with the palm. The system can perform key-point localization to determine precise three-dimensional coordinates for the hand landmark positions. In this manner, the machine-learned system can accurately track a hand depicted in the sequence of images using the precise three-dimensional coordinates for the hand landmark positions.
    Type: Application
    Filed: September 1, 2023
    Publication date: December 21, 2023
    Inventors: Valentin Bazarevsky, Fan Zhang, Andrei Tkachenka, Andrei Vakunov, Matthias Grundmann
  • Patent number: 11783496
    Abstract: Example aspects of the present disclosure are directed to computing systems and methods for hand tracking using a machine-learned system for palm detection and key-point localization of hand landmarks. In particular, example aspects of the present disclosure are directed to a multi-model hand tracking system that performs both palm detection and hand landmark detection. Given a sequence of image frames, for example, the hand tracking system can detect one or more palms depicted in each image frame. For each palm detected within an image frame, the machine-learned system can determine a plurality of hand landmark positions of a hand associated with the palm. The system can perform key-point localization to determine precise three-dimensional coordinates for the hand landmark positions. In this manner, the machine-learned system can accurately track a hand depicted in the sequence of images using the precise three-dimensional coordinates for the hand landmark positions.
    Type: Grant
    Filed: November 16, 2021
    Date of Patent: October 10, 2023
    Assignee: GOOGLE LLC
    Inventors: Valentin Bazarevsky, Fan Zhang, Andrei Vakunov, Andrei Tkachenka, Matthias Grundmann
  • Patent number: 11694087
    Abstract: A computing system is disclosed including a convolutional neural configured to receive an input that describes a facial image and generate a facial object recognition output that describes one or more facial feature locations with respect to the facial image. The convolutional neural network can include a plurality of convolutional blocks. At least one of the convolutional blocks can include one or more separable convolutional layers configured to apply a depthwise convolution and a pointwise convolution during processing of an input to generate an output. The depthwise convolution can be applied with a kernel size that is greater than 3×3. At least one of the convolutional blocks can include a residual shortcut connection from its input to its output.
    Type: Grant
    Filed: September 19, 2022
    Date of Patent: July 4, 2023
    Assignee: GOOGLE LLC
    Inventors: Valentin Bazarevsky, Yury Kartynnik, Andrei Vakunov, Karthik Raveendran, Matthias Grundmann
  • Publication number: 20230017459
    Abstract: A computing system is disclosed including a convolutional neural configured to receive an input that describes a facial image and generate a facial object recognition output that describes one or more facial feature locations with respect to the facial image. The convolutional neural network can include a plurality of convolutional blocks. At least one of the convolutional blocks can include one or more separable convolutional layers configured to apply a depthwise convolution and a pointwise convolution during processing of an input to generate an output. The depthwise convolution can be applied with a kernel size that is greater than 3×3. At least one of the convolutional blocks can include a residual shortcut connection from its input to its output.
    Type: Application
    Filed: September 19, 2022
    Publication date: January 19, 2023
    Inventors: Valentin Bazarevsky, Yury Kartynnik, Andrei Vakunov, Karthik Raveendran, Matthias Grundmann
  • Patent number: 11449714
    Abstract: A computing system is disclosed including a convolutional neural configured to receive an input that describes a facial image and generate a facial object recognition output that describes one or more facial feature locations with respect to the facial image. The convolutional neural network can include a plurality of convolutional blocks. At least one of the convolutional blocks can include one or more separable convolutional layers configured to apply a depthwise convolution and a pointwise convolution during processing of an input to generate an output. The depthwise convolution can be applied with a kernel size that is greater than 3×3. At least one of the convolutional blocks can include a residual shortcut connection from its input to its output.
    Type: Grant
    Filed: October 30, 2019
    Date of Patent: September 20, 2022
    Assignee: GOOGLE LLC
    Inventors: Valentin Bazarevsky, Yury Kartynnik, Andrei Vakunov, Karthik Raveendran, Matthias Grundmann
  • Publication number: 20220076433
    Abstract: Example aspects of the present disclosure are directed to computing systems and methods for hand tracking using a machine-learned system for palm detection and key-point localization of hand landmarks. In particular, example aspects of the present disclosure are directed to a multi-model hand tracking system that performs both palm detection and hand landmark detection. Given a sequence of image frames, for example, the hand tracking system can detect one or more palms depicted in each image frame. For each palm detected within an image frame, the machine-learned system can determine a plurality of hand landmark positions of a hand associated with the palm. The system can perform key-point localization to determine precise three-dimensional coordinates for the hand landmark positions. In this manner, the machine-learned system can accurately track a hand depicted in the sequence of images using the precise three-dimensional coordinates for the hand landmark positions.
    Type: Application
    Filed: November 16, 2021
    Publication date: March 10, 2022
    Inventors: Valentin Bazarevsky, Fan Zhang, Andrei Vakunov, Andrei Tkachenka, Matthias Grundmann
  • Patent number: 11182909
    Abstract: Example aspects of the present disclosure are directed to computing systems and methods for hand tracking using a machine-learned system for palm detection and key-point localization of hand landmarks. In particular, example aspects of the present disclosure are directed to a multi-model hand tracking system that performs both palm detection and hand landmark detection. Given a sequence of image frames, for example, the hand tracking system can detect one or more palms depicted in each image frame. For each palm detected within an image frame, the machine-learned system can determine a plurality of hand landmark positions of a hand associated with the palm. The system can perform key-point localization to determine precise three-dimensional coordinates for the hand landmark positions. In this manner, the machine-learned system can accurately track a hand depicted in the sequence of images using the precise three-dimensional coordinates for the hand landmark positions.
    Type: Grant
    Filed: December 10, 2019
    Date of Patent: November 23, 2021
    Assignee: Google LLC
    Inventors: Valentin Bazarevsky, Fan Zhang, Andrei Vakunov, Andrei Tkachenka, Matthias Grundmann
  • Publication number: 20210174519
    Abstract: Example aspects of the present disclosure are directed to computing systems and methods for hand tracking using a machine-learned system for palm detection and key-point localization of hand landmarks. In particular, example aspects of the present disclosure are directed to a multi-model hand tracking system that performs both palm detection and hand landmark detection. Given a sequence of image frames, for example, the hand tracking system can detect one or more palms depicted in each image frame. For each palm detected within an image frame, the machine-learned system can determine a plurality of hand landmark positions of a hand associated with the palm. The system can perform key-point localization to determine precise three-dimensional coordinates for the hand landmark positions. In this manner, the machine-learned system can accurately track a hand depicted in the sequence of images using the precise three-dimensional coordinates for the hand landmark positions.
    Type: Application
    Filed: December 10, 2019
    Publication date: June 10, 2021
    Inventors: Valentin Bazarevsky, Fan Zhang, Andrei Vakunov, Andrei Tkachenka, Matthias Grundmann
  • Publication number: 20210133508
    Abstract: A computing system is disclosed including a convolutional neural configured to receive an input that describes a facial image and generate a facial object recognition output that describes one or more facial feature locations with respect to the facial image. The convolutional neural network can include a plurality of convolutional blocks. At least one of the convolutional blocks can include one or more separable convolutional layers configured to apply a depthwise convolution and a pointwise convolution during processing of an input to generate an output. The depthwise convolution can be applied with a kernel size that is greater than 3×3. At least one of the convolutional blocks can include a residual shortcut connection from its input to its output.
    Type: Application
    Filed: October 30, 2019
    Publication date: May 6, 2021
    Inventors: Valentin Bazarevsky, Yury Kartynnik, Andrei Vakunov, Karthik Raveendran, Matthias Grundmann