Patents by Inventor Maxime Meilland

Maxime Meilland has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11928779
    Abstract: Various implementations disclosed herein generate a mesh representing the surfaces in a physical environment. The mesh is generated using multi-resolution voxels based on detected depth information, e.g., from a depth camera. The techniques may use multiple hash tables to store the multi-resolution voxel data. For example, the hash tables may store each voxel's 3D position and a truncated signed distance field (TSDF) value corresponding to each voxels' distance to a nearest surface. Each of the multiple hash tables may include data corresponding to a different level of resolution and those resolutions may depend upon distance/noise or other factors. For example, voxels close to a depth camera may have a finer resolution and smaller size compared to voxels that are further from the depth camera. Techniques disclosed herein may involve using a meshing algorithm that combines multi-resolution voxel information stored in multiple hash tables to generate a single mesh.
    Type: Grant
    Filed: April 11, 2022
    Date of Patent: March 12, 2024
    Assignee: Apple Inc.
    Inventors: Maxime Meilland, Andrew Predoehl, Kyle L. Simek, Ming Chuang, Pedro A. Pinies Rodriguez
  • Publication number: 20240078692
    Abstract: In one implementation, a method of performing perspective correction is performed by a device including an image sensor, a display, one or more processors, and non-transitory memory. The method includes capturing, using the image sensor, an image of a physical environment. The method includes obtaining a depth map including a plurality of depths respectively associated with a plurality of pixels of the image of the physical environment, wherein the depth map includes, for a particular pixel at a particular pixel location representing a dynamic object in the physical environment, a particular depth corresponding to a distance between the image sensor and a static object in the physical environment behind the dynamic object. The method includes transforming, using the one or more processors, the image of the physical environment based on the depth map. The method includes displaying, on the display, the transformed image.
    Type: Application
    Filed: September 5, 2023
    Publication date: March 7, 2024
    Inventors: Emmanuel Piuze-Phaneuf, Maxime Meilland
  • Publication number: 20240005536
    Abstract: In one implementation, a method of determining a display location is performed by a device including one or more processors and non-transitory memory. The method includes obtaining a camera set of two-dimensional coordinates of a user input object in a physical environment. The method includes obtaining depth information of the physical environment excluding the user input object. The method includes transforming the camera set of two-dimensional coordinates into a display set of two-dimensional coordinates based on the depth information of the physical environment excluding the user input object.
    Type: Application
    Filed: June 21, 2023
    Publication date: January 4, 2024
    Inventors: Maxime Meilland, Emmanuel Piuze-Phaneuf
  • Publication number: 20230215079
    Abstract: In one implementation, a method includes: obtaining a representation for a volumetric region and obtaining SR content with a first set of dimensions; adapting the SR content by modifying one or more dimensions of the SR content from the first set of dimensions to a second set of dimensions based on one or more portions of the representation of the volumetric region; and causing presentation of the adapted SR content with the second set of dimensions via the display device.
    Type: Application
    Filed: July 14, 2022
    Publication date: July 6, 2023
    Inventors: Ian M. Richter, Maxime Meilland, Patrick W. O'Keefe
  • Patent number: 11423604
    Abstract: In one implementation, a method includes: obtaining locality data characterizing objects and relative spatial information of a volumetric region around a user; synthesizing a mesh map of the volumetric region based on the locality data; selecting synthesized reality (SR) content based on the mesh map, wherein the SR content satisfies a dimensional variance threshold relative to one or more portions of the mesh map; compositing at least a portion of the SR content with the mesh map in order to generate composite SR content; and presenting the composite SR content to the user in order to occlude at least a portion of a visual presentation of the volumetric region. In some implementations, the SR content is adapted to fit the one or more portions of the mesh map. In some implementations, the SR content is updated as the user location changes or the user interacts with the SR content.
    Type: Grant
    Filed: August 9, 2021
    Date of Patent: August 23, 2022
    Assignee: APPLE INC.
    Inventors: Ian M. Richter, Maxime Meilland, Patrick W. O'Keefe
  • Patent number: 11328481
    Abstract: Various implementations disclosed herein generate a mesh representing the surfaces in a physical environment. The mesh is generated using multi-resolution voxels based on detected depth information, e.g., from a depth camera. The techniques may use multiple hash tables to store the multi-resolution voxel data. For example, the hash tables may store each voxel's 3D position and a truncated signed distance field (TSDF) value corresponding to each voxels' distance to a nearest surface. Each of the multiple hash tables may include data corresponding to a different level of resolution and those resolutions may depend upon distance/noise or other factors. For example, voxels close to a depth camera may have a finer resolution and smaller size compared to voxels that are further from the depth camera. Techniques disclosed herein may involve using a meshing algorithm that combines multi-resolution voxel information stored in multiple hash tables to generate a single mesh.
    Type: Grant
    Filed: January 13, 2021
    Date of Patent: May 10, 2022
    Assignee: Apple Inc.
    Inventors: Maxime Meilland, Andrew Predoehl, Kyle L. Simek, Ming Chuang, Pedro A. Pinies Rodriguez
  • Publication number: 20210366179
    Abstract: In one implementation, a method includes: obtaining locality data characterizing objects and relative spatial information of a volumetric region around a user; synthesizing a mesh map of the volumetric region based on the locality data; selecting synthesized reality (SR) content based on the mesh map, wherein the SR content satisfies a dimensional variance threshold relative to one or more portions of the mesh map; compositing at least a portion of the SR content with the mesh map in order to generate composite SR content; and presenting the composite SR content to the user in order to occlude at least a portion of a visual presentation of the volumetric region. In some implementations, the SR content is adapted to fit the one or more portions of the mesh map. In some implementations, the SR content is updated as the user location changes or the user interacts with the SR content.
    Type: Application
    Filed: August 9, 2021
    Publication date: November 25, 2021
    Inventors: Ian M. Richter, Maxime Meilland, Patrick W. O'Keefe
  • Patent number: 11120612
    Abstract: In one implementation, a method includes: obtaining locality data characterizing objects and relative spatial information of a volumetric region around a user; synthesizing a mesh map of the volumetric region based on the locality data; selecting synthesized reality (SR) content based on the mesh map, wherein the SR content satisfies a dimensional variance threshold relative to one or more portions of the mesh map; compositing at least a portion of the SR content with the mesh map in order to generate composite SR content; and presenting the composite SR content to the user in order to occlude at least a portion of a visual presentation of the volumetric region. In some implementations, the SR content is adapted to fit the one or more portions of the mesh map. In some implementations, the SR content is updated as the user location changes or the user interacts with the SR content.
    Type: Grant
    Filed: January 18, 2019
    Date of Patent: September 14, 2021
    Assignee: Apple Inc.
    Inventors: Ian M. Richter, Maxime Meilland, Patrick W. O'Keefe
  • Patent number: 11082633
    Abstract: A method for estimating the speed of movement of a first video camera when it captures a current image of a three-dimensional scene, includes storing a reference image corresponding to an image of the same scene captured by a second video camera in a different pose, the reference image including pixels. The method also includes storing the current image, the current image including pixels containing the measurement of a physical quantity measured by that pixel, that physical quantity being the same as the physical quantity measured by the pixels of the reference image. The method further includes storing for each pixel of the reference image or of the current image the measurement of a depth that separates that pixel from the point of the scene photographed by that pixel, estimating the pose of the first camera, and estimating the speed of movement of the first camera.
    Type: Grant
    Filed: February 6, 2019
    Date of Patent: August 3, 2021
    Assignee: PIXMAP
    Inventors: Andrew Comport, Maxime Meilland
  • Publication number: 20210225074
    Abstract: Various implementations disclosed herein generate a mesh representing the surfaces in a physical environment. The mesh is generated using multi-resolution voxels based on detected depth information, e.g., from a depth camera. The techniques may use multiple hash tables to store the multi-resolution voxel data. For example, the hash tables may store each voxel's 3D position and a truncated signed distance field (TSDF) value corresponding to each voxels' distance to a nearest surface. Each of the multiple hash tables may include data corresponding to a different level of resolution and those resolutions may depend upon distance/noise or other factors. For example, voxels close to a depth camera may have a finer resolution and smaller size compared to voxels that are further from the depth camera. Techniques disclosed herein may involve using a meshing algorithm that combines multi-resolution voxel information stored in multiple hash tables to generate a single mesh.
    Type: Application
    Filed: January 13, 2021
    Publication date: July 22, 2021
    Inventors: Maxime Meilland, Andrew Predoehl, Kyle L. Simek, Ming Chuang, Pedro A. Pinies Rodriguez
  • Patent number: 10984583
    Abstract: Various implementations are disclosed of producing a 3-dimensional model of a scene. Various method, electronic device, or system implementations use RGB-D camera to provide RGB-D video content or periodic aligned RGB images and depth images to localize camera spatial position(s) defined in a three dimensional (3D) coordinate system or reconstruct a 3D virtual representation of a current camera frame in the 3D coordinate system, each in real time.
    Type: Grant
    Filed: February 28, 2019
    Date of Patent: April 20, 2021
    Assignee: Apple Inc.
    Inventor: Maxime Meilland
  • Publication number: 20210056749
    Abstract: In one implementation, a method includes: obtaining locality data characterizing objects and relative spatial information of a volumetric region around a user; synthesizing a mesh map of the volumetric region based on the locality data; selecting synthesized reality (SR) content based on the mesh map, wherein the SR content satisfies a dimensional variance threshold relative to one or more portions of the mesh map; compositing at least a portion of the SR content with the mesh map in order to generate composite SR content; and presenting the composite SR content to the user in order to occlude at least a portion of a visual presentation of the volumetric region. In some implementations, the SR content is adapted to fit the one or more portions of the mesh map. In some implementations, the SR content is updated as the user location changes or the user interacts with the SR content.
    Type: Application
    Filed: January 18, 2019
    Publication date: February 25, 2021
    Inventors: Ian M. Richter, Maxime Meilland, Patrick W. O'Keefe
  • Patent number: 10636151
    Abstract: A method for estimating the speed a first video camera when it captures a current image of a three-dimensional scene, the current image including pixels. The method includes storing a reference image corresponding to an image of the same scene captured by a second video camera in a different pose, the reference image including pixels. The method also includes storing the current image, containing for each pixel of the current image the measurement of a physical magnitude measured by that pixel, which is the same as the physical magnitude measured by the pixels of the reference image. The method further includes storing for each pixel of the reference image or of the current image the measurement of a depth that separates that pixel from the point of the scene photographed by that pixel, estimating the pose and speed of the first video camera.
    Type: Grant
    Filed: April 9, 2018
    Date of Patent: April 28, 2020
    Assignee: PIXMAP
    Inventors: Andrew Comport, Maxime Meilland
  • Publication number: 20190304170
    Abstract: Various implementations are disclosed of producing a 3-dimensional model of a scene. Various method, electronic device, or system implementations use RGB-D camera to provide RGB-D video content or periodic aligned RGB images and depth images to localize camera spatial position(s) defined in a three dimensional (3D) coordinate system or reconstruct a 3D virtual representation of a current camera frame in the 3D coordinate system, each in real time.
    Type: Application
    Filed: February 28, 2019
    Publication date: October 3, 2019
    Inventor: Maxime Meilland
  • Publication number: 20190182433
    Abstract: A method for estimating the speed of movement of a first video camera when it captures a current image of a three-dimensional scene, includes storing a reference image corresponding to an image of the same scene captured by a second video camera in a different pose, the reference image including pixels. The method also includes storing the current image, the current image including pixels containing the measurement of a physical quantity measured by that pixel, that physical quantity being the same as the physical quantity measured by the pixels of the reference image. The method further includes storing for each pixel of the reference image or of the current image the measurement of a depth that separates that pixel from the point of the scene photographed by that pixel, estimating the pose of the first camera, and estimating the speed of movement of the first camera.
    Type: Application
    Filed: February 6, 2019
    Publication date: June 13, 2019
    Applicant: PIXMAP
    Inventors: Andrew Comport, Maxime Meilland
  • Publication number: 20180308240
    Abstract: A method for estimating the speed a first video camera when it captures a current image of a three-dimensional scene, the current image including pixels. The method includes storing a reference image corresponding to an image of the same scene captured by a second video camera in a different pose, the reference image including pixels. The method also includes storing the current image, containing for each pixel of the current image the measurement of a physical magnitude measured by that pixel, which is the same as the physical magnitude measured by the pixels of the reference image. The method further includes storing for each pixel of the reference image or of the current image the measurement of a depth that separates that pixel from the point of the scene photographed by that pixel, estimating the pose and speed of the first video camera.
    Type: Application
    Filed: April 9, 2018
    Publication date: October 25, 2018
    Applicant: PIXMAP
    Inventors: Andrew Comport, Maxime Meilland
  • Patent number: 9940725
    Abstract: A method for estimating the speed a first video camera when it captures a current image of a three-dimensional scene, the current image including pixels. The method includes storing a reference image corresponding to an image of the same scene captured by a second video camera in a different pose, the reference image including pixels. The method also includes storing the current image, containing for each pixel of the current image the measurement of a physical magnitude measured by that pixel, which is the same as the physical magnitude measured by the pixels of the reference image. The method further includes storing for each pixel of the reference image or of the current image the measurement of a depth that separates that pixel from the point of the scene photographed by that pixel, estimating the pose and speed of the first video camera.
    Type: Grant
    Filed: November 17, 2014
    Date of Patent: April 10, 2018
    Assignee: PIXMAP
    Inventors: Andrew Comport, Maxime Meilland
  • Publication number: 20160292882
    Abstract: A method for estimating the speed a first video camera when it captures a current image of a three-dimensional scene, the current image including pixels. The method includes storing a reference image corresponding to an image of the same scene captured by a second video camera in a different pose, the reference image including pixels. The method also includes storing the current image, containing for each pixel of the current image the measurement of a physical magnitude measured by that pixel, which is the same as the physical magnitude measured by the pixels of the reference image. The method further includes storing for each pixel of the reference image or of the current image the measurement of a depth that separates that pixel from the point of the scene photographed by that pixel, estimating the pose and speed of the first video camera.
    Type: Application
    Filed: November 17, 2014
    Publication date: October 6, 2016
    Applicants: Universite de Nice (UNS), Centre National de la Recherche Scientifique
    Inventors: Andrew Comport, Maxime Meilland
  • Publication number: 20160292883
    Abstract: This method comprises the estimation of the speed XvR of displacement of a camera by searching for the speed XvR which minimizes a discrepancy directly between: -a first value of a physical quantity at the level of a first point (p*) of a reference image, and -a second value of the same physical quantity at the level of a second point (p v2) of a current image, the first value of the physical quantity at the level of the first point (p*) of the reference image being constructed: -by selecting neighbour points of the first point (p*) as a function of the speed XvR and of a time to equal to the exposure time of the first camera, then -by averaging the values of the physical quantity at the level of the neighbour points selected and of the first point in such a way as to generate a new value of the physical quantity at the level of the first point.
    Type: Application
    Filed: November 17, 2014
    Publication date: October 6, 2016
    Applicants: Universite de Nice (UNS), Centre National de la Recherche Scientifique
    Inventors: Andrew Comport, Maxime Meilland