Solid Modelling Patents (Class 345/420)
  • Patent number: 11386297
    Abstract: This learning data generation device (10) is provided with: an identification unit (11) which identifies a subject included in a first captured image, and generates an identification result in which information indicating the type and existence of the identified subject or the motion of the identified subject is associated with the first captured image; and a generation unit (12) which generates learning data on the basis of the identification result and a second captured image, which is associated with the first captured image but is different in type from the first captured image.
    Type: Grant
    Filed: July 10, 2020
    Date of Patent: July 12, 2022
    Assignee: NEC CORPORATION
    Inventor: Soma Shiraishi
  • Patent number: 11386593
    Abstract: The present invention relates to a method and system for automatically setting a scan range. The method comprises: receiving an RGB image and a depth image of an object positioned on a scan table, respectively, by an RGB image prediction model and a depth image prediction model; generating an RGB prediction result based on the RGB image and a depth prediction result based on the depth image with respect to predetermined key points of the object, respectively, by the RGB image prediction model and the depth image prediction model; selecting a prediction result for setting the scan range from the RGB prediction result and the depth prediction result; and automatically setting the scan range based on the selected prediction result.
    Type: Grant
    Filed: June 2, 2020
    Date of Patent: July 12, 2022
    Assignee: GE Precision Healthcare LLC
    Inventors: Yanran Xu, Fanbo Meng, Yu Huang
  • Patent number: 11386524
    Abstract: A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. The encoder is configured to convert a point cloud into an image based representation. The encoder packs patch images into an image frame and fills empty spaces in the image frame with a padding, wherein pixel values for the padding are determined based on neighboring pixels values such that the padding is smoothed in the image frame. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud.
    Type: Grant
    Filed: September 27, 2019
    Date of Patent: July 12, 2022
    Assignee: Apple Inc.
    Inventors: Khaled Mammou, Alexandros Tourapis, Jungsun Kim, Yeping Su, David W. Singer, Valery G. Valentin, Fabrice A. Robinet
  • Patent number: 11378390
    Abstract: Using the same image sensor to capture both a two-dimensional (2D) image of a three-dimensional (3D) object and 3D depth measurements for the object. A laser point-scans the surface of the object with light spots, which are detected by a pixel array in the image sensor to generate the 3D depth profile of the object using triangulation. Each row of pixels in the pixel array forms an epipolar line of the corresponding laser scan line. Timestamping provides a correspondence between the pixel location of a captured light spot and the respective scan angle of the laser to remove any ambiguity in triangulation. An Analog-to-Digital Converter (ADC) in the image sensor generates a multi-bit output in the 2D mode and a binary output in the 3D mode to generate timestamps. Strong ambient light is rejected by switching the image sensor to a 3D logarithmic mode from a 3D linear mode.
    Type: Grant
    Filed: October 21, 2020
    Date of Patent: July 5, 2022
    Inventor: Yibing Michelle Wang
  • Patent number: 11380078
    Abstract: System and method are provided for scaling a 3-D representation of a building structure. The method includes obtaining images of the building structure, including non-camera anchors. The method also includes identifying reference poses for images based on the non-camera anchors. The method also includes obtaining world map data including real-world poses for the images. The method also includes selecting candidate poses from the real-world poses based on corresponding reference poses. The method also includes calculating a scaling factor for a 3-D representation of the building structure based on correlating the reference poses with the selected candidate poses. Some implementations use structure from motion techniques or LiDAR, in addition to augmented reality frameworks, for scaling the 3-D representations of the building structure. In some implementations, the world map data includes environmental data, such as illumination data, and the method includes generating or displaying the 3-D representation.
    Type: Grant
    Filed: December 10, 2020
    Date of Patent: July 5, 2022
    Assignee: HOVER, INC.
    Inventors: Manish Upendran, William Castillo, Jena Dzitsiuk, Yunwen Zhou, Matthew Thomas
  • Patent number: 11380073
    Abstract: A method of aligning a digital model of a structure with a displayed portion of the structure within a video stream is disclosed. An approximate position of the camera device in the digital model is determined. A position and an orientation are determined for a plurality of digital surfaces within the digital model visible from the approximate position of the camera device. A position and an orientation of a plurality of object surfaces visible in a video stream are determined. A 3D translation, a 3D scale, and a 3D rotation that maximize an alignment of the position and orientation of the plurality of digital surfaces with the position and orientation of the plurality of object surfaces are determined. The 3D translation, the 3D scale, and the 3D rotation are applied to the digital model and the digital model is displayed contemporaneously with a display of the video stream.
    Type: Grant
    Filed: February 3, 2021
    Date of Patent: July 5, 2022
    Assignee: Unity IPR ApS
    Inventor: Francis Ruben Malka
  • Patent number: 11379996
    Abstract: Various implementations disclosed herein include devices, systems, and methods that use event camera data to track deformable objects such as faces, hands, and other body parts. One exemplary implementation involves receiving a stream of pixel events output by an event camera. The device tracks the deformable object using this data. Various implementations do so by generating a dynamic representation of the object and modifying the dynamic representation of the object in response to obtaining additional pixel events output by the event camera. In some implementations, generating the dynamic representation of the object involves identifying features disposed on the deformable surface of the object using the stream of pixel events. The features are determined by identifying patterns of pixel events. As new event stream data is received, the patterns of pixel events are recognized in the new data and used to modify the dynamic representation of the object.
    Type: Grant
    Filed: November 13, 2018
    Date of Patent: July 5, 2022
    Assignee: Apple Inc.
    Inventors: Peter Kaufmann, Daniel Kurz, Brian Amberg, Yanghai Tsin
  • Patent number: 11373382
    Abstract: A method is described for implementing augmented reality by means of transferring annotations from a model image to a video flow or to a sequence of destination images. The method entails the creation of a model image for each point of interest (or several points of interest), in addition to the annotation of said model image with point annotations on the objects points of interest that are intended to be augmented and made user-interactable. The method will automatically try to understand if the video flow or the sequence of destination images contains the object contained in the model image and, if so, it will automatically transfer the annotations made from the model image to the video flow or to the sequence of destination images, maintaining the location of the points relative to the object.
    Type: Grant
    Filed: April 1, 2020
    Date of Patent: June 28, 2022
    Assignee: PIKKART S.R.L.
    Inventor: Davide Baltieri
  • Patent number: 11375107
    Abstract: Disclosed herein are an apparatus and method for guiding multi-view capture. The apparatus for guiding multi-view capture includes one or more processors and an execution memory for storing at least one program that is executed by the one or more processors, wherein the at least one program is configured to receive a single-view two-dimensional (2D) image obtained by capturing an image of an object of interest through a camera, generate an orthographic projection image and a perspective projection image for the object of interest from the single-view 2D image using an image conversion parameter that is previously learned from multi-view 2D images for the object of interest, generate a 3D silhouette model for the object of interest using the orthographic projection image and the perspective projection image, and output the 3D silhouette model and a guidance interface for the 3D silhouette model.
    Type: Grant
    Filed: October 30, 2020
    Date of Patent: June 28, 2022
    Assignee: Electronics and Telecommunications Research Institute
    Inventors: Seong-Jae Lim, Ki-Nam Kim, Tae-Joon Kim, Seung-Uk Yoon, Seung-Wook Lee, Bon-Woo Hwang
  • Patent number: 11373369
    Abstract: Systems and methods are disclosed for polygonal mesh geometry extraction for a bevel operation in a modeling application. One method comprises receiving an original shape outline and determining a straight skeleton graph of the original shape outline, the straight skeleton graph comprising a plurality of edges. One or more inverse offset outlines of the original shape outline may then be determined based on the straight skeleton graph. The method may further comprise determining one or more polygons based on a union of the straight skeleton graph, the original shape outline, and the one or more inverse offset outlines. The one or more polygons may include one or more graph polygons and one or more sub-polygons. A beveled shape of the original shape outline may be generated based on the one or more polygons.
    Type: Grant
    Filed: September 2, 2020
    Date of Patent: June 28, 2022
    Assignee: MAXON Computer GmbH
    Inventor: Fritz Kemmler
  • Patent number: 11372474
    Abstract: Systems and methods are provided to create training data, validate, deploy and test artificial intelligence (AI) systems in a virtual development environment, incorporating virtual spaces, objects, machinery, devices, subsystems, and actual human action and behavior.
    Type: Grant
    Filed: July 2, 2020
    Date of Patent: June 28, 2022
    Assignee: SAEC/Kinetic Vision, Inc.
    Inventors: Richard Raymond Schweet, Bendenetto Christopher Ruggiero, Kyle Robert Hartshorn, Gregory Ryan Sweeney, Kyle Dean Cypher, Melissa Yenni Scharf, Emily Ann Meyer, Alec Brenders Lisy, Jeremy David Jarrett, Matthew David Fye
  • Patent number: 11364635
    Abstract: There is provided a method of controlling a robot within an environment comprising: i) receiving, from a 3D scanner, data relating to at least a portion of the environment for constructing a 3D point cloud representing at least a portion of the environment; ii) comparing the 3D point cloud to a virtual 3D model of the environment and, based upon the comparison, determining a position of the robot; then iii) determining a movement trajectory for the robot based upon the determined position of the robot. Also provided is a control apparatus and a robot control system.
    Type: Grant
    Filed: February 8, 2019
    Date of Patent: June 21, 2022
    Assignee: Rolls-Royce plc
    Inventors: James Kell, Bilal M. Nasser, Amir Rabani
  • Patent number: 11363240
    Abstract: Disclosed herein are systems and methods for augmented reality multi-view telepresence. An embodiment takes the form of a method that includes obtaining a session geometry of a multi-location telepresence session that includes a first-location participant at a first location and a second-location participant at a second location, each location having respective pluralities of cameras and display segments. The method includes selecting, according to the session geometry, both a first-to-second-viewpoint second-location camera from the plurality of second-location cameras as well as a first-to-second-viewpoint first-location display segment from the plurality of first-location display segments.
    Type: Grant
    Filed: June 5, 2020
    Date of Patent: June 14, 2022
    Assignee: PCMS HOLDINGS, INC.
    Inventor: Seppo T. Valli
  • Patent number: 11360970
    Abstract: A computer-implemented method includes accessing, by a processing unit, an existing layer representing geospatial-temporal data at a selected timestamp. A first overview layer of the existing layer is generated by iteratively aggregating each cluster of cells of the existing layer into a corresponding lower-resolution cell of the first overview layer. The first overview layer therefore has a lower resolution than the existing layer. A query is received related to the geospatial-temporal data in the existing layer, and the query is processed with reference to the first overview layer.
    Type: Grant
    Filed: November 13, 2018
    Date of Patent: June 14, 2022
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Marcus Oliver Freitag, Conrad M. Albrecht, Fernando Jimenez Marianno, Siyuan Lu, Hendrik F. Hamann, Johannes Walter Schmude
  • Patent number: 11361484
    Abstract: Systems and methods for segmenting scan data are disclosed. The methods include receiving scan data representing a plurality of points in an environment associated with a ground surface and one or more objects, and creating a graph from the scan data. The graph includes a plurality of vertices corresponding to the plurality of points. The method further includes assigning a unary potential to each of the plurality of vertices that is a cost of assigning that vertex to a ground label or a non-ground label, and assigning a pairwise potential to each pair of neighboring vertices in the graph that is the cost of assigning different labels to neighboring vertices. The methods include using the unary potentials and the pairwise potentials to identify labels for each of the plurality of points, and segmenting the scan data to identify points associated with the ground based on the identified labels.
    Type: Grant
    Filed: December 4, 2020
    Date of Patent: June 14, 2022
    Assignee: Argo AI, LLC
    Inventor: John Peterson
  • Patent number: 11360574
    Abstract: A system comprising an object; a computer device with a digital model containing a virtual object corresponding to the object; a position sensor for detecting the position of the object; and a communication means to transmit the position of the object to the computer device. The object is stretchable and the system is further provided with sensors for detecting the stretched condition of the object and with communication means to transmit said condition to the computer device. A system with actuators to modify the position and/or the stretched condition of the object and use of such systems for a collaborative complex problem-solving procedure in the context of urban planning.
    Type: Grant
    Filed: August 29, 2018
    Date of Patent: June 14, 2022
    Assignee: Luxembourg Institute of Science and Technology (LIST)
    Inventors: Lou Schwartz, Valérie Maquil, Thibaud Latour, Annie Guerriero, Séverine Mignon
  • Patent number: 11344790
    Abstract: A system manages playing cards to which information is given by an RFID tag and gaming chips to which information is given by an RFID tag. A system is provided with a storage box integrally configured to contain a card drawer for storing playing cards and a chip drawer for storing gaming chips, a card RFID reader for reading information of an RFID tag attached to the playing cards stored in the card drawer, and a chip RFID reader for reading information of the RFID tag attached to the chips stored in the chip drawer. A card RFID reader and a chip RFID reader read information of each RFID tag using different frequencies.
    Type: Grant
    Filed: September 8, 2017
    Date of Patent: May 31, 2022
    Assignee: ANGEL GROUP CO., LTD.
    Inventor: Yasushi Shigeta
  • Patent number: 11348260
    Abstract: This method for encoding a 3D point cloud comprises: —segmenting the point cloud into first patches (2); —determining inverse projection parameters of the patches (4); —projecting the first patches to planes using the inverse projection parameters (10); —encoding the planes (12); —determining sparse points in the point cloud not belonging to any patch (HC, 18); and—encoding the sparse points using an octree-decomposition (20).
    Type: Grant
    Filed: June 18, 2018
    Date of Patent: May 31, 2022
    Assignee: InterDigital VC Holdings, Inc.
    Inventors: Kangying Cai, Sebastien Lasserre
  • Patent number: 11341296
    Abstract: A computer-implemented method computes an unfolded part of a modeled bended 3D object in a 3D scene of a computer-aided design system. The method a) provides the 3D object; b) selects a fixed portion (FP) of the 3D object; c) selects a mobile portion (MP) of the 3D object; d) determines a 1D interface (INT) forming an intersection between the fixed portion (FP) and the mobile portion; e) computes a transformed portion resulting from a linear transformation of the mobile portion (MP) according to an drawing direction (DD); f) trims the transformed portion in the vicinity of the 1D interface (INT), thereby forming a trimmed transformed portion (TTP); g) creates a fillet (FI) between the 1D interface (INT) and the trimmed transformed portion (TTP); and h) defines the unfolded part as an union of the fixed portion (FP), the trimmed transformed portion (TTP) and the created fillet (FI).
    Type: Grant
    Filed: December 10, 2018
    Date of Patent: May 24, 2022
    Assignee: DASSAULT SYSTEMES
    Inventor: Pierre Denis Voirin
  • Patent number: 11343484
    Abstract: A display device is switchable between 3D image display and 2D image display. The display device includes a first display panel, a second display panel, a controller, and an optical system. The first and second display panels each include subpixels arranged in a grid. The controller performs switching and/or is configured to perform switching between multiple display modes including a first display mode for displaying a 2D image and a second display mode for displaying a parallax image on the first display panel, and switches and/or is configured to switch a drive mode of the second display panel between multiple drive modes including a first drive mode corresponding to the first display mode and a second drive mode corresponding to the second display mode.
    Type: Grant
    Filed: July 25, 2019
    Date of Patent: May 24, 2022
    Assignee: KYOCERA Corporation
    Inventors: Kaoru Kusafuka, Mitsuhiro Murata
  • Patent number: 11341702
    Abstract: An animation system wherein a machine learning model is adopted to learn a transformation relationship between facial muscle movements and skin surface movements. For example, for the skin surface representing “smile,” the transformation model derives movement vectors relating to what facial muscles are activated, what are the muscle strains, what is the joint movement, and/or the like. Such derived movement vectors may be used to simulate the skin surface “smile.
    Type: Grant
    Filed: October 28, 2020
    Date of Patent: May 24, 2022
    Assignee: UNITY TECHNOLOGIES SF
    Inventor: Byung Kuk Choi
  • Patent number: 11343591
    Abstract: The present application discloses a method of presenting moving images or videos corresponding to still images. The method includes: storing a still image and a moving image or video corresponding to the still image into a cloud storage; extracting feature points of the still image stored in the cloud storage, and storing the feature points in the cloud storage in a manner which associates the feature points with the still image; when a device obtains a first still image through scanning, extracting feature points from the first still image, comparing and judging whether the extracted feature points match feature points of each still image stored in the cloud storage to determine a second still image whose feature points match the feature points of the first still image; rendering a moving image or video corresponding to the second still image stored in the cloud storage at the position of the first still image.
    Type: Grant
    Filed: July 2, 2019
    Date of Patent: May 24, 2022
    Assignee: SAMSUNG ELECTRONICS CO., LTD.
    Inventors: Xiao Ling, Junqing Ma
  • Patent number: 11335065
    Abstract: A method of construction of a computer-generated image, the computer-generated image comprising a representation of an environment, called setting image, from a point of view, and at least one location of the setting for insertion of an image of an object, comprises: a step of creating of a setting image applying the laws of perspective, a step of generating at least one rendered or captured image of at least one object, the representation of the object in each image obeying the laws of perspective so during the generation of the object image, the generating step being independent from the creating step, a step of incorporation of the object image into at least one location of the setting image, wherein the application of the laws of perspective on the object image creates an error, discernible to the human eye, with the law of perspective applied in the point of view, whereby the constructed computer-generated image presents an overall perceived photorealism.
    Type: Grant
    Filed: December 5, 2018
    Date of Patent: May 17, 2022
    Assignee: DIAKSE
    Inventors: Julien Berthomier, Emmanuel Naxos
  • Patent number: 11334232
    Abstract: An interactive map may comprise image data associated with an area. First image data and second image data may comprise an indication of a subsection of the area at different times. At least a portion of the first image data and second image data may be output via the interactive map. The subsection of the area in the first image data and in the second image data may be output at corresponding positions. The interactive map may comprise an interactive icon. A user input may be received, via the interactive map, indicative of a movement of the interactive icon. Responsive to the user input, a part of the first image data or the second image data may be revealed and a part of the second image data or the first image data may be obscured in a direction corresponding to a direction of the movement of the interactive icon.
    Type: Grant
    Filed: February 26, 2021
    Date of Patent: May 17, 2022
    Assignee: United Services Automobile Association (USAA)
    Inventors: John D. Clark, Jess W. Gingrich, Kristina Tomasetti
  • Patent number: 11334690
    Abstract: A method of determining a modified shape of a computer-aided design model of an object or a partition thereof is described. With the method, at least two reference landmarks are placed at a reference position with respect to the shape of the object or to the partition. An adjustment landmark is placed at the reference position of the corresponding reference landmark. The adjustment landmarks are displaced from the reference position to a modified position, and a displacement between the modified position and the reference position is obtained. The modified shape of the object or the partition is generated by modifying the shape of the object or the partition by applying a homogeneous function of the displacement to the computer-aided design model of the shape of the object or the partition.
    Type: Grant
    Filed: August 17, 2018
    Date of Patent: May 17, 2022
    Assignee: UNIVERSITĂ„T STUTTGART
    Inventors: Steffen Reichert, Long Nguyen, Achim Menges
  • Patent number: 11328401
    Abstract: A stationary object detecting method, a stationary object apparatus, and an electronic device are disclosed in embodiments of the present disclosure, the method includes: obtaining point cloud data of a scene and feature information of each of data points in the point cloud data; performing a triangulation network connection on each of the data points in the point cloud data to generate a triangular network model, and taking a picture of the triangular network model using a preset camera to obtain an image; obtaining a first data point corresponding to each pixel point in the image, and obtaining a feature map of the point cloud data according to feature information of each first data point; inputting the feature map into a classification model to obtain data points corresponding to the stationary object in the point cloud data.
    Type: Grant
    Filed: July 22, 2019
    Date of Patent: May 10, 2022
    Inventors: Weixin Lu, Shiyu Song, Fangfang Dong, Shengpan Xu
  • Patent number: 11328474
    Abstract: Some embodiments of an example method disclosed herein may include receiving point cloud data representing one or more three-dimensional objects; receiving a viewpoint of the point cloud data; selecting a selected object from the one or more three-dimensional objects using the viewpoint; retrieving a neural network model for the selected object; generating a level of detail data for the selected object using the neural network model; and replacing, within the point cloud data, points corresponding to the selected object with the level of detail data.
    Type: Grant
    Filed: March 12, 2019
    Date of Patent: May 10, 2022
    Assignee: INTERDIGITAL MADISON PATENT HOLDINGS, SAS
    Inventor: Tatu V. J. Harviainen
  • Patent number: 11328523
    Abstract: The present disclosure relates to an image composite system that employs a generative adversarial network to generate realistic composite images. For example, in one or more embodiments, the image composite system trains a geometric prediction neural network using an adversarial discrimination neural network to learn warp parameters that provide correct geometric alignment of foreground objects with respect to a background image. Once trained, the determined warp parameters provide realistic geometric corrections to foreground objects such that the warped foreground objects appear to blend into background images naturally when composited together.
    Type: Grant
    Filed: June 9, 2020
    Date of Patent: May 10, 2022
    Assignee: Adobe Inc.
    Inventors: Elya Shechtman, Oliver Wang, Mehmet Yumer, Chen-Hsuan Lin
  • Patent number: 11328404
    Abstract: An evaluation apparatus (10) includes an acquisition unit (110), an index computation unit (120), and a suitability computation unit (140). The acquisition unit (110) acquires arrangement information and environmental information. The arrangement information indicates at least any of position, orientation and field angle of an image capturing apparatus. The environmental information indicates an environment of a target region. The index computation unit (120) computes a surveillance index by using the arrangement information and the environmental information, the surveillance index indicating difficulty or easiness of surveilling a target object. The suitability computation unit (140) computes suitability of arrangement of the image capturing apparatus indicated by the arrangement information, based on the distribution of the surveillance index in the target region.
    Type: Grant
    Filed: July 31, 2018
    Date of Patent: May 10, 2022
    Assignee: NEC CORPORATION
    Inventor: Ryoma Oami
  • Patent number: 11321903
    Abstract: A technique for performing ray tracing operations is provided. The technique includes receiving a ray for an intersection test, testing the ray against boxes specified in a bounding volume hierarchy to eliminate one or more boxes or triangles from consideration, unpacking a triangle from a compressed triangle block of the bounding volume hierarchy, the compressed triangle block including two or more triangles that share at least one vertex, and testing the ray for intersection against at least one of the unpacked triangles.
    Type: Grant
    Filed: March 27, 2020
    Date of Patent: May 3, 2022
    Assignee: Advanced Micro Devices, Inc.
    Inventors: Skyler Jonathon Saleh, Young In Yeo
  • Patent number: 11318607
    Abstract: An amusement park attraction includes a ride vehicle configured to travel along a path of the amusement park attraction, an attraction controller configured to generate virtual elements, and a robot assembly disposed on the ride vehicle. The head of the robot assembly is configured to adjust positions relative to the ride vehicle. The amusement park attraction further includes a headset communicatively coupled to the attraction controller and a camera disposed in the head of the robot assembly. The headset is configured to be disposed on the head of the robot assembly. The headset is also configured to display the virtual elements generated by the attraction controller based on a headset position in the amusement park attraction. The camera is configured to acquire data indicative of the virtual elements displayed by the headset.
    Type: Grant
    Filed: March 19, 2019
    Date of Patent: May 3, 2022
    Assignee: Universal City Studios LLC
    Inventors: Martin Evan Graham, Patrick John Goergen, Tomas Manuel Trujillo
  • Patent number: 11320262
    Abstract: A system having: a terrestrial scanning device, a topographic benchmark secured permanently to a support, including a code that is readable automatically at a distance of at least 3 m by the scanning device, this code ss providing information as to the position of the benchmark in a given frame of reference and/or having an identifier listed in a database in which the position of the benchmark in said frame of reference is also recorded.
    Type: Grant
    Filed: June 18, 2019
    Date of Patent: May 3, 2022
    Assignee: SOLETANCHE FREYSSINET
    Inventors: Yohan Rabot, Gilles Hovhanessian, Maxime Trocme, Jerome Stubler
  • Patent number: 11320803
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for determining a suitable building layout for a property. One of the operations is performed by optionally, receiving building module parameters for section of qualifying building modules. displaying, via a user interface, a graphical representation of two or more building modules. A user may customize a building layout from the two or more building modules. The system displays an assembled building layout including two or more building modules. The system receives a confirmation of the assembled building layout for 3D printing. Based on the assembled building layout, instructions are generated and transmitted to one or more 3D printers to manufacture one or more building structures based on the selected building layout with potential subsequent modification of the preconfigured positions of windows, doors and walls of the building.
    Type: Grant
    Filed: March 5, 2021
    Date of Patent: May 3, 2022
    Assignee: Mighty Buildings, Inc.
    Inventor: Aleksei Dubov
  • Patent number: 11321768
    Abstract: Disclosed are systems and methods for augmenting a customer image with at least one virtual object. One or more customer images depicting a customer environment are received from a customer computing device and 3D feature data for the customer images is determined. A virtual object is determined which corresponds to a desired merchant item. A positioning signal is received which corresponds to a desired location of virtual object in the customer environment, and a corresponding first location in each customer image at which to overlay the virtual object using the positioning signal is determined. An appropriate size and orientation of the virtual object is determined for each customer image based on corresponding 3D feature data, causing an overlay of an appropriately sized and oriented virtual object at the corresponding first location in customer images.
    Type: Grant
    Filed: December 21, 2018
    Date of Patent: May 3, 2022
    Assignee: Shopify Inc.
    Inventor: Daniel Beauchamp
  • Patent number: 11321896
    Abstract: Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing at least one program and a method for rendering an avatar. A first avatar having a first level of detail is stored in a database, the first avatar including a first plurality of components. A level of detail of each of the first plurality of components is reduced separately from each other. A second plurality of components comprising the reduced level of detail of each of the first plurality of components is stored. A request is received for the first avatar in a second level of detail that comprises a lower level of detail than the first level of detail. In response to receiving the request, the second plurality of components is assembled to generate a second avatar having the second level of detail.
    Type: Grant
    Filed: November 18, 2020
    Date of Patent: May 3, 2022
    Assignee: Snap Inc.
    Inventors: Rahul Bhupendra Sheth, Maoning Guo, William Eastcott
  • Patent number: 11315327
    Abstract: In one embodiment, a method includes determining a first time interval for rendering pixel values for a first portion of a display, where the first portion of the display is one of a plurality of portions of the display, accessing a representation of virtual objects in a virtual scene, and determining that the first time interval is insufficient for rendering pixel values for the first portion of the display based on the representation of virtual objects. The method further includes, for each of the portions of the display, determining a current viewpoint of a user, rendering, based on the current viewpoint of the user, pixel values for that portion of the display, and sending the rendered pixel values to be output by the display, where the rendered pixel values for the first portion of the display are rendered based on a simplified representation of virtual objects in the virtual scene.
    Type: Grant
    Filed: October 30, 2019
    Date of Patent: April 26, 2022
    Assignee: Facebook Technologies, LLC.
    Inventors: Larry Seiler, Alexander Nankervis, John Adrian Arthur Johnston, Jeremy Freeman
  • Patent number: 11315282
    Abstract: A system and method for documenting a scene having evidence markers is provided. The method includes placing the evidence markers in the scene. A plurality of 3D coordinates of points are measured on surfaces in the environment, a first portion of the plurality of 3D coordinates being on a first evidence marker of the evidence markers, the evidence marker having a photogrammetric symbol on one surface. A point cloud is generated from the plurality of 3D coordinates. The first evidence marker in the point cloud based is automatically identified at least in part on the photogrammetric symbol. The location and at least one attribute of the evidence marker are stored.
    Type: Grant
    Filed: August 20, 2020
    Date of Patent: April 26, 2022
    Assignee: FARO TECHNOLOGIES, INC.
    Inventors: David J. Dustin, Dennis C. Sweet
  • Patent number: 11313951
    Abstract: Embodiments of the present disclosure provide a ground detection method and apparatus, an electronic device, a vehicle, and a storage medium. The method includes: projecting a laser point cloud obtained to a high resolution mesh and a low resolution mesh respectively; filtering out candidate ground points in the high resolution mesh; and performing ground fitting based on the candidate ground points in the low resolution mesh.
    Type: Grant
    Filed: September 10, 2019
    Date of Patent: April 26, 2022
    Assignee: BAIDU ONLINE NETWORK TECHNOLOGY (BEIJING) CO., LTD.
    Inventors: Dongming Chen, Hao Wang, Liang Wang
  • Patent number: 11315255
    Abstract: The present disclosure includes methods and systems for identifying and manipulating a segment of a three-dimensional digital model based on soft classification of the three-dimensional digital model. In particular, one or more embodiments of the disclosed systems and methods identify a soft classification of a digital model and utilize the soft classification to tune segmentation algorithms. For example, the disclosed systems and methods can utilize a soft classification to select a segmentation algorithm from a plurality of segmentation algorithms, to combine segmentation parameters from a plurality of segmentation algorithms, and/or to identify input parameters for a segmentation algorithm. The disclosed systems and methods can utilize the tuned segmentation algorithms to accurately and efficiently identify a segment of a three-dimensional digital model.
    Type: Grant
    Filed: June 22, 2020
    Date of Patent: April 26, 2022
    Assignee: Adobe Inc.
    Inventors: Vladimir Kim, Aaron Hertzmann, Mehmet Yumer
  • Patent number: 11311794
    Abstract: Presented herein is an electronic gaming device comprising a first axle with a first center cubelet rotatably coupled to the first axle, a first sensor operatively coupled to the first axle to detect rotation of the first cubelet, and responsive to detecting rotation of the first cubelet, transmit a first signal, a second axle with a second center cubelet rotatably coupled to the second axle, a second sensor operatively coupled to the second axle to detect rotation of the second cubelet, and responsive to detecting rotation of the second cubelet, transmit a second signal, a third axle with a third center cubelet rotatably coupled to the third axle, a third sensor operatively coupled to the third axle to detect rotation of the third cubelet, and responsive to detecting rotation of the third cubelet, transmit a third signal, a plurality of interchangeable cubelets positioned about the first, second, and third axle such that mechanical rotation of a set of the interchangeable cubelets having a common plane cause
    Type: Grant
    Filed: July 17, 2020
    Date of Patent: April 26, 2022
    Assignee: Virtual Vectors LLC.
    Inventor: Rajesh A. Limaye
  • Patent number: 11315263
    Abstract: A medical image processing apparatus having a processor configured to detect at least four reference landmarks among the left eye, the right eye, the diencephalon, the fornix, the corpus callosum, the left hippocampus, and the right hippocampus from a brain image, performs first registration including registration by similarity transformation using reference landmarks between the brain image and a standard brain image, and perform second registration by nonlinear transformation between the brain image and the standard brain image after the first registration.
    Type: Grant
    Filed: January 30, 2020
    Date of Patent: April 26, 2022
    Assignee: FUJIFILM Corporation
    Inventor: Caihua Wang
  • Patent number: 11308752
    Abstract: The present disclosure describes systems and methods for simulating gameplay of a live event and placing wagers or non-wager submissions concerning an outcome of a simulation. The systems incorporate statistical data, event information, and user modifications to create the simulation.
    Type: Grant
    Filed: August 20, 2020
    Date of Patent: April 19, 2022
    Inventor: Vinod Khosla
  • Patent number: 11309087
    Abstract: Two methods are provided for the automated derivation of Voronoi diagrams in 3D. The invention, implementable via various means such as a processing system, method, or data structure in a recording medium such as memory or as a self-contained electronic circuit, has wide ranging applicability to numerous fields such as big data analysis, computer graphics and animation, route planning, collision avoidance, computer vision, robotic vision, and etc. The first method of the invention details steps necessary to segment data according to a set of generators so as to produce a Voronoi partitioning of the data. The second method of the invention includes steps associated with the derivation of the mathematical specification of a 3D Voronoi diagram.
    Type: Grant
    Filed: November 26, 2014
    Date of Patent: April 19, 2022
    Inventor: Jeffrey W. Holcomb
  • Patent number: 11308679
    Abstract: The technique of this disclosure suppresses a reduction in visibility of a predetermined object in virtual viewpoint image data. An image processing apparatus includes: an image capturing information acquisition unit configured to acquire image capturing information indicating a position and orientation of each of a plurality of image capturing apparatuses; an object information acquisition unit configured to acquire object information indicating a position and orientation of an object to be captured by the image capturing apparatuses, the object having a specific viewing angle; and a determination unit configured to determine, based on the acquired image capturing information and the position and orientation of the object indicated by the acquired object information, an image to be used for generating a virtual viewpoint image according to a position and orientation of a virtual viewpoint among a plurality of images based on capturing by the image capturing apparatuses.
    Type: Grant
    Filed: May 29, 2020
    Date of Patent: April 19, 2022
    Assignee: CANON KABUSHIKI KAISHA
    Inventor: Daichi Adachi
  • Patent number: 11303915
    Abstract: A three-dimensional data encoding method includes: generating first information in which an N-ary tree structure of a plurality of three-dimensional points included in three-dimensional data is expressed using a first formula, where N is an integer of 2 or higher; and generating a bitstream including the first information. The first information includes pieces of three-dimensional point information each associated with a corresponding one of the plurality of three-dimensional points. The pieces of three-dimensional point information each include indexes each associated with a corresponding one of a plurality of levels in the N-ary tree structure. The indexes each indicate a subblock, among N subblocks belonging to a corresponding one of the plurality of levels, to which a corresponding one of the plurality of three-dimensional points belongs.
    Type: Grant
    Filed: May 20, 2020
    Date of Patent: April 12, 2022
    Assignee: PANASONIC INTELLECTUAL PROPERTY CORPORATION OF AMERICA
    Inventors: Chi Wang, Pongsak Lasang, Toshiyasu Sugio, Tatsuya Koyama
  • Patent number: 11302082
    Abstract: Provided herein are exemplary embodiments directed to a method for creating digital media, including the placing the digital media in a computer graphics environment, the computer graphics environment further comprising visually perceptible elements appearing as real objects placed in a real world setting, and viewing the digital media when at the real world setting. Various exemplary systems include an augmented reality and virtual reality server connected to a network, and a client device connected to the network, the client device having an augmented reality and virtual reality application. Further exemplary systems include a body or motion sensor connected to the client device and/or an augmented reality and virtual reality interface connected to the client device.
    Type: Grant
    Filed: July 17, 2020
    Date of Patent: April 12, 2022
    Assignee: tagSpace Pty Ltd
    Inventor: Paul Simon Martin
  • Patent number: 11300664
    Abstract: A LiDAR odometry method based on the directed geometric point and sparse frame includes: obtaining original three-dimensional point cloud data of a surrounding environment in the coordinate system of the current frame of the LiDAR sensor; constructing and converting the first directed geometric point set and first environmental directed geometric point set into the second directed geometric point set and second environmental directed geometric point set in the world coordinate system, respectively; calculating candidate associated points in the second environmental directed geometric point set in association with each point in the second directed geometric point set, and determining the best associated point of each point in the second directed geometric point set; constructing and solving a pose optimization function to obtain an optimized pose from the coordinate system of the current frame of the LiDAR sensor to the world coordinate system; and updating directed geometric point sets of sparse frames.
    Type: Grant
    Filed: October 13, 2021
    Date of Patent: April 12, 2022
    Assignees: INSTITUTE OF AUTOMATION, CHINESE ACADEMY OF SCIENCES, Beijing EnGo Technology Co., Ltd.
    Inventors: Zhiqiang Cao, Zhonghui Li, Jinli Kang, Min Tan, Shuang Liang, Chengpeng Wang
  • Patent number: 11302061
    Abstract: An image processing apparatus includes an acquisition unit configured to acquire a three-dimensional shape data of an object based on images captured by a plurality of cameras, a generation unit configured to generate information based on a relationship between the three-dimensional shape data acquired by the acquisition unit and positions of the plurality of cameras, and a correction unit configured to correct the three-dimensional shape data based on the information generated by the generation unit.
    Type: Grant
    Filed: January 18, 2019
    Date of Patent: April 12, 2022
    Assignee: Canon Kabushiki Kaisha
    Inventor: Mitsuru Maeda
  • Patent number: 11295508
    Abstract: A bounding volume is used to approximate the space an object occupies. If a more precise understanding beyond an approximation is required, the object itself is then inspected to determine what space it occupies. Often, a simple volume (such as an axis-aligned box) is used as bounding volume to approximate the space occupied by an object. But objects can be arbitrary, complicated shapes. So a simple volume often does not fit the object very well. That causes a lot of space that is not occupied by the object to be included in the approximation of the space being occupied by the object. Hardware-based techniques are disclosed herein, for example, for efficiently using multiple bounding volumes (such as axis-aligned bounding boxes) to represent, in effect, an arbitrarily shaped bounding volume to better fit the object, and for using such arbitrary bounding volumes to improve performance in applications such as ray tracing.
    Type: Grant
    Filed: June 10, 2020
    Date of Patent: April 5, 2022
    Assignee: NVIDIA Corporation
    Inventors: Gregory Muthler, John Burgess
  • Patent number: 11295040
    Abstract: The disclosure notably relates to a computer-implemented method for designing, with a CAD system, a 3D modeled object representing a mechanical part. The method includes providing a B-rep representing the mechanical part. The B-rep has faces, edges and vertices. The method includes providing a sharp edge. The method further includes, automatically by the CAD system, identifying a set of edges. The set of edges includes the provided sharp edge. Each first edge of the set of edges is directed similarly to at least one second edge of the set of edges, the first edge and the second edge sharing a face. The method further includes, automatically by the CAD system, selecting sharp edges within the set of edges. Such a method constitutes an improved method for designing a mechanical part.
    Type: Grant
    Filed: December 26, 2019
    Date of Patent: April 5, 2022
    Assignee: DASSAULT SYSTEMES
    Inventors: Laurent Marini, Quentin Freger