Patents by Inventor David Thomasson

David Thomasson has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11953879
    Abstract: An agent engine allocates a collection of agents to scan the surface of an object model. Each agent operates autonomously and implements particular behaviors based on the actions of nearby agents. Accordingly, the collection of agents exhibits swarm-like behavior. Over a sequence of time steps, the agents traverse the surface of the object model. Each agent acts to avoid other agents, thereby maintaining a relatively consistent distribution of agents across the surface of the object model over all time steps. At a given time step, the agent engine generates a slice through the object model that intersects each agent in a group of agents. The slice associated with a given time step represents a set of locations where material should be deposited to fabricate a 3D object. Based on a set of such slices, a robot engine causes a robot to fabricate the 3D object.
    Type: Grant
    Filed: September 8, 2020
    Date of Patent: April 9, 2024
    Assignee: AUTODESK, INC.
    Inventors: Evan Patrick Atherton, David Thomasson, Maurice Ugo Conti, Heather Kerrick, Nicholas Cote, Hui Li
  • Publication number: 20240065672
    Abstract: An acoustic coupling device includes a base tape member having a hole, and a sensor side and a skin side each having an adhesive material thereon. The device further includes a tray removably attached to cover the skin side of the base tape member. The tray includes an indentation, which defines a concave face facing the hole, and a flat portion surrounding the indentation. The device further includes an ultrasound transferring gel disposed in the indentation, and a removable protective liner configured to cover the sensor side of the base tape member.
    Type: Application
    Filed: August 30, 2022
    Publication date: February 29, 2024
    Applicant: RESPINOR AS
    Inventors: David Gerez, Marit MELLEMSETER, Jessica THOMASSON
  • Patent number: 11609547
    Abstract: A robot system is configured to identify gestures performed by an end-user proximate to a work piece. The robot system then determines a set of modifications to be made to the work piece based on the gestures. A projector coupled to the robot system projects images onto the work piece that represent the modification to be made and/or a CAD model of the work piece. The robot system then performs the modifications.
    Type: Grant
    Filed: December 19, 2016
    Date of Patent: March 21, 2023
    Assignee: AUTODESK, INC.
    Inventors: Evan Patrick Atherton, David Thomasson, Maurice Ugo Conti, Heather Kerrick
  • Patent number: 11556108
    Abstract: A robot is configured to assist an end-user with creative tasks. While the end-user modifies the work piece, the robot observes the modifications made by the end-user and determines one or more objectives that the end-user may endeavor to accomplish. The robot then determines a set of actions to perform that assist the end-user with accomplishing the objectives.
    Type: Grant
    Filed: May 4, 2020
    Date of Patent: January 17, 2023
    Assignee: AUTODESK, INC.
    Inventors: Evan Patrick Atherton, David Thomasson, Maurice Ugo Conti, Heather Kerrick
  • Patent number: 11295400
    Abstract: One embodiment of the present invention sets forth a technique for performing tasks associated with a construction project. The technique includes transmitting to a worker, via a mobile computing device worn by the worker, a first instruction related to performing a first task included in a plurality of tasks associated with a construction project, and transmitting to a light-emitting device a command to provide a visual indicator to the worker that facilitates performing the first task, based on an input received from the mobile computing device, determining that the worker has completed the first task of the construction project, selecting, from a database that tracks eligibility of each of the plurality of tasks, a second task included in the plurality of tasks that the worker is eligible to perform, and transmitting to the worker, via the mobile computing device, a second instruction related to performing the second task.
    Type: Grant
    Filed: November 22, 2016
    Date of Patent: April 5, 2022
    Assignee: AUTODESK, INC.
    Inventors: Tovi Grossman, George Fitzmaurice, Anderson Nogueira, Nick Beirne, Justin Frank Matejka, Danil Nagy, Steven Li, Benjamin LaFreniere, Heather Kerrick, Thomas White, Fraser Anderson, Evan Atherton, David Thomasson, Arthur Harsuvanakit, Maurice Ugo Conti
  • Patent number: 11181886
    Abstract: A robot system is configured to fabricate three-dimensional (3D) objects using closed-loop, computer vision-based control. The robot system initiates fabrication based on a set of fabrication paths along which material is to be deposited. During deposition of material, the robot system captures video data and processes that data to determine the specific locations where the material is deposited. Based on these locations, the robot system adjusts future deposition locations to compensate for deviations from the fabrication paths. Additionally, because the robot system includes a 6-axis robotic arm, the robot system can deposit material at any locations, along any pathway, or across any surface. Accordingly, the robot system is capable of fabricating a 3D object with multiple non-parallel, non-horizontal, and/or non-planar layers.
    Type: Grant
    Filed: April 24, 2017
    Date of Patent: November 23, 2021
    Assignee: AUTODESK, INC.
    Inventors: Evan Atherton, David Thomasson, Maurice Ugo Conti, Heather Kerrick, Nicholas Cote
  • Patent number: 11179793
    Abstract: A control application implements computer vision techniques to cause a positioning robot and a welding robot to perform fabrication operations. The control application causes the positioning robot to place elements of a structure at certain positions based on real-time visual feedback captured by the positioning robot. The control application also causes the welding robot to weld those elements into place based on real-time visual feedback captured by the welding robot. By analyzing the real-time visual feedback captured by both robots, the control application adjusts the positioning and welding operations in real time.
    Type: Grant
    Filed: September 12, 2017
    Date of Patent: November 23, 2021
    Assignee: AUTODESK, INC.
    Inventors: Evan Atherton, David Thomasson, Heather Kerrick, Hui Li
  • Patent number: 11072071
    Abstract: A robot system models the behavior of a user when the user occupies an operating zone associated with a robot. The robot system predicts future behaviors of the user, and then determines whether those predicted behaviors interfere with anticipated behaviors of the robot. When such interference may occur, the robot system generates dynamics adjustments that can be implemented by the robot to avoid such interference. The robot system may also generate dynamics adjustments that can be implemented by the user to avoid such interference.
    Type: Grant
    Filed: September 19, 2017
    Date of Patent: July 27, 2021
    Assignee: AUTODESK, INC.
    Inventors: Evan Atherton, David Thomasson, Heather Kerrick, Hui Li
  • Publication number: 20210208563
    Abstract: A robot system is configured to fabricate three-dimensional (3D) objects using closed-loop, computer vision-based control. The robot system initiates fabrication based on a set of fabrication paths along which material is to be deposited. During deposition of material, the robot system captures video data and processes that data to determine the specific locations where the material is deposited. Based on these locations, the robot system adjusts future deposition locations to compensate for deviations from the fabrication paths. Additionally, because the robot system includes a 6-axis robotic arm, the robot system can deposit material at any locations, along any pathway, or across any surface. Accordingly, the robot system is capable of fabricating a 3D object with multiple non-parallel, non-horizontal, and/or non-planar layers.
    Type: Application
    Filed: March 22, 2021
    Publication date: July 8, 2021
    Inventors: Evan ATHERTON, David THOMASSON, Maurice Ugo CONTI, Heather KERRICK, Nicholas COTE
  • Patent number: 10955814
    Abstract: A robot system is configured to fabricate three-dimensional (3D) objects using closed-loop, computer vision-based control. The robot system initiates fabrication based on a set of fabrication paths along which material is to be deposited. During deposition of material, the robot system captures video data and processes that data to determine the specific locations where the material is deposited. Based on these locations, the robot system adjusts future deposition locations to compensate for deviations from the fabrication paths. Additionally, because the robot system includes a 6-axis robotic arm, the robot system can deposit material at any locations, along any pathway, or across any surface. Accordingly, the robot system is capable of fabricating a 3D object with multiple non-parallel, non-horizontal, and/or non-planar layers.
    Type: Grant
    Filed: April 24, 2017
    Date of Patent: March 23, 2021
    Assignee: AUTODESK, INC.
    Inventors: Evan Atherton, David Thomasson, Maurice Ugo Conti, Heather Kerrick, Nicholas Cote
  • Patent number: 10956739
    Abstract: A technique for displaying a representative path associated with a robotic device. The technique includes detecting at least one reference point within a first image of a workspace, generating the representative path based on path instructions associated with the robotic device and the at least one reference point, and displaying the representative path within the workspace.
    Type: Grant
    Filed: June 27, 2016
    Date of Patent: March 23, 2021
    Assignee: AUTODESK, INC.
    Inventors: David Thomasson, Evan Patrick Atherton, Maurice Ugo Conti, Heather Kerrick
  • Publication number: 20210073445
    Abstract: A robotic assembly cell is configured to generate a physical mesh of physical polygons based on a simulated mesh of simulated triangles. A control application configured to operate the assembly cell selects a simulated polygon in the simulated mesh and then causes a positioning robot in the cell to obtain a physical polygon that is similar to the simulated polygon. The positioning robot positions the polygon on the physical mesh, and a welding robot in the cell then welds the polygon to the mesh. The control application captures data that reflects how the physical polygon is actually positioned on the physical mesh, and then updates the simulated mesh to be geometrically consistent with the physical mesh. In doing so, the control application may execute a multi-objective solver to generate an updated simulated mesh that meets specific design criteria.
    Type: Application
    Filed: November 24, 2020
    Publication date: March 11, 2021
    Inventors: Evan Patrick Atherton, David Thomasson, Maurice Ugo Conti, Heather Kerrick, Nicholas Cote
  • Publication number: 20200401105
    Abstract: An agent engine allocates a collection of agents to scan the surface of an object model. Each agent operates autonomously and implements particular behaviors based on the actions of nearby agents. Accordingly, the collection of agents exhibits swarm-like behavior. Over a sequence of time steps, the agents traverse the surface of the object model. Each agent acts to avoid other agents, thereby maintaining a relatively consistent distribution of agents across the surface of the object model over all time steps. At a given time step, the agent engine generates a slice through the object model that intersects each agent in a group of agents. The slice associated with a given time step represents a set of locations where material should be deposited to fabricate a 3D object. Based on a set of such slices, a robot engine causes a robot to fabricate the 3D object.
    Type: Application
    Filed: September 8, 2020
    Publication date: December 24, 2020
    Inventors: Evan Patrick ATHERTON, David THOMASSON, Maurice Ugo CONTI, Heather KERRICK, Nicholas COTE, Hui LI
  • Patent number: 10853539
    Abstract: A robotic assembly cell is configured to generate a physical mesh of physical polygons based on a simulated mesh of simulated triangles. A control application configured to operate the assembly cell selects a simulated polygon in the simulated mesh and then causes a positioning robot in the cell to obtain a physical polygon that is similar to the simulated polygon. The positioning robot positions the polygon on the physical mesh, and a welding robot in the cell then welds the polygon to the mesh. The control application captures data that reflects how the physical polygon is actually positioned on the physical mesh, and then updates the simulated mesh to be geometrically consistent with the physical mesh. In doing so, the control application may execute a multi-objective solver to generate an updated simulated mesh that meets specific design criteria.
    Type: Grant
    Filed: May 26, 2017
    Date of Patent: December 1, 2020
    Assignee: AUTODESK, INC.
    Inventors: Evan Patrick Atherton, David Thomasson, Maurice Ugo Conti, Heather Kerrick, Nicholas Cote
  • Patent number: 10768606
    Abstract: An agent engine allocates a collection of agents to scan the surface of an object model. Each agent operates autonomously and implements particular behaviors based on the actions of nearby agents. Accordingly, the collection of agents exhibits swarm-like behavior. Over a sequence of time steps, the agents traverse the surface of the object model. Each agent acts to avoid other agents, thereby maintaining a relatively consistent distribution of agents across the surface of the object model over all time steps. At a given time step, the agent engine generates a slice through the object model that intersects each agent in a group of agents. The slice associated with a given time step represents a set of locations where material should be deposited to fabricate a 3D object. Based on a set of such slices, a robot engine causes a robot to fabricate the 3D object.
    Type: Grant
    Filed: June 2, 2017
    Date of Patent: September 8, 2020
    Assignee: AUTODESK, INC.
    Inventors: Evan Patrick Atherton, David Thomasson, Maurice Ugo Conti, Heather Kerrick, Nicholas Cote, Hui Li
  • Publication number: 20200264583
    Abstract: A robot is configured to assist an end-user with creative tasks. While the end-user modifies the work piece, the robot observes the modifications made by the end-user and determines one or more objectives that the end-user may endeavor to accomplish. The robot then determines a set of actions to perform that assist the end-user with accomplishing the objectives.
    Type: Application
    Filed: May 4, 2020
    Publication date: August 20, 2020
    Inventors: Evan Patrick Atherton, David Thomasson, Maurice Ugo Conti, Heather Kerrick
  • Patent number: 10708479
    Abstract: One embodiment of the present invention sets forth a technique for determining a location of an object that is being manipulated or processed by a robot. The technique includes capturing a digital image of the object while the object is disposed by the robot within an imaging space, wherein the digital image includes a direct view of the object and a reflected view of the object, detecting a visible feature of the object in the direct view and the visible feature of the object in the reflected view, and computing a first location of the visible feature in a first direction based on a position of the visible feature in the direct view. The technique further includes computing a second location of the visible feature in a second direction based on a position of the visible feature in the reflected view and causing the robot to move the object to a processing station based at least in part on the first location and the second location.
    Type: Grant
    Filed: July 16, 2019
    Date of Patent: July 7, 2020
    Assignee: Autodesk, Inc.
    Inventors: Evan Atherton, David Thomasson, Heather Kerrick, Maurice Conti
  • Patent number: 10642244
    Abstract: A robot is configured to assist an end-user with creative tasks. While the end-user modifies the work piece, the robot observes the modifications made by the end-user and determines one or more objectives that the end-user may endeavor to accomplish. The robot then determines a set of actions to perform that assist the end-user with accomplishing the objectives.
    Type: Grant
    Filed: December 19, 2016
    Date of Patent: May 5, 2020
    Assignee: Autodesk, Inc.
    Inventors: Evan Patrick Atherton, David Thomasson, Maurice Ugo Conti, Heather Kerrick
  • Patent number: 10579046
    Abstract: A robot system is configured to fabricate three-dimensional (3D) objects using closed-loop, computer vision-based control. The robot system initiates fabrication based on a set of fabrication paths along which material is to be deposited. During deposition of material, the robot system captures video data and processes that data to determine the specific locations where the material is deposited. Based on these locations, the robot system adjusts future deposition locations to compensate for deviations from the fabrication paths. Additionally, because the robot system includes a 6-axis robotic arm, the robot system can deposit material at any locations, along any pathway, or across any surface. Accordingly, the robot system is capable of fabricating a 3D object with multiple non-parallel, non-horizontal, and/or non-planar layers.
    Type: Grant
    Filed: April 24, 2017
    Date of Patent: March 3, 2020
    Assignee: AUTODESK, INC.
    Inventors: Evan Atherton, David Thomasson, Maurice Ugo Conti, Heather Kerrick, Nicholas Cote
  • Publication number: 20200030986
    Abstract: A motion capture setup records the movements of an operator, and a control engine then translates those movements into control signals for controlling a robot. The control engine may directly translate the operator movements into analogous movements to be performed by the robot, or the control engine may compute robot dynamics that cause a portion of the robot to mimic a corresponding portion of the operator.
    Type: Application
    Filed: September 30, 2019
    Publication date: January 30, 2020
    Inventors: Evan Patrick Atherton, David Thomasson, Maurice Ugo Conti, Heather Kerrick