Patents by Inventor Arthur William Juliani, JR.

Arthur William Juliani, JR. has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11900233
    Abstract: In example embodiments, a method of interactive imitation learning method is disclosed. An input is received from an input device. The input includes data describing a first set of example actions defining a behavior for a virtual character. Inverse reinforcement learning is used to estimate a reward function for the set of example actions. The reward function and the set of example actions is used as input to a reinforcement learning model to train a machine learning agent to mimic the behavior in a training environment. Based on a measure of failure of the training of the machine learning agent reaching a threshold, the training of the machine learning agent is paused to request a second set of example actions from the input device. The second set of example actions is used in addition to the first set of example actions to estimate a new reward function.
    Type: Grant
    Filed: May 26, 2022
    Date of Patent: February 13, 2024
    Assignee: Unity IPR ApS
    Inventors: Arthur William Juliani, Jr., Mohamed Marwan A. Mattar
  • Publication number: 20220355205
    Abstract: In example embodiments, a method of interactive imitation learning method is disclosed. An input is received from an input device. The input includes data describing a first set of example actions defining a behavior for a virtual character. Inverse reinforcement learning is used to estimate a reward function for the set of example actions. The reward function and the set of example actions is used as input to a reinforcement learning model to train a machine learning agent to mimic the behavior in a training environment. Based on a measure of failure of the training of the machine learning agent reaching a threshold, the training of the machine learning agent is paused to request a second set of example actions from the input device. The second set of example actions is used in addition to the first set of example actions to estimate a new reward function.
    Type: Application
    Filed: May 26, 2022
    Publication date: November 10, 2022
    Inventors: Arthur William Juliani, JR., Mohamed Marwan A. Mattar
  • Patent number: 11369879
    Abstract: In example embodiments, a method of interactive imitation learning method is disclosed. An input is received from an input device. The input includes data describing a first set of example actions defining a behavior for a virtual character. Inverse reinforcement learning is used to estimate a reward function for the set of example actions. The reward function and the set of example actions is used as input to a reinforcement learning model to train a machine learning agent to mimic the behavior in a training environment. Based on a measure of failure of the training of the machine learning agent reaching a threshold, the training of the machine learning agent is paused to request a second set of example actions from the input device. The second set of example actions is used in addition to the first set of example actions to estimate a new reward function.
    Type: Grant
    Filed: October 18, 2019
    Date of Patent: June 28, 2022
    Assignee: Unity IPR ApS
    Inventors: Arthur William Juliani, Jr., Mohamed Marwan A. Mattar
  • Publication number: 20200122039
    Abstract: A method of behavior generation is disclosed. Planning state data in a planning domain language format is received and a state description and an associated action description based on the planning state data are generated. The state description and the associated action description are parsed into a series of tokens for a machine learning encoded state and associated ML encoded action. The series of tokens describe the state and the action. The ML encoded state and ML encoded action is processed with a recurrent neural network to generate an estimate of a value of the state description and the action description. Output of the RNN is taken as input into a neural network to generate a value estimate for a state-action pair. A plan that includes a plurality of sequential actions for an agent is generated. The plurality of sequential actions is chosen based on at least the value estimate.
    Type: Application
    Filed: October 22, 2019
    Publication date: April 23, 2020
    Inventors: Nicolas Francois Xavier Meuleau, Vincent-Pierre Serge Mary Berges, Amir Pascal Ebrahimi, Arthur William Juliani, JR., Trevor Joseph Santarra
  • Publication number: 20200122040
    Abstract: In example embodiments, a method of interactive imitation learning method is disclosed. An input is received from an input device. The input includes data describing a first set of example actions defining a behavior for a virtual character. Inverse reinforcement learning is used to estimate a reward function for the set of example actions. The reward function and the set of example actions is used as input to a reinforcement learning model to train a machine learning agent to mimic the behavior in a training environment. Based on a measure of failure of the training of the machine learning agent reaching a threshold, the training of the machine learning agent is paused to request a second set of example actions from the input device. The second set of example actions is used in addition to the first set of example actions to estimate a new reward function.
    Type: Application
    Filed: October 18, 2019
    Publication date: April 23, 2020
    Inventors: Arthur William Juliani, JR., Mohamed Marwan A. Mattar