Patents by Inventor Seyed Mohammad Khansari Zadeh

Seyed Mohammad Khansari Zadeh has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20240118667
    Abstract: Implementations disclosed herein relate to mitigating the reality gap through training a simulation-to-real machine learning model (“Sim2Real” model) using a vision-based robot task machine learning model. The vision-based robot task machine learning model can be, for example, a reinforcement learning (“RL”) neural network model (RL-network), such as an RL-network that represents a Q-function.
    Type: Application
    Filed: May 15, 2020
    Publication date: April 11, 2024
    Inventors: Kanishka Rao, Chris Harris, Julian Ibarz, Alexander Irpan, Seyed Mohammad Khansari Zadeh, Sergey Levine
  • Patent number: 11887363
    Abstract: Training a machine learning model (e.g., a neural network model such as a convolutional neural network (CNN) model) so that, when trained, the model can be utilized in processing vision data (e.g., from a vision component of a robot), that captures an object, to generate a rich object-centric embedding for the vision data. The generated embedding can enable differentiation of even subtle variations of attributes of the object captured by the vision data.
    Type: Grant
    Filed: September 27, 2019
    Date of Patent: January 30, 2024
    Assignee: GOOGLE LLC
    Inventors: Soeren Pirk, Yunfei Bai, Pierre Sermanet, Seyed Mohammad Khansari Zadeh, Harrison Lynch
  • Patent number: 11872699
    Abstract: Generating a robot control policy that regulates both motion control and interaction with an environment and/or includes a learned potential function and/or dissipative field. Some implementations relate to resampling temporally distributed data points to generate spatially distributed data points, and generating the control policy using the spatially distributed data points. Some implementations additionally or alternatively relate to automatically determining a potential gradient for data points, and generating the control policy using the automatically determined potential gradient. Some implementations additionally or alternatively relate to determining and assigning a prior weight to each of the data points of multiple groups, and generating the control policy using the weights.
    Type: Grant
    Filed: January 13, 2023
    Date of Patent: January 16, 2024
    Assignee: GOOGLE LLC
    Inventor: Seyed Mohammad Khansari Zadeh
  • Publication number: 20230381970
    Abstract: Implementations described herein relate to training and refining robotic control policies using imitation learning techniques. A robotic control policy can be initially trained based on human demonstrations of various robotic tasks. Further, the robotic control policy can be refined based on human interventions while a robot is performing a robotic task. In some implementations, the robotic control policy may determine whether the robot will fail in performance of the robotic task, and prompt a human to intervene in performance of the robotic task. In additional or alternative implementations, a representation of the sequence of actions can be visually rendered for presentation to the human can proactively intervene in performance of the robotic task.
    Type: Application
    Filed: August 11, 2023
    Publication date: November 30, 2023
    Inventors: Seyed Mohammad Khansari Zadeh, Eric Jang, Daniel Lam, Daniel Kappler, Matthew Bennice, Brent Austin, Yunfei Bai, Sergey Levine, Alexander Irpan, Nicolas Sievers, Chelsea Finn
  • Patent number: 11772272
    Abstract: Implementations described herein relate to training and refining robotic control policies using imitation learning techniques. A robotic control policy can be initially trained based on human demonstrations of various robotic tasks. Further, the robotic control policy can be refined based on human interventions while a robot is performing a robotic task. In some implementations, the robotic control policy may determine whether the robot will fail in performance of the robotic task, and prompt a human to intervene in performance of the robotic task. In additional or alternative implementations, a representation of the sequence of actions can be visually rendered for presentation to the human can proactively intervene in performance of the robotic task.
    Type: Grant
    Filed: March 16, 2021
    Date of Patent: October 3, 2023
    Assignee: GOOGLE LLC
    Inventors: Seyed Mohammad Khansari Zadeh, Eric Jang, Daniel Lam, Daniel Kappler, Matthew Bennice, Brent Austin, Yunfei Bai, Sergey Levine, Alexander Irpan, Nicolas Sievers, Chelsea Finn
  • Publication number: 20230150126
    Abstract: Generating a robot control policy that regulates both motion control and interaction with an environment and/or includes a learned potential function and/or dissipative field. Some implementations relate to resampling temporally distributed data points to generate spatially distributed data points, and generating the control policy using the spatially distributed data points. Some implementations additionally or alternatively relate to automatically determining a potential gradient for data points, and generating the control policy using the automatically determined potential gradient. Some implementations additionally or alternatively relate to determining and assigning a prior weight to each of the data points of multiple groups, and generating the control policy using the weights.
    Type: Application
    Filed: January 13, 2023
    Publication date: May 18, 2023
    Inventor: Seyed Mohammad Khansari Zadeh
  • Patent number: 11607802
    Abstract: Generating and utilizing action image(s) that represent a candidate pose (e.g., a candidate end effector pose), in determining whether to utilize the candidate pose in performance of a robotic task. The action image(s) and corresponding current image(s) can be processed, using a trained critic network, to generate a value that indicates a probability of success of the robotic task if component(s) of the robot are traversed to the particular pose. When the value satisfies one or more conditions (e.g., satisfies a threshold), the robot can be controlled to cause the component(s) to traverse to the particular pose in performing the robotic task.
    Type: Grant
    Filed: May 28, 2020
    Date of Patent: March 21, 2023
    Assignee: X DEVELOPMENT LLC
    Inventors: Seyed Mohammad Khansari Zadeh, Daniel Kappler, Jianlan Luo, Jeffrey Bingham, Mrinal Kalakrishnan
  • Patent number: 11607807
    Abstract: Training and/or use of a machine learning model for placement of an object secured by an end effector of a robot. A trained machine learning model can be used to process: (1) a current image, captured by a vision component of a robot, that captures an end effector securing an object; (2) a candidate end effector action that defines a candidate motion of the end effector; and (3) a target placement input that indicates a target placement location for the object. Based on the processing, a prediction can be generated that indicates likelihood of successful placement of the object in the target placement location with application of the motion defined by the candidate end effector action. At many iterations, the candidate end effector action with the highest probability is selected and control commands provided to cause the end effector to move in conformance with the corresponding end effector action.
    Type: Grant
    Filed: April 14, 2021
    Date of Patent: March 21, 2023
    Assignee: X DEVELOPMENT LLC
    Inventors: Seyed Mohammad Khansari Zadeh, Mrinal Kalakrishnan, Paul Wohlhart
  • Patent number: 11554485
    Abstract: Generating a robot control policy that regulates both motion control and interaction with an environment and/or includes a learned potential function and/or dissipative field. Some implementations relate to resampling temporally distributed data points to generate spatially distributed data points, and generating the control policy using the spatially distributed data points. Some implementations additionally or alternatively relate to automatically determining a potential gradient for data points, and generating the control policy using the automatically determined potential gradient. Some implementations additionally or alternatively relate to determining and assigning a prior weight to each of the data points of multiple groups, and generating the control policy using the weights.
    Type: Grant
    Filed: July 25, 2019
    Date of Patent: January 17, 2023
    Assignee: X DEVELOPMENT LLC
    Inventor: Seyed Mohammad Khansari Zadeh
  • Patent number: 11554483
    Abstract: Deep machine learning methods and apparatus, some of which are related to determining a grasp outcome prediction for a candidate grasp pose of an end effector of a robot. Some implementations are directed to training and utilization of both a geometry network and a grasp outcome prediction network. The trained geometry network can be utilized to generate, based on two-dimensional or two-and-a-half-dimensional image(s), geometry output(s) that are: geometry-aware, and that represent (e.g., high-dimensionally) three-dimensional features captured by the image(s). In some implementations, the geometry output(s) include at least an encoding that is generated based on a trained encoding neural network trained to generate encodings that represent three-dimensional features (e.g., shape). The trained grasp outcome prediction network can be utilized to generate, based on applying the geometry output(s) and additional data as input(s) to the network, a grasp outcome prediction for a candidate grasp pose.
    Type: Grant
    Filed: November 10, 2020
    Date of Patent: January 17, 2023
    Assignee: GOOGLE LLC
    Inventors: James Davidson, Xinchen Yan, Yunfei Bai, Honglak Lee, Abhinav Gupta, Seyed Mohammad Khansari Zadeh, Arkanath Pathak, Jasmine Hsu
  • Publication number: 20220410380
    Abstract: Utilizing an initial set of offline positive-only robotic demonstration data for pre-training an actor network and a critic network for robotic control, followed by further training of the networks based on online robotic episodes that utilize the network(s). Implementations enable the actor network to be effectively pre-trained, while mitigating occurrences of and/or the extent of forgetting when further trained based on episode data. Implementations additionally or alternatively enable the actor network to be trained to a given degree of effectiveness in fewer training steps. In various implementations, one or more adaptation techniques are utilized in performing the robotic episodes and/or in performing the robotic training. The adaptation techniques can each, individually, result in one or more corresponding advantages and, when used in any combination, the corresponding advantages can accumulate.
    Type: Application
    Filed: June 17, 2022
    Publication date: December 29, 2022
    Inventors: Yao Lu, Mengyuan Yan, Seyed Mohammad Khansari Zadeh, Alexander Herzog, Eric Jang, Karol Hausman, Yevgen Chebotar, Sergey Levine, Alexander Irpan
  • Publication number: 20220331962
    Abstract: Training and/or using a machine learning model for performing robotic tasks is disclosed herein. In many implementations, an environment-conditioned action sequence prediction model is used to determine a set of actions as well as a corresponding particular order for the actions for the robot to perform to complete the task. In many implementations, each action in the set of actions has a corresponding action network used to control the robot in performing the action.
    Type: Application
    Filed: September 9, 2020
    Publication date: October 20, 2022
    Inventors: Soeren Pirk, Seyed Mohammad Khansari Zadeh, Karol Hausman, Alexander Toshev
  • Publication number: 20220297303
    Abstract: Implementations described herein relate to training and refining robotic control policies using imitation learning techniques. A robotic control policy can be initially trained based on human demonstrations of various robotic tasks. Further, the robotic control policy can be refined based on human interventions while a robot is performing a robotic task. In some implementations, the robotic control policy may determine whether the robot will fail in performance of the robotic task, and prompt a human to intervene in performance of the robotic task. In additional or alternative implementations, a representation of the sequence of actions can be visually rendered for presentation to the human can proactively intervene in performance of the robotic task.
    Type: Application
    Filed: March 16, 2021
    Publication date: September 22, 2022
    Inventors: Seyed Mohammad Khansari Zadeh, Eric Jang, Daniel Lam, Daniel Kappler, Matthew Bennice, Brent Austin, Yunfei Bai, Sergey Levine, Alexander Irpan, Nicolas Sievers, Chelsea Finn
  • Publication number: 20220105624
    Abstract: Techniques are disclosed that enable training a meta-learning model, for use in causing a robot to perform a task, using imitation learning as well as reinforcement learning. Some implementations relate to training the meta-learning model using imitation learning based on one or more human guided demonstrations of the task. Additional or alternative implementations relate to training the meta-learning model using reinforcement learning based on trials of the robot attempting to perform the task. Further implementations relate to using the trained meta-learning model to few shot (or one shot) learn a new task based on a human guided demonstration of the new task.
    Type: Application
    Filed: January 23, 2020
    Publication date: April 7, 2022
    Inventors: Mrinal Kalakrishnan, Yunfei Bai, Paul Wohlhart, Eric Jang, Chelsea Finn, Seyed Mohammad Khansari Zadeh, Sergey Levine, Allan Zhou, Alexander Herzog, Daniel Kappler
  • Publication number: 20220076099
    Abstract: Methods, systems, and apparatus, including computer programs encoded on computer storage media, for controlling an agent. One of the methods includes controlling the agent using a policy neural network that processes a policy input that includes (i) a current observation, (ii) a goal observation, and (iii) a selected latent plan to generate a current action output that defines an action to be performed in response to the current observation.
    Type: Application
    Filed: February 19, 2020
    Publication date: March 10, 2022
    Inventors: Pierre Sermanet, Seyed Mohammad Khansari Zadeh, Harrison Corey Lynch
  • Publication number: 20210334599
    Abstract: Training a machine learning model (e.g., a neural network model such as a convolutional neural network (CNN) model) so that, when trained, the model can be utilized in processing vision data (e.g., from a vision component of a robot), that captures an object, to generate a rich object-centric embedding for the vision data. The generated embedding can enable differentiation of even subtle variations of attributes of the object captured by the vision data.
    Type: Application
    Filed: September 27, 2019
    Publication date: October 28, 2021
    Inventors: Soeren Pirk, Yunfei Bai, Pierre Sermanet, Seyed Mohammad Khansari Zadeh, Harrison Lynch
  • Publication number: 20210229276
    Abstract: Training and/or use of a machine learning model for placement of an object secured by an end effector of a robot. A trained machine learning model can be used to process: (1) a current image, captured by a vision component of a robot, that captures an end effector securing an object; (2) a candidate end effector action that defines a candidate motion of the end effector; and (3) a target placement input that indicates a target placement location for the object. Based on the processing, a prediction can be generated that indicates likelihood of successful placement of the object in the target placement location with application of the motion defined by the candidate end effector action. At many iterations, the candidate end effector action with the highest probability is selected and control commands provided to cause the end effector to move in conformance with the corresponding end effector action.
    Type: Application
    Filed: April 14, 2021
    Publication date: July 29, 2021
    Inventors: Seyed Mohammad Khansari Zadeh, Mrinal Kalakrishnan, Paul Wohlhart
  • Patent number: 11007642
    Abstract: Training and/or use of a machine learning model for placement of an object secured by an end effector of a robot. A trained machine learning model can be used to process: (1) a current image, captured by a vision component of a robot, that captures an end effector securing an object; (2) a candidate end effector action that defines a candidate motion of the end effector; and (3) a target placement input that indicates a target placement location for the object. Based on the processing, a prediction can be generated that indicates likelihood of successful placement of the object in the target placement location with application of the motion defined by the candidate end effector action. At many iterations, the candidate end effector action with the highest probability is selected and control commands provided to cause the end effector to move in conformance with the corresponding end effector action.
    Type: Grant
    Filed: October 23, 2018
    Date of Patent: May 18, 2021
    Assignee: X DEVELOPMENT LLC
    Inventors: Seyed Mohammad Khansari Zadeh, Mrinal Kalakrishnan, Paul Wohlhart
  • Publication number: 20210101286
    Abstract: Implementations relate to training a point cloud prediction model that can be utilized to process a single-view two-and-a-half-dimensional (2.5D) observation of an object, to generate a domain-invariant three-dimensional (3D) representation of the object. Implementations additionally or alternatively relate to utilizing the domain-invariant 3D representation to train a robotic manipulation policy model using, as at least part of the input to the robotic manipulation policy model during training, the domain-invariant 3D representations of simulated objects to be manipulated. Implementations additionally or alternatively relate to utilizing the trained robotic manipulation policy model in control of a robot based on output generated by processing generated domain-invariant 3D representations utilizing the robotic manipulation policy model.
    Type: Application
    Filed: February 28, 2020
    Publication date: April 8, 2021
    Inventors: Honglak Lee, Xinchen Yan, Soeren Pirk, Yunfei Bai, Seyed Mohammad Khansari Zadeh, Yuanzheng Gong, Jasmine Hsu
  • Publication number: 20210078167
    Abstract: Generating and utilizing action image(s) that represent a candidate pose (e.g., a candidate end effector pose), in determining whether to utilize the candidate pose in performance of a robotic task. The action image(s) and corresponding current image(s) can be processed, using a trained critic network, to generate a value that indicates a probability of success of the robotic task if component(s) of the robot are traversed to the particular pose. When the value satisfies one or more conditions (e.g., satisfies a threshold), the robot can be controlled to cause the component(s) to traverse to the particular pose in performing the robotic task.
    Type: Application
    Filed: May 28, 2020
    Publication date: March 18, 2021
    Inventors: Seyed Mohammad Khansari Zadeh, Daniel Kappler, Jianlan Luo, Jeffrey Bingham, Mrinal Kalakrishnan