Patents by Inventor Yihao Feng

Yihao Feng has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20250139411
    Abstract: Embodiments described herein provide a large language model (LLM) based AI agent that adopts Monte-Carlo Tree Search (MCTS) to execute a task. The LLM is prompted with a task description and it responds with its first attempted list of actions. Based on the success or failure of the first attempt, the LLM is prompted with an updated prompt which includes feedback from the first attempt based on a determined reward. The prompt may include a relative “score” for each action taken at each step. A numeric score may be mapped to a set of pre-defined text labels, such as “high” or “low” value putting the score in a form more suited for an LLM prompt. In this way, the LLM is iteratively given prompts which are updated with the scores from each action taken at each previous iterations so that it traverses different paths on the tree in each iteration.
    Type: Application
    Filed: October 31, 2023
    Publication date: May 1, 2025
    Inventors: Rithesh Murthy, Shelby Heinecke, Juan Carlos Niebles Duque, Zhiwei Liu, Le Xue, Weiran Yao, Yihao Feng, Zeyuan Chen, Akash Gokul, Devansh Arpit, Ran Xu, Lik Mui, Huan Wang, Caiming Xiong, Silvio Savarese
  • Publication number: 20250086402
    Abstract: Methods, systems, apparatuses, devices, and computer program products are described. A flow generation service may receive a natural language input that indicates instructions for automating a task according to a first process flow. Using a large language model (LLM), the flow generation service may decompose the natural language input into a set of elements (e.g., logical actions) and connectors, where the LLM may be trained on first metadata corresponding to a second process flow that is created manually by a user. In addition, using the LLM, the flow generation service may generate second metadata corresponding to each of the set of elements based on decomposing the natural language input. The flow generation service may sequence and merge the set of elements to generate the first process flow. In some examples, the flow generation service may send, for display to a user interface of a user device, the first process flow.
    Type: Application
    Filed: January 17, 2024
    Publication date: March 13, 2025
    Inventors: Ran Xu, Zeyuan Chen, Yihao Feng, Krithika Ramakrishnan, Congying Xia, Juan Carlos Niebles Duque, Vetter Serdikova, Huan Wang, Yuxi Zhang, Kexin Xie, Donglin Hu, Bo Wang, Ajaay Ravi, Matthew David Trepina, Sam Bailey, Abhishek Das, Yuliya Feldman, Pawan Agarwal
  • Publication number: 20250068901
    Abstract: Embodiments described herein provide a diffusion-based framework that is trained on a dataset with limited text labels, to generate a distribution of data samples in the dataset given a specific text description label. Specifically, firstly, unlabeled data is used to train the diffusion model to generate a data distribution of data samples given a specific text description label. Then text-labeled data samples are used to finetune the diffusion model to generate data distribution given a specific text description label, thus enhancing controllability of training.
    Type: Application
    Filed: January 25, 2024
    Publication date: February 27, 2025
    Inventors: Shiyu Wang, Yihao Feng, Tian Lan, Ning Yu, Yu Bai, Ran Xu, Huan Wang, Caiming Xiong, Silvio Savarese
  • Publication number: 20250053793
    Abstract: Embodiments described herein provide a method of predicting an action by a plurality of language model augmented agents (LAAs). In at least one embodiment, a controller receives a task instruction to be performed using an environment. The controller receives an observation of a first state from the environment. The controller selects a LAA from the plurality of LAAs based on the task instruction and the observation. The controller obtains an output from the selected LAA generated using an input combining the task instruction, the observation, and an LAA-specific prompt template. The controller determines the action based on the output. The controller causes the action to be performed on the environment thereby causing the first state of the environment to change to a second state.
    Type: Application
    Filed: October 25, 2023
    Publication date: February 13, 2025
    Inventors: Zhiwei Liu, Weiran Yao, Jianguo Zhang, Le Xue, Shelby Heinecke, Rithesh Murthy, Yihao Feng, Zeyuan Chen, Juan Carlos Niebles Duque, Devansh Arpit, Ran Xu, Lik Mui, Huan Wang, Caiming Xiong, Silvio Savarese
  • Publication number: 20250045567
    Abstract: Embodiments described herein provide for optimizing a language model (LM) agent. In at least one embodiment, and LM agent comprises an “actor” LM and a “retrospective LM which provides reflections on attempts by the actor LM. The reflections are used to update subsequent prompts to the actor LM. Optimizing the LM agent comprises fine-tuning parameters of the retrospective LM while keeping parameters of the actor LM frozen. A gradient may be determined by a change in reward from the environment based on actions taken by the actor LM with and without a reflection of the retrospective LM. Using this gradient, parameters of the retrospective LM may be updated via backpropagation.
    Type: Application
    Filed: October 31, 2023
    Publication date: February 6, 2025
    Inventors: Weiran Yao, Shelby Heinecke, Juan Carlos Niebles Duque, Zhiwei Liu, Yihao Feng, Le Xue, Rithesh Murthy, Zeyuan Chen, Jianguo Zhang, Devansh Arpit, Ran Xu, Lik Mui, Huan Wang, Caiming Xiong, Silvio Savarese
  • Publication number: 20240386623
    Abstract: Embodiments described herein provide a method of image generation. The method includes a fixed diffusion model, and a trainable diffusion model. The fixed diffusion model may be pretrained on a large training corpus. The trainable diffusion model may be used to control the image generation of the fixed diffusion model by modifying internal representations of the fixed diffusion model. A task instruction may be provided in addition to a text prompt, and the task instruction may guide the trainable diffusion model together with the visual conditions. The visual conditions may be adapted according to the task instruction. During training, a fixed number of task instructions may be used. At inference, unseen task instructions may be used by combining convolutional kernels of the visual condition adapter.
    Type: Application
    Filed: September 29, 2023
    Publication date: November 21, 2024
    Inventors: Ning YU, Can QIN, Shu ZHANG, Yihao FENG, Xinyi YANG, Ran XU
  • Publication number: 20240303882
    Abstract: Embodiments described herein provide a feedback based instructional image editing framework that employs a diffusion process to follow user instruction for image editing. A diffusion model is fine-tuned using a reward model, which may be trained via human annotation. The training of the reward model may be done by having the image editing model output a number of images, which a human annotator ranks based on their alignment with the original image and a given instruction.
    Type: Application
    Filed: July 12, 2023
    Publication date: September 12, 2024
    Inventors: Shu Zhang, Xinyi Yang, Yihao Feng, Ran Xu, Ning Yu, Chia-Chih Chen
  • Publication number: 20210374524
    Abstract: Some embodiments of the current disclosure disclose methods and systems for detecting out-of-distribution (ODD) data. For example, a method for detecting ODD data includes obtaining, at a neural network composed of a plurality of layers, a set of training data generated according to a distribution. Further, the method comprises generating, via a processor, a feature map by combining mapping functions corresponding to the plurality of layers into a vector of mapping function elements and mapping, by the feature map, the set of training data to a set of feature space training data in a feature space. Further, the method comprises identifying, via the processor, a hyper-ellipsoid in the feature space enclosing the feature space training data based on the generated feature map. In addition, the method comprises determining, via the processor, the first test data sample is OOD data when a mapped first test data sample in the feature space is outside the hyper-ellipsoid.
    Type: Application
    Filed: November 13, 2020
    Publication date: December 2, 2021
    Inventors: Yihao Feng, Caiming Xiong