Patents by Inventor Ruiyi Zhang
Ruiyi Zhang has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Publication number: 20250103813Abstract: This disclosure describes one or more implementations of systems, non-transitory computer-readable media, and methods that train a named entity recognition (NER) model with noisy training data through a self-cleaning discriminator model. For example, the disclosed systems utilize a self-cleaning guided denoising framework to improve NER learning on noisy training data via a guidance training set. In one or more implementations, the disclosed systems utilize, within the denoising framework, an auxiliary discriminator model to correct noise in the noisy training data while training an NER model through the noisy training data. For example, while training the NER model to predict labels from the noisy training data, the disclosed systems utilize a discriminator model to detect noisy NER labels and reweight the noisy NER labels provided for training in the NER model.Type: ApplicationFiled: September 22, 2023Publication date: March 27, 2025Inventors: Ruiyi Zhang, Zhendong Chu, Vlad Morariu, Tong Yu, Rajiv Jain, Nedim Lipka, Jiuxiang Gu
-
Publication number: 20250078200Abstract: The present disclosure relates to systems, non-transitory computer-readable media, and methods that implement a neural network framework for interactive multi-round image generation from natural language inputs. Specifically, the disclosed systems provide an intelligent framework (i.e., a text-based interactive image generation model) that facilitates a multi-round image generation and editing workflow that comports with arbitrary input text and synchronous interaction. In particular embodiments, the disclosed systems utilize natural language feedback for conditioning a generative neural network that performs text-to-image generation and text-guided image modification. For example, the disclosed systems utilize a trained model to inject textual features from natural language feedback into a unified joint embedding space for generating text-informed style vectors. In turn, the disclosed systems can generate an image with semantically meaningful features that map to the natural language feedback.Type: ApplicationFiled: November 19, 2024Publication date: March 6, 2025Inventors: Ruiyi Zhang, Yufan Zhou, Christopher Tensmeyer, Jiuxiang Gu, Tong Yu, Tong Sun
-
Publication number: 20250028751Abstract: Dialogue skeleton assisted prompt transfer for dialogue summarization techniques are described that support training of a language model to perform dialogue summarization in a few-shot scenario. A processing device, for instance, receives a training dataset that includes training dialogues. The processing device then generates dialogue skeletons based on the training dialogues using one or more perturbation-based probes. The processing device trains a language model using prompt transfer between a source task, e.g., dialogue state tracking, and a target task, e.g., dialogue summarization, using the dialogue skeletons as supervision. The processing device then receives an input dialogue and uses the trained language model to generate a summary of the input dialogue.Type: ApplicationFiled: July 20, 2023Publication date: January 23, 2025Applicant: Adobe Inc.Inventors: Tong Yu, Kaige Xie, Haoliang Wang, Junda Wu, Handong Zhao, Ruiyi Zhang, Kanak Vivek Mahadik, Ani Nenkova
-
Publication number: 20250005289Abstract: Dialogue state aware dialogue summarization techniques are described that enable generation of dialogue summaries from target domains with limited training data. A content processing system, for instance, generates one or more clusters based on training dialogues from one or more source domains. The clusters represent domain-specific features of the training dialogues and are further based on dialogue states of the training dialogues. The content processing system trains a machine learning model to generate summaries of dialogues by using the one or more clusters as prefixes in a prefix-tuning approach. The content processing system receives an input that includes a dialogue from a target domain. The content processing system generates an input prompt based on the dialogue and the one or more clusters, and the model generates a summary of the dialogue based on the input prompt.Type: ApplicationFiled: June 28, 2023Publication date: January 2, 2025Applicant: Adobe Inc.Inventors: Haoliang Wang, Kaige Xie, Tong Yu, Junda Wu, Handong Zhao, Ruiyi Zhang, Kanak Vivek Mahadik, Ani Nenkova
-
Publication number: 20240427998Abstract: Contextual query generation techniques are described that enable generation of a contextual query for output to a question-answering (QA) model. A content processing system, for instance, configures a language model using in-context learning to generate queries based on semantic contexts of input documents, e.g., based on one or more linguistic cues from text of the input documents. The content processing system receives an input that includes a document having text and a reference query. The content processing system leverages the language model to generate a contextual query based on a semantic context of the text of the document and the reference query. The content processing system then outputs the contextual query and the document to a QA model. Using the QA model, the content processing system generates a response as an answer to the contextual query based on the contextual query and the document.Type: ApplicationFiled: June 22, 2023Publication date: December 26, 2024Applicant: Adobe Inc.Inventors: Haoliang Wang, Tong Yu, Sungchul Kim, Ruiyi Zhang, Paiheng Xu, Junda Wu, Handong Zhao, Ani Nenkova
-
Publication number: 20240386621Abstract: Techniques and systems for training and/or implementing a text-to-image generation model are provided. A pre-trained multimodal model is leveraged for avoiding slower and more labor-intensive methodologies for training a text-to-image generation model. Accordingly, images without associated text (i.e., bare images) are provided to the pre-trained multimodal model so that it can produce generated text-image pairs. The generated text-image pairs are provided to the text-to-image generation model for training and/or implementing the text-to-image generation model.Type: ApplicationFiled: May 17, 2023Publication date: November 21, 2024Applicant: Adobe Inc.Inventors: Ruiyi Zhang, Yufan Zhou, Tong Yu, Tong Sun, Rajiv Jain, Jiuxiang Gu, Christopher Alan Tensmeyer
-
Patent number: 12148119Abstract: The present disclosure relates to systems, non-transitory computer-readable media, and methods that implement a neural network framework for interactive multi-round image generation from natural language inputs. Specifically, the disclosed systems provide an intelligent framework (i.e., a text-based interactive image generation model) that facilitates a multi-round image generation and editing workflow that comports with arbitrary input text and synchronous interaction. In particular embodiments, the disclosed systems utilize natural language feedback for conditioning a generative neural network that performs text-to-image generation and text-guided image modification. For example, the disclosed systems utilize a trained model to inject textual features from natural language feedback into a unified joint embedding space for generating text-informed style vectors. In turn, the disclosed systems can generate an image with semantically meaningful features that map to the natural language feedback.Type: GrantFiled: January 14, 2022Date of Patent: November 19, 2024Assignee: Adobe Inc.Inventors: Ruiyi Zhang, Yufan Zhou, Christopher Tensmeyer, Jiuxiang Gu, Tong Yu, Tong Sun
-
Publication number: 20240273296Abstract: Embodiments of the technology described herein describe a machine classifier capable of continually learning new classes through a continual few-shot learning approach. A natural language processing (NLP) machine classifier may initially be trained to identify a plurality of other classes through a conventional training process. In order to learn a new class, natural-language training data for a new class is generated. The training data for the new class may be few-shot training data. The training also uses synthetic training data that represents each of the plurality of other classes. The synthetic training data may be generated through a model inversion of the original classifier. The synthetic training data and the natural-language training data are used to retrain the NLP classifier to identify text in the plurality of other classes and the new class using.Type: ApplicationFiled: April 3, 2024Publication date: August 15, 2024Inventors: Sungchul KIM, Subrata MITRA, Ruiyi Zhang, Rui Wang, Handong ZHAO, Tong YU
-
Publication number: 20240232525Abstract: Systems and methods for document classification are described. Embodiments of the present disclosure generate classification data for a plurality of samples using a neural network trained to identify a plurality of known classes; select a set of samples for annotation from the plurality of samples using an open-set metric based on the classification data, wherein the annotation includes an unknown class; and train the neural network to identify the unknown class based on the annotation of the set of samples.Type: ApplicationFiled: October 24, 2022Publication date: July 11, 2024Inventors: Rajiv Bhawanji Jain, Michelle Yuan, Vlad Ion Morariu, Ani Nenkova Nenkova, Smitha Bangalore Naresh, Nikolaos Barmpalios, Ruchi Deshpande, Ruiyi Zhang, Jiuxiang Gu, Varun Manjunatha, Nedim Lipka, Andrew Marc Greene
-
Patent number: 11995403Abstract: Embodiments of the technology described herein describe a machine classifier capable of continually learning new classes through a continual few-shot learning approach. A natural language processing (NLP) machine classifier may initially be trained to identify a plurality of other classes through a conventional training process. In order to learn a new class, natural-language training data for a new class is generated. The training data for the new class may be few-shot training data. The training also uses synthetic training data that represents each of the plurality of other classes. The synthetic training data may be generated through a model inversion of the original classifier. The synthetic training data and the natural-language training data are used to retrain the NLP classifier to identify text in the plurality of other classes and the new class using.Type: GrantFiled: November 11, 2021Date of Patent: May 28, 2024Assignee: ADOBE INC.Inventors: Sungchul Kim, Subrata Mitra, Ruiyi Zhang, Rui Wang, Handong Zhao, Tong Yu
-
Publication number: 20240135096Abstract: Systems and methods for document classification are described. Embodiments of the present disclosure generate classification data for a plurality of samples using a neural network trained to identify a plurality of known classes; select a set of samples for annotation from the plurality of samples using an open-set metric based on the classification data, wherein the annotation includes an unknown class; and train the neural network to identify the unknown class based on the annotation of the set of samples.Type: ApplicationFiled: October 23, 2022Publication date: April 25, 2024Inventors: Rajiv Bhawanji Jain, Michelle Yuan, Vlad Ion Morariu, Ani Nenkova Nenkova, Smitha Bangalore Naresh, Nikolaos Barmpalios, Ruchi Deshpande, Ruiyi Zhang, Jiuxiang Gu, Varun Manjunatha, Nedim Lipka, Andrew Marc Greene
-
Publication number: 20240086457Abstract: A content analysis system provides content understanding for a content item using an attention aware multi-modal model. Given a content item, feature extractors extract features from content components of the content item in which the content components comprise multiple modalities. A cross-modal attention encoder of the attention aware multi-modal model generates an embedding of the content item using features extracted from the content components. A decoder of the attention aware multi-modal model generates an action-reason statement using the embedding of the content item from the cross-modal attention encoder.Type: ApplicationFiled: September 14, 2022Publication date: March 14, 2024Inventors: Yaman KUMAR, Vaibhav AHLAWAT, Ruiyi ZHANG, Milan AGGARWAL, Ganesh Karbhari PALWE, Balaji KRISHNAMURTHY, Varun KHURANA
-
Publication number: 20230368003Abstract: The technology described herein is directed to an adaptive sparse attention pattern that is learned during fine-tuning and deployed in a machine-learning model. In aspects, a row or a column in an attention matrix with an importance score for a task that is above a threshold importance score is identified. The important row or the column is included in an adaptive attention pattern used with a machine-learning model having a self-attention operation. In response to an input, a task-specific inference is generated for the input using the machine-learning model with the adaptive attention pattern.Type: ApplicationFiled: May 10, 2022Publication date: November 16, 2023Inventors: Jiuxiang Gu, Zihan Wang, Jason Wen Yong Kuen, Handong Zhao, Vlad Ion Morariu, Ruiyi Zhang, Ani Nenkova Nenkova, Tong Sun
-
Publication number: 20230230198Abstract: The present disclosure relates to systems, non-transitory computer-readable media, and methods that implement a neural network framework for interactive multi-round image generation from natural language inputs. Specifically, the disclosed systems provide an intelligent framework (i.e., a text-based interactive image generation model) that facilitates a multi-round image generation and editing workflow that comports with arbitrary input text and synchronous interaction. In particular embodiments, the disclosed systems utilize natural language feedback for conditioning a generative neural network that performs text-to-image generation and text-guided image modification. For example, the disclosed systems utilize a trained model to inject textual features from natural language feedback into a unified joint embedding space for generating text-informed style vectors. In turn, the disclosed systems can generate an image with semantically meaningful features that map to the natural language feedback.Type: ApplicationFiled: January 14, 2022Publication date: July 20, 2023Inventors: Ruiyi Zhang, Yufan Zhou, Christopher Tensmeyer, Jiuxiang Gu, Tong Yu, Tong Sun
-
Publication number: 20230143721Abstract: Embodiments of the technology described herein describe a machine classifier capable of continually learning new classes through a continual few-shot learning approach. A natural language processing (NLP) machine classifier may initially be trained to identify a plurality of other classes through a conventional training process. In order to learn a new class, natural-language training data for a new class is generated. The training data for the new class may be few-shot training data. The training also uses synthetic training data that represents each of the plurality of other classes. The synthetic training data may be generated through a model inversion of the original classifier. The synthetic training data and the natural-language training data are used to retrain the NLP classifier to identify text in the plurality of other classes and the new class using.Type: ApplicationFiled: November 11, 2021Publication date: May 11, 2023Inventors: Sungchul Kim, Subrata Mitra, Ruiyi Zhang, Rui Wang, Handong Zhao, Tong Yu