Patents by Inventor TRUNG HUU BUI

TRUNG HUU BUI has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11967128
    Abstract: The present disclosure describes a model for large scale color prediction of objects identified in images. Embodiments of the present disclosure include an object detection network, an attention network, and a color classification network. The object detection network generates object features for an object in an image and may include a convolutional neural network (CNN), region proposal network, or a ResNet. The attention network generates an attention vector for the object based on the object features, wherein the attention network takes a query vector based on the object features, and a plurality of key vector and a plurality of value vectors corresponding to a plurality of colors as input. The color classification network generates a color attribute vector based on the attention vector, wherein the color attribute vector indicates a probability of the object including each of the plurality of colors.
    Type: Grant
    Filed: May 28, 2021
    Date of Patent: April 23, 2024
    Assignee: ADOBE INC.
    Inventors: Qiuyu Chen, Quan Hung Tran, Kushal Kafle, Trung Huu Bui, Franck Dernoncourt, Walter Chang
  • Patent number: 11960843
    Abstract: Techniques and systems are provided for training a machine learning model using different datasets to perform one or more tasks. The machine learning model can include a first sub-module configured to perform a first task and a second sub-module configured to perform a second task. The first sub-module can be selected for training using a first training dataset based on a format of the first training dataset. The first sub-module can then be trained using the first training dataset to perform the first task. The second sub-module can be selected for training using a second training dataset based on a format of the second training dataset. The second sub-module can then be trained using the second training dataset to perform the second task.
    Type: Grant
    Filed: May 2, 2019
    Date of Patent: April 16, 2024
    Assignee: Adobe Inc.
    Inventors: Zhe Lin, Trung Huu Bui, Scott Cohen, Mingyang Ling, Chenyun Wu
  • Publication number: 20240037906
    Abstract: Systems and methods for color prediction are described. Embodiments of the present disclosure receive an image that includes an object including a color, generate a color vector based on the image using a color classification network, where the color vector includes a color value corresponding to each of a set of colors, generate a bias vector by comparing the color vector to teach of a set of center vectors, where each of the set of center vectors corresponds to a color of the set of colors, and generate an unbiased color vector based on the color vector and the bias vector, where the unbiased color vector indicates the color of the object.
    Type: Application
    Filed: July 26, 2022
    Publication date: February 1, 2024
    Inventors: Qiuyu Chen, Quan Hung Tran, Kushal Kafle, Trung Huu Bui, Franck Dernoncourt, Walter W. Chang
  • Publication number: 20240020337
    Abstract: Systems and methods for intent discovery and video summarization are described. Embodiments of the present disclosure receive a video and a transcript of the video, encode the video to obtain a sequence of video encodings, encode the transcript to obtain a sequence of text encodings, apply a visual gate to the sequence of text encodings based on the sequence of video encodings to obtain gated text encodings, and generate an intent label for the transcript based on the gated text encodings.
    Type: Application
    Filed: July 12, 2022
    Publication date: January 18, 2024
    Inventors: Adyasha Maharana, Quan Hung Tran, Seunghyun Yoon, Franck Dernoncourt, Trung Huu Bui, Walter W. Chang
  • Publication number: 20230419164
    Abstract: Multitask machine-learning model training and training data augmentation techniques are described. In one example, training is performed for multiple tasks simultaneously as part of training a multitask machine-learning model using question pairs. Examples of the multiple tasks include question summarization and recognizing question entailment. Further, a loss function is described that incorporates a parameter sharing loss that is configured to adjust an amount that parameters are shared between corresponding layers trained for the first and second tasks, respectively. In an implementation, training data augmentation techniques are also employed by synthesizing question pairs, automatically and without user intervention, to improve accuracy in model training.
    Type: Application
    Filed: June 22, 2022
    Publication date: December 28, 2023
    Applicant: Adobe Inc.
    Inventors: Khalil Mrini, Franck Dernoncourt, Seunghyun Yoon, Trung Huu Bui, Walter W. Chang, Emilia Farcas, Ndapandula T. Nakashole
  • Publication number: 20230418868
    Abstract: Systems and methods for text processing are described. Embodiments of the present disclosure receive a query comprising a natural language expression; extract a plurality of mentions from the query; generate a relation vector between a pair of the plurality of mentions using a relation encoder network, wherein the relation encoder network is trained using a contrastive learning process where mention pairs from a same document are labeled as positive samples and mention pairs from different documents are labeled as negative samples; combine the plurality of mentions with the relation vector to obtain a virtual knowledge graph of the query; identify a document corresponding to the query by comparing the virtual knowledge graph of the query to a virtual knowledge graph of the document; and transmit a response to the query, wherein the response includes a reference to the document.
    Type: Application
    Filed: June 24, 2022
    Publication date: December 28, 2023
    Inventors: Yeon Seonwoo, Seunghyun Yoon, Trung Huu Bui, Franck Dernoncourt, Roger K. Brooks, Mihir Naware
  • Publication number: 20230403175
    Abstract: Systems and methods for coreference resolution are provided. One aspect of the systems and methods includes inserting a speaker tag into a transcript, wherein the speaker tag indicates that a name in the transcript corresponds to a speaker of a portion of the transcript; encoding a plurality of candidate spans from the transcript based at least in part on the speaker tag to obtain a plurality of span vectors; extracting a plurality of entity mentions from the transcript based on the plurality of span vectors, wherein each of the plurality of entity mentions corresponds to one of the plurality of candidate spans; and generating coreference information for the transcript based on the plurality of entity mentions, wherein the coreference information indicates that a pair of candidate spans of the plurality of candidate spans corresponds to a pair of entity mentions that refer to a same entity.
    Type: Application
    Filed: June 14, 2022
    Publication date: December 14, 2023
    Inventors: Tuan Manh Lai, Trung Huu Bui, Doo Soon Kim
  • Publication number: 20230386208
    Abstract: Systems and methods for video segmentation and summarization are described. Embodiments of the present disclosure receive a video and a transcript of the video; generate visual features representing frames of the video using an image encoder; generate language features representing the transcript using a text encoder, wherein the image encoder and the text encoder are trained based on a correlation between training visual features and training language features; and segment the video into a plurality of video segments based on the visual features and the language features.
    Type: Application
    Filed: May 31, 2022
    Publication date: November 30, 2023
    Inventors: Hailin Jin, Jielin Qiu, Zhaowen Wang, Trung Huu Bui, Franck Dernoncourt
  • Patent number: 11769111
    Abstract: The present invention is directed towards providing automated workflows for the identification of a reading order from text segments extracted from a document. Ordering the text segments is based on trained natural language models. In some embodiments, the workflows are enabled to perform a method for identifying a sequence associated with a portable document. The methods includes iteratively generating a probabilistic language model, receiving the portable document, and selectively extracting features (such as but not limited to text segments) from the document. The method may generate pairs of features (or feature pair from the extracted features). The method may further generate a score for each of the pairs based on the probabilistic language model and determine an order to features based on the scores. The method may provide the extracted features in the determined order.
    Type: Grant
    Filed: June 18, 2020
    Date of Patent: September 26, 2023
    Assignee: Adobe Inc.
    Inventors: Trung Huu Bui, Hung Hai Bui, Shawn Alan Gaither, Walter Wei-Tuh Chang, Michael Frank Kraley, Pranjal Daga
  • Publication number: 20230297603
    Abstract: Systems and methods for natural language processing are described. Embodiments of the present disclosure identify a task set including a plurality of pseudo tasks, wherein each of the plurality of pseudo tasks includes a support set corresponding to a first natural language processing (NLP) task and a query set corresponding to a second NLP task; update a machine learning model in an inner loop based on the support set; update the machine learning model in an outer loop based on the query set; and perform the second NLP task using the machine learning model.
    Type: Application
    Filed: March 18, 2022
    Publication date: September 21, 2023
    Inventors: Meryem M'hamdi, Doo Soon Kim, Franck Dernoncourt, Trung Huu Bui
  • Publication number: 20230267726
    Abstract: Embodiments of the disclosure provide a machine learning model for generating a predicted executable command for an image. The learning model includes an interface configured to obtain an utterance indicating a request associated with the image, an utterance sub-model, a visual sub-model, an attention network, and a selection gate. The machine learning model generates a segment of the predicted executable command from weighted probabilities of each candidate token in a predetermined vocabulary determined based on the visual features, the concept features, current command features, and the utterance features extracted from the utterance or the image.
    Type: Application
    Filed: February 18, 2022
    Publication date: August 24, 2023
    Inventors: Seunghyun Yoon, Trung Huu Bui, Franck Dernoncourt, Hyounghun Kim, Doo Soon Kim
  • Publication number: 20230259708
    Abstract: Systems and methods for key-phrase extraction are described. The systems and methods include receiving a transcript including a text paragraph and generating key-phrase data for the text paragraph using a key-phrase extraction network. The key-phrase extraction network is trained to identify domain-relevant key-phrase data based on domain data obtained using a domain discriminator network. The systems and methods further include generating meta-data for the transcript based on the key-phrase data.
    Type: Application
    Filed: February 14, 2022
    Publication date: August 17, 2023
    Inventors: Amir Pouran Ben Veyseh, Franck Dernoncourt, Walter W. Chang, Trung Huu Bui, Hanieh Deilamsalehy, Seunghyun Yoon, Rajiv Bhawanji Jain, Quan Hung Tran, Varun Manjunatha
  • Publication number: 20230259718
    Abstract: Techniques for training a language model for code switching content are disclosed. Such techniques include, in some embodiments, generating a dataset, which includes identifying one or more portions within textual content in a first language, the identified one or more portions each including one or more of offensive content or non-offensive content; translating the identified one or more salient portions to a second language; and reintegrating the translated one or more portions into the textual content to generate code-switched textual content. In some cases, the textual content in the first language includes offensive content and non-offensive content, the identified one or more portions include the offensive content, and the translated one or more portions include a translated version of the offensive content. In some embodiments, the code-switched textual content is at least part of a synthetic dataset usable to train a language model, such as a multilingual classification model.
    Type: Application
    Filed: February 17, 2022
    Publication date: August 17, 2023
    Inventors: Cesa Salaam, Seunghyun Yoon, Trung Huu Bui, Franck Dernoncourt
  • Publication number: 20230237093
    Abstract: Systems and methods for item recommendation are described. Embodiments of the present disclosure receive input indicating a relationship between a user and a first content item; generate a knowledge graph based on the input, wherein the knowledge graph comprises relationship information between the user and a plurality of content items; generate a first feature embedding representing the user and a second feature embedding representing a second content item of the plurality of content items based on the knowledge graph, wherein the second feature embedding is generated using a first modality for a query vector of an attention mechanism and a second modality for a key vector and a value vector of the attention mechanism; compare the first feature embedding to the second feature embedding to obtain a similarity score; and recommend the second content item for the user based on the similarity score.
    Type: Application
    Filed: January 27, 2022
    Publication date: July 27, 2023
    Inventors: Yifan Li, Trung Huu Bui, Timothy Jeewun Ganter, David Fox
  • Publication number: 20230153522
    Abstract: Systems and methods for image captioning are described. One or more aspects of the systems and methods include generating a training caption for a training image using an image captioning network; encoding the training caption using a multi-modal encoder to obtain an encoded training caption; encoding the training image using the multi-modal encoder to obtain an encoded training image; computing a reward function based on the encoded training caption and the encoded training image; and updating parameters of the image captioning network based on the reward function.
    Type: Application
    Filed: November 18, 2021
    Publication date: May 18, 2023
    Inventors: Jaemin Cho, Seunghyun Yoon, Ajinkya Gorakhnath Kale, Trung Huu Bui, Franck Dernoncourt
  • Patent number: 11651211
    Abstract: Techniques for training a first neural network (NN) model using a pre-trained second NN model are disclosed. In an example, training data is input to the first and second models. The training data includes masked tokens and unmasked tokens. In response, the first model generates a first prediction associated with a masked token and a second prediction associated with an unmasked token, and the second model generates a third prediction associated with the masked token and a fourth prediction associated with the unmasked token. The first model is trained, based at least in part on the first, second, third, and fourth predictions. In another example, a prediction associated with a masked token, a prediction associated with an unmasked token, and a prediction associated with whether two sentences of training data are adjacent sentences are received from each of the first and second models. The first model is trained using the predictions.
    Type: Grant
    Filed: December 17, 2019
    Date of Patent: May 16, 2023
    Assignee: Adobe Inc.
    Inventors: Tuan Manh Lai, Trung Huu Bui, Quan Hung Tran
  • Publication number: 20230136527
    Abstract: Systems and methods for natural language processing are described. One or more aspects of a method, apparatus, and non-transitory computer readable medium include receiving a text phrase; encoding the text phrase using an encoder to obtain a hidden representation of the text phrase, wherein the encoder is trained during a first training phrase using self-supervised learning based on a first contrastive loss and during a second training phrase using supervised learning based on a second contrastive learning loss; identifying an intent of the text phrase from a predetermined set of intent labels using a classification network, wherein the classification network is jointly trained with the encoder in the second training phase; and generating a response to the text phrase based on the intent.
    Type: Application
    Filed: November 4, 2021
    Publication date: May 4, 2023
    Inventors: Jianguo Zhang, Trung Huu Bui, Seunghyun Yoon, Xiang Chen, Quan Hung Tran, Walter W. Chang
  • Publication number: 20220414338
    Abstract: System and methods for a text summarization system are described. In one example, a text summarization system receives an input utterance and determines whether the utterance should be included in a summary of the text. The text summarization system includes an embedding network, a convolution network, an encoding component, and a summary component. The embedding network generates a semantic embedding of an utterance. The convolution network generates a plurality of feature vectors based on the semantic embedding. The encoding component identifies a plurality of latent codes respectively corresponding to the plurality of feature vectors. The summary component identifies a prominent code among the latent codes and to select the utterance as a summary utterance based on the prominent code.
    Type: Application
    Filed: June 29, 2021
    Publication date: December 29, 2022
    Inventors: SANGWOO CHO, Franck Dernoncourt, Timothy Jeewun Ganter, Trung Huu Bui, Nedim Lipka, Varun Manjunatha, Walter Chang, Hailin Jin, Jonathan Brandt
  • Patent number: 11538463
    Abstract: Methods and systems are provided for generating a customized speech recognition neural network system comprised of an adapted automatic speech recognition neural network and an adapted language model neural network. The automatic speech recognition neural network is first trained in a generic domain and then adapted to a target domain. The language model neural network is first trained in a generic domain and then adapted to a target domain. Such a customized speech recognition neural network system can be used to understand input vocal commands.
    Type: Grant
    Filed: April 12, 2019
    Date of Patent: December 27, 2022
    Assignee: Adobe Inc.
    Inventors: Trung Huu Bui, Subhadeep Dey, Franck Dernoncourt
  • Publication number: 20220383031
    Abstract: The present disclosure describes a model for large scale color prediction of objects identified in images. Embodiments of the present disclosure include an object detection network, an attention network, and a color classification network. The object detection network generates object features for an object in an image and may include a convolutional neural network (CNN), region proposal network, or a ResNet. The attention network generates an attention vector for the object based on the object features, wherein the attention network takes a query vector based on the object features, and a plurality of key vector and a plurality of value vectors corresponding to a plurality of colors as input. The color classification network generates a color attribute vector based on the attention vector, wherein the color attribute vector indicates a probability of the object including each of the plurality of colors.
    Type: Application
    Filed: May 28, 2021
    Publication date: December 1, 2022
    Inventors: Qiuyu Chen, Quan Hung Tran, Kushal Kafle, Trung Huu Bui, Franck Dernoncourt, Walter Chang