Patents by Inventor Caiming Xiong

Caiming Xiong has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 11087177
    Abstract: Approaches to zero-shot learning include partitioning training data into first and second sets according to classes assigned to the training data, training a prediction module based on the first set to predict a cluster center based on a class label, training a correction module based on the second set and each of the class labels in the first set to generate a correction to a cluster center predicted by the prediction module, presenting a new class label for a new class to the prediction module to predict a new cluster center, presenting the new class label, the predicted new cluster center, and each of the class labels in the first set to the correction module to generate a correction for the predicted new cluster center, augmenting a classifier based on the corrected cluster center for the new class, and classifying input data into the new class using the classifier.
    Type: Grant
    Filed: October 31, 2018
    Date of Patent: August 10, 2021
    Assignee: salesforce.com, inc.
    Inventors: Lily Hu, Caiming Xiong, Richard Socher
  • Patent number: 11080595
    Abstract: The technology disclosed provides a quasi-recurrent neural network (QRNN) encoder-decoder model that alternates convolutional layers, which apply in parallel across timesteps, and minimalist recurrent pooling layers that apply in parallel across feature dimensions.
    Type: Grant
    Filed: January 31, 2017
    Date of Patent: August 3, 2021
    Assignee: salesforce.com, inc.
    Inventors: James Bradbury, Stephen Joseph Merity, Caiming Xiong, Richard Socher
  • Publication number: 20210216728
    Abstract: Approaches for the translation of structured text include an embedding module for encoding and embedding source text in a first language, an encoder for encoding output of the embedding module, a decoder for iteratively decoding output of the encoder based on generated tokens in translated text from previous iterations, a beam module for constraining output of the decoder with respect to possible embedded tags to include in the translated text for a current iteration using a beam search, and a layer for selecting a token to be included in the translated text for the current iteration. The translated text is in a second language different from the first language. In some embodiments, the approach further includes scoring and pointer modules for selecting the token based on the output of the beam module or copied from the source text or reference text from a training pair best matching the source text.
    Type: Application
    Filed: March 26, 2021
    Publication date: July 15, 2021
    Inventors: Kazuma HASHIMOTO, Raffaella BUSCHIAZZO, James BRADBURY, Teresa MARSHALL, Caiming XIONG, Richard SOCHER
  • Publication number: 20210216828
    Abstract: A method for generating a neural network for detecting one or more objects in images includes generating one or more self-supervised proposal learning losses based on the one or more proposal features and corresponding proposal feature predictions. One or more consistency-based proposal learning losses are generated based on noisy proposal feature predictions and the corresponding proposal predictions without noise. A combined loss is generated using the one or more self-supervised proposal learning losses and one or more consistency-based proposal learning losses. The neural network is updated based on the combined loss.
    Type: Application
    Filed: October 26, 2020
    Publication date: July 15, 2021
    Inventors: Chetan Ramaiah, Peng Tang, Caiming Xiong
  • Patent number: 11056099
    Abstract: The disclosed technology teaches a deep end-to-end speech recognition model, including using multi-objective learning criteria to train a deep end-to-end speech recognition model on training data comprising speech samples temporally labeled with ground truth transcriptions.
    Type: Grant
    Filed: September 5, 2019
    Date of Patent: July 6, 2021
    Assignee: salesforce.com, inc.
    Inventors: Yingbo Zhou, Caiming Xiong
  • Patent number: 11042796
    Abstract: The technology disclosed provides a so-called “joint many-task neural network model” to solve a variety of increasingly complex natural language processing (NLP) tasks using growing depth of layers in a single end-to-end model. The model is successively trained by considering linguistic hierarchies, directly connecting word representations to all model layers, explicitly using predictions in lower tasks, and applying a so-called “successive regularization” technique to prevent catastrophic forgetting. Three examples of lower level model layers are part-of-speech (POS) tagging layer, chunking layer, and dependency parsing layer. Two examples of higher level model layers are semantic relatedness layer and textual entailment layer. The model achieves the state-of-the-art results on chunking, dependency parsing, semantic relatedness and textual entailment.
    Type: Grant
    Filed: January 31, 2017
    Date of Patent: June 22, 2021
    Assignee: salesforce.com, inc.
    Inventors: Kazuma Hashimoto, Caiming Xiong, Richard Socher
  • Publication number: 20210174026
    Abstract: Embodiments described in this disclosure illustrate the use of self-/semi supervised approaches for label-efficient DST in task-oriented dialogue systems. Conversational behavior is modeled by next response generation and turn utterance generation tasks. Prediction consistency is strengthened by augmenting data with stochastic word dropout and label guessing. Experimental results show that by exploiting self-supervision the joint goal accuracy can be boosted with limited labeled data.
    Type: Application
    Filed: May 8, 2020
    Publication date: June 10, 2021
    Inventors: Chien-Sheng Wu, Chu Hong Hoi, Caiming Xiong
  • Publication number: 20210174204
    Abstract: A method for using a neural network model for natural language processing (NLP) includes receiving training data associated with a source domain and a target domain; and generating one or more query batches. Each query batch includes one or more source tasks associated with the source domain and one or more target tasks associated with the target domain. For each query batch, class representations are generated for each class in the source domain and the target domain. A query batch loss for the query batch is generated based on the corresponding class representations. An optimization is performed on the neural network model by adjusting its network parameters based on the query batch loss. The optimized neural network model is used to perform one or more new NLP tasks.
    Type: Application
    Filed: November 9, 2020
    Publication date: June 10, 2021
    Inventors: Wenpeng Yin, Nazneen Rajani, Richard Socher, Caiming Xiong
  • Publication number: 20210174028
    Abstract: A method for maintaining a dialogue state associated with a dialogue between a user and a digital system includes receiving, by a dialogue state tracker associated with the digital system, a representation of a user communication, updating, by the dialogue state tracker, the dialogue state and providing a system response based on the updated dialogue state. The dialogue state is updated by evaluating, based on the representation of the user communication, a plurality of member scores corresponding to a plurality of ontology members of an ontology set, and selecting, based on the plurality of member scores, zero or more of the plurality of ontology members to add to or remove from the dialogue state.
    Type: Application
    Filed: February 19, 2021
    Publication date: June 10, 2021
    Inventors: Victor Zhong, Caiming Xiong
  • Publication number: 20210174798
    Abstract: Embodiments described in this disclosure illustrate the use of self-/semi supervised approaches for label-efficient DST in task-oriented dialogue systems. Conversational behavior is modeled by next response generation and turn utterance generation tasks. Prediction consistency is strengthened by augmenting data with stochastic word dropout and label guessing. Experimental results show that by exploiting self-supervision the joint goal accuracy can be boosted with limited labeled data.
    Type: Application
    Filed: May 8, 2020
    Publication date: June 10, 2021
    Inventors: Chien-Sheng Wu, Chu Hong Hoi, Caiming Xiong
  • Patent number: 11029694
    Abstract: An agent for navigating a mobile automated system is disclosed herein. The navigation agent receives a navigation instruction and visual information for one or more observed images. The navigation agent is provided or equipped with self-awareness, which provides or supports the following abilities: identifying which direction to go or proceed by determining the part of the instruction that corresponds to the observed images (visual grounding), and identifying which part of the instruction has been completed or ongoing and which part is potentially needed for the next action selection (textual grounding). In some embodiments, the navigation agent applies regularization to ensures that the grounded instruction can correctly be used to estimate the progress made towards the navigation goal (progress monitoring).
    Type: Grant
    Filed: October 31, 2018
    Date of Patent: June 8, 2021
    Assignee: salesforce.com, inc.
    Inventors: Chih-Yao Ma, Caiming Xiong
  • Publication number: 20210150340
    Abstract: Embodiments described herein provides a training mechanism that transfers the knowledge from a trained BERT model into a much smaller model to approximate the behavior of BERT. Specifically, the BERT model may be treated as a teacher model, and a much smaller student model may be trained using the same inputs to the teacher model and the output from the teacher model. In this way, the student model can be trained within a much shorter time than the BERT teacher model, but with comparable performance with BERT.
    Type: Application
    Filed: May 18, 2020
    Publication date: May 20, 2021
    Inventors: Wenhao Liu, Ka Chun Au, Shashank Harinath, Bryan McCann, Govardana Sachithanandam Ramachandran, Alexis Roos, Caiming Xiong
  • Publication number: 20210150282
    Abstract: Computing systems may support image classification and image detection services, and these services may utilize object detection/image classification machine learning models. The described techniques provide for normalization of confidence scores corresponding to manipulated target images and for non-max suppression within the range of confidence scores for manipulated images. In one example, the techniques provide for generating different scales of a test image, and the system performs normalization of confidence scores corresponding to each scaled image and non-max suppression per scaled image These techniques may be used to provide more accurate image detection (e.g., object detection and/or image classification) and may be used with models that are not trained on modified image sets. The model may be trained on a standard (e.g. non-manipulated) image set but used with manipulated target images and the described techniques to provide accurate object detection.
    Type: Application
    Filed: November 15, 2019
    Publication date: May 20, 2021
    Inventors: Ankit Chadha, Caiming Xiong, Ran Xu
  • Publication number: 20210150366
    Abstract: An embodiment proposed herein uses sparsification techniques to train the neural network with a high feature dimension that may yield desirable in-domain detection accuracy but may prune away dimensions in the output that are less important. Specifically, a sparsification vector is generated based on Gaussian distribution (or other probabilistic distribution) and is used to multiply with the higher dimension output to reduce the number of feature dimensions. The pruned output may be then used for the neural network to learn the sparsification vector. In this way, out-of-distribution detection accuracy can be improved.
    Type: Application
    Filed: May 18, 2020
    Publication date: May 20, 2021
    Inventors: Govardana Sachithanandam Ramachandran, Ka Chun Au, Shashank Harinath, Wenhao Liu, Alexis Roos, Caiming Xiong
  • Publication number: 20210150365
    Abstract: An embodiment provided herein preprocesses the input samples to the classification neural network, e.g., by adding Gaussian noise to word/sentence representations to make the function of the neural network satisfy Lipschitz property such that a small change in the input does not cause much change to the output if the input sample is in-distribution. Method to induce properties in the feature representation of neural network such that for out-of-distribution examples the feature representation magnitude is either close to zero or the feature representation is orthogonal to all class representations. Method to generate examples that are structurally similar to in-domain and semantically out-of domain for use in out-of-domain classification training. Method to prune feature representation dimension to mitigate long tail error of unused dimension in out-of-domain classification. Using these techniques, the accuracy of both in-domain and out-of-distribution identification can be improved.
    Type: Application
    Filed: May 18, 2020
    Publication date: May 20, 2021
    Inventors: Govardana Sachithanandam Ramachandran, Ka Chun Au, Shashank Harinath, Wenhao Liu, Alexis Roos, Caiming Xiong
  • Publication number: 20210152534
    Abstract: A system authenticates users using voice-based conversations. The system allows the authentication process to be customized using an authentication plan. For example, the system may be a multi-tenant system that allows customization of the authentication process for each tenant. The authentication plan is represented as an expression of phrase types, each phrase type associated with a phrase verification method. The system authenticates a user by executing the expression of an authentication plan for that user in response to a request from the user. The system performs a conversation with the user according to the authentication plan. The system determines whether to allow or deny the user request based on the result of evaluation of the expression of the authentication plan.
    Type: Application
    Filed: November 15, 2019
    Publication date: May 20, 2021
    Inventors: Tian Xie, Caiming Xiong
  • Publication number: 20210141781
    Abstract: Embodiments described herein provide unsupervised density-based clustering to infer table structure from document. Specifically, a number of words are identified from a block of text in an noneditable document, and the spatial coordinates of each word relative to the rectangular region are identified. Based on the word density of the rectangular region, the words are grouped into clusters using a heuristic radius search method. Words that are grouped into the same cluster are determined to be the element that belong to the same cell. In this way, the cells of the table structure can be identified. Once the cells are identified based on the word density of the block of text, the identified cells can be expanded horizontally or grouped vertically to identify rows or columns of the table structure.
    Type: Application
    Filed: November 11, 2019
    Publication date: May 13, 2021
    Inventors: Ankit CHADHA, Zeyuan CHEN, Caiming XIONG, Ran XU, Richard SOCHER
  • Publication number: 20210141865
    Abstract: A multi-tenant system performs custom configuration of a tenant-specific chatbot to process and act upon natural language requests. The multi-tenant system configures the tenant-specific chatbots without requiring tenant-specific training. The multi-tenant system providing a user interface for configuring a tenant-specific set of permitted actions. The multi-tenant system determines a set of example phrases for each of the selected permitted actions. The multi-tenant system receives a natural language request from a user and identifies the action that the user wants to perform. The multi-tenant system uses a neural network to compare the natural language request with example phrases to identify an example phrase that matches the natural language request. The multi-tenant system performs the action corresponding to the matching example phrase.
    Type: Application
    Filed: November 11, 2019
    Publication date: May 13, 2021
    Inventors: Michael Machado, James Douglas Harrison, Caiming Xiong, Xinyi Yang, Thomas Archie Cook, Roojuta Lalani, Jean-Marc Soumet, Karl Ryszard Skucha, Juan Manuel Rodriguez, Manju Vijayakumar, Vishal Motwani, Tian Xie, Bryan McCann, Nitish Shirish Keskar, Armen Abrahamyan, Zhihao Zou, Chitra Gulabrani, Minal Khodani, Adarsha Badarinath, Rohiniben Thakar, Srikanth Kollu, Kevin Schoen, Qiong Liu, Amit Hetawal, Kevin Zhang, Kevin Zhang, Victor Brouk, Johnson Liu, Rafael Amsili
  • Publication number: 20210142103
    Abstract: An online system that allows users to interact with it using expressions in natural language form includes an intent inference module allowing it to infer an intent represented by a user expression. The intent inference module has a set of possible intents, along with a small set of example natural language expressions known to represent that intent. When a user interacts with the system using a natural language expression for which the intent is not already known, the intent inference module applies a natural language inference model to compute scores indicating whether the user expression textually entails the various example natural language expressions. Based on the scores, the intent inference module determines an intent that is most applicable for the expression. If an intent cannot be determined with sufficient confidence, the intent inference module may further attempt to determine whether the various example natural language expressions textually entail the user expression.
    Type: Application
    Filed: December 18, 2019
    Publication date: May 13, 2021
    Inventors: Tian Xie, Kazuma Hashimoto, Xinyi Yang, Caiming Xiong
  • Publication number: 20210142164
    Abstract: Systems and methods are provided that employ knowledge distillation under a multi-task learning setting. In some embodiments, the systems and methods are implemented with a larger teacher model and a smaller student model, each of which comprise one or more shared layers and a plurality of task layers for performing multiple tasks. During training of the teacher model, its shared layers are initialized, and then the teacher model is multi-task refined. The teacher model predicts teacher logits. During training of the student model, its shared layers are initialized. Knowledge distillation is employed to transfer knowledge from the teacher model to the student model by the student model updating its shared layers and task layers, for example, according to the teacher logits of the teacher model. Other features are also provided.
    Type: Application
    Filed: December 16, 2019
    Publication date: May 13, 2021
    Inventors: Linqing LIU, Caiming XIONG