Patents by Inventor Haoyu Wang

Haoyu Wang has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Publication number: 20210221613
    Abstract: A management method applied to the goods-to-person system includes: in response to a task of putting products on shelves, calculating popularity of a product to be put on shelf according to historical sales order data of the product to be put on shelf, and matching the popularity of the product to be put on shelf with popularity of a shelf to determine a shelf area; selecting a goods location with space randomly in the determined shelf area, wherein the goods location is used for storing the product to be put on shelf; and controlling a mobile robot to transport a shelf where the goods location is located to a work station.
    Type: Application
    Filed: May 30, 2019
    Publication date: July 22, 2021
    Applicant: BEIJING GEEKPLUS TECHNOLOGY CO., LTD.
    Inventors: Kai LIU, Haoyu WANG, Xun WU, Kai SUN
  • Patent number: 11057330
    Abstract: A deep learning module classifies messages received from a plurality of entities into one or more conversation threads. In response to receiving a subsequent message, the deep learning module determines which of the one or more conversation threads and a new conversation thread is contextually a best fit for the subsequent message. The subsequent message is added to the determined conversation thread.
    Type: Grant
    Filed: August 26, 2019
    Date of Patent: July 6, 2021
    Assignee: INTERNATIONAL BUSINESS MACHINES CORPORATION
    Inventors: Ming Tan, Haoyu Wang, Dakuo Wang, Chuang Gan
  • Patent number: 11023683
    Abstract: A computer-implemented method includes obtaining a training data set including text data indicating one or more phrases or sentences. The computer-implemented method includes training a classifier using supervised machine learning based on the training data set and additional text data indicating one or more out-of-domain phrases or sentences. The computer-implemented method includes training an autoencoder using unsupervised machine learning based on the training data. The computer-implemented method further includes combining the classifier and the autoencoder to generate the out-of-domain sentence detector configured to generate an output indicating a classification of whether input text data corresponds to an out-of-domain sentence. The output is based on a combination of a first output of the classifier and a second output of the autoencoder.
    Type: Grant
    Filed: March 6, 2019
    Date of Patent: June 1, 2021
    Assignee: International Business Machines Corporation
    Inventors: Inkit Padhi, Ruijian Wang, Haoyu Wang, Saloni Potdar
  • Publication number: 20210141860
    Abstract: Provided is a method, system, and computer program product for context-dependent spellchecking. The method comprises receiving context data to be used in spell checking. The method further comprises receiving a user input. The method further comprises identifying an out-of-vocabulary (OOV) word in the user input. An initial suggestion pool of candidate words is identified based, at least in part, on the context data. The method then comprises using a noisy channel approach to evaluate a probability that one or more of the candidate words of the initial suggestion pool is an intended word and should be used as a candidate for replacement of the OOV word. The method further comprises selecting one or more candidate words for replacement of the OOV word. The method further comprises outputting the one or more candidates.
    Type: Application
    Filed: November 11, 2019
    Publication date: May 13, 2021
    Inventors: Panos Karagiannis, Ladislav Kunc, Saloni Potdar, Haoyu Wang, Navneet N. Rao
  • Patent number: 10984075
    Abstract: A computer transforms high-dimensional data into low-dimensional data. A distance is computed between a selected observation vector and each observation vector of a plurality of observation vectors, a nearest neighbors are selected using the computed distances, and a first sigmoid function is applied to compute a distance similarity value between the selected observation vector and each of the selected nearest neighbors where each of the computed distance similarity values is added to a first matrix. The process is repeated with each observation vector of the plurality of observation vectors as the selected observation vector. An optimization method is executed with an initial matrix, the first matrix, and a gradient of a second sigmoid function that computes a second distance similarity value between the selected observation vector and each of the nearest neighbors to transform each observation vector of the plurality of observation vectors into the low-dimensional space.
    Type: Grant
    Filed: October 13, 2020
    Date of Patent: April 20, 2021
    Assignee: SAS Institute Inc.
    Inventors: Yu Liang, Arin Chaudhuri, Haoyu Wang
  • Publication number: 20210110266
    Abstract: A computer system identifies threads in a communication session. A feature vector is generated for a message in a communication session, wherein the feature vector includes elements for features and contextual information of the message. The message feature vector and feature vectors for a plurality of threads are processed using machine learning models each associated with a corresponding thread to determine a set of probability values for classifying the message into at least one thread, wherein the threads include one or more pre-existing threads and a new thread. A classification of the message into at least one of the threads is indicated based on the set of probability values. Classification of one or more prior messages is adjusted based on the message's classification. Embodiments of the present invention further include a method and program product for identifying threads in a communication session in substantially the same manner described above.
    Type: Application
    Filed: October 10, 2019
    Publication date: April 15, 2021
    Inventors: Dakuo Wang, Ming Tan, Mo Yu, Haoyu Wang, Yupeng Gao, Chuang Gan
  • Publication number: 20210109991
    Abstract: Domain specific model compression by providing a weighting parameter for a candidate operation of a neural network, applying the weighting parameter to an output vector of the candidate operation, performing a regularization of the weighting parameter output vector combination, compressing the neural network model according to the results of the regularization, and providing the neural network model after compression.
    Type: Application
    Filed: October 10, 2019
    Publication date: April 15, 2021
    Inventors: Haoyu Wang, Yang Yu, Ming Tan, Saloni Potdar
  • Patent number: 10977445
    Abstract: A computer-implemented method includes obtaining a training data set including a plurality of training examples. The method includes generating, for each training example, multiple feature vectors corresponding, respectively, to multiple feature types. The method includes applying weighting factors to feature vectors corresponding to a subset of the feature types. The weighting factors are determined based on one or more of: a number of training examples, a number of classes associated with the training data set, an average number of training examples per class, a language of the training data set, a vocabulary size of the training data set, or a commonality of the vocabulary with a public corpus. The method includes concatenating the feature vectors of a particular training example to form an input vector and providing the input vector as training data to a machine-learning intent classification model to train the model to determine intent based on text input.
    Type: Grant
    Filed: February 1, 2019
    Date of Patent: April 13, 2021
    Assignee: International Business Machines Corporation
    Inventors: Yang Yu, Ladislav Kunc, Haoyu Wang, Ming Tan, Saloni Potdar
  • Publication number: 20210103636
    Abstract: Systems and methods provide for automated messaging summarization and ranking. The systems and methods may use an integrated machine learning model to perform thread detection, thread summarization, and summarization ranking. The messages may be received from a team chat application, organized, summarized and ranked by the machine learning model, and the results may be returned to the team chat application. In some cases, the ranking may be different for different users of the team chat application.
    Type: Application
    Filed: October 8, 2019
    Publication date: April 8, 2021
    Inventors: Dakuo Wang, Ming Tan, Chuang Gan, Haoyu Wang
  • Publication number: 20210067477
    Abstract: A deep learning module classifies messages received from a plurality of entities into one or more conversation threads. In response to receiving a subsequent message, the deep learning module determines which of the one or more conversation threads and a new conversation thread is contextually a best fit for the subsequent message. The subsequent message is added to the determined conversation thread.
    Type: Application
    Filed: August 26, 2019
    Publication date: March 4, 2021
    Inventors: Ming Tan, Haoyu Wang, Dakuo Wang, Chuang Gan
  • Publication number: 20210064666
    Abstract: An artificial intelligence (AI) interaction method, system, and computer program product include selecting an artificial intelligence model to respond to a query to generating a response to the query using the selected artificial intelligence model, and receiving the response to the query from the selected artificial intelligence model.
    Type: Application
    Filed: August 26, 2019
    Publication date: March 4, 2021
    Inventors: Dakuo Wang, Ming Tan, Chuang Gan, Haoyu Wang, Mo Yu
  • Patent number: 10937416
    Abstract: A method includes providing input text to a plurality of multi-task learning (MTL) models corresponding to a plurality of domains. Each MTL model is trained to generate an embedding vector based on the input text. The method further includes providing the input text to a domain identifier that is trained to generate a weight vector based on the input text. The weight vector indicates a classification weight for each domain of the plurality of domains. The method further includes scaling each embedding vector based on a corresponding classification weight of the weight vector to generate a plurality of scaled embedding vectors, generating a feature vector based on the plurality of scaled embedding vectors, and providing the feature vector to an intent classifier that is trained to generate, based on the feature vector, an intent classification result associated with the input text.
    Type: Grant
    Filed: February 1, 2019
    Date of Patent: March 2, 2021
    Assignee: International Business Machines Corporation
    Inventors: Ming Tan, Haoyu Wang, Ladislav Kunc, Yang Yu, Saloni Potdar
  • Publication number: 20210049502
    Abstract: A method includes determining, based on an input data sample, a set of probabilities. Each probability of the set of probabilities is associated with a respective label of a set of labels. A particular probability associated with a particular label indicates an estimated likelihood that the input data sample is associated with the particular label. The method includes modifying the set of probabilities based on a set of adjustment factors to generate a modified set of probabilities. The set of adjustment factors is based on a first relative frequency distribution and a second relative frequency distribution. The first relative frequency distribution indicates for each label of the set of labels, a frequency of occurrence of the label among training data. The second relative frequency distribution indicates for each label of the set of labels, a frequency of occurrence of the label among post-training data provided to the trained classifier.
    Type: Application
    Filed: August 16, 2019
    Publication date: February 18, 2021
    Inventors: Haoyu Wang, Ming Tan, Dakuo Wang, Chuang Gan, Saloni Potdar
  • Publication number: 20210034965
    Abstract: A computer-implemented method includes using an embedding network to generate prototypical vectors. Each prototypical vector is based on a corresponding label associated with a first domain. The computer-implemented method also includes using the embedding network to generate an in-domain test vector based on at least one data sample from a particular label associated with the first domain and using the embedding network to generate an out-of-domain test vector based on at least one other data sample associated with a different domain. The computer-implemented method also includes comparing the prototypical vectors to the in-domain test vector to generate in-domain comparison values and comparing the prototypical vectors to the out-of-domain test vector to generate out-of-domain comparison values. The computer-implemented method also includes modifying, based on the in-domain comparison values and the out-of-domain comparison values, one or more parameters of the embedding network.
    Type: Application
    Filed: August 2, 2019
    Publication date: February 4, 2021
    Inventors: Ming Tan, Dakuo Wang, Mo Yu, Haoyu Wang, Yang Yu, Shiyu Chang, Saloni Potdar
  • Publication number: 20200364300
    Abstract: A computing device receives a natural language input from a user. The computing device routes the natural language input from an active domain node of multiple domain nodes of a multi-domain context-based hierarchy to a leaf node of the domain nodes by selecting a parent domain node in the hierarchy until an off-topic classifier labels the natural language input as in-domain and then selecting a subdomain node in the hierarchy until an in-domain classifier labels the natural language input with a classification label, each of the plurality of domain nodes comprising a respective off-topic classifier and a respective in-domain classifier trained for a respective domain node. The computing device outputs the classification label determined by the leaf node.
    Type: Application
    Filed: May 13, 2019
    Publication date: November 19, 2020
    Inventors: MING TAN, LADISLAV KUNC, YANG YU, HAOYU WANG, SALONI POTDAR
  • Publication number: 20200356629
    Abstract: Techniques facilitating detection of conversation threads in unstructured channels are provided. A system can comprise a memory that stores computer executable components and a processor that executes the computer executable components stored in the memory. The computer executable components can comprise an extraction component that employs a model to detect conversation messages based on a defined confidence level and assigns the conversation messages to respective conversation thread categories. The computer executable components also can comprise a model component that trains the model on conversation messages that comprise respective text data, wherein the model is trained to detect the respective text data to the defined confidence level.
    Type: Application
    Filed: May 6, 2019
    Publication date: November 12, 2020
    Inventors: Ming Tan, Dakuo Wang, Mo Yu, Chuang Gan, Haoyu Wang, Shiyu Chang
  • Publication number: 20200327445
    Abstract: A method of text classification includes generating a text embedding vector representing a text sample and applying weights of a regression layer to the text embedding vector to generate a first data model output vector. The method also includes generating a plurality of prototype embedding vectors associated with a respective classification labels and comparing the plurality of prototype embedding vectors to the text embedding vector to generate a second data model output vector.
    Type: Application
    Filed: April 9, 2019
    Publication date: October 15, 2020
    Inventors: Yang Yu, Ming Tan, Ravi Nair, Haoyu Wang, Saloni Potdar
  • Publication number: 20200285702
    Abstract: A computer-implemented method includes obtaining a training data set including text data indicating one or more phrases or sentences. The computer-implemented method includes training a classifier using supervised machine learning based on the training data set and additional text data indicating one or more out-of-domain phrases or sentences. The computer-implemented method includes training an autoencoder using unsupervised machine learning based on the training data. The computer-implemented method further includes combining the classifier and the autoencoder to generate the out-of-domain sentence detector configured to generate an output indicating a classification of whether input text data corresponds to an out-of-domain sentence. The output is based on a combination of a first output of the classifier and a second output of the autoencoder.
    Type: Application
    Filed: March 6, 2019
    Publication date: September 10, 2020
    Inventors: Inkit Padhi, Ruijian Wang, Haoyu Wang, Saloni Potdar
  • Publication number: 20200251100
    Abstract: A method includes providing input text to a plurality of multi-task learning (MTL) models corresponding to a plurality of domains. Each MTL model is trained to generate an embedding vector based on the input text. The method further includes providing the input text to a domain identifier that is trained to generate a weight vector based on the input text. The weight vector indicates a classification weight for each domain of the plurality of domains. The method further includes scaling each embedding vector based on a corresponding classification weight of the weight vector to generate a plurality of scaled embedding vectors, generating a feature vector based on the plurality of scaled embedding vectors, and providing the feature vector to an intent classifier that is trained to generate, based on the feature vector, an intent classification result associated with the input text.
    Type: Application
    Filed: February 1, 2019
    Publication date: August 6, 2020
    Inventors: Ming Tan, Haoyu Wang, Ladislav Kunc, Yang Yu, Saloni Potdar
  • Publication number: 20200250274
    Abstract: An online version of a sentence representation generation module updated by training a first sentence representation generation module using first labeled data of a first corpus. After training the first sentence representation generation module using the first labeled data, a second corpus of second labeled data is obtained. The second corpus is distinct from the first corpus. A subset of the first labeled data is identified based on similarities between the first corpus and the second corpus. A second sentence representation generation module is trained using the second labeled data of the second corpus and the subset of the first labeled data.
    Type: Application
    Filed: February 5, 2019
    Publication date: August 6, 2020
    Inventors: Ming Tan, Ladislav Kunc, Yang Yu, Haoyu Wang, Saloni Potdar