Patents by Inventor Kuzman Ganchev

Kuzman Ganchev has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 10289952
    Abstract: A computer-implemented technique can include receiving, at a server, labeled training data including a plurality of groups of words, each group of words having a predicate word, each word having generic word embeddings. The technique can include extracting, at the server, the plurality of groups of words in a syntactic context of their predicate words. The technique can include concatenating, at the server, the generic word embeddings to create a high dimensional vector space representing features for each word. The technique can include obtaining, at the server, a model having a learned mapping from the high dimensional vector space to a low dimensional vector space and learned embeddings for each possible semantic frame in the low dimensional vector space. The technique can also include outputting, by the server, the model for storage, the model being configured to identify a specific semantic frame for an input.
    Type: Grant
    Filed: January 28, 2016
    Date of Patent: May 14, 2019
    Assignee: Google LLC
    Inventors: Dipanjan Das, Kuzman Ganchev, Jason Weston, Karl Moritz Hermann
  • Patent number: 9779087
    Abstract: A computer-implemented method can include obtaining (i) an aligned bi-text for a source language and a target language, and (ii) a supervised sequence model for the source language. The method can include labeling a source side of the aligned bi-text using the supervised sequence model and projecting labels from the labeled source side to a target side of the aligned bi-text to obtain a labeled target side of the aligned bi-text. The method can include filtering the labeled target side based on a task of a natural language processing (NLP) system configured to utilize a sequence model for the target language to obtain a filtered target side of the aligned bi-text. The method can also include training the sequence model for the target language using posterior regularization with soft constraints on the filtered target side to obtain a trained sequence model for the target language.
    Type: Grant
    Filed: December 13, 2013
    Date of Patent: October 3, 2017
    Assignee: GOOGLE INC.
    Inventors: Dipanjan Das, Kuzman Ganchev
  • Publication number: 20170270407
    Abstract: A method includes training a neural network having parameters on training data, in which the neural network receives an input state and processes the input state to generate a respective score for each decision in a set of decisions. The method includes receiving training data including training text sequences and, for each training text sequence, a corresponding gold decision sequence. The method includes training the neural network on the training data to determine trained values of parameters of the neural network. Training the neural network includes for each training text sequence: maintaining a beam of candidate decision sequences for the training text sequence, updating each candidate decision sequence by adding one decision at a time, determining that a gold candidate decision sequence matching a prefix of the gold decision sequence has dropped out of the beam, and in response, performing an iteration of gradient descent to optimize an objective function.
    Type: Application
    Filed: January 17, 2017
    Publication date: September 21, 2017
    Inventors: Christopher Alberti, Aliaksei Severyn, Daniel Andor, Slav Petrov, Kuzman Ganchev Ganchev, David Joseph Weiss, Michael John Collins, Alessandro Presta
  • Publication number: 20160239739
    Abstract: A computer-implemented technique can include receiving, at a server, labeled training data including a plurality of groups of words, each group of words having a predicate word, each word having generic word embeddings. The technique can include extracting, at the server, the plurality of groups of words in a syntactic context of their predicate words. The technique can include concatenating, at the server, the generic word embeddings to create a high dimensional vector space representing features for each word. The technique can include obtaining, at the server, a model having a learned mapping from the high dimensional vector space to a low dimensional vector space and learned embeddings for each possible semantic frame in the low dimensional vector space. The technique can also include outputting, by the server, the model for storage, the model being configured to identify a specific semantic frame for an input.
    Type: Application
    Filed: January 28, 2016
    Publication date: August 18, 2016
    Applicant: Google Inc.
    Inventors: Dipanjan Das, Kuzman Ganchev, Jason Weston, Karl Moritz Hermann
  • Patent number: 9262406
    Abstract: A computer-implemented technique can include receiving, at a server, labeled training data including a plurality of groups of words, each group of words having a predicate word, each word having generic word embeddings. The technique can include extracting, at the server, the plurality of groups of words in a syntactic context of their predicate words. The technique can include concatenating, at the server, the generic word embeddings to create a high dimensional vector space representing features for each word. The technique can include obtaining, at the server, a model having a learned mapping from the high dimensional vector space to a low dimensional vector space and learned embeddings for each possible semantic frame in the low dimensional vector space. The technique can also include outputting, by the server, the model for storage, the model being configured to identify a specific semantic frame for an input.
    Type: Grant
    Filed: May 7, 2014
    Date of Patent: February 16, 2016
    Assignee: Google Inc.
    Inventors: Dipanjan Das, Kuzman Ganchev, Jason Weston, Karl Moritz Hermann
  • Publication number: 20150169549
    Abstract: A computer-implemented method can include obtaining (i) an aligned bi-text for a source language and a target language, and (ii) a supervised sequence model for the source language. The method can include labeling a source side of the aligned bi-text using the supervised sequence model and projecting labels from the labeled source side to a target side of the aligned bi-text to obtain a labeled target side of the aligned bi-text. The method can include filtering the labeled target side based on a task of a natural language processing (NLP) system configured to utilize a sequence model for the target language to obtain a filtered target side of the aligned bi-text. The method can also include training the sequence model for the target language using posterior regularization with soft constraints on the filtered target side to obtain a trained sequence model for the target language.
    Type: Application
    Filed: December 13, 2013
    Publication date: June 18, 2015
    Applicant: Google Inc.
    Inventors: Dipanjan Das, Kuzman Ganchev