Patents by Inventor Ebru Arisoy

Ebru Arisoy has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 9934778
    Abstract: Techniques for conversion of non-back-off language models for use in speech decoders. For example, an apparatus for conversion of non-back-off language models for use in speech decoders. For example, an apparatus is configured convert a non-back-off language model to a back-off language model. The converted back-off language model is pruned. The converted back-off language model is usable for decoding speech.
    Type: Grant
    Filed: August 1, 2016
    Date of Patent: April 3, 2018
    Assignee: International Business Machines Corporation
    Inventors: Ebru Arisoy, Bhuvana Ramabhadran, Abhinav Sethy, Stanley Chen
  • Publication number: 20170193391
    Abstract: A plurality of corpora is received from one or more sources. A separate model is trained on each corpus of the plurality of corpora. The models for the plurality of corpora are merged into a joint model using parameter interpolation. The models for each corpus of the plurality of corpora are retrained separately using the joint model. A single model is created based on the retrained models.
    Type: Application
    Filed: December 31, 2015
    Publication date: July 6, 2017
    Inventors: Stanley Chen, Bhuvana Ramabhadran, Ebru Arisoy-Saraclar, Abhinav Sethy
  • Patent number: 9514744
    Abstract: Techniques for conversion of non-back-off language models for use in speech decoders. For example, an apparatus for conversion of non-back-off language models for use in speech decoders. For example, an apparatus is configured convert a non-back-off language model to a back-off language model. The converted back-off language model is pruned. The converted back-off language model is usable for decoding speech.
    Type: Grant
    Filed: August 12, 2013
    Date of Patent: December 6, 2016
    Assignee: International Business Machines Corporation
    Inventors: Ebru Arisoy, Bhuvana Ramabhadran, Abhinav Sethy, Stanley Chen
  • Publication number: 20160343369
    Abstract: Techniques for conversion of non-back-off language models for use in speech decoders. For example, an apparatus for conversion of non-back-off language models for use in speech decoders. For example, an apparatus is configured convert a non-back-off language model to a back-off language model. The converted back-off language model is pruned. The converted back-off language model is usable for decoding speech.
    Type: Application
    Filed: August 1, 2016
    Publication date: November 24, 2016
    Inventors: Ebru Arisoy, Bhuvana Ramabhadran, Abhinav Sethy, Stanley Chen
  • Patent number: 9484023
    Abstract: Techniques for conversion of non-back-off language models for use in speech decoders. For example, a method comprises the following step. A non-back-off language model is converted to a back-off language model. The converted back-off language model is pruned. The converted back-off language model is usable for decoding speech.
    Type: Grant
    Filed: February 22, 2013
    Date of Patent: November 1, 2016
    Assignee: International Business Machines Corporation
    Inventors: Ebru Arisoy, Bhuvana Ramabhadran, Abhinav Sethy, Stanley Chen
  • Publication number: 20140244261
    Abstract: Techniques for conversion of non-back-off language models for use in speech decoders. For example, an apparatus for conversion of non-back-off language models for use in speech decoders. For example, an apparatus is configured convert a non-back-off language model to a back-off language model. The converted back-off language model is pruned. The converted back-off language model is usable for decoding speech.
    Type: Application
    Filed: August 12, 2013
    Publication date: August 28, 2014
    Inventors: Ebru Arisoy, Bhuvana Ramabhadran, Abhinav Sethy, Stanley Chen
  • Publication number: 20140244248
    Abstract: Techniques for conversion of non-back-off language models for use in speech decoders. For example, a method comprises the following step. A non-back-off language model is converted to a back-off language model. The converted back-off language model is pruned. The converted back-off language model is usable for decoding speech.
    Type: Application
    Filed: February 22, 2013
    Publication date: August 28, 2014
    Applicant: International Business Machines Corporation
    Inventors: Ebru Arisoy, Bhuvana Ramabhadran, Abhinav Sethy, Stanley Chen
  • Publication number: 20140156575
    Abstract: Deep belief networks are usually associated with a large number of parameters and high computational complexity. The large number of parameters results in a long and computationally consuming training phase. According to at least one example embodiment, low-rank matrix factorization is used to approximate at least a first set of parameters, associated with an output layer, with a second and a third set of parameters. The total number of parameters in the second and third sets of parameters is smaller than the number of sets of parameters in the first set. An architecture of a resulting artificial neural network, when employing low-rank matrix factorization, may be characterized with a low-rank layer, not employing activation function(s), and defined by a relatively small number of nodes and the second set of parameters. By using low rank matrix factorization, training is faster, leading to rapid deployment of the respective system.
    Type: Application
    Filed: November 30, 2012
    Publication date: June 5, 2014
    Applicant: NUANCE COMMUNICATIONS, INC.
    Inventors: Tara N. Sainath, Ebru Arisoy, Bhuvana Ramabhadran