Patents by Inventor Girija Yegnanarayanan

Girija Yegnanarayanan has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 9520124
    Abstract: A system is provided for training an acoustic model for use in speech recognition. In particular, such a system may be used to perform training based on a spoken audio stream and a non-literal transcript of the spoken audio stream. Such a system may identify text in the non-literal transcript which represents concepts having multiple spoken forms. The system may attempt to identify the actual spoken form in the audio stream which produced the corresponding text in the non-literal transcript, and thereby produce a revised transcript which more accurately represents the spoken audio stream. The revised, and more accurate, transcript may be used to train the acoustic model using discriminative training techniques, thereby producing a better acoustic model than that which would be produced using conventional techniques, which perform training based directly on the original non-literal transcript.
    Type: Grant
    Filed: November 16, 2015
    Date of Patent: December 13, 2016
    Assignee: MModal IP LLC
    Inventors: Lambert Mathias, Girija Yegnanarayanan, Juergen Fritsch
  • Publication number: 20160196821
    Abstract: A system is provided for training an acoustic model for use in speech recognition. In particular, such a system may be used to perform training based on a spoken audio stream and a non-literal transcript of the spoken audio stream. Such a system may identify text in the non-literal transcript which represents concepts having multiple spoken forms. The system may attempt to identify the actual spoken form in the audio stream which produced the corresponding text in the non-literal transcript, and thereby produce a revised transcript which more accurately represents the spoken audio stream. The revised, and more accurate, transcript may be used to train the acoustic model, thereby producing a better acoustic model than that which would be produced using conventional techniques, which perform training based directly on the original non-literal transcript.
    Type: Application
    Filed: March 10, 2016
    Publication date: July 7, 2016
    Applicant: MModal IP LLC
    Inventors: Girija Yegnanarayanan, Michael Finke, Juergen Fritsch, Detlef Koll, Monika Woszczyna
  • Publication number: 20160078861
    Abstract: A system is provided for training an acoustic model for use in speech recognition. In particular, such a system may be used to perform training based on a spoken audio stream and a non-literal transcript of the spoken audio stream. Such a system may identify text in the non-literal transcript which represents concepts having multiple spoken forms. The system may attempt to identify the actual spoken form in the audio stream which produced the corresponding text in the non-literal transcript, and thereby produce a revised transcript which more accurately represents the spoken audio stream. The revised, and more accurate, transcript may be used to train the acoustic model using discriminative training techniques, thereby producing a better acoustic model than that which would be produced using conventional techniques, which perform training based directly on the original non-literal transcript.
    Type: Application
    Filed: November 16, 2015
    Publication date: March 17, 2016
    Applicant: MMODAL IP LLC
    Inventors: Lambert Mathias, Girija Yegnanarayanan, Juergen Fritsch
  • Patent number: 9286896
    Abstract: A system is provided for training an acoustic model for use in speech recognition. In particular, such a system may be used to perform training based on a spoken audio stream and a non-literal transcript of the spoken audio stream. Such a system may identify text in the non-literal transcript which represents concepts having multiple spoken forms. The system may attempt to identify the actual spoken form in the audio stream which produced the corresponding text in the non-literal transcript, and thereby produce a revised transcript which more accurately represents the spoken audio stream. The revised, and more accurate, transcript may be used to train the acoustic model, thereby producing a better acoustic model than that which would be produced using conventional techniques, which perform training based directly on the original non-literal transcript.
    Type: Grant
    Filed: May 16, 2014
    Date of Patent: March 15, 2016
    Assignee: MModal IP LLC
    Inventors: Girija Yegnanarayanan, Michael Finke, Juergen Fritsch, Detlef Koll, Monika Woszczyna
  • Patent number: 9190050
    Abstract: A system is provided for training an acoustic model for use in speech recognition. In particular, such a system may be used to perform training based on a spoken audio stream and a non-literal transcript of the spoken audio stream. Such a system may identify text in the non-literal transcript which represents concepts having multiple spoken forms. The system may attempt to identify the actual spoken form in the audio stream which produced the corresponding text in the non-literal transcript, and thereby produce a revised transcript which more accurately represents the spoken audio stream. The revised, and more accurate, transcript may be used to train the acoustic model using discriminative training techniques, thereby producing a better acoustic model than that which would be produced using conventional techniques, which perform training based directly on the original non-literal transcript.
    Type: Grant
    Filed: April 3, 2014
    Date of Patent: November 17, 2015
    Assignee: MModal IP LLC
    Inventors: Lambert Mathias, Girija Yegnanarayanan, Juergen Fritsch
  • Patent number: 9135571
    Abstract: Techniques for entity detection include matching a token from at least a portion of a text string with a matching concept in an ontology, wherein the at least a portion of the text string has been labeled as corresponding to a particular entity type. A first concept may be identified as being hierarchically related to the matching concept within the ontology, and a second concept may be identified as being hierarchically related to the first concept within the ontology. Based at least in part on the labeling of the at least a portion of the text string as corresponding to the particular entity type, a statistical model may be trained to associate the first concept with a first probability of corresponding to the particular entity type and the second concept with a second probability of corresponding to the particular entity type.
    Type: Grant
    Filed: March 12, 2013
    Date of Patent: September 15, 2015
    Assignee: Nuance Communications, Inc.
    Inventors: Brian W. Delaney, Girija Yegnanarayanan
  • Patent number: 9129013
    Abstract: Techniques for entity detection include matching a token from at least a portion of a text string with a matching concept in an ontology. A first concept may be identified as being hierarchically related to the matching concept within the ontology, and a second concept may be identified as being hierarchically related to the first concept within the ontology. The first and second concepts may be included in a set of features of the token. Based at least in part on the set of features of the token, a measure related to a likelihood that the at least a portion of the text string corresponds to a particular entity type may be determined.
    Type: Grant
    Filed: March 12, 2013
    Date of Patent: September 8, 2015
    Assignee: Nuance Communications, Inc.
    Inventors: Brian W. Delaney, Girija Yegnanarayanan
  • Publication number: 20150006199
    Abstract: Cascaded models may be applied to extract facts from a medical text. A first model may be applied to at least a portion of the medical text. The first model extracts at least one first medical fact. The at least one first medical fact is linked to at least first text in the at least a portion of the medical text. A second model may be applied to the first text. The second model extracts at least one second fact that is an attribute of the at least one first medical fact.
    Type: Application
    Filed: June 26, 2013
    Publication date: January 1, 2015
    Inventors: Neal E. Snider, Brian William Delaney, Girija Yegnanarayanan, Radu Florian, Martin Franz, Scott McCarley, John F. Pitrelli, Imed Zitouni, Salim E. Roukos
  • Publication number: 20140343939
    Abstract: A system is provided for training an acoustic model for use in speech recognition. In particular, such a system may be used to perform training based on a spoken audio stream and a non-literal transcript of the spoken audio stream. Such a system may identify text in the non-literal transcript which represents concepts having multiple spoken forms. The system may attempt to identify the actual spoken form in the audio stream which produced the corresponding text in the non-literal transcript, and thereby produce a revised transcript which more accurately represents the spoken audio stream. The revised, and more accurate, transcript may be used to train the acoustic model using discriminative training techniques, thereby producing a better acoustic model than that which would be produced using conventional techniques, which perform training based directly on the original non-literal transcript.
    Type: Application
    Filed: April 3, 2014
    Publication date: November 20, 2014
    Applicant: MModal IP LLC
    Inventors: Lambert Mathias, Girija Yegnanarayanan, Juergen Fritsch
  • Publication number: 20140280353
    Abstract: Techniques for entity detection include matching a token from at least a portion of a text string with a matching concept in an ontology. A first concept may be identified as being hierarchically related to the matching concept within the ontology, and a second concept may be identified as being hierarchically related to the first concept within the ontology. The first and second concepts may be included in a set of features of the token. Based at least in part on the set of features of the token, a measure related to a likelihood that the at least a portion of the text string corresponds to a particular entity type may be determined.
    Type: Application
    Filed: March 12, 2013
    Publication date: September 18, 2014
    Applicant: Nuance Communications, Inc.
    Inventors: Brian W. Delaney, Girija Yegnanarayanan
  • Publication number: 20140279729
    Abstract: Techniques for entity detection include matching a token from at least a portion of a text string with a matching concept in an ontology, wherein the at least a portion of the text string has been labeled as corresponding to a particular entity type. A first concept may be identified as being hierarchically related to the matching concept within the ontology, and a second concept may be identified as being hierarchically related to the first concept within the ontology. Based at least in part on the labeling of the at least a portion of the text string as corresponding to the particular entity type, a statistical model may be trained to associate the first concept with a first probability of corresponding to the particular entity type and the second concept with a second probability of corresponding to the particular entity type.
    Type: Application
    Filed: March 12, 2013
    Publication date: September 18, 2014
    Inventors: Brian W. Delaney, Girija Yegnanarayanan
  • Publication number: 20140249818
    Abstract: A system is provided for training an acoustic model for use in speech recognition. In particular, such a system may be used to perform training based on a spoken audio stream and a non-literal transcript of the spoken audio stream. Such a system may identify text in the non-literal transcript which represents concepts having multiple spoken forms. The system may attempt to identify the actual spoken form in the audio stream which produced the corresponding text in the non-literal transcript, and thereby produce a revised transcript which more accurately represents the spoken audio stream. The revised, and more accurate, transcript may be used to train the acoustic model, thereby producing a better acoustic model than that which would be produced using conventional techniques, which perform training based directly on the original non-literal transcript.
    Type: Application
    Filed: May 16, 2014
    Publication date: September 4, 2014
    Applicant: MModal IP LLC
    Inventors: Girija Yegnanarayanan, Michael Finke, Juergen Fritsch, Detlef Koll, Monika Woszczyna
  • Patent number: 8768723
    Abstract: An original text that is a representation of a narration of a patient encounter provided by a clinician may be received and re-formatted to produce a formatted text. One or more clinical facts may be extracted from the formatted text. A first fact of the clinical facts may be extracted from a first portion of the formatted text, and the first portion of the formatted text may be a formatted version of a first portion of the original text. A linkage may be maintained between the first fact and the first portion of the original text.
    Type: Grant
    Filed: February 18, 2011
    Date of Patent: July 1, 2014
    Assignee: Nuance Communications, Inc.
    Inventors: Frank Montyne, David Decraene, Joeri Van der Vloet, Johan Raedemaeker, Ignace Desimpel, Frederik Coppens, Tom Deray, James R. Flanagan, Mariana Casella dos Santos, Marnix Holvoet, Maria van Gurp, David Hellman, Girija Yegnanarayanan, Karen Anne Doyle
  • Patent number: 8756079
    Abstract: A method for applying a user correction to medical fact extraction may include extracting, using at least one processor, a first set of one or more medical facts from a first portion of a text documenting a patient encounter. A correction to the first set of medical facts may be received from a user documenting the patient encounter. A second set of one or more medical facts may be extracted from a second portion of the text based at least in part on the user's correction to the first set of medical facts. An apparatus for applying a user correction to medical fact extraction may include at least one processor and at least one computer-readable storage medium storing processor-executable instructions that, when executed by the at least one processor, perform the above-described method. At least one computer-readable storage medium may be encoded with computer-executable instructions that perform the above-described method.
    Type: Grant
    Filed: June 8, 2012
    Date of Patent: June 17, 2014
    Assignee: Nuance Communications, Inc.
    Inventor: Girija Yegnanarayanan
  • Publication number: 20140164023
    Abstract: Techniques for applying user corrections to medical fact extraction may include extracting a first set of one or more medical facts from a first portion of text documenting a patient encounter. A correction to the first set of medical facts may be received from a user. The correction may identify a fact that should be associated with the first portion of the text. A second set of one or more medical facts may be extracted from a second portion of the text based at least in part on the user's correction to the first set of medical facts. Extracting the second set of facts may include extracting one or more facts similar to the identified fact from the second portion of the text.
    Type: Application
    Filed: February 14, 2014
    Publication date: June 12, 2014
    Applicant: Nuance Communications, Inc.
    Inventor: Girija Yegnanarayanan
  • Patent number: 8731920
    Abstract: A system is provided for training an acoustic model for use in speech recognition. In particular, such a system may be used to perform training based on a spoken audio stream and a non-literal transcript of the spoken audio stream. Such a system my identify text in the non-literal transcript which represents concepts having multiple spoken forms. The system may attempt to identify the actual spoken form in the audio stream which produced the corresponding text in the non-literal transcript, and thereby produce a revised transcript which more accurately represents the spoken audio stream. The revised, and more accurate, transcript may be used to train the acoustic model, thereby producing a better acoustic model than that which would be produced using conventional techniques, which perform training based directly on the original non-literal transcript.
    Type: Grant
    Filed: November 30, 2012
    Date of Patent: May 20, 2014
    Assignee: MModal IP LLC
    Inventors: Girija Yegnanarayanan, Michael Finke, Juergen Fritsch, Detlef Koll, Monika Woszczyna
  • Patent number: 8694312
    Abstract: A system is provided for training an acoustic model for use in speech recognition. In particular, such a system may be used to perform training based on a spoken audio stream and a non-literal transcript of the spoken audio stream. Such a system may identify text in the non-literal transcript which represents concepts having multiple spoken forms. The system may attempt to identify the actual spoken form in the audio stream which produced the corresponding text in the non-literal transcript, and thereby produce a revised transcript which more accurately represents the spoken audio stream. The revised, and more accurate, transcript may be used to train the acoustic model using discriminative training techniques, thereby producing a better acoustic model than that which would be produced using conventional techniques, which perform training based directly on the original non-literal transcript.
    Type: Grant
    Filed: February 22, 2013
    Date of Patent: April 8, 2014
    Assignee: MModal IP LLC
    Inventors: Lambert Mathias, Girija Yegnanarayanan, Juergen Fritsch
  • Patent number: 8694335
    Abstract: Techniques for applying user corrections to medical fact extraction may include extracting a first set of one or more medical facts from a first portion of text documenting a patient encounter. A correction to the first set of medical facts may be received from a user. The correction may identify a fact that should be associated with the first portion of the text. A second set of one or more medical facts may be extracted from a second portion of the text based at least in part on the user's correction to the first set of medical facts. Extracting the second set of facts may include extracting one or more facts similar to the identified fact from the second portion of the text.
    Type: Grant
    Filed: October 12, 2012
    Date of Patent: April 8, 2014
    Assignee: Nuance Communications, Inc.
    Inventor: Girija Yegnanarayanan
  • Publication number: 20130304453
    Abstract: Techniques are disclosed for automatically generating structured documents based on speech, including identification of relevant concepts and their interpretation. In one embodiment, a structured document generator uses an integrated process to generate a structured textual document (such as a structured textual medical report) based on a spoken audio stream. The spoken audio stream may be recognized using a language model which includes a plurality of sub-models arranged in a hierarchical structure. Each of the sub-models may correspond to a concept that is expected to appear in the spoken audio stream. Different portions of the spoken audio stream may be recognized using different sub-models. The resulting structured textual document may have a hierarchical structure that corresponds to the hierarchical structure of the language sub-models that were used to generate the structured textual document.
    Type: Application
    Filed: May 22, 2009
    Publication date: November 14, 2013
    Inventors: Juergen Fritsch, Michael Finke, Detlef Koll, Monika Woszczyna, Girija Yegnanarayanan
  • Patent number: 8412521
    Abstract: A system is provided for training an acoustic model for use in speech recognition. In particular, such a system may be used to perform training based on a spoken audio stream and a non-literal transcript of the spoken audio stream. Such a system may identify text in the non-literal transcript which represents concepts having multiple spoken forms. The system may attempt to identify the actual spoken form in the audio stream which produced the corresponding text in the non-literal transcript, and thereby produce a revised transcript which more accurately represents the spoken audio stream. The revised, and more accurate, transcript may be used to train the acoustic model using discriminative training techniques, thereby producing a better acoustic model than that which would be produced using conventional techniques, which perform training based directly on the original non-literal transcript.
    Type: Grant
    Filed: September 16, 2005
    Date of Patent: April 2, 2013
    Assignee: Multimodal Technologies, LLC
    Inventors: Lambert Mathias, Girija Yegnanarayanan, Juergen Fritsch