Patents by Inventor Michael Levit

Michael Levit has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 9460081
    Abstract: Examples of the present disclosure describe generation of a multi-arc confusion network to improve, for example, an ability to return alternatives to output generated. A confusion network comprising token representations of lexicalized hypotheses and normalized hypotheses is generated. Each arc of the confusion network represents a token of a lexicalized hypothesis or a normalized hypothesis. The confusion network is transformed into a multi-arc confusion network, wherein the transforming comprising realigning at least one token of the confusion network to span multiple arcs of the confusion network. Other examples are also described.
    Type: Grant
    Filed: June 2, 2016
    Date of Patent: October 4, 2016
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Michael Levit, Umut Ozertem, Sarangarajan Parthasarathy, Padma Varadharajan, Karthik Raghunathan, Issac Alphonso
  • Publication number: 20160275071
    Abstract: Examples of the present disclosure describe generation of a multi-arc confusion network to improve, for example, an ability to return alternatives to output generated. A confusion network comprising token representations of lexicalized hypotheses and normalized hypotheses is generated. Each arc of the confusion network represents a token of a lexicalized hypothesis or a normalized hypothesis. The confusion network is transformed into a multi-arc confusion network, wherein the transforming comprising realigning at least one token of the confusion network to span multiple arcs of the confusion network. Other examples are also described.
    Type: Application
    Filed: June 2, 2016
    Publication date: September 22, 2016
    Applicant: Microsoft Technology Licensing, LLC
    Inventors: Michael Levit, Umut Ozertem, Sarangarajan Parthasarathy, Padma Varadharajan, Karthik Raghunathan, Issac Alphonso
  • Publication number: 20160267905
    Abstract: Optimized language models are provided for in-domain applications through an iterative, joint-modeling approach that interpolates a language model (LM) from a number of component LMs according to interpolation weights optimized for a target domain. The component LMs may include class-based LMs, and the interpolation may be context-specific or context-independent. Through iterative processes, the component LMs may be interpolated and used to express training material as alternative representations or parses of tokens. Posterior probabilities may be determined for these parses and used for determining new (or updated) interpolation weights for the LM components, such that a combination or interpolation of component LMs is further optimized for the domain. The component LMs may be merged, according to the optimized weights, into a single, combined LM, for deployment in an application scenario.
    Type: Application
    Filed: March 11, 2015
    Publication date: September 15, 2016
    Inventors: Michael Levit, Sarangarajan Parthasarathy, Andreas Stolcke, Shuangyu Chang
  • Publication number: 20160217125
    Abstract: Examples of the present disclosure describe generation of a multi-arc confusion network to improve, for example, an ability to return alternatives to output generated. A confusion network comprising token representations of lexicalized hypotheses and normalized hypotheses is generated. Each arc of the confusion network represents a token of a lexicalized hypothesis or a normalized hypothesis. The confusion network is transformed into a multi-arc confusion network, wherein the transforming comprising realigning at least one token of the confusion network to span multiple arcs of the confusion network. Other examples are also described.
    Type: Application
    Filed: January 27, 2015
    Publication date: July 28, 2016
    Applicant: Microsoft Technology Licensing, LLC
    Inventors: Michael Levit, Umut Ozertem, Sarangarajan Parthasarathy, Padma Varadharajan, Karthik Raghunathan, Issac Alphonso
  • Patent number: 9384188
    Abstract: Examples of the present disclosure describe generation of a multi-arc confusion network to improve, for example, an ability to return alternatives to output generated. A confusion network comprising token representations of lexicalized hypotheses and normalized hypotheses is generated. Each arc of the confusion network represents a token of a lexicalized hypothesis or a normalized hypothesis. The confusion network is transformed into a multi-arc confusion network, wherein the transforming comprising realigning at least one token of the confusion network to span multiple arcs of the confusion network. Other examples are also described.
    Type: Grant
    Filed: January 27, 2015
    Date of Patent: July 5, 2016
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Michael Levit, Umut Ozertem, Sarangarajan Parthasarathy, Padma Varadharajan, Karthik Raghunathan, Issac Alphonso
  • Patent number: 9299342
    Abstract: Query history expansion may be provided. Upon receiving a spoken query from a user, an adapted language model may be applied to convert the spoken query to text. The adapted language model may comprise a plurality of queries interpolated from the user's previous queries and queries associated with other users. The spoken query may be executed and the results of the spoken query may be provided to the user.
    Type: Grant
    Filed: July 23, 2015
    Date of Patent: March 29, 2016
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Shuangyu Chang, Michael Levit, Bruce Melvin Buntschuh
  • Publication number: 20150325235
    Abstract: Systems and methods are provided for optimizing language models for in-domain applications through an iterative, joint-modeling approach that expresses training material as alternative representations of higher-level tokens, such as named entities and carrier phrases. From a first language model, an in-domain training corpus may be represented as a set of alternative parses of tokens. Statistical information determined from these parsed representations may be used to produce a second (or updated) language model, which is further optimized for the domain. The second language model may be used to determine another alternative parsed representation of the corpus for a next iteration, and the statistical information determined from this representation may be used to produce a third (or further updated) language model. Through each iteration, a language model may be determined that is further optimized for the domain.
    Type: Application
    Filed: May 7, 2014
    Publication date: November 12, 2015
    Applicant: Microsoft Corporation
    Inventors: Michael Levit, Sarangarajan Parthasarathy, Andreas Stolcke
  • Publication number: 20150325236
    Abstract: The customization of recognition of speech utilizing context-specific language model scale factors is provided. Training audio may be received from a source in a training phase. The received training audio may be recognized utilizing acoustic and language models being combined utilizing static scale factors. A comparison may then be made of the recognition results to a transcription of the training audio. The recognition results may include one or more hypotheses for recognizing speech. Context specific scale factors may then be generated based on the comparison. The context specific scale factors may then be applied for use in the speech recognition of audio signals in an application phase.
    Type: Application
    Filed: May 8, 2014
    Publication date: November 12, 2015
    Applicant: MICROSOFT CORPORATION
    Inventors: MICHAEL LEVIT, SHUANGYU CHANG, ZHIHENG HUANG
  • Publication number: 20150325237
    Abstract: Query history expansion may be provided. Upon receiving a spoken query from a user, an adapted language model may be applied to convert the spoken query to text. The adapted language model may comprise a plurality of queries interpolated from the user's previous queries and queries associated with other users. The spoken query may be executed and the results of the spoken query may be provided to the user.
    Type: Application
    Filed: July 23, 2015
    Publication date: November 12, 2015
    Applicant: MICROSOFT TECHNOLOGY LICENSING, LLC
    Inventors: Shuangyu Chang, Michael Levit, Bruce Melvin Buntschuh
  • Publication number: 20150278191
    Abstract: The customization of language modeling components for speech recognition is provided. A list of language modeling components may be made available by a computing device. A hint may then be sent to a recognition service provider for combining the multiple language modeling components from the list. The hint may be based on a number of different domains. A customized combination of the language modeling components based on the hint may then be received from the recognition service provider.
    Type: Application
    Filed: March 27, 2014
    Publication date: October 1, 2015
    Applicant: Microsoft Corporation
    Inventors: Michael Levit, Hernan Guelman, Shuangyu Chang, Sarangarajan Parthasarathy, Benoit Dumoulin
  • Publication number: 20150269949
    Abstract: An incremental speech recognition system. The incremental speech recognition system incrementally decodes a spoken utterance using an additional utterance decoder only when the additional utterance decoder is likely to add significant benefit to the combined result. The available utterance decoders are ordered in a series based on accuracy, performance, diversity, and other factors. A recognition management engine coordinates decoding of the spoken utterance by the series of utterance decoders, combines the decoded utterances, and determines whether additional processing is likely to significantly improve the recognition result. If so, the recognition management engine engages the next utterance decoder and the cycle continues. If the accuracy cannot be significantly improved, the result is accepted and decoding stops.
    Type: Application
    Filed: March 19, 2014
    Publication date: September 24, 2015
    Applicant: MICROSOFT CORPORATION
    Inventors: Shuangyu Chang, Michael Levit, Abhik Lahiri, Barlas Oguz, Benoit Dumoulin
  • Publication number: 20150269136
    Abstract: Various components provide options to re-format an input based on one or more contexts. The input is received that has been submitted to an application (e.g., messaging application, mobile application, word-processing application, web browser, search tool, etc.), and one or more outputs are identified that are possibilities to be provided as options for re-formatting. A respective score of each output is determined by applying a statistical model to a respective combination of the input and each output, the respective score comprising a plurality of context scores that quantify a plurality of contexts of the respective combination. Exemplary contexts include historical-user contexts, domain contexts, and general contexts. One or more suggested outputs are selected from among the one or more outputs based on the respective scores and are provided as options to re-format the input.
    Type: Application
    Filed: March 20, 2014
    Publication date: September 24, 2015
    Applicant: MICROSOFT CORPORATION
    Inventors: ISSAC ALPHONSO, NICK KIBRE, MICHAEL LEVIT, SARANGARAJAN PARTHASARATHY
  • Patent number: 9129606
    Abstract: Query history expansion may be provided. Upon receiving a spoken query from a user, an adapted language model may be applied to convert the spoken query to text. The adapted language model may comprise a plurality of queries interpolated from the user's previous queries and queries associated with other users. The spoken query may be executed and the results of the spoken query may be provided to the user.
    Type: Grant
    Filed: September 23, 2011
    Date of Patent: September 8, 2015
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Shuangyu Chang, Michael Levit, Bruce Melvin Buntschuh
  • Patent number: 9053087
    Abstract: A semantic error rate calculation may be provided. After receiving a spoken query from a user, the spoken query may be converted to text according to a first speech recognition hypothesis. A plurality of results associated with the converted query may be received and compared to a second plurality of results associated with the converted query.
    Type: Grant
    Filed: September 23, 2011
    Date of Patent: June 9, 2015
    Assignee: Microsoft Technology Licensing, LLC
    Inventors: Michael Levit, Shuangyu Chang, Bruce Melvin Buntschuh, Nick Kibre
  • Publication number: 20150046163
    Abstract: On a computing device a speech utterance is received from a user. The speech utterance is a section of a speech dialog that includes a plurality of speech utterances. One or more features from the speech utterance are identified. Each identified feature from the speech utterance is a specific characteristic of the speech utterance. One or more features from the speech dialog are identified. Each identified feature from the speech dialog is associated with one or more events in the speech dialog. The one or more events occur prior to the speech utterance. One or more identified features from the speech utterance and one or more identified features from the speech dialog are used to calculate a confidence score for the speech utterance.
    Type: Application
    Filed: October 23, 2014
    Publication date: February 12, 2015
    Applicant: Microsoft Corporation
    Inventors: Michael Levit, Bruce Melvin Buntschuh
  • Patent number: 8930179
    Abstract: Architecture that employs an overall grammar as a set of context-specific grammars for recognition of an input, each responsible for a specific context, such as subtask category, geographic region, etc. The grammars together cover the entire domain. Moreover, multiple recognitions can be run in parallel against the same input, where each recognition uses one or more of the context-specific grammars. The multiple intermediate recognition results from the different recognizer-grammars are reconciled by running re-recognition using a dynamically composed grammar based on the multiple recognition results and potentially other domain knowledge, or selecting the winner using a statistical classifier operating on classification features extracted from the multiple recognition results and other domain knowledge.
    Type: Grant
    Filed: June 4, 2009
    Date of Patent: January 6, 2015
    Assignee: Microsoft Corporation
    Inventors: Shuangyu Chang, Michael Levit, Bruce Buntschuh
  • Publication number: 20140365218
    Abstract: A received utterance is recognized using different language models. For example, recognition of the utterance is independently performed using a baseline language model (BLM) and using an adapted language model (ALM). A determination is made as to what results from the different language model are more likely to be accurate. Different features may be used to assist in making the determination (e.g. language model scores, recognition confidences, acoustic model scores, quality measurements, . . . ) may be used. A classifier may be trained and then used in determining whether to select the results using the BLM or to select the results using the ALM. A language model may be automatically trained or re-trained that adjusts a weight of the training data used in training the model in response to differences between the two results obtained from applying the different language models.
    Type: Application
    Filed: June 7, 2013
    Publication date: December 11, 2014
    Inventors: Shuangyu Chang, Michael Levit
  • Publication number: 20140350931
    Abstract: A Statistical Machine Translation (SMT) model is trained using pairs of sentences that include content obtained from one or more content sources (e.g. feed(s)) with corresponding queries that have been used to access the content. A query click graph may be used to assist in determining candidate pairs for the SMT training data. All/portion of the candidate pairs may be used to train the SMT model. After training the SMT model using the SMT training data, the SMT model is applied to content to determine predicted queries that may be used to search for the content. The predicted queries are used to train a language model, such as a query language model. The query language model may be interpolated other language models, such as a background language model, as well as a feed language model trained using the content used in determining the predicted queries.
    Type: Application
    Filed: May 24, 2013
    Publication date: November 27, 2014
    Applicant: Microsoft Corporation
    Inventors: Michael Levit, Dilek Hakkani-Tur, Gokhan Tur
  • Patent number: 8886532
    Abstract: On a computing device a speech utterance is received from a user. The speech utterance is a section of a speech dialog that includes a plurality of speech utterances. One or more features from the speech utterance are identified. Each identified feature from the speech utterance is a specific characteristic of the speech utterance. One or more features from the speech dialog are identified. Each identified feature from the speech dialog is associated with one or more events in the speech dialog. The one or more events occur prior to the speech utterance. One or more identified features from the speech utterance and one or more identified features from the speech dialog are used to calculate a confidence score for the speech utterance.
    Type: Grant
    Filed: October 27, 2010
    Date of Patent: November 11, 2014
    Assignee: Microsoft Corporation
    Inventors: Michael Levit, Bruce Melvin Buntschuh
  • Publication number: 20130080150
    Abstract: A semantic error rate calculation may be provided. After receiving a spoken query from a user, the spoken query may be converted to text according to a first speech recognition hypothesis. A plurality of results associated with the converted query may be received and compared to a second plurality of results associated with the converted query.
    Type: Application
    Filed: September 23, 2011
    Publication date: March 28, 2013
    Applicant: Microsoft Corporation
    Inventors: Michael Levit, Shuangyu Chang, Bruce Melvin Buntschuh, Nick Kibre