Patents by Inventor Jerome Bellegarda

Jerome Bellegarda has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 9760559
    Abstract: Systems and processes for predictive text input are provided. In one example process, a text input can be received. The text input can be associated with an input context. A frequency of occurrence of an m-gram with respect to a subset of a corpus can be determined using a language model. The subset can be associated with a context. A weighting factor can be determined based on a degree of similarity between the input context and the context. A weighted probability of a predicted text given the text input can be determined based on the frequency of occurrence of the m-gram and the weighting factor. The m-gram can include at least one word in the text input and at least one word in the predicted text.
    Type: Grant
    Filed: May 22, 2015
    Date of Patent: September 12, 2017
    Assignee: Apple Inc.
    Inventors: Jannes Dolfing, Brent Ramerth, Douglas Davidson, Jerome Bellegarda, Jennifer Moore, Andreas Eminidis, Joshua Shaffer
  • Publication number: 20150347382
    Abstract: Systems and processes for predictive text input are provided. In one example process, a text input can be received. The text input can be associated with an input context. A frequency of occurrence of an m-gram with respect to a subset of a corpus can be determined using a language model. The subset can be associated with a context. A weighting factor can be determined based on a degree of similarity between the input context and the context. A weighted probability of a predicted text given the text input can be determined based on the frequency of occurrence of the m-gram and the weighting factor. The m-gram can include at least one word in the text input and at least one word in the predicted text.
    Type: Application
    Filed: May 22, 2015
    Publication date: December 3, 2015
    Inventors: Jannes DOLFING, Brent RAMERTH, Douglas DAVIDSON, Jerome BELLEGARDA, Jennifer MOORE, Andreas EMINIDIS, Joshua SHAFFER
  • Patent number: 9053089
    Abstract: Methods and apparatuses to assign part-of-speech tags to words are described. An input sequence of words is received. A global fabric of a corpus having training sequences of words may be analyzed in a vector space. A global semantic information associated with the input sequence of words may be extracted based on the analyzing. A part-of-speech tag may be assigned to a word of the input sequence based on POS tags from pertinent words in relevant training sequences identified using the global semantic information. The input sequence may be mapped into a vector space. A neighborhood associated with the input sequence may be formed in the vector space wherein the neighborhood represents one or more training sequences that are globally relevant to the input sequence.
    Type: Grant
    Filed: October 2, 2007
    Date of Patent: June 9, 2015
    Assignee: Apple Inc.
    Inventor: Jerome Bellegarda
  • Patent number: 8935167
    Abstract: Methods, systems, and computer-readable media related to selecting observation-specific training data (also referred to as “observation-specific exemplars”) from a general training corpus, and then creating, from the observation-specific training data, a focused, observation-specific acoustic model for recognizing the observation in an output domain are disclosed. In one aspect, a global speech recognition model is established based on an initial set of training data; a plurality of input speech segments to be recognized in an output domain are received; and for each of the plurality of input speech segments: a respective set of focused training data relevant to the input speech segment is identified in the global speech recognition model; a respective focused speech recognition model is generated based on the respective set of focused training data; and the respective focused speech recognition model is provided to a recognition device for recognizing the input speech segment in the output domain.
    Type: Grant
    Filed: September 25, 2012
    Date of Patent: January 13, 2015
    Assignee: Apple Inc.
    Inventor: Jerome Bellegarda
  • Patent number: 8712776
    Abstract: Algorithms for synthesizing speech used to identify media assets are provided. Speech may be selectively synthesized form text strings associated with media assets. A text string may be normalized and its native language determined for obtaining a target phoneme for providing human-sounding speech in a language (e.g., dialect or accent) that is familiar to a user. The algorithms may be implemented on a system including several dedicated render engines. The system may be part of a back end coupled to a front end including storage for media assets and associated synthesized speech, and a request processor for receiving and processing requests that result in providing the synthesized speech. The front end may communicate media assets and associated synthesized speech content over a network to host devices coupled to portable electronic devices on which the media assets and synthesized speech are played back.
    Type: Grant
    Filed: September 29, 2008
    Date of Patent: April 29, 2014
    Assignee: Apple Inc.
    Inventors: Jerome Bellegarda, Devang Naik, Kim Silverman
  • Publication number: 20140088964
    Abstract: Methods, systems, and computer-readable media related to selecting observation-specific training data (also referred to as “observation-specific exemplars”) from a general training corpus, and then creating, from the observation-specific training data, a focused, observation-specific acoustic model for recognizing the observation in an output domain are disclosed. In one aspect, a global speech recognition model is established based on an initial set of training data; a plurality of input speech segments to be recognized in an output domain are received; and for each of the plurality of input speech segments: a respective set of focused training data relevant to the input speech segment is identified in the global speech recognition model; a respective focused speech recognition model is generated based on the respective set of focused training data; and the respective focused speech recognition model is provided to a recognition device for recognizing the input speech segment in the output domain.
    Type: Application
    Filed: September 25, 2012
    Publication date: March 27, 2014
    Applicant: APPLE INC.
    Inventor: Jerome Bellegarda
  • Patent number: 8682649
    Abstract: A semantically organized domain space is created from a training corpus. Affective data are mapped onto the domain space to generate affective anchors for the domain space. A sentiment associated with an input text is determined based the affective anchors. A speech output may be generated from the input text based on the determined sentiment.
    Type: Grant
    Filed: November 12, 2009
    Date of Patent: March 25, 2014
    Assignee: Apple Inc.
    Inventor: Jerome Bellegarda
  • Patent number: 8626930
    Abstract: Methods and apparatuses to filter multimedia content are described. The multimedia content in one embodiment is analyzed for one or more parameters. The multimedia content in one embodiment is filtered based on the one or more parameters using a latent semantic mapping (“LSM”) filter. In one embodiment, the one or more parameters include information about a structure of the multimedia content. A tag that encapsulates the one or more parameters may be generated. Then, the tag is input into the latent semantic mapping filter. In one embodiment, the LSM filter is trained to recognize the multimedia content based on the one or more parameters. In one embodiment, more than two categories are provided for a multimedia content. The multimedia content is classified in more than two categories using the LSM filter. The multimedia content may be blocked based on the classifying.
    Type: Grant
    Filed: March 15, 2007
    Date of Patent: January 7, 2014
    Assignee: Apple Inc.
    Inventors: Giovanni Donelli, Jerome Bellegarda, Steve Ko, John Scalo
  • Patent number: 8620662
    Abstract: Methods and apparatuses to perform context-aware unit selection for natural language processing are described. Streams of information associated with input units are received. The streams of information are analyzed in a context associated with first candidate units to determine a first set of weights of the streams of information. A first candidate unit is selected from the first candidate units based on the first set of weights of the streams of information. The streams of information are analyzed in the context associated with second candidate units to determine a second set of weights of the streams of information. A second candidate unit is selected from second candidate units to concatenate with the first candidate unit based on the second set of weights of the streams of information.
    Type: Grant
    Filed: November 20, 2007
    Date of Patent: December 31, 2013
    Assignee: Apple Inc.
    Inventor: Jerome Bellegarda
  • Patent number: 8355919
    Abstract: Algorithms for synthesizing speech used to identify media assets are provided. Speech may be selectively synthesized form text strings associated with media assets. A text string may be normalized and its native language determined for obtaining a target phoneme for providing human-sounding speech in a language (e.g., dialect or accent) that is familiar to a user. The algorithms may be implemented on a system including several dedicated render engines. The system may be part of a back end coupled to a front end including storage for media assets and associated synthesized speech, and a request processor for receiving and processing requests that result in providing the synthesized speech. The front end may communicate media assets and associated synthesized speech content over a network to host devices coupled to portable electronic devices on which the media assets and synthesized speech are played back.
    Type: Grant
    Filed: September 29, 2008
    Date of Patent: January 15, 2013
    Assignee: Apple Inc.
    Inventors: Kim Silverman, Devang Naik, Jerome Bellegarda, Kevin Lenzo
  • Patent number: 8352268
    Abstract: Algorithms for synthesizing speech used to identify media assets are provided. Speech may be selectively synthesized form text strings associated with media assets. A text string may be normalized and its native language determined for obtaining a target phoneme for providing human-sounding speech in a language (e.g., dialect or accent) that is familiar to a user. The algorithms may be implemented on a system including several dedicated render engines. The system may be part of a back end coupled to a front end including storage for media assets and associated synthesized speech, and a request processor for receiving and processing requests that result in providing the synthesized speech. The front end may communicate media assets and associated synthesized speech content over a network to host devices coupled to portable electronic devices on which the media assets and synthesized speech are played back.
    Type: Grant
    Filed: September 29, 2008
    Date of Patent: January 8, 2013
    Assignee: Apple Inc.
    Inventors: DeVang Naik, Kim Silverman, Jerome Bellegarda
  • Publication number: 20110112825
    Abstract: A semantically organized domain space is created from a training corpus. Affective data are mapped onto the domain space to generate affective anchors for the domain space. A sentiment associated with an input text is determined based the affective anchors. A speech output may be generated from the input text based on the determined sentiment.
    Type: Application
    Filed: November 12, 2009
    Publication date: May 12, 2011
    Inventor: Jerome Bellegarda
  • Publication number: 20100082344
    Abstract: Algorithms for synthesizing speech used to identify media assets are provided. Speech may be selectively synthesized form text strings associated with media assets. A text string may be normalized and its native language determined for obtaining a target phoneme for providing human-sounding speech in a language (e.g., dialect or accent) that is familiar to a user. The algorithms may be implemented on a system including several dedicated render engines. The system may be part of a back end coupled to a front end including storage for media assets and associated synthesized speech, and a request processor for receiving and processing requests that result in providing the synthesized speech. The front end may communicate media assets and associated synthesized speech content over a network to host devices coupled to portable electronic devices on which the media assets and synthesized speech are played back.
    Type: Application
    Filed: September 29, 2008
    Publication date: April 1, 2010
    Applicant: Apple, Inc.
    Inventors: Devang Naik, Kim Silverman, Jerome Bellegarda
  • Publication number: 20100082327
    Abstract: Algorithms for synthesizing speech used to identify media assets are provided. Speech may be selectively synthesized form text strings associated with media assets. A text string may be normalized and its native language determined for obtaining a target phoneme for providing human-sounding speech in a language (e.g., dialect or accent) that is familiar to a user. The algorithms may be implemented on a system including several dedicated render engines. The system may be part of a back end coupled to a front end including storage for media assets and associated synthesized speech, and a request processor for receiving and processing requests that result in providing the synthesized speech. The front end may communicate media assets and associated synthesized speech content over a network to host devices coupled to portable electronic devices on which the media assets and synthesized speech are played back.
    Type: Application
    Filed: September 29, 2008
    Publication date: April 1, 2010
    Applicant: Apple Inc.
    Inventors: Matthew Rogers, Kim Silverman, Devang Naik, Kevin Lenzo, Jerome Bellegarda
  • Publication number: 20100082349
    Abstract: Algorithms for synthesizing speech used to identify media assets are provided. Speech may be selectively synthesized form text strings associated with media assets. A text string may be normalized and its native language determined for obtaining a target phoneme for providing human-sounding speech in a language (e.g., dialect or accent) that is familiar to a user. The algorithms may be implemented on a system including several dedicated render engines. The system may be part of a back end coupled to a front end including storage for media assets and associated synthesized speech, and a request processor for receiving and processing requests that result in providing the synthesized speech. The front end may communicate media assets and associated synthesized speech content over a network to host devices coupled to portable electronic devices on which the media assets and synthesized speech are played back.
    Type: Application
    Filed: September 29, 2008
    Publication date: April 1, 2010
    Applicant: Apple Inc.
    Inventors: Jerome Bellegarda, Devang Naik, Kim Silverman
  • Publication number: 20100082348
    Abstract: Algorithms for synthesizing speech used to identify media assets are provided. Speech may be selectively synthesized form text strings associated with media assets. A text string may be normalized and its native language determined for obtaining a target phoneme for providing human-sounding speech in a language (e.g., dialect or accent) that is familiar to a user. The algorithms may be implemented on a system including several dedicated render engines. The system may be part of a back end coupled to a front end including storage for media assets and associated synthesized speech, and a request processor for receiving and processing requests that result in providing the synthesized speech. The front end may communicate media assets and associated synthesized speech content over a network to host devices coupled to portable electronic devices on which the media assets and synthesized speech are played back.
    Type: Application
    Filed: September 29, 2008
    Publication date: April 1, 2010
    Applicant: Apple Inc.
    Inventors: Kim Silverman, Devang Naik, Jerome Bellegarda, Kevin Lenzo
  • Publication number: 20090132253
    Abstract: Methods and apparatuses to perform context-aware unit selection for natural language processing are described. Streams of information associated with input units are received. The streams of information are analyzed in a context associated with first candidate units to determine a first set of weights of the streams of information. A first candidate unit is selected from the first candidate units based on the first set of weights of the streams of information. The streams of information are analyzed in the context associated with second candidate units to determine a second set of weights of the streams of information. A second candidate unit is selected from second candidate units to concatenate with the first candidate unit based on the second set of weights of the streams of information.
    Type: Application
    Filed: November 20, 2007
    Publication date: May 21, 2009
    Inventor: Jerome Bellegarda
  • Publication number: 20090089058
    Abstract: Methods and apparatuses to assign part-of-speech tags to words are described. An input sequence of words is received. A global fabric of a corpus having training sequences of words may be analyzed in a vector space. A global semantic information associated with the input sequence of words may be extracted based on the analyzing. A part-of-speech tag may be assigned to a word of the input sequence based on POS tags from pertinent words in relevant training sequences identified using the global semantic information. The input sequence may be mapped into a vector space. A neighborhood associated with the input sequence may be formed in the vector space wherein the neighborhood represents one or more training sequences that are globally relevant to the input sequence.
    Type: Application
    Filed: October 2, 2007
    Publication date: April 2, 2009
    Inventor: Jerome Bellegarda
  • Publication number: 20080228928
    Abstract: Methods and apparatuses to filter multimedia content are described. The multimedia content in one embodiment is analyzed for one or more parameters. The multimedia content in one embodiment is filtered based on the one or more parameters using a latent semantic mapping (“LSM”) filter. In one embodiment, the one or more parameters include information about a structure of the multimedia content. A tag that encapsulates the one or more parameters may be generated. Then, the tag is input into the latent semantic mapping filter. In one embodiment, the LSM filter is trained to recognize the multimedia content based on the one or more parameters. In one embodiment, more than two categories are provided for a multimedia content. The multimedia content is classified in more than two categories using the LSM filter. The multimedia content may be blocked based on the classifying.
    Type: Application
    Filed: March 15, 2007
    Publication date: September 18, 2008
    Inventors: Giovanni Donelli, Jerome Bellegarda, Steve Ko, John Scalo
  • Publication number: 20080091430
    Abstract: A method and apparatus is provided for generating speech that sounds more natural. In one embodiment, word prominence and latent semantic analysis are used to generate more natural sounding speech. A method for generating speech that sounds more natural may comprise generating synthesized speech having certain word prominence characteristics and applying a semantically-driven word prominence assignment model to specify word prominence consistent with the way humans assign word prominence.
    Type: Application
    Filed: December 4, 2007
    Publication date: April 17, 2008
    Inventors: Jerome Bellegarda, Kim Silverman