Patents by Inventor Mazin G. Rahim

Mazin G. Rahim has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 7567906
    Abstract: Systems and methods for generating an annotation guide. Speech data is organized and presented to a user. After the user selects some of the utterances in the speech data, the selected utterances are included in a class and/or call type. Additional utterances that belong to the class and/or call type can be found in the speech data using relevance feedback, data mining, data clustering, support vector machines, and the like. After a call type is complete, it is committed to the annotation guide. After all call types are completed, the annotation guide is generated.
    Type: Grant
    Filed: April 3, 2007
    Date of Patent: July 28, 2009
    Assignee: AT&T Intellectual Property II, L.P.
    Inventors: Lee Begeja, Harris Drucker, David Crawford Gibbon, Allen Louis Gorin, Patrick Guy Haffner, Steven H Lewis, Zhu Liu, Mazin G Rahim, Bernard S. Renger, Behzad Shahraray
  • Patent number: 7562014
    Abstract: A large amount of human labor is required to transcribe and annotate a training corpus that is needed to create and update models for automatic speech recognition (ASR) and spoken language understanding (SLU). Active learning enables a reduction in the amount of transcribed and annotated data required to train ASR and SLU models. In one aspect of the present invention, an active learning ASR process and active learning SLU process are coupled, thereby enabling further efficiencies to be gained relative to a process that maintains an isolation of data in both the ASR and SLU domains.
    Type: Grant
    Filed: September 26, 2007
    Date of Patent: July 14, 2009
    Assignee: AT&T Intellectual Property II, L.P.
    Inventors: Dilek Z Hakkani-Tur, Mazin G Rahim, Giuseppe Riccardi, Gokhan Tur
  • Patent number: 7562017
    Abstract: An active labeling process is provided that aims to minimize the number of utterances to be checked again by automatically selecting the ones that are likely to be erroneous or inconsistent with the previously labeled examples. In one embodiment, the errors and inconsistencies are identified based on the confidences obtained from a previously trained classifier model. In a second embodiment, the errors and inconsistencies are identified based on an unsupervised learning process. In both embodiments, the active labeling process is not dependent upon the particular classifier model.
    Type: Grant
    Filed: September 27, 2007
    Date of Patent: July 14, 2009
    Assignee: AT&T Intellectual Property II, L.P.
    Inventors: Dilek Z. Hakkani-Tur, Mazin G. Rahim, Gokhan Tur
  • Publication number: 20090112585
    Abstract: Recognizing a stream of speech received as speech vectors over a lossy communications link includes constructing for a speech recognizer a series of speech vectors from packets received over a lossy packetized transmission link, wherein some of the packets associated with each speech vector are lost or corrupted during transmission. Each constructed speech vector is multi-dimensional and includes associated features. After waiting for a predetermined time, speech vectors are generated and potentially corrupted features within the speech vector are indicated to the speech recognizer when present. Speech recognition is attempted at the speech recognizer on the speech vectors when corrupted features are present. This recognition may be based only on certain or valid features within each speech vector. Retransmission of a missing or corrupted packet is requested when corrupted values are indicated by the indicating step and when the attempted recognition step fails.
    Type: Application
    Filed: December 29, 2008
    Publication date: April 30, 2009
    Applicant: AT&T Corp.
    Inventors: Richard Vandervoort Cox, Stephen Michael Marcus, Mazin G. Rahim, Nambirajan Seshadri, Robert Douglas Sharp
  • Publication number: 20090070113
    Abstract: A voice-enabled help desk service is disclosed. The service comprises an automatic speech recognition module for recognizing speech from a user, a spoken language understanding module for understanding the output from the automatic speech recognition module, a dialog management module for generating a response to speech from the user, a natural voices text-to-speech synthesis module for synthesizing speech to generate the response to the user, and a frequently asked questions module. The frequently asked questions module handles frequently asked questions from the user by changing voices and providing predetermined prompts to answer the frequently asked question.
    Type: Application
    Filed: November 7, 2008
    Publication date: March 12, 2009
    Applicant: AT&T Corp.
    Inventors: Narendra K. Gupta, Mazin G. Rahim, Giuseppe Riccardi
  • Patent number: 7496503
    Abstract: Recognizing a stream of speech received as speech vectors over a lossy communications link includes constructing for a speech recognizer a series of speech vectors from packets received over a lossy packetized transmission link, wherein some of the packets associated with each speech vector are lost or corrupted during transmission. Each constructed speech vector is multi-dimensional and includes associated features. After waiting for a predetermined time, speech vectors are generated and potentially corrupted features within the speech vector are indicated to the speech recognizer when present. Speech recognition is attempted at the speech recognizer on the speech vectors when corrupted features are present. This recognition may be based only on certain or valid features within each speech vector. Retransmission of a missing or corrupted packet is requested when corrupted values are indicated by the indicating step and when the attempted recognition step fails.
    Type: Grant
    Filed: December 18, 2006
    Date of Patent: February 24, 2009
    Assignee: AT&T Intellectual Property II, L.P.
    Inventors: Richard Vandervoort Cox, Stephen Michael Marcus, Mazin G. Rahim, Nambirajan Seshadri, Robert Douglas Sharp
  • Patent number: 7451089
    Abstract: A voice-enabled help desk service is disclosed. The service comprises an automatic speech recognition module for recognizing speech from a user, a spoken language understanding module for understanding the output from the automatic speech recognition module, a dialog management module for generating a response to speech from the user, a natural voices text-to-speech synthesis module for synthesizing speech to generate the response to the user, and a frequently asked questions module. The frequently asked questions module handles frequently asked questions from the user by changing voices and providing predetermined prompts to answer the frequently asked question.
    Type: Grant
    Filed: February 15, 2007
    Date of Patent: November 11, 2008
    Assignee: AT&T Intellectual Property II, L.P.
    Inventors: Narendra K. Gupta, Mazin G. Rahim, Giuseppe Riccardi
  • Publication number: 20080270130
    Abstract: Systems and methods for annotating speech data. The present invention reduces the time required to annotate speech data by selecting utterances for annotation that will be of greatest benefit. A selection module uses speech models, including speech recognition models and spoken language understanding models, to identify utterances that should be annotated based on criteria such as confidence scores generated by the models. These utterances are placed in an annotation list along with a type of annotation to be performed for the utterances and an order in which the annotation should proceed. The utterances in the annotation list can be annotated for speech recognition purposes, spoken language understanding purposes, labeling purposes, etc. The selection module can also select utterances for annotation based on previously annotated speech data and deficiencies in the various models.
    Type: Application
    Filed: July 1, 2008
    Publication date: October 30, 2008
    Applicant: AT&T Corp.
    Inventors: Tirso M. Alonso, Ilana Bromberg, Dilek Z. Hakkani-Tur, Barbara B. Hollister, Mazin G. Rahim, Giuseppe Riccardi, Lawrence Lyon Rose, Daniel Leon Stern, Gokhan Tur, James M. Wilson
  • Patent number: 7412383
    Abstract: Systems and methods for annotating speech data. The present invention reduces the time required to annotate speech data by selecting utterances for annotation that will be of greatest benefit. A selection module uses speech models, including speech recognition models and spoken language understanding models, to identify utterances that should be annotated based on criteria such as confidence scores generated by the models. These utterances are placed in an annotation list along with a type of annotation to be performed for the utterances and an order in which the annotation should proceed. The utterances in the annotation list can be annotated for speech recognition purposes, spoken language understanding purposes, labeling purposes, etc. The selection module can also select utterances for annotation based on previously annotated speech data and deficiencies in the various models.
    Type: Grant
    Filed: April 4, 2003
    Date of Patent: August 12, 2008
    Assignee: AT&T Corp
    Inventors: Tirso M. Alonso, Ilana Bromberg, Dilek Z. Hakkani-Tur, Barbara B. Hollister, Mazin G. Rahim, Giuseppe Riccardi, Lawrence Lyon Rose, Daniel Leon Stern, Gokhan Tur, James M. Wilson
  • Patent number: 7373300
    Abstract: Disclosed is a system and method for generating a spoken dialog service from website data. Spoken dialog components typically include an automatic speech recognition module, a language understanding module, a dialog management module, a language generation module and a test-to-speech module. These components are capable of being automatically trained from processed website data. A website analyzer converts a website into structured text data set and a structured task knowledge base. The website analyzer further extracts linguistic items from the website data. The dialog components are automatically trained from the structured text data set, structured task knowledge base and linguistic items.
    Type: Grant
    Filed: December 18, 2003
    Date of Patent: May 13, 2008
    Assignee: AT&T Corp.
    Inventors: Srinivas Bangalore, Junlan Feng, Mazin G. Rahim
  • Patent number: 7366655
    Abstract: A method is disclosed for designing a labeling guide for use by a labeler in labeling data used for training a spoken language understanding (SLU) module for an application. The method comprises a labeling guide designer selecting domain-independent actions applicable to an application, selecting domain-dependent objects according to characteristics of the application, and generating a labeling guide using the selected domain-independent actions and selected domain-dependent objects. An advantage of the labeling guide generated in this manner is that the labeling guide designer can easily port the labeling guide to a new application by selecting a set of domain-independent action and then selecting the domain-dependent objects related to the new application.
    Type: Grant
    Filed: April 2, 2003
    Date of Patent: April 29, 2008
    Assignee: AT&T Corp.
    Inventors: Narendra K. Gupta, Barbara B. Hollister, Mazin G Rahim, Giuseppe Riccardi
  • Patent number: 7328146
    Abstract: A system for understanding entries, such as speech, develops a classifier by employing prior knowledge with which a given corpus of training entries is enlarged threefold. A rule is created for each of the labels employed in the classifier, and the created rules are applied to the given corpus to create a corpus of attachments by appending a weight of ?p(x), or 1??p(x), to labels of entries that meet, or fail to meet, respectively, conditions of the labels' rules, and to also create a corpus of non-attachments by appending a weight of 1??p(x), or ?p(x), to labels of entries that meet, or fail to meet conditions of the labels' rules.
    Type: Grant
    Filed: July 11, 2006
    Date of Patent: February 5, 2008
    Assignee: AT&T Corp.
    Inventors: Hiyan Alshawi, Giuseppe DiFabrizzio, Narendra K. Gupta, Mazin G. Rahim, Robert E. Schapire, Yoram Singer
  • Patent number: 7295981
    Abstract: A method of generating a natural language model for use in a spoken dialog system is disclosed. The method comprises using sample utterances and creating a number of hand crafted rules for each call-type defined in a labeling guide. A first NLU model is generated and tested using the hand crafted rules and sample utterances. A second NLU model is built using the sample utterances as new training data and using the hand crafted rules. The second NLU model is tested for performance using a first batch of labeled data. A series of NLU models are built by adding a previous batch of labeled data to training data and using a new batch of labeling data as test data to generate the series of NLU models with training data that increases constantly. If not all the labeling data is received, the method comprises repeating the step of building a series of NLU models until all labeling data is received.
    Type: Grant
    Filed: January 9, 2004
    Date of Patent: November 13, 2007
    Assignee: AT&T Corp.
    Inventors: Narendra K. Gupta, Mazin G. Rahim, Gokhan Tur, Antony Van der Mude
  • Patent number: 7292982
    Abstract: An active labeling process is provided that aims to minimize the number of utterances to be checked again by automatically selecting the ones that are likely to be erroneous or inconsistent with the previously labeled examples. In one embodiment, the errors and inconsistencies are identified based on the confidences obtained from a previously trained classifier model. In a second embodiment, the errors and inconsistencies are identified based on an unsupervised learning process. In both embodiments, the active labeling process is not dependent upon the particular classifier model.
    Type: Grant
    Filed: May 29, 2003
    Date of Patent: November 6, 2007
    Assignee: AT&T Corp.
    Inventors: Dilek Z. Hakkani-Tur, Mazin G. Rahim, Gokhan Tur
  • Patent number: 7292976
    Abstract: A large amount of human labor is required to transcribe and annotate a training corpus that is needed to create and update models for automatic speech recognition (ASR) and spoken language understanding (SLU). Active learning enables a reduction in the amount of transcribed and annotated data required to train ASR and SLU models. In one aspect of the present invention, an active learning ASR process and active learning SLU process are coupled, thereby enabling further efficiencies to be gained relative to a process that maintains an isolation of data in both the ASR and SLU domains.
    Type: Grant
    Filed: May 29, 2003
    Date of Patent: November 6, 2007
    Assignee: AT&T Corp.
    Inventors: Dilek Z. Hakkani-Tur, Mazin G. Rahim, Giuseppe Riccardi, Gokhan Tur
  • Patent number: 7280965
    Abstract: Systems and methods for monitoring labelers of speech data. To test or train labelers, a labeler is presented with utterances that have already been identified as belonging to a particular class or call type. The labeler is asked to assign a call type to the utterances. The performance of the labeler is measured by comparing the call types assigned by the labeler with the existing call types of the utterances. The performance of a labeler can also be monitored as the labeler labels speech data by occasionally having the labeler label an utterance that is already labeled and by storing the results.
    Type: Grant
    Filed: April 4, 2003
    Date of Patent: October 9, 2007
    Assignee: AT&T Corp.
    Inventors: Lee Begeja, Richard Vandervoort Cox, Harris Drucker, David Crawford Gibbon, Allen Louis Gorin, Patrick Guy Haffner, Steven H. Lewis, Zhu Liu, Mazin G. Rahim, Bernard S. Renger, Behzad Shahraray
  • Patent number: 7219054
    Abstract: Systems and methods for generating an annotation guide. Speech data is organized and presented to a user. After the user selects some of the utterances in the speech data, the selected utterances are included in a class and/or call type. Additional utterances that belong to the class and/or call type can be found in the speech data using relevance feedback, data mining, data clustering, support vector machines, and the like. After a call type is complete, it is committed to the annotation guide. After all call types are completed, the annotation guide is generated.
    Type: Grant
    Filed: April 4, 2003
    Date of Patent: May 15, 2007
    Assignee: AT&T Corp.
    Inventors: Lee Begeja, Harris Drucker, David Crawford Gibbon, Allen Louis Gorin, Patrick Guy Haffner, Steven H Lewis, Zhu Liu, Mazin G Rahim, Bernard S. Renger, Behzad Shahraray
  • Patent number: 7197460
    Abstract: A voice-enabled help desk service is disclosed. The service comprises an automatic speech recognition module for recognizing speech from a user, a spoken language understanding module for understanding the output from the automatic speech recognition module, a dialog management module for generating a response to speech from the user, a natural voices text-to-speech synthesis module for synthesizing speech to generate the response to the user, and a frequently asked questions module. The frequently asked questions module handles frequently asked questions from the user by changing voices and providing predetermined prompts to answer the frequently asked question.
    Type: Grant
    Filed: December 19, 2002
    Date of Patent: March 27, 2007
    Assignee: AT&T Corp.
    Inventors: Narendra K. Gupta, Mazin G Rahim, Giuseppe Riccardi
  • Patent number: 7181399
    Abstract: A system for recognizing connected digits in natural spoken dialogue includes a speech recognition processor that receives unconstrained fluent input speech and produces a string of words that can include a numeric language, and a numeric understanding processor that converts the string of words into a sequence of digits based on a set of rules. An acoustic model database utilized by the speech recognition processor includes a first set of hidden Markov models that characterize the acoustic features of numeric words and phrases, a second set of hidden Markov models that characterize the acoustic features of the remaining vocabulary words, and a filler model that characterizes the acoustic features of out-of-vocabulary utterances. An utterance verification processor verifies the accuracy of the string of words. A validation database stores a grammar, and a string validation processor outputs validity information based on a comparison of the sequence of digits with the grammar.
    Type: Grant
    Filed: May 19, 1999
    Date of Patent: February 20, 2007
    Assignee: AT&T Corp.
    Inventors: Mazin G. Rahim, Giuseppe Riccardi, Jeremy Huntley Wright, Bruce Melvin Buntschuh, Allen Louis Gorin
  • Patent number: 7171359
    Abstract: Recognizing a stream of speech received as speech vectors over a lossy communications link includes constructing for a speech recognizer a series of speech vectors from packets received over a lossy packetized transmission link, wherein some of the packets associated with each speech vector are lost or corrupted during transmission. Each constructed speech vector is multi-dimensional and includes associated features. Potentially corrupted features within the speech vector are indicated to the speech recognizer when present. Speech recognition is attempted at the speech recognizer on the speech vectors when corrupted features are present. This recognition may be based only on certain or valid features within each speech vector. Retransmission of a missing or corrupted packet is requested when corrupted values are indicated by the indicating step and when the attempted recognition step fails.
    Type: Grant
    Filed: July 29, 2004
    Date of Patent: January 30, 2007
    Assignee: AT&T Corp.
    Inventors: Richard Vandervoort Cox, Stephen Michael Marcus, Mazin G. Rahim, Nambirajan Seshadri, Robert Douglas Sharp