Patents by Inventor Clement Yip
Clement Yip has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).
-
Patent number: 8374879Abstract: Systems and methods are described for speech systems that utilize an interaction manager to manage interactions—also known as dialogues—from one or more applications. The interactions are managed properly even if multiple applications use different grammars. The interaction manager maintains an interaction list. An application wishing to utilize the speech system submits one or more interactions to the interaction manager. Interactions are normally processed in the order in which they are received. An exception to this rule is an interaction that is configured by an application to be processed immediately, which causes the interaction manager to place the interaction at the front of the interaction list of interactions. If an application has designated an interaction to interrupt a currently processing interaction, then the newly submitted application will interrupt any interaction currently being processed and, therefore, it will be processed immediately.Type: GrantFiled: December 16, 2005Date of Patent: February 12, 2013Assignee: Microsoft CorporationInventors: Stephen Russell Falcon, Clement Yip, Dan Banay, David Miller
-
Patent number: 7742925Abstract: Systems and methods are described for a speech system that includes one or more speech controls incorporated into one or more speech-enabled applications that run on the speech system. The controls allow applications to be developed with minimal programming effort to incorporate common speech-enabled application functions. A question control provides a customizable template for requesting information from a user. An announcer control allows a speech-enabled application to provide a user with information without having to re-create an entire announcer process each time it is used. A command control provides a simple way to attach command and control functions to speech-enabled applications. A word trainer control provides a way to associate user-selected voice tags with certain information. Providing the controls for use with speech-enabled applications ensures standardized user interfaces across multiple speech-enabled applications.Type: GrantFiled: December 19, 2005Date of Patent: June 22, 2010Assignee: Microsoft CorporationInventors: Stephen R Falcon, Clement Yip, Dan Banay, David M Miller
-
Patent number: 7720678Abstract: Systems and methods are described for a speech system that manages multiple grammars from one or more speech-enabled applications. The speech system includes a speech server that supports different grammars and different types of grammars by exposing several methods to the speech-enabled applications. The speech server supports static grammars that do not change and dynamic grammars that may change after a commit. The speech server provides persistence by supporting persistent grammars that enable a user to issue a command to an application even when the application is not loaded. In such a circumstance, the application is automatically launched and the command is processed. The speech server may enable or disable a grammar in order to limit confusion between grammars. Global and yielding grammars are also supported by the speech server. Global grammars are always active (e.g., “call 9-1-1”) while yielding grammars may be deactivated when an interaction whose grammar requires priority is active.Type: GrantFiled: December 16, 2005Date of Patent: May 18, 2010Assignee: Microsoft CorporationInventors: Steve Falcon, Clement Yip, David Miller, Dan Banay
-
Publication number: 20070143115Abstract: Systems and methods are described for speech systems that utilize an interaction manager to manage interactions—also known as dialogues—from one or more applications. The interactions are managed properly even if multiple applications use different grammars. The interaction manager maintains an interaction list. An application wishing to utilize the speech system submits one or more interactions to the interaction manager. Interactions are normally processed in the order in which they are received. An exception to this rule is an interaction that is configured by an application to be processed immediately, which causes the interaction manager to place the interaction at the front of the interaction list of interactions. If an application has designated an interaction to interrupt a currently processing interaction, then the newly submitted application will interrupt any interaction currently being processed and, therefore, it will be processed immediately.Type: ApplicationFiled: December 16, 2005Publication date: June 21, 2007Applicant: Microsoft CorporationInventors: Stephen Falcon, Clement Yip, Dan Banay, David Miller
-
Publication number: 20060161429Abstract: Systems and methods are described for a speech system that manages multiple grammars from one or more speech-enabled applications. The speech system includes a speech server that supports different grammars and different types of grammars by exposing several methods to the speech-enabled applications. The speech server supports static grammars that do not change and dynamic grammars that may change after a commit. The speech server provides persistence by supporting persistent grammars that enable a user to issue a command to an application even when the application is not loaded. In such a circumstance, the application is automatically launched and the command is processed. The speech server may enable or disable a grammar in order to limit confusion between grammars. Global and yielding grammars are also supported by the speech server. Global grammars are always active (e.g., “call 9-1-1”) while yielding grammars may be deactivated when an interaction whose grammar requires priority is active.Type: ApplicationFiled: December 16, 2005Publication date: July 20, 2006Applicant: Microsoft CorporationInventors: Stephen Falcon, Clement Yip, David Miller, Dan Banay
-
Publication number: 20060106617Abstract: Systems and methods are described for a speech system that includes one or more speech controls incorporated into one or more speech-enabled applications that run on the speech system. The controls allow applications to be developed with minimal programming effort to incorporate common speech-enabled application functions. A question control provides a customizable template for requesting information from a user. An announcer control allows a speech-enabled application to provide a user with information without having to re-create an entire announcer process each time it is used. A command control provides a simple way to attach command and control functions to speech-enabled applications. A word trainer control provides a way to associate user-selected voice tags with certain information. Providing the controls for use with speech-enabled applications ensures standardized user interfaces across multiple speech-enabled applications.Type: ApplicationFiled: December 19, 2005Publication date: May 18, 2006Applicant: Microsoft CorporationInventors: Stephen Falcon, Clement Yip, Dan Banay, David Miller
-
Publication number: 20060069571Abstract: Systems and methods are described for speech systems that utilize an interaction manager to manage interactions—also known as dialogues—from one or more applications. The interactions are managed properly even if multiple applications use different grammars. The interaction manager maintains an interaction list. An application wishing to utilize the speech system submits one or more interactions to the interaction manager. Interactions are normally processed in the order in which they are received. An exception to this rule is an interaction that is configured by an application to be processed immediately, which causes the interaction manager to place the interaction at the front of the interaction list of interactions. If an application has designated an interaction to interrupt a currently processing interaction, then the newly submitted application will interrupt any interaction currently being processed and, therefore, it will be processed immediately.Type: ApplicationFiled: November 1, 2005Publication date: March 30, 2006Applicant: Microsoft CorporationInventors: Stephen Falcon, Clement Yip, Dan Banay, David Miller
-
Publication number: 20060069573Abstract: Systems and methods are described for a speech system that includes one or more speech controls incorporated into one or more speech-enabled applications that run on the speech system. The controls allow applications to be developed with minimal programming effort to incorporate common speech-enabled application functions. A question control provides a customizable template for requesting information from a user. An announcer control allows a speech-enabled application to provide a user with information without having to re-create an entire announcer process each time it is used. A command control provides a simple way to attach command and control functions to speech-enabled applications. A word trainer control provides a way to associate user-selected voice tags with certain information. Providing the controls for use with speech-enabled applications ensures standardized user interfaces across multiple speech-enabled applications.Type: ApplicationFiled: November 2, 2005Publication date: March 30, 2006Applicant: Microsoft CorporationInventors: Stephen Falcon, Clement Yip, Dan Banay, David Miller
-
Publication number: 20060053016Abstract: Systems and methods are described for a speech system that manages multiple grammars from one or more speech-enabled applications. The speech system includes a speech server that supports different grammars and different types of grammars by exposing several methods to the speech-enabled applications. The speech server supports static grammars that do not change and dynamic grammars that may change after a commit. The speech server provides persistence by supporting persistent grammars that enable a user to issue a command to an application even when the application is not loaded. In such a circumstance, the application is automatically launched and the command is processed. The speech server may enable or disable a grammar in order to limit confusion between grammars. Global and yielding grammars are also supported by the speech server. Global grammars are always active (e.g., “call 9-1-1”) while yielding grammars may be deactivated when an interaction whose grammar requires priority is active.Type: ApplicationFiled: November 4, 2005Publication date: March 9, 2006Applicant: Microsoft CorporationInventors: Stephen Falcon, Clement Yip, Dan Banay, David Miller
-
Publication number: 20060010499Abstract: Methods and arrangements are provided to verify if a requesting computer application is authorized to change a controlled parameter associated with a computer controlled device and/or function. To accomplish this, one or verification functions are employed to analyze a security code or absence thereof, as identified by a requesting application. If the security code, which may be encrypted, matches a known or calculated valid security code, then the requesting application is deemed to be authorized to change the controlled parameter and/or modify certain limitations associated with an acceptable range for the controlled parameter. If the security code does not match a known or calculated valid security code, then the requesting application is deemed to be unauthorized to change the controlled parameter outside of a previously established acceptable range for the controlled parameter.Type: ApplicationFiled: September 6, 2005Publication date: January 12, 2006Applicant: Microsoft CorporationInventors: Stephen Falcon, Clement Yip