Patents by Inventor Jean-Claude Junqua

Jean-Claude Junqua has filed for patents to protect the following inventions. This listing includes patent applications that are pending as well as patents that have already been granted by the United States Patent and Trademark Office (USPTO).

  • Patent number: 6970820
    Abstract: The speech synthesizer is personalized to sound like or mimic the speech characteristics of an individual speaker. The individual speaker provides a quantity of enrollment data, which can be extracted from a short quantity of speech, and the system modifies the base synthesis parameters to more closely resemble those of the new speaker. More specifically, the synthesis parameters may be decomposed into speaker dependent parameters, such as context-independent parameters, and speaker independent parameters, such as context dependent parameters. The speaker dependent parameters are adapted using enrollment data from the new speaker. After adaptation, the speaker dependent parameters are combined with the speaker independent parameters to provide a set of personalized synthesis parameters.
    Type: Grant
    Filed: February 26, 2001
    Date of Patent: November 29, 2005
    Assignee: Matsushita Electric Industrial Co., Ltd.
    Inventors: Jean-Claude Junqua, Florent Perronnin, Roland Kuhn, Patrick Nguyen
  • Publication number: 20050228663
    Abstract: A media production system includes a textual alignment module aligning multiple speech recordings to textual lines of a script based on speech recognition results. A navigation module responds to user navigation selections respective of the textual lines of the script by communicating to the user corresponding, line-specific portions of the multiple speech recordings. An editing module responds to user associations of multiple speech recordings with textual lines by accumulating line-specific portions of the multiple speech recordings in a combination recording based on at least one of relationships of textual lines in the script to the combination recording, and temporal alignments between the multiple speech recordings and the combination recording.
    Type: Application
    Filed: March 31, 2004
    Publication date: October 13, 2005
    Inventors: Robert Boman, Patrick Nguyen, Jean-Claude Junqua
  • Patent number: 6952164
    Abstract: A wearable, computerized apparatus for use with law enforcement has an evidence collector adapted to collect evidentiary information of a type collected according to law enforcement procedures and useful for identification of a suspect. It further has a safety monitor adapted to collect safety information relating to well-being of an officer. A wireless communications link communicates the evidentiary information and the safety information to a centralized component of a distributed communications system to assist in identifying suspects and dispatching assistance.
    Type: Grant
    Filed: November 5, 2002
    Date of Patent: October 4, 2005
    Assignee: Matsushita Electric Industrial Co., Ltd.
    Inventor: Jean-Claude Junqua
  • Publication number: 20050182630
    Abstract: A multilingual text-to-speech system includes a source datastore of primary source parameters providing information about a speaker of a primary language. A plurality of primary filter parameters provides information about sounds in the primary language. A plurality of secondary filter parameters provides information about sounds in a secondary language. One or more secondary filter parameters is normalized to the primary filter parameters and mapped to a primary source parameter.
    Type: Application
    Filed: February 2, 2004
    Publication date: August 18, 2005
    Inventors: Xavier Miro, Peter Veprek, Jean-Claude Junqua
  • Publication number: 20050171774
    Abstract: A speaker authentication system includes an input receptive of user speech from a user. An extraction module extracts acoustic correlates of aspects of the user's physiology from the user speech, including at least one of glottal source parameters, formant related parameters, timing characteristics, and pitch related qualities. An output communicates the acoustic correlates to an authentication module adapted to authenticate the user by comparing the acoustic correlates to predefined acoustic correlates in a datastore.
    Type: Application
    Filed: January 30, 2004
    Publication date: August 4, 2005
    Inventors: Ted Applebaum, Steven Pearson, Philippe Morin, Jean-Claude Junqua
  • Publication number: 20050154265
    Abstract: A robotic nursing system for use with a patient comprises a nursing robot having at least one patient condition sensor, a transmitter, and a receiver mounted therein. A display device for displays data sensed by the patient condition sensor. The display device includes a receiver in communication with the nursing robot. The nursing robot senses patient physiological conditions using the patient condition sensor and transmits the physiological conditions to the display device using the transmitter. The display device then displays the physiological conditions for review by a user. One or another or both. The nursing robot also transmits the physiological conditions to a patient database for storage.
    Type: Application
    Filed: January 12, 2004
    Publication date: July 14, 2005
    Inventors: Xavier Miro, Jean-Claude Junqua
  • Patent number: 6915259
    Abstract: Linear approximation of the background noise is applied after feature extraction and prior to speaker adaptation to allow the speaker adaptation system to adapt the speech models to the enrolling user without distortion from background noise. The linear approximation is applied in the feature domain, such as in the cepstral domain. Any adaptation technique that is commutative in the feature domain may be used.
    Type: Grant
    Filed: May 24, 2001
    Date of Patent: July 5, 2005
    Assignee: Matsushita Electric Industrial Co., Ltd.
    Inventors: Luca Rigazio, Patrick Nguyen, David Kryze, Jean-Claude Junqua
  • Patent number: 6907397
    Abstract: An embedded device for playing media files is capable of generating a play list of media files based on input speech from a user. It includes an indexer generating a plurality of speech recognition grammars. According to one aspect of the invention, the indexer generates speech recognition grammars based on contents of a media file header of the media file. According to another aspect of the invention, the indexer generates speech recognition grammars based on categories in a file path for retrieving the media file to a user location. When a speech recognizer receives an input speech from a user while in a selection mode, a media file selector compares the input speech received while in the selection mode to the plurality of speech recognition grammars, thereby selecting the media file.
    Type: Grant
    Filed: September 16, 2002
    Date of Patent: June 14, 2005
    Assignee: Matsushita Electric Industrial Co., Ltd.
    Inventors: David Kryze, Luca Rigazio, Patrick Nguyen, Jean-Claude Junqua
  • Patent number: 6901364
    Abstract: An e-mail message process is provided for use with a personal digital assistant which allows for the use of input speech messaging which is converted to text using a focused language model which is downloaded by a cellular phone connection to an Internet server which provides the focused language model based upon a topic for the intended e-mail message. The text that is generated from the input speech method can be summarized by the e-mail message processor and can be edited by the user. The generated e-mail message can then be transmitted again via cellular connection to an Internet e-mail server for transmitting the e-mail message to a recipient.
    Type: Grant
    Filed: September 13, 2001
    Date of Patent: May 31, 2005
    Assignee: Matsushita Electric Industrial Co., Ltd.
    Inventors: Patrick Nguyen, Luca Rigazio, Jean-Claude Junqua
  • Patent number: 6901366
    Abstract: The system retrieves information from the internet using multiple search engines that are simultaneously launched by the search engine commander. The commander is responsive to a speech-enabled system including a speech recognizer and natural language parser. The user speaks to the system in natural language requests, and the parser extracts the semantic content from the user's speech, based on a set of goal oriented grammars. The preferred system includes a fixed grammar and an updatable or downloaded grammar, allowing the system to be used without extensive training and yet capable of being customized for a particular user's purposes. Results obtained from the search engines are filtered based on information extracted from an electronic program guide and from prestored user profile data. The results may be displayed on screen or through synthesized speech.
    Type: Grant
    Filed: August 26, 1999
    Date of Patent: May 31, 2005
    Assignee: Matsushita Electric Industrial Co., Ltd.
    Inventors: Roland Kuhn, Jean-Claude Junqua, Tony Davis, Weiying Li, Yi Zhao
  • Patent number: 6895376
    Abstract: A reduced dimensionality eigenvoice analytical technique is used during training to develop context-dependent acoustic models for allophones. Re-estimation processes are performed to more strongly separate speaker-dependent and speaker-independent components of the speech model. The eigenvoice technique is also used during run time upon the speech of a new speaker. The technique removes individual speaker idiosyncrasies, to produce more universally applicable and robust allophone models. In one embodiment the eigenvoice technique is used to identify the centroid of each speaker, which may then be “subtracted out” of the recognition equation.
    Type: Grant
    Filed: May 4, 2001
    Date of Patent: May 17, 2005
    Assignee: Matsushita Electric Industrial Co., Ltd.
    Inventors: Florent Perronnin, Roland Kuhn, Patrick Nguyen, Jean-Claude Junqua
  • Patent number: 6895257
    Abstract: Personalized agent services are provided in a personal messaging device, such as a cellular telephone or personal digital assistant, through services of a speech recognizer that converts speech into text and a text-to-speech synthesizer that converts text to speech. Both recognizer and synthesizer may be server-based or locally deployed within the device. The user dictates an e-mail message which is converted to text and stored. The stored text is sent back to the user as text or as synthesized speech, to allow the user to edit the message and correct transcription errors before sending as e-mail. The system includes a summarization module that prepares short summaries of incoming e-mail and voice mail. The user may access these summaries, and retrieve and organize email and voice mail using speech commands.
    Type: Grant
    Filed: February 18, 2002
    Date of Patent: May 17, 2005
    Assignee: Matsushita Electric Industrial Co., Ltd.
    Inventors: Robert Boman, Kirill Stoimenov, Roland Kuhn, Jean-Claude Junqua
  • Publication number: 20050075881
    Abstract: A media capture device has an audio input receptive of user speech relating to a media capture activity in close temporal relation to the media capture activity. A plurality of focused speech recognition lexica respectively relating to media capture activities are stored on the device, and a speech recognizer recognizes the user speech based on a selected one of the focused speech recognition lexica. A media tagger tags captured media with generated speech recognition text, and a media annotator annotates the captured media with a sample of the user speech that is suitable for input to a speech recognizer. Tagging and annotating are based on close temporal relation between receipt of the user speech and capture of the captured media. Annotations may be converted to tags during post processing, employed to edit a lexicon using letter-to-sound rules and spelled word input, or matched directly to speech to retrieve captured media.
    Type: Application
    Filed: October 2, 2003
    Publication date: April 7, 2005
    Inventors: Luca Rigazio, Robert Boman, Patrick Nguyen, Jean-Claude Junqua
  • Publication number: 20050049860
    Abstract: A method for improving recognition results of a speech recognizer uses supplementary information to confirm recognition results. A user inputs speech to a speech recognizer. The speech recognizer resides on a mobile device or on a server at a remote location. The speech recognizer determines a recognition result based on the input speech. A confidence measure is calculated for the recognition result. If the confidence measure is below a threshold, the user is prompted for supplementary data. The supplementary data is determined dynamically based on ambiguities between the input speech and the recognition result, wherein the supplementary data will distinguish the input speech over potential incorrect results. The supplementary data may be a subset of alphanumeric characters that comprise the input speech, or other data associated with a desired result, such as an area code or location. The user may provide the supplementary data verbally, or manually using a keypad, touchpad, touchscreen, or stylus pen.
    Type: Application
    Filed: August 29, 2003
    Publication date: March 3, 2005
    Inventors: Jean-Claude Junqua, Roland Kuhn, Matteo Contolini, Rathinavelu Chengalvarayan
  • Publication number: 20050010411
    Abstract: A speech data mining system for use in generating a rich transcription having utility in call center management includes a speech differentiation module differentiating between speech of interacting speakers, and a speech recognition module improving automatic recognition of speech of one speaker based on interaction with another speaker employed as a reference speaker. A transcript generation module generates a rich transcript based on recognized speech of the speakers. Focused, interactive language models improve recognition of a customer on a low quality channel using context extracted from speech of a call center operator on a high quality channel with a speech model adapted to the operator. Mined speech data includes number of interaction turns, customer frustration phrases, operator polity, interruptions, and/or contexts extracted from speech recognition results, such as topics, complaints, solutions, and resolutions.
    Type: Application
    Filed: July 9, 2003
    Publication date: January 13, 2005
    Inventors: Luca Rigazio, Patrick Nguyen, Jean-Claude Junqua, Robert Boman
  • Publication number: 20040246127
    Abstract: A wearable, computerized apparatus for use with law enforcement has an evidence collector adapted to collect evidentiary information of a type collected according to law enforcement procedures and useful for identification of a suspect. It further has a safety monitor adapted to collect safety information relating to well-being of an officer. A wireless communications link communicates the evidentiary information and the safety information to a centralized component of a distributed communications system to assist in identifying suspects and dispatching assistance.
    Type: Application
    Filed: July 7, 2004
    Publication date: December 9, 2004
    Applicant: MATSUSHITA ELECTRIC INDUSTRIAL CO., LTD.
    Inventor: Jean-Claude Junqua
  • Publication number: 20040249636
    Abstract: Unstructured voice information from an incoming caller is processed by automatic speech recognition and semantic categorization system to convert the information into structured data that may then be used to access one or more databases to retrieve associated supplemental data. The structured data and associated supplemental data are then made available through a presentation system that provides information to the call center agent and, optionally, to the incoming caller. The system thus allows a call center information processing system to handle unstructured voice input for use by the live agent in handling the incoming call and for storage and retrieval at a later time. The semantic analysis system may be implemented by a global parser or by an information retrieval technique, such as latent semantic analysis. Co-occurrence of keywords may be used to associate prior calls with an incoming call to assist in understanding the purpose of the incoming call.
    Type: Application
    Filed: June 4, 2003
    Publication date: December 9, 2004
    Inventors: Ted Applebaum, Jean-Claude Junqua
  • Publication number: 20040236778
    Abstract: Program content, recorded to a storage medium such as disk recorder, optical recorder or random access memory, is indexed by the replay file system. The file system maintains a storage location and program I.D. record for each recorded program. The file system further maintains other data obtained from an electronic program guide that may be accessed by downloading from the cable or satellite infrastructure or over the internet. The file system also may store additional user data, such as the date and time the program was last viewed, together with any user-recorded indexes. The file system may be accessed through natural language input speech. The system includes a speech recognizer and natural language parser, coupled to a dialog system that engages the user in a dialog to determine what the user is interested in accessing from the storage medium. The natural language parser operates with a task-based grammar that is keyed to the electronic program guide data and user data maintained by the file system.
    Type: Application
    Filed: July 7, 2004
    Publication date: November 25, 2004
    Applicant: MATSUSHITA ELECTRIC INDUSTRIAL CO., LTD.
    Inventors: Jean-Claude Junqua, Roland Kuhn, Tony Davis, Yi Zhao, Weiying Li
  • Publication number: 20040225504
    Abstract: A portable device increases user access to equipment utilizing a communications interface providing communication with the equipment in accordance with various, combinable embodiments. In one embodiment, a speech generator generates speech based on commands relating to equipment operation, which may be received from the equipment via the communications interface. A selection mechanism allows the user to select commands and thereby operate the equipment. In another embodiment, a command navigator navigates commands based on user input by shifting focus between commands, communicates a command having the focus to the speech generator, and allows the user to select a command. In a further embodiment, a phoneticizer converts the commands and/or predetermined navigation and selection options into a dynamic speech lexicon, and a speech recognizer uses the lexicon to recognize a user navigation input and/or user selection of a command.
    Type: Application
    Filed: May 9, 2003
    Publication date: November 11, 2004
    Inventors: Jean-Claude Junqua, Eugene J. Seagriff
  • Publication number: 20040210443
    Abstract: The system assesses a measure of quality associated with the user's query, which may be based on the query itself or upon the results returned from a first search space. If the measure of quality is low, the system accesses one or more second knowledge sources and retrieves intermediate results that belong to the vocabulary of the first search space. A second query is then constructed using the intermediate results, and based on further input from the user as needed. The second query is then used to search the first search space with results returned to the user.
    Type: Application
    Filed: April 17, 2003
    Publication date: October 21, 2004
    Inventors: Roland Kuhn, Jean-Claude Junqua, Patrick Nguyen